{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 315.7894736842105, "eval_steps": 500, "global_step": 30000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.10526315789473684, "grad_norm": 6.911316871643066, "learning_rate": 1.8e-07, "loss": 0.9931, "step": 10 }, { "epoch": 0.21052631578947367, "grad_norm": 6.736523151397705, "learning_rate": 3.8e-07, "loss": 1.0041, "step": 20 }, { "epoch": 0.3157894736842105, "grad_norm": 5.807806491851807, "learning_rate": 5.8e-07, "loss": 0.9723, "step": 30 }, { "epoch": 0.42105263157894735, "grad_norm": 6.1695075035095215, "learning_rate": 7.8e-07, "loss": 0.954, "step": 40 }, { "epoch": 0.5263157894736842, "grad_norm": 4.514342784881592, "learning_rate": 9.8e-07, "loss": 0.8, "step": 50 }, { "epoch": 0.631578947368421, "grad_norm": 3.0857925415039062, "learning_rate": 1.18e-06, "loss": 0.7067, "step": 60 }, { "epoch": 0.7368421052631579, "grad_norm": 2.41701602935791, "learning_rate": 1.3800000000000001e-06, "loss": 0.5401, "step": 70 }, { "epoch": 0.8421052631578947, "grad_norm": 1.1335846185684204, "learning_rate": 1.5800000000000003e-06, "loss": 0.4601, "step": 80 }, { "epoch": 0.9473684210526315, "grad_norm": 0.8202275633811951, "learning_rate": 1.7800000000000001e-06, "loss": 0.3297, "step": 90 }, { "epoch": 1.0526315789473684, "grad_norm": 0.645756185054779, "learning_rate": 1.98e-06, "loss": 0.2693, "step": 100 }, { "epoch": 1.1578947368421053, "grad_norm": 0.5055798292160034, "learning_rate": 2.1800000000000003e-06, "loss": 0.2264, "step": 110 }, { "epoch": 1.263157894736842, "grad_norm": 0.3649744987487793, "learning_rate": 2.38e-06, "loss": 0.1918, "step": 120 }, { "epoch": 1.368421052631579, "grad_norm": 0.3170933127403259, "learning_rate": 2.5800000000000003e-06, "loss": 0.1727, "step": 130 }, { "epoch": 1.4736842105263157, "grad_norm": 0.3381025195121765, "learning_rate": 2.78e-06, "loss": 0.1606, "step": 140 }, { "epoch": 1.5789473684210527, "grad_norm": 0.30769628286361694, "learning_rate": 2.9800000000000003e-06, "loss": 0.1513, "step": 150 }, { "epoch": 1.6842105263157894, "grad_norm": 0.489881306886673, "learning_rate": 3.1800000000000005e-06, "loss": 0.1443, "step": 160 }, { "epoch": 1.7894736842105263, "grad_norm": 0.3954871892929077, "learning_rate": 3.38e-06, "loss": 0.1317, "step": 170 }, { "epoch": 1.8947368421052633, "grad_norm": 0.32977819442749023, "learning_rate": 3.58e-06, "loss": 0.1224, "step": 180 }, { "epoch": 2.0, "grad_norm": 0.5767644643783569, "learning_rate": 3.7800000000000002e-06, "loss": 0.1165, "step": 190 }, { "epoch": 2.1052631578947367, "grad_norm": 0.27882176637649536, "learning_rate": 3.98e-06, "loss": 0.1088, "step": 200 }, { "epoch": 2.2105263157894735, "grad_norm": 0.2769348621368408, "learning_rate": 4.18e-06, "loss": 0.0993, "step": 210 }, { "epoch": 2.3157894736842106, "grad_norm": 0.28452855348587036, "learning_rate": 4.38e-06, "loss": 0.0978, "step": 220 }, { "epoch": 2.4210526315789473, "grad_norm": 0.29632964730262756, "learning_rate": 4.58e-06, "loss": 0.0931, "step": 230 }, { "epoch": 2.526315789473684, "grad_norm": 0.2746676802635193, "learning_rate": 4.780000000000001e-06, "loss": 0.0877, "step": 240 }, { "epoch": 2.6315789473684212, "grad_norm": 0.27415645122528076, "learning_rate": 4.98e-06, "loss": 0.0852, "step": 250 }, { "epoch": 2.736842105263158, "grad_norm": 0.3377937376499176, "learning_rate": 5.18e-06, "loss": 0.0798, "step": 260 }, { "epoch": 2.8421052631578947, "grad_norm": 0.2798013389110565, "learning_rate": 5.38e-06, "loss": 0.0753, "step": 270 }, { "epoch": 2.9473684210526314, "grad_norm": 0.24506144225597382, "learning_rate": 5.580000000000001e-06, "loss": 0.0702, "step": 280 }, { "epoch": 3.0526315789473686, "grad_norm": 0.24791370332241058, "learning_rate": 5.78e-06, "loss": 0.0682, "step": 290 }, { "epoch": 3.1578947368421053, "grad_norm": 0.255985289812088, "learning_rate": 5.98e-06, "loss": 0.0683, "step": 300 }, { "epoch": 3.263157894736842, "grad_norm": 0.26971709728240967, "learning_rate": 6.18e-06, "loss": 0.067, "step": 310 }, { "epoch": 3.3684210526315788, "grad_norm": 0.27581414580345154, "learning_rate": 6.38e-06, "loss": 0.0621, "step": 320 }, { "epoch": 3.473684210526316, "grad_norm": 0.23893456161022186, "learning_rate": 6.58e-06, "loss": 0.0616, "step": 330 }, { "epoch": 3.5789473684210527, "grad_norm": 0.26243728399276733, "learning_rate": 6.78e-06, "loss": 0.0611, "step": 340 }, { "epoch": 3.6842105263157894, "grad_norm": 0.30480292439460754, "learning_rate": 6.98e-06, "loss": 0.0567, "step": 350 }, { "epoch": 3.7894736842105265, "grad_norm": 0.21191826462745667, "learning_rate": 7.180000000000001e-06, "loss": 0.0538, "step": 360 }, { "epoch": 3.8947368421052633, "grad_norm": 0.333941251039505, "learning_rate": 7.3800000000000005e-06, "loss": 0.0536, "step": 370 }, { "epoch": 4.0, "grad_norm": 0.3113187253475189, "learning_rate": 7.580000000000001e-06, "loss": 0.0529, "step": 380 }, { "epoch": 4.105263157894737, "grad_norm": 0.28991782665252686, "learning_rate": 7.78e-06, "loss": 0.0508, "step": 390 }, { "epoch": 4.2105263157894735, "grad_norm": 0.24123762547969818, "learning_rate": 7.98e-06, "loss": 0.0509, "step": 400 }, { "epoch": 4.315789473684211, "grad_norm": 0.3945069909095764, "learning_rate": 8.18e-06, "loss": 0.0487, "step": 410 }, { "epoch": 4.421052631578947, "grad_norm": 0.4084774851799011, "learning_rate": 8.380000000000001e-06, "loss": 0.0447, "step": 420 }, { "epoch": 4.526315789473684, "grad_norm": 0.28505757451057434, "learning_rate": 8.580000000000001e-06, "loss": 0.0485, "step": 430 }, { "epoch": 4.631578947368421, "grad_norm": 0.28456640243530273, "learning_rate": 8.78e-06, "loss": 0.045, "step": 440 }, { "epoch": 4.7368421052631575, "grad_norm": 0.3525456488132477, "learning_rate": 8.98e-06, "loss": 0.0447, "step": 450 }, { "epoch": 4.842105263157895, "grad_norm": 0.34427133202552795, "learning_rate": 9.180000000000002e-06, "loss": 0.0428, "step": 460 }, { "epoch": 4.947368421052632, "grad_norm": 0.2925397753715515, "learning_rate": 9.38e-06, "loss": 0.0426, "step": 470 }, { "epoch": 5.052631578947368, "grad_norm": 0.28694477677345276, "learning_rate": 9.58e-06, "loss": 0.0404, "step": 480 }, { "epoch": 5.157894736842105, "grad_norm": 0.3347450792789459, "learning_rate": 9.78e-06, "loss": 0.041, "step": 490 }, { "epoch": 5.2631578947368425, "grad_norm": 0.34103500843048096, "learning_rate": 9.980000000000001e-06, "loss": 0.0404, "step": 500 }, { "epoch": 5.368421052631579, "grad_norm": 0.360219806432724, "learning_rate": 1.018e-05, "loss": 0.0387, "step": 510 }, { "epoch": 5.473684210526316, "grad_norm": 0.29291853308677673, "learning_rate": 1.038e-05, "loss": 0.0383, "step": 520 }, { "epoch": 5.578947368421053, "grad_norm": 0.31220418214797974, "learning_rate": 1.058e-05, "loss": 0.037, "step": 530 }, { "epoch": 5.684210526315789, "grad_norm": 0.32007700204849243, "learning_rate": 1.0780000000000002e-05, "loss": 0.0361, "step": 540 }, { "epoch": 5.7894736842105265, "grad_norm": 0.2351897805929184, "learning_rate": 1.098e-05, "loss": 0.0337, "step": 550 }, { "epoch": 5.894736842105263, "grad_norm": 0.27419722080230713, "learning_rate": 1.118e-05, "loss": 0.0359, "step": 560 }, { "epoch": 6.0, "grad_norm": 0.3414967656135559, "learning_rate": 1.1380000000000001e-05, "loss": 0.0351, "step": 570 }, { "epoch": 6.105263157894737, "grad_norm": 0.3224303424358368, "learning_rate": 1.1580000000000001e-05, "loss": 0.0331, "step": 580 }, { "epoch": 6.2105263157894735, "grad_norm": 0.2965313494205475, "learning_rate": 1.178e-05, "loss": 0.0344, "step": 590 }, { "epoch": 6.315789473684211, "grad_norm": 0.3003818690776825, "learning_rate": 1.198e-05, "loss": 0.0323, "step": 600 }, { "epoch": 6.421052631578947, "grad_norm": 0.1877259761095047, "learning_rate": 1.2180000000000002e-05, "loss": 0.0336, "step": 610 }, { "epoch": 6.526315789473684, "grad_norm": 0.3015790283679962, "learning_rate": 1.238e-05, "loss": 0.0326, "step": 620 }, { "epoch": 6.631578947368421, "grad_norm": 0.3456159234046936, "learning_rate": 1.258e-05, "loss": 0.0322, "step": 630 }, { "epoch": 6.7368421052631575, "grad_norm": 0.3246683180332184, "learning_rate": 1.278e-05, "loss": 0.0318, "step": 640 }, { "epoch": 6.842105263157895, "grad_norm": 0.24540579319000244, "learning_rate": 1.2980000000000001e-05, "loss": 0.0311, "step": 650 }, { "epoch": 6.947368421052632, "grad_norm": 0.3442283570766449, "learning_rate": 1.3180000000000001e-05, "loss": 0.0303, "step": 660 }, { "epoch": 7.052631578947368, "grad_norm": 0.26126688718795776, "learning_rate": 1.338e-05, "loss": 0.0308, "step": 670 }, { "epoch": 7.157894736842105, "grad_norm": 0.28024911880493164, "learning_rate": 1.358e-05, "loss": 0.0307, "step": 680 }, { "epoch": 7.2631578947368425, "grad_norm": 0.20184384286403656, "learning_rate": 1.3780000000000002e-05, "loss": 0.0288, "step": 690 }, { "epoch": 7.368421052631579, "grad_norm": 0.31081923842430115, "learning_rate": 1.3980000000000002e-05, "loss": 0.0279, "step": 700 }, { "epoch": 7.473684210526316, "grad_norm": 0.298738956451416, "learning_rate": 1.4180000000000001e-05, "loss": 0.0293, "step": 710 }, { "epoch": 7.578947368421053, "grad_norm": 0.3353920578956604, "learning_rate": 1.4380000000000001e-05, "loss": 0.0302, "step": 720 }, { "epoch": 7.684210526315789, "grad_norm": 0.29159170389175415, "learning_rate": 1.4580000000000003e-05, "loss": 0.0286, "step": 730 }, { "epoch": 7.7894736842105265, "grad_norm": 0.390815794467926, "learning_rate": 1.4779999999999999e-05, "loss": 0.0305, "step": 740 }, { "epoch": 7.894736842105263, "grad_norm": 0.30330944061279297, "learning_rate": 1.4979999999999999e-05, "loss": 0.0271, "step": 750 }, { "epoch": 8.0, "grad_norm": 0.38654837012290955, "learning_rate": 1.518e-05, "loss": 0.0267, "step": 760 }, { "epoch": 8.105263157894736, "grad_norm": 0.26975253224372864, "learning_rate": 1.538e-05, "loss": 0.0261, "step": 770 }, { "epoch": 8.210526315789474, "grad_norm": 0.2753315567970276, "learning_rate": 1.558e-05, "loss": 0.0277, "step": 780 }, { "epoch": 8.31578947368421, "grad_norm": 0.24002665281295776, "learning_rate": 1.578e-05, "loss": 0.0271, "step": 790 }, { "epoch": 8.421052631578947, "grad_norm": 0.31318843364715576, "learning_rate": 1.598e-05, "loss": 0.0247, "step": 800 }, { "epoch": 8.526315789473685, "grad_norm": 0.26542964577674866, "learning_rate": 1.618e-05, "loss": 0.026, "step": 810 }, { "epoch": 8.631578947368421, "grad_norm": 0.3313523828983307, "learning_rate": 1.6380000000000002e-05, "loss": 0.0256, "step": 820 }, { "epoch": 8.736842105263158, "grad_norm": 0.34340304136276245, "learning_rate": 1.658e-05, "loss": 0.0276, "step": 830 }, { "epoch": 8.842105263157894, "grad_norm": 0.29216575622558594, "learning_rate": 1.6780000000000002e-05, "loss": 0.0259, "step": 840 }, { "epoch": 8.947368421052632, "grad_norm": 0.3387986123561859, "learning_rate": 1.698e-05, "loss": 0.0257, "step": 850 }, { "epoch": 9.052631578947368, "grad_norm": 0.291605681180954, "learning_rate": 1.718e-05, "loss": 0.0244, "step": 860 }, { "epoch": 9.157894736842104, "grad_norm": 0.2703651785850525, "learning_rate": 1.7380000000000003e-05, "loss": 0.0263, "step": 870 }, { "epoch": 9.263157894736842, "grad_norm": 0.22250217199325562, "learning_rate": 1.758e-05, "loss": 0.0256, "step": 880 }, { "epoch": 9.368421052631579, "grad_norm": 0.27727240324020386, "learning_rate": 1.7780000000000003e-05, "loss": 0.0252, "step": 890 }, { "epoch": 9.473684210526315, "grad_norm": 0.37269163131713867, "learning_rate": 1.798e-05, "loss": 0.0245, "step": 900 }, { "epoch": 9.578947368421053, "grad_norm": 0.32362499833106995, "learning_rate": 1.818e-05, "loss": 0.0253, "step": 910 }, { "epoch": 9.68421052631579, "grad_norm": 0.2753606140613556, "learning_rate": 1.838e-05, "loss": 0.0242, "step": 920 }, { "epoch": 9.789473684210526, "grad_norm": 0.3391495645046234, "learning_rate": 1.858e-05, "loss": 0.024, "step": 930 }, { "epoch": 9.894736842105264, "grad_norm": 0.272231787443161, "learning_rate": 1.878e-05, "loss": 0.0237, "step": 940 }, { "epoch": 10.0, "grad_norm": 0.40862220525741577, "learning_rate": 1.898e-05, "loss": 0.0242, "step": 950 }, { "epoch": 10.105263157894736, "grad_norm": 0.29129934310913086, "learning_rate": 1.918e-05, "loss": 0.0229, "step": 960 }, { "epoch": 10.210526315789474, "grad_norm": 0.27182066440582275, "learning_rate": 1.938e-05, "loss": 0.0235, "step": 970 }, { "epoch": 10.31578947368421, "grad_norm": 0.29875871539115906, "learning_rate": 1.9580000000000002e-05, "loss": 0.0225, "step": 980 }, { "epoch": 10.421052631578947, "grad_norm": 0.25254395604133606, "learning_rate": 1.978e-05, "loss": 0.0233, "step": 990 }, { "epoch": 10.526315789473685, "grad_norm": 0.25928860902786255, "learning_rate": 1.9980000000000002e-05, "loss": 0.0239, "step": 1000 }, { "epoch": 10.631578947368421, "grad_norm": 0.24945499002933502, "learning_rate": 2.0180000000000003e-05, "loss": 0.027, "step": 1010 }, { "epoch": 10.736842105263158, "grad_norm": 0.3269650340080261, "learning_rate": 2.038e-05, "loss": 0.0215, "step": 1020 }, { "epoch": 10.842105263157894, "grad_norm": 0.2274831384420395, "learning_rate": 2.0580000000000003e-05, "loss": 0.0225, "step": 1030 }, { "epoch": 10.947368421052632, "grad_norm": 0.316244512796402, "learning_rate": 2.078e-05, "loss": 0.0229, "step": 1040 }, { "epoch": 11.052631578947368, "grad_norm": 0.3142685890197754, "learning_rate": 2.098e-05, "loss": 0.0239, "step": 1050 }, { "epoch": 11.157894736842104, "grad_norm": 0.274091511964798, "learning_rate": 2.118e-05, "loss": 0.0219, "step": 1060 }, { "epoch": 11.263157894736842, "grad_norm": 0.29383450746536255, "learning_rate": 2.138e-05, "loss": 0.0216, "step": 1070 }, { "epoch": 11.368421052631579, "grad_norm": 0.2674098312854767, "learning_rate": 2.158e-05, "loss": 0.0221, "step": 1080 }, { "epoch": 11.473684210526315, "grad_norm": 0.23531341552734375, "learning_rate": 2.178e-05, "loss": 0.022, "step": 1090 }, { "epoch": 11.578947368421053, "grad_norm": 0.2695291042327881, "learning_rate": 2.198e-05, "loss": 0.0219, "step": 1100 }, { "epoch": 11.68421052631579, "grad_norm": 0.3937045633792877, "learning_rate": 2.218e-05, "loss": 0.0229, "step": 1110 }, { "epoch": 11.789473684210526, "grad_norm": 0.273936927318573, "learning_rate": 2.2380000000000003e-05, "loss": 0.0215, "step": 1120 }, { "epoch": 11.894736842105264, "grad_norm": 0.30022338032722473, "learning_rate": 2.258e-05, "loss": 0.0212, "step": 1130 }, { "epoch": 12.0, "grad_norm": 0.3432502746582031, "learning_rate": 2.2780000000000002e-05, "loss": 0.0199, "step": 1140 }, { "epoch": 12.105263157894736, "grad_norm": 0.2106420248746872, "learning_rate": 2.298e-05, "loss": 0.0198, "step": 1150 }, { "epoch": 12.210526315789474, "grad_norm": 0.25277647376060486, "learning_rate": 2.318e-05, "loss": 0.022, "step": 1160 }, { "epoch": 12.31578947368421, "grad_norm": 0.33950966596603394, "learning_rate": 2.3380000000000003e-05, "loss": 0.0206, "step": 1170 }, { "epoch": 12.421052631578947, "grad_norm": 0.30961674451828003, "learning_rate": 2.358e-05, "loss": 0.0223, "step": 1180 }, { "epoch": 12.526315789473685, "grad_norm": 0.25990673899650574, "learning_rate": 2.3780000000000003e-05, "loss": 0.0215, "step": 1190 }, { "epoch": 12.631578947368421, "grad_norm": 0.3362879455089569, "learning_rate": 2.398e-05, "loss": 0.0228, "step": 1200 }, { "epoch": 12.736842105263158, "grad_norm": 0.28244519233703613, "learning_rate": 2.418e-05, "loss": 0.0228, "step": 1210 }, { "epoch": 12.842105263157894, "grad_norm": 0.20395860075950623, "learning_rate": 2.438e-05, "loss": 0.0214, "step": 1220 }, { "epoch": 12.947368421052632, "grad_norm": 0.29801928997039795, "learning_rate": 2.4580000000000002e-05, "loss": 0.0207, "step": 1230 }, { "epoch": 13.052631578947368, "grad_norm": 0.3196837306022644, "learning_rate": 2.478e-05, "loss": 0.0205, "step": 1240 }, { "epoch": 13.157894736842104, "grad_norm": 0.35226601362228394, "learning_rate": 2.498e-05, "loss": 0.0224, "step": 1250 }, { "epoch": 13.263157894736842, "grad_norm": 0.2687443792819977, "learning_rate": 2.5180000000000003e-05, "loss": 0.0215, "step": 1260 }, { "epoch": 13.368421052631579, "grad_norm": 0.29817506670951843, "learning_rate": 2.5380000000000004e-05, "loss": 0.0205, "step": 1270 }, { "epoch": 13.473684210526315, "grad_norm": 0.271478533744812, "learning_rate": 2.5580000000000002e-05, "loss": 0.0192, "step": 1280 }, { "epoch": 13.578947368421053, "grad_norm": 0.24432797729969025, "learning_rate": 2.5779999999999997e-05, "loss": 0.0199, "step": 1290 }, { "epoch": 13.68421052631579, "grad_norm": 0.24896180629730225, "learning_rate": 2.598e-05, "loss": 0.0211, "step": 1300 }, { "epoch": 13.789473684210526, "grad_norm": 0.33776551485061646, "learning_rate": 2.618e-05, "loss": 0.0198, "step": 1310 }, { "epoch": 13.894736842105264, "grad_norm": 0.1854528784751892, "learning_rate": 2.6379999999999998e-05, "loss": 0.0193, "step": 1320 }, { "epoch": 14.0, "grad_norm": 0.27419474720954895, "learning_rate": 2.658e-05, "loss": 0.0197, "step": 1330 }, { "epoch": 14.105263157894736, "grad_norm": 0.2922714948654175, "learning_rate": 2.678e-05, "loss": 0.0191, "step": 1340 }, { "epoch": 14.210526315789474, "grad_norm": 0.2375408560037613, "learning_rate": 2.698e-05, "loss": 0.0187, "step": 1350 }, { "epoch": 14.31578947368421, "grad_norm": 0.2544850707054138, "learning_rate": 2.718e-05, "loss": 0.019, "step": 1360 }, { "epoch": 14.421052631578947, "grad_norm": 0.2569688558578491, "learning_rate": 2.738e-05, "loss": 0.019, "step": 1370 }, { "epoch": 14.526315789473685, "grad_norm": 0.24661923944950104, "learning_rate": 2.758e-05, "loss": 0.0185, "step": 1380 }, { "epoch": 14.631578947368421, "grad_norm": 0.21666888892650604, "learning_rate": 2.778e-05, "loss": 0.0197, "step": 1390 }, { "epoch": 14.736842105263158, "grad_norm": 0.3165898621082306, "learning_rate": 2.798e-05, "loss": 0.0197, "step": 1400 }, { "epoch": 14.842105263157894, "grad_norm": 0.3912632465362549, "learning_rate": 2.818e-05, "loss": 0.0214, "step": 1410 }, { "epoch": 14.947368421052632, "grad_norm": 0.3478437066078186, "learning_rate": 2.8380000000000003e-05, "loss": 0.0195, "step": 1420 }, { "epoch": 15.052631578947368, "grad_norm": 0.2942185401916504, "learning_rate": 2.858e-05, "loss": 0.0188, "step": 1430 }, { "epoch": 15.157894736842104, "grad_norm": 0.38490423560142517, "learning_rate": 2.8780000000000002e-05, "loss": 0.0179, "step": 1440 }, { "epoch": 15.263157894736842, "grad_norm": 0.37049296498298645, "learning_rate": 2.898e-05, "loss": 0.0197, "step": 1450 }, { "epoch": 15.368421052631579, "grad_norm": 0.378836989402771, "learning_rate": 2.9180000000000002e-05, "loss": 0.0194, "step": 1460 }, { "epoch": 15.473684210526315, "grad_norm": 0.3994239866733551, "learning_rate": 2.9380000000000003e-05, "loss": 0.019, "step": 1470 }, { "epoch": 15.578947368421053, "grad_norm": 0.2906571924686432, "learning_rate": 2.958e-05, "loss": 0.0187, "step": 1480 }, { "epoch": 15.68421052631579, "grad_norm": 0.2748374938964844, "learning_rate": 2.9780000000000003e-05, "loss": 0.0183, "step": 1490 }, { "epoch": 15.789473684210526, "grad_norm": 0.2856259047985077, "learning_rate": 2.998e-05, "loss": 0.0194, "step": 1500 }, { "epoch": 15.894736842105264, "grad_norm": 0.21892808377742767, "learning_rate": 3.0180000000000002e-05, "loss": 0.0181, "step": 1510 }, { "epoch": 16.0, "grad_norm": 0.2924973964691162, "learning_rate": 3.0380000000000004e-05, "loss": 0.0186, "step": 1520 }, { "epoch": 16.105263157894736, "grad_norm": 0.2638249099254608, "learning_rate": 3.058e-05, "loss": 0.0175, "step": 1530 }, { "epoch": 16.210526315789473, "grad_norm": 0.38046738505363464, "learning_rate": 3.078e-05, "loss": 0.0206, "step": 1540 }, { "epoch": 16.31578947368421, "grad_norm": 0.3159891366958618, "learning_rate": 3.0980000000000005e-05, "loss": 0.0186, "step": 1550 }, { "epoch": 16.42105263157895, "grad_norm": 0.24289682507514954, "learning_rate": 3.118e-05, "loss": 0.0177, "step": 1560 }, { "epoch": 16.526315789473685, "grad_norm": 0.2969236671924591, "learning_rate": 3.138e-05, "loss": 0.0188, "step": 1570 }, { "epoch": 16.63157894736842, "grad_norm": 0.22618618607521057, "learning_rate": 3.1580000000000006e-05, "loss": 0.0173, "step": 1580 }, { "epoch": 16.736842105263158, "grad_norm": 0.27922213077545166, "learning_rate": 3.1780000000000004e-05, "loss": 0.0178, "step": 1590 }, { "epoch": 16.842105263157894, "grad_norm": 0.23072963953018188, "learning_rate": 3.198e-05, "loss": 0.0183, "step": 1600 }, { "epoch": 16.94736842105263, "grad_norm": 0.2684817314147949, "learning_rate": 3.218e-05, "loss": 0.0175, "step": 1610 }, { "epoch": 17.05263157894737, "grad_norm": 0.2677648663520813, "learning_rate": 3.238e-05, "loss": 0.0185, "step": 1620 }, { "epoch": 17.157894736842106, "grad_norm": 0.4198419153690338, "learning_rate": 3.2579999999999996e-05, "loss": 0.0172, "step": 1630 }, { "epoch": 17.263157894736842, "grad_norm": 0.21918287873268127, "learning_rate": 3.278e-05, "loss": 0.0187, "step": 1640 }, { "epoch": 17.36842105263158, "grad_norm": 0.23956555128097534, "learning_rate": 3.298e-05, "loss": 0.0173, "step": 1650 }, { "epoch": 17.473684210526315, "grad_norm": 0.2776199281215668, "learning_rate": 3.318e-05, "loss": 0.0169, "step": 1660 }, { "epoch": 17.57894736842105, "grad_norm": 0.3476629853248596, "learning_rate": 3.338e-05, "loss": 0.0192, "step": 1670 }, { "epoch": 17.68421052631579, "grad_norm": 0.322711318731308, "learning_rate": 3.358e-05, "loss": 0.019, "step": 1680 }, { "epoch": 17.789473684210527, "grad_norm": 0.37305283546447754, "learning_rate": 3.378e-05, "loss": 0.0173, "step": 1690 }, { "epoch": 17.894736842105264, "grad_norm": 0.2622014582157135, "learning_rate": 3.398e-05, "loss": 0.0195, "step": 1700 }, { "epoch": 18.0, "grad_norm": 0.3414655029773712, "learning_rate": 3.418e-05, "loss": 0.0185, "step": 1710 }, { "epoch": 18.105263157894736, "grad_norm": 0.24055714905261993, "learning_rate": 3.438e-05, "loss": 0.018, "step": 1720 }, { "epoch": 18.210526315789473, "grad_norm": 0.2899917662143707, "learning_rate": 3.4580000000000004e-05, "loss": 0.017, "step": 1730 }, { "epoch": 18.31578947368421, "grad_norm": 0.3021065592765808, "learning_rate": 3.478e-05, "loss": 0.0189, "step": 1740 }, { "epoch": 18.42105263157895, "grad_norm": 0.2968493700027466, "learning_rate": 3.498e-05, "loss": 0.0176, "step": 1750 }, { "epoch": 18.526315789473685, "grad_norm": 0.23692794144153595, "learning_rate": 3.518e-05, "loss": 0.0173, "step": 1760 }, { "epoch": 18.63157894736842, "grad_norm": 0.29372721910476685, "learning_rate": 3.5380000000000003e-05, "loss": 0.0162, "step": 1770 }, { "epoch": 18.736842105263158, "grad_norm": 0.3291381001472473, "learning_rate": 3.558e-05, "loss": 0.0169, "step": 1780 }, { "epoch": 18.842105263157894, "grad_norm": 0.3133411407470703, "learning_rate": 3.578e-05, "loss": 0.017, "step": 1790 }, { "epoch": 18.94736842105263, "grad_norm": 0.2656151056289673, "learning_rate": 3.5980000000000004e-05, "loss": 0.0159, "step": 1800 }, { "epoch": 19.05263157894737, "grad_norm": 0.28110626339912415, "learning_rate": 3.618e-05, "loss": 0.0178, "step": 1810 }, { "epoch": 19.157894736842106, "grad_norm": 0.2630675435066223, "learning_rate": 3.638e-05, "loss": 0.0173, "step": 1820 }, { "epoch": 19.263157894736842, "grad_norm": 0.22150804102420807, "learning_rate": 3.6580000000000006e-05, "loss": 0.0164, "step": 1830 }, { "epoch": 19.36842105263158, "grad_norm": 0.27160438895225525, "learning_rate": 3.6780000000000004e-05, "loss": 0.0171, "step": 1840 }, { "epoch": 19.473684210526315, "grad_norm": 0.25728437304496765, "learning_rate": 3.698e-05, "loss": 0.0166, "step": 1850 }, { "epoch": 19.57894736842105, "grad_norm": 0.2922554910182953, "learning_rate": 3.7180000000000007e-05, "loss": 0.0195, "step": 1860 }, { "epoch": 19.68421052631579, "grad_norm": 0.3068928122520447, "learning_rate": 3.7380000000000005e-05, "loss": 0.0157, "step": 1870 }, { "epoch": 19.789473684210527, "grad_norm": 0.26313072443008423, "learning_rate": 3.758e-05, "loss": 0.0166, "step": 1880 }, { "epoch": 19.894736842105264, "grad_norm": 0.27856892347335815, "learning_rate": 3.778000000000001e-05, "loss": 0.0157, "step": 1890 }, { "epoch": 20.0, "grad_norm": 0.33739784359931946, "learning_rate": 3.7980000000000006e-05, "loss": 0.016, "step": 1900 }, { "epoch": 20.105263157894736, "grad_norm": 0.24668458104133606, "learning_rate": 3.818e-05, "loss": 0.0159, "step": 1910 }, { "epoch": 20.210526315789473, "grad_norm": 0.2568504214286804, "learning_rate": 3.838e-05, "loss": 0.0168, "step": 1920 }, { "epoch": 20.31578947368421, "grad_norm": 0.31129711866378784, "learning_rate": 3.858e-05, "loss": 0.0181, "step": 1930 }, { "epoch": 20.42105263157895, "grad_norm": 0.3404821753501892, "learning_rate": 3.878e-05, "loss": 0.0172, "step": 1940 }, { "epoch": 20.526315789473685, "grad_norm": 0.20277900993824005, "learning_rate": 3.898e-05, "loss": 0.0157, "step": 1950 }, { "epoch": 20.63157894736842, "grad_norm": 0.20442187786102295, "learning_rate": 3.918e-05, "loss": 0.0168, "step": 1960 }, { "epoch": 20.736842105263158, "grad_norm": 0.2047983556985855, "learning_rate": 3.938e-05, "loss": 0.0153, "step": 1970 }, { "epoch": 20.842105263157894, "grad_norm": 0.21016128361225128, "learning_rate": 3.958e-05, "loss": 0.0158, "step": 1980 }, { "epoch": 20.94736842105263, "grad_norm": 0.3434202969074249, "learning_rate": 3.978e-05, "loss": 0.0162, "step": 1990 }, { "epoch": 21.05263157894737, "grad_norm": 0.24346482753753662, "learning_rate": 3.998e-05, "loss": 0.0169, "step": 2000 }, { "epoch": 21.157894736842106, "grad_norm": 0.3217323124408722, "learning_rate": 4.018e-05, "loss": 0.018, "step": 2010 }, { "epoch": 21.263157894736842, "grad_norm": 0.2823413908481598, "learning_rate": 4.038e-05, "loss": 0.0153, "step": 2020 }, { "epoch": 21.36842105263158, "grad_norm": 0.26961958408355713, "learning_rate": 4.058e-05, "loss": 0.0158, "step": 2030 }, { "epoch": 21.473684210526315, "grad_norm": 0.2138289362192154, "learning_rate": 4.078e-05, "loss": 0.0161, "step": 2040 }, { "epoch": 21.57894736842105, "grad_norm": 0.21254433691501617, "learning_rate": 4.0980000000000004e-05, "loss": 0.0149, "step": 2050 }, { "epoch": 21.68421052631579, "grad_norm": 0.2876909375190735, "learning_rate": 4.118e-05, "loss": 0.0161, "step": 2060 }, { "epoch": 21.789473684210527, "grad_norm": 0.21695412695407867, "learning_rate": 4.138e-05, "loss": 0.0148, "step": 2070 }, { "epoch": 21.894736842105264, "grad_norm": 0.263871431350708, "learning_rate": 4.1580000000000005e-05, "loss": 0.0153, "step": 2080 }, { "epoch": 22.0, "grad_norm": 0.3634055554866791, "learning_rate": 4.178e-05, "loss": 0.0152, "step": 2090 }, { "epoch": 22.105263157894736, "grad_norm": 0.24590721726417542, "learning_rate": 4.198e-05, "loss": 0.0153, "step": 2100 }, { "epoch": 22.210526315789473, "grad_norm": 0.28125500679016113, "learning_rate": 4.2180000000000006e-05, "loss": 0.0145, "step": 2110 }, { "epoch": 22.31578947368421, "grad_norm": 0.23269900679588318, "learning_rate": 4.2380000000000004e-05, "loss": 0.0165, "step": 2120 }, { "epoch": 22.42105263157895, "grad_norm": 0.2539200782775879, "learning_rate": 4.258e-05, "loss": 0.0158, "step": 2130 }, { "epoch": 22.526315789473685, "grad_norm": 0.26024892926216125, "learning_rate": 4.278e-05, "loss": 0.016, "step": 2140 }, { "epoch": 22.63157894736842, "grad_norm": 0.31408730149269104, "learning_rate": 4.2980000000000005e-05, "loss": 0.0162, "step": 2150 }, { "epoch": 22.736842105263158, "grad_norm": 0.2947801351547241, "learning_rate": 4.318e-05, "loss": 0.0164, "step": 2160 }, { "epoch": 22.842105263157894, "grad_norm": 0.25068870186805725, "learning_rate": 4.338e-05, "loss": 0.0148, "step": 2170 }, { "epoch": 22.94736842105263, "grad_norm": 0.2528790533542633, "learning_rate": 4.3580000000000006e-05, "loss": 0.017, "step": 2180 }, { "epoch": 23.05263157894737, "grad_norm": 0.23629875481128693, "learning_rate": 4.3780000000000004e-05, "loss": 0.0156, "step": 2190 }, { "epoch": 23.157894736842106, "grad_norm": 0.2674585282802582, "learning_rate": 4.398e-05, "loss": 0.0152, "step": 2200 }, { "epoch": 23.263157894736842, "grad_norm": 0.24131807684898376, "learning_rate": 4.418000000000001e-05, "loss": 0.0157, "step": 2210 }, { "epoch": 23.36842105263158, "grad_norm": 0.20623131096363068, "learning_rate": 4.438e-05, "loss": 0.0145, "step": 2220 }, { "epoch": 23.473684210526315, "grad_norm": 0.2313234806060791, "learning_rate": 4.458e-05, "loss": 0.0166, "step": 2230 }, { "epoch": 23.57894736842105, "grad_norm": 0.2362811267375946, "learning_rate": 4.478e-05, "loss": 0.0157, "step": 2240 }, { "epoch": 23.68421052631579, "grad_norm": 0.24219168722629547, "learning_rate": 4.498e-05, "loss": 0.0162, "step": 2250 }, { "epoch": 23.789473684210527, "grad_norm": 0.21615122258663177, "learning_rate": 4.518e-05, "loss": 0.0157, "step": 2260 }, { "epoch": 23.894736842105264, "grad_norm": 0.2494129091501236, "learning_rate": 4.538e-05, "loss": 0.0172, "step": 2270 }, { "epoch": 24.0, "grad_norm": 0.2897922694683075, "learning_rate": 4.558e-05, "loss": 0.0164, "step": 2280 }, { "epoch": 24.105263157894736, "grad_norm": 0.2386951595544815, "learning_rate": 4.578e-05, "loss": 0.015, "step": 2290 }, { "epoch": 24.210526315789473, "grad_norm": 0.2254834920167923, "learning_rate": 4.5980000000000004e-05, "loss": 0.015, "step": 2300 }, { "epoch": 24.31578947368421, "grad_norm": 0.31651771068573, "learning_rate": 4.618e-05, "loss": 0.0148, "step": 2310 }, { "epoch": 24.42105263157895, "grad_norm": 0.24229343235492706, "learning_rate": 4.638e-05, "loss": 0.0155, "step": 2320 }, { "epoch": 24.526315789473685, "grad_norm": 0.20597727596759796, "learning_rate": 4.6580000000000005e-05, "loss": 0.0144, "step": 2330 }, { "epoch": 24.63157894736842, "grad_norm": 0.19301161170005798, "learning_rate": 4.678e-05, "loss": 0.0159, "step": 2340 }, { "epoch": 24.736842105263158, "grad_norm": 0.25149571895599365, "learning_rate": 4.698e-05, "loss": 0.0143, "step": 2350 }, { "epoch": 24.842105263157894, "grad_norm": 0.22302471101284027, "learning_rate": 4.718e-05, "loss": 0.0159, "step": 2360 }, { "epoch": 24.94736842105263, "grad_norm": 0.2506723701953888, "learning_rate": 4.7380000000000004e-05, "loss": 0.0148, "step": 2370 }, { "epoch": 25.05263157894737, "grad_norm": 0.22094213962554932, "learning_rate": 4.758e-05, "loss": 0.0138, "step": 2380 }, { "epoch": 25.157894736842106, "grad_norm": 0.25187990069389343, "learning_rate": 4.778e-05, "loss": 0.0147, "step": 2390 }, { "epoch": 25.263157894736842, "grad_norm": 0.21639500558376312, "learning_rate": 4.7980000000000005e-05, "loss": 0.0138, "step": 2400 }, { "epoch": 25.36842105263158, "grad_norm": 0.26217934489250183, "learning_rate": 4.818e-05, "loss": 0.0141, "step": 2410 }, { "epoch": 25.473684210526315, "grad_norm": 0.23881463706493378, "learning_rate": 4.838e-05, "loss": 0.0142, "step": 2420 }, { "epoch": 25.57894736842105, "grad_norm": 0.2230776846408844, "learning_rate": 4.8580000000000006e-05, "loss": 0.0146, "step": 2430 }, { "epoch": 25.68421052631579, "grad_norm": 0.18248969316482544, "learning_rate": 4.8780000000000004e-05, "loss": 0.0134, "step": 2440 }, { "epoch": 25.789473684210527, "grad_norm": 0.229995995759964, "learning_rate": 4.898e-05, "loss": 0.014, "step": 2450 }, { "epoch": 25.894736842105264, "grad_norm": 0.2897026538848877, "learning_rate": 4.918000000000001e-05, "loss": 0.0151, "step": 2460 }, { "epoch": 26.0, "grad_norm": 0.3858569860458374, "learning_rate": 4.9380000000000005e-05, "loss": 0.0142, "step": 2470 }, { "epoch": 26.105263157894736, "grad_norm": 0.22265145182609558, "learning_rate": 4.958e-05, "loss": 0.0137, "step": 2480 }, { "epoch": 26.210526315789473, "grad_norm": 0.2364322394132614, "learning_rate": 4.978e-05, "loss": 0.0151, "step": 2490 }, { "epoch": 26.31578947368421, "grad_norm": 0.19527968764305115, "learning_rate": 4.9980000000000006e-05, "loss": 0.016, "step": 2500 }, { "epoch": 26.42105263157895, "grad_norm": 0.2821948230266571, "learning_rate": 5.0180000000000004e-05, "loss": 0.0141, "step": 2510 }, { "epoch": 26.526315789473685, "grad_norm": 0.2636605203151703, "learning_rate": 5.038e-05, "loss": 0.0142, "step": 2520 }, { "epoch": 26.63157894736842, "grad_norm": 0.2681495249271393, "learning_rate": 5.058000000000001e-05, "loss": 0.0131, "step": 2530 }, { "epoch": 26.736842105263158, "grad_norm": 0.24584391713142395, "learning_rate": 5.0780000000000005e-05, "loss": 0.0138, "step": 2540 }, { "epoch": 26.842105263157894, "grad_norm": 0.2135368436574936, "learning_rate": 5.098e-05, "loss": 0.0143, "step": 2550 }, { "epoch": 26.94736842105263, "grad_norm": 0.24466530978679657, "learning_rate": 5.118000000000001e-05, "loss": 0.0134, "step": 2560 }, { "epoch": 27.05263157894737, "grad_norm": 0.2009442001581192, "learning_rate": 5.1380000000000006e-05, "loss": 0.0131, "step": 2570 }, { "epoch": 27.157894736842106, "grad_norm": 0.18206191062927246, "learning_rate": 5.1580000000000004e-05, "loss": 0.0136, "step": 2580 }, { "epoch": 27.263157894736842, "grad_norm": 0.24507564306259155, "learning_rate": 5.178000000000001e-05, "loss": 0.0133, "step": 2590 }, { "epoch": 27.36842105263158, "grad_norm": 0.27838847041130066, "learning_rate": 5.198000000000001e-05, "loss": 0.0132, "step": 2600 }, { "epoch": 27.473684210526315, "grad_norm": 0.27153679728507996, "learning_rate": 5.2180000000000005e-05, "loss": 0.0139, "step": 2610 }, { "epoch": 27.57894736842105, "grad_norm": 0.1971493810415268, "learning_rate": 5.238000000000001e-05, "loss": 0.0137, "step": 2620 }, { "epoch": 27.68421052631579, "grad_norm": 0.2915213704109192, "learning_rate": 5.258000000000001e-05, "loss": 0.0127, "step": 2630 }, { "epoch": 27.789473684210527, "grad_norm": 0.24552352726459503, "learning_rate": 5.2780000000000006e-05, "loss": 0.0138, "step": 2640 }, { "epoch": 27.894736842105264, "grad_norm": 0.26575133204460144, "learning_rate": 5.2980000000000004e-05, "loss": 0.0135, "step": 2650 }, { "epoch": 28.0, "grad_norm": 0.24953310191631317, "learning_rate": 5.318000000000001e-05, "loss": 0.0138, "step": 2660 }, { "epoch": 28.105263157894736, "grad_norm": 0.2487332969903946, "learning_rate": 5.338000000000001e-05, "loss": 0.0134, "step": 2670 }, { "epoch": 28.210526315789473, "grad_norm": 0.23603233695030212, "learning_rate": 5.3580000000000005e-05, "loss": 0.0144, "step": 2680 }, { "epoch": 28.31578947368421, "grad_norm": 0.2551063597202301, "learning_rate": 5.378e-05, "loss": 0.0135, "step": 2690 }, { "epoch": 28.42105263157895, "grad_norm": 0.278890460729599, "learning_rate": 5.3979999999999995e-05, "loss": 0.0141, "step": 2700 }, { "epoch": 28.526315789473685, "grad_norm": 0.263301819562912, "learning_rate": 5.418e-05, "loss": 0.0137, "step": 2710 }, { "epoch": 28.63157894736842, "grad_norm": 0.2546515166759491, "learning_rate": 5.438e-05, "loss": 0.0123, "step": 2720 }, { "epoch": 28.736842105263158, "grad_norm": 0.24411241710186005, "learning_rate": 5.4579999999999996e-05, "loss": 0.0126, "step": 2730 }, { "epoch": 28.842105263157894, "grad_norm": 0.25948888063430786, "learning_rate": 5.478e-05, "loss": 0.0133, "step": 2740 }, { "epoch": 28.94736842105263, "grad_norm": 0.1948682814836502, "learning_rate": 5.498e-05, "loss": 0.0129, "step": 2750 }, { "epoch": 29.05263157894737, "grad_norm": 0.18054303526878357, "learning_rate": 5.518e-05, "loss": 0.0136, "step": 2760 }, { "epoch": 29.157894736842106, "grad_norm": 0.30704236030578613, "learning_rate": 5.538e-05, "loss": 0.0142, "step": 2770 }, { "epoch": 29.263157894736842, "grad_norm": 0.23288966715335846, "learning_rate": 5.558e-05, "loss": 0.0134, "step": 2780 }, { "epoch": 29.36842105263158, "grad_norm": 0.2203456610441208, "learning_rate": 5.578e-05, "loss": 0.0136, "step": 2790 }, { "epoch": 29.473684210526315, "grad_norm": 0.2856433391571045, "learning_rate": 5.5979999999999996e-05, "loss": 0.0129, "step": 2800 }, { "epoch": 29.57894736842105, "grad_norm": 0.2155352383852005, "learning_rate": 5.618e-05, "loss": 0.0131, "step": 2810 }, { "epoch": 29.68421052631579, "grad_norm": 0.21510446071624756, "learning_rate": 5.638e-05, "loss": 0.0142, "step": 2820 }, { "epoch": 29.789473684210527, "grad_norm": 0.19845904409885406, "learning_rate": 5.658e-05, "loss": 0.0138, "step": 2830 }, { "epoch": 29.894736842105264, "grad_norm": 0.17803670465946198, "learning_rate": 5.678e-05, "loss": 0.0122, "step": 2840 }, { "epoch": 30.0, "grad_norm": 0.20965130627155304, "learning_rate": 5.698e-05, "loss": 0.0113, "step": 2850 }, { "epoch": 30.105263157894736, "grad_norm": 0.1633261889219284, "learning_rate": 5.718e-05, "loss": 0.0135, "step": 2860 }, { "epoch": 30.210526315789473, "grad_norm": 0.2481870800256729, "learning_rate": 5.738e-05, "loss": 0.0139, "step": 2870 }, { "epoch": 30.31578947368421, "grad_norm": 0.19258445501327515, "learning_rate": 5.758e-05, "loss": 0.0128, "step": 2880 }, { "epoch": 30.42105263157895, "grad_norm": 0.22927923500537872, "learning_rate": 5.778e-05, "loss": 0.0131, "step": 2890 }, { "epoch": 30.526315789473685, "grad_norm": 0.23867081105709076, "learning_rate": 5.7980000000000004e-05, "loss": 0.0127, "step": 2900 }, { "epoch": 30.63157894736842, "grad_norm": 0.23130963742733002, "learning_rate": 5.818e-05, "loss": 0.0145, "step": 2910 }, { "epoch": 30.736842105263158, "grad_norm": 0.20021697878837585, "learning_rate": 5.838e-05, "loss": 0.0134, "step": 2920 }, { "epoch": 30.842105263157894, "grad_norm": 0.19835348427295685, "learning_rate": 5.858e-05, "loss": 0.0151, "step": 2930 }, { "epoch": 30.94736842105263, "grad_norm": 0.25451764464378357, "learning_rate": 5.878e-05, "loss": 0.0127, "step": 2940 }, { "epoch": 31.05263157894737, "grad_norm": 0.20969824492931366, "learning_rate": 5.898e-05, "loss": 0.0134, "step": 2950 }, { "epoch": 31.157894736842106, "grad_norm": 0.2281201183795929, "learning_rate": 5.918e-05, "loss": 0.0131, "step": 2960 }, { "epoch": 31.263157894736842, "grad_norm": 0.26159822940826416, "learning_rate": 5.9380000000000004e-05, "loss": 0.012, "step": 2970 }, { "epoch": 31.36842105263158, "grad_norm": 0.3159657120704651, "learning_rate": 5.958e-05, "loss": 0.0137, "step": 2980 }, { "epoch": 31.473684210526315, "grad_norm": 0.1764010637998581, "learning_rate": 5.978e-05, "loss": 0.0129, "step": 2990 }, { "epoch": 31.57894736842105, "grad_norm": 0.2075349986553192, "learning_rate": 5.9980000000000005e-05, "loss": 0.0132, "step": 3000 }, { "epoch": 31.68421052631579, "grad_norm": 0.22571231424808502, "learning_rate": 6.018e-05, "loss": 0.0144, "step": 3010 }, { "epoch": 31.789473684210527, "grad_norm": 0.20289911329746246, "learning_rate": 6.038e-05, "loss": 0.0125, "step": 3020 }, { "epoch": 31.894736842105264, "grad_norm": 0.21673227846622467, "learning_rate": 6.0580000000000006e-05, "loss": 0.013, "step": 3030 }, { "epoch": 32.0, "grad_norm": 0.2718662619590759, "learning_rate": 6.0780000000000004e-05, "loss": 0.0123, "step": 3040 }, { "epoch": 32.10526315789474, "grad_norm": 0.2036024034023285, "learning_rate": 6.098e-05, "loss": 0.0127, "step": 3050 }, { "epoch": 32.21052631578947, "grad_norm": 0.2117195874452591, "learning_rate": 6.118000000000001e-05, "loss": 0.0122, "step": 3060 }, { "epoch": 32.31578947368421, "grad_norm": 0.24152909219264984, "learning_rate": 6.138e-05, "loss": 0.0125, "step": 3070 }, { "epoch": 32.421052631578945, "grad_norm": 0.21338534355163574, "learning_rate": 6.158e-05, "loss": 0.0132, "step": 3080 }, { "epoch": 32.526315789473685, "grad_norm": 0.26010361313819885, "learning_rate": 6.178000000000001e-05, "loss": 0.0127, "step": 3090 }, { "epoch": 32.63157894736842, "grad_norm": 0.2264317274093628, "learning_rate": 6.198e-05, "loss": 0.0131, "step": 3100 }, { "epoch": 32.73684210526316, "grad_norm": 0.2858884632587433, "learning_rate": 6.218e-05, "loss": 0.013, "step": 3110 }, { "epoch": 32.8421052631579, "grad_norm": 0.1834568828344345, "learning_rate": 6.238000000000001e-05, "loss": 0.0115, "step": 3120 }, { "epoch": 32.94736842105263, "grad_norm": 0.21643531322479248, "learning_rate": 6.258e-05, "loss": 0.0125, "step": 3130 }, { "epoch": 33.05263157894737, "grad_norm": 0.21884675323963165, "learning_rate": 6.278e-05, "loss": 0.0116, "step": 3140 }, { "epoch": 33.1578947368421, "grad_norm": 0.18483570218086243, "learning_rate": 6.298000000000001e-05, "loss": 0.0131, "step": 3150 }, { "epoch": 33.26315789473684, "grad_norm": 0.20479287207126617, "learning_rate": 6.318e-05, "loss": 0.0117, "step": 3160 }, { "epoch": 33.36842105263158, "grad_norm": 0.24097992479801178, "learning_rate": 6.338e-05, "loss": 0.0137, "step": 3170 }, { "epoch": 33.473684210526315, "grad_norm": 0.21518687903881073, "learning_rate": 6.358000000000001e-05, "loss": 0.0109, "step": 3180 }, { "epoch": 33.578947368421055, "grad_norm": 0.20799541473388672, "learning_rate": 6.378e-05, "loss": 0.0122, "step": 3190 }, { "epoch": 33.68421052631579, "grad_norm": 0.2067592442035675, "learning_rate": 6.398000000000001e-05, "loss": 0.0121, "step": 3200 }, { "epoch": 33.78947368421053, "grad_norm": 0.1966487616300583, "learning_rate": 6.418000000000001e-05, "loss": 0.0129, "step": 3210 }, { "epoch": 33.89473684210526, "grad_norm": 0.2280348837375641, "learning_rate": 6.438e-05, "loss": 0.014, "step": 3220 }, { "epoch": 34.0, "grad_norm": 0.3025597333908081, "learning_rate": 6.458000000000001e-05, "loss": 0.0129, "step": 3230 }, { "epoch": 34.10526315789474, "grad_norm": 0.19219127297401428, "learning_rate": 6.478000000000001e-05, "loss": 0.013, "step": 3240 }, { "epoch": 34.21052631578947, "grad_norm": 0.2089925855398178, "learning_rate": 6.498e-05, "loss": 0.0123, "step": 3250 }, { "epoch": 34.31578947368421, "grad_norm": 0.24766287207603455, "learning_rate": 6.518000000000001e-05, "loss": 0.0125, "step": 3260 }, { "epoch": 34.421052631578945, "grad_norm": 0.25073468685150146, "learning_rate": 6.538000000000001e-05, "loss": 0.0116, "step": 3270 }, { "epoch": 34.526315789473685, "grad_norm": 0.2205936312675476, "learning_rate": 6.558e-05, "loss": 0.0129, "step": 3280 }, { "epoch": 34.63157894736842, "grad_norm": 0.14812502264976501, "learning_rate": 6.578000000000001e-05, "loss": 0.0123, "step": 3290 }, { "epoch": 34.73684210526316, "grad_norm": 0.2422351986169815, "learning_rate": 6.598e-05, "loss": 0.0123, "step": 3300 }, { "epoch": 34.8421052631579, "grad_norm": 0.2280186414718628, "learning_rate": 6.618e-05, "loss": 0.0118, "step": 3310 }, { "epoch": 34.94736842105263, "grad_norm": 0.16804343461990356, "learning_rate": 6.638e-05, "loss": 0.0115, "step": 3320 }, { "epoch": 35.05263157894737, "grad_norm": 0.22532235085964203, "learning_rate": 6.658e-05, "loss": 0.0118, "step": 3330 }, { "epoch": 35.1578947368421, "grad_norm": 0.20278091728687286, "learning_rate": 6.678e-05, "loss": 0.0127, "step": 3340 }, { "epoch": 35.26315789473684, "grad_norm": 0.2387356162071228, "learning_rate": 6.698e-05, "loss": 0.0122, "step": 3350 }, { "epoch": 35.36842105263158, "grad_norm": 0.16018961369991302, "learning_rate": 6.718e-05, "loss": 0.0122, "step": 3360 }, { "epoch": 35.473684210526315, "grad_norm": 0.22274169325828552, "learning_rate": 6.738e-05, "loss": 0.0126, "step": 3370 }, { "epoch": 35.578947368421055, "grad_norm": 0.24608154594898224, "learning_rate": 6.758e-05, "loss": 0.0124, "step": 3380 }, { "epoch": 35.68421052631579, "grad_norm": 0.1784745156764984, "learning_rate": 6.778e-05, "loss": 0.0125, "step": 3390 }, { "epoch": 35.78947368421053, "grad_norm": 0.24177123606204987, "learning_rate": 6.798e-05, "loss": 0.0127, "step": 3400 }, { "epoch": 35.89473684210526, "grad_norm": 0.19756901264190674, "learning_rate": 6.818e-05, "loss": 0.0119, "step": 3410 }, { "epoch": 36.0, "grad_norm": 0.24723011255264282, "learning_rate": 6.838e-05, "loss": 0.0131, "step": 3420 }, { "epoch": 36.10526315789474, "grad_norm": 0.18414437770843506, "learning_rate": 6.858e-05, "loss": 0.0121, "step": 3430 }, { "epoch": 36.21052631578947, "grad_norm": 0.18900494277477264, "learning_rate": 6.878e-05, "loss": 0.0123, "step": 3440 }, { "epoch": 36.31578947368421, "grad_norm": 0.25005725026130676, "learning_rate": 6.898e-05, "loss": 0.013, "step": 3450 }, { "epoch": 36.421052631578945, "grad_norm": 0.21239832043647766, "learning_rate": 6.918e-05, "loss": 0.013, "step": 3460 }, { "epoch": 36.526315789473685, "grad_norm": 0.23622478544712067, "learning_rate": 6.938e-05, "loss": 0.0125, "step": 3470 }, { "epoch": 36.63157894736842, "grad_norm": 0.19594675302505493, "learning_rate": 6.958e-05, "loss": 0.011, "step": 3480 }, { "epoch": 36.73684210526316, "grad_norm": 0.22064080834388733, "learning_rate": 6.978e-05, "loss": 0.0117, "step": 3490 }, { "epoch": 36.8421052631579, "grad_norm": 0.19189271330833435, "learning_rate": 6.998e-05, "loss": 0.0123, "step": 3500 }, { "epoch": 36.94736842105263, "grad_norm": 0.16470250487327576, "learning_rate": 7.018e-05, "loss": 0.0114, "step": 3510 }, { "epoch": 37.05263157894737, "grad_norm": 0.1845826655626297, "learning_rate": 7.038e-05, "loss": 0.0129, "step": 3520 }, { "epoch": 37.1578947368421, "grad_norm": 0.1780155748128891, "learning_rate": 7.058e-05, "loss": 0.0114, "step": 3530 }, { "epoch": 37.26315789473684, "grad_norm": 0.17184217274188995, "learning_rate": 7.078e-05, "loss": 0.0112, "step": 3540 }, { "epoch": 37.36842105263158, "grad_norm": 0.17569538950920105, "learning_rate": 7.098e-05, "loss": 0.0094, "step": 3550 }, { "epoch": 37.473684210526315, "grad_norm": 0.16804102063179016, "learning_rate": 7.118e-05, "loss": 0.0121, "step": 3560 }, { "epoch": 37.578947368421055, "grad_norm": 0.17033886909484863, "learning_rate": 7.138e-05, "loss": 0.0111, "step": 3570 }, { "epoch": 37.68421052631579, "grad_norm": 0.17091251909732819, "learning_rate": 7.158e-05, "loss": 0.0116, "step": 3580 }, { "epoch": 37.78947368421053, "grad_norm": 0.1946047991514206, "learning_rate": 7.178000000000001e-05, "loss": 0.0126, "step": 3590 }, { "epoch": 37.89473684210526, "grad_norm": 0.2501821517944336, "learning_rate": 7.198e-05, "loss": 0.0106, "step": 3600 }, { "epoch": 38.0, "grad_norm": 0.18594461679458618, "learning_rate": 7.218e-05, "loss": 0.0119, "step": 3610 }, { "epoch": 38.10526315789474, "grad_norm": 0.21163001656532288, "learning_rate": 7.238000000000001e-05, "loss": 0.0113, "step": 3620 }, { "epoch": 38.21052631578947, "grad_norm": 0.24333113431930542, "learning_rate": 7.258e-05, "loss": 0.0114, "step": 3630 }, { "epoch": 38.31578947368421, "grad_norm": 0.2021602839231491, "learning_rate": 7.278e-05, "loss": 0.0117, "step": 3640 }, { "epoch": 38.421052631578945, "grad_norm": 0.19318395853042603, "learning_rate": 7.298000000000001e-05, "loss": 0.0111, "step": 3650 }, { "epoch": 38.526315789473685, "grad_norm": 0.176956444978714, "learning_rate": 7.318e-05, "loss": 0.0112, "step": 3660 }, { "epoch": 38.63157894736842, "grad_norm": 0.24039044976234436, "learning_rate": 7.338e-05, "loss": 0.0117, "step": 3670 }, { "epoch": 38.73684210526316, "grad_norm": 0.2641764283180237, "learning_rate": 7.358000000000001e-05, "loss": 0.0119, "step": 3680 }, { "epoch": 38.8421052631579, "grad_norm": 0.17854046821594238, "learning_rate": 7.378e-05, "loss": 0.0109, "step": 3690 }, { "epoch": 38.94736842105263, "grad_norm": 0.23439641296863556, "learning_rate": 7.398e-05, "loss": 0.0114, "step": 3700 }, { "epoch": 39.05263157894737, "grad_norm": 0.2854568362236023, "learning_rate": 7.418000000000001e-05, "loss": 0.0121, "step": 3710 }, { "epoch": 39.1578947368421, "grad_norm": 0.17908264696598053, "learning_rate": 7.438e-05, "loss": 0.0118, "step": 3720 }, { "epoch": 39.26315789473684, "grad_norm": 0.22197091579437256, "learning_rate": 7.458000000000001e-05, "loss": 0.0114, "step": 3730 }, { "epoch": 39.36842105263158, "grad_norm": 0.14508722722530365, "learning_rate": 7.478e-05, "loss": 0.0105, "step": 3740 }, { "epoch": 39.473684210526315, "grad_norm": 0.16710858047008514, "learning_rate": 7.498e-05, "loss": 0.0109, "step": 3750 }, { "epoch": 39.578947368421055, "grad_norm": 0.20797361433506012, "learning_rate": 7.518000000000001e-05, "loss": 0.0125, "step": 3760 }, { "epoch": 39.68421052631579, "grad_norm": 0.2588461935520172, "learning_rate": 7.538e-05, "loss": 0.0111, "step": 3770 }, { "epoch": 39.78947368421053, "grad_norm": 0.17533957958221436, "learning_rate": 7.558e-05, "loss": 0.0115, "step": 3780 }, { "epoch": 39.89473684210526, "grad_norm": 0.2226632833480835, "learning_rate": 7.578000000000001e-05, "loss": 0.0114, "step": 3790 }, { "epoch": 40.0, "grad_norm": 0.2700129449367523, "learning_rate": 7.598e-05, "loss": 0.0132, "step": 3800 }, { "epoch": 40.10526315789474, "grad_norm": 0.2296552211046219, "learning_rate": 7.618e-05, "loss": 0.0128, "step": 3810 }, { "epoch": 40.21052631578947, "grad_norm": 0.2377881556749344, "learning_rate": 7.638000000000001e-05, "loss": 0.0118, "step": 3820 }, { "epoch": 40.31578947368421, "grad_norm": 0.15254360437393188, "learning_rate": 7.658e-05, "loss": 0.0115, "step": 3830 }, { "epoch": 40.421052631578945, "grad_norm": 0.17628058791160583, "learning_rate": 7.678000000000001e-05, "loss": 0.0121, "step": 3840 }, { "epoch": 40.526315789473685, "grad_norm": 0.24458418786525726, "learning_rate": 7.698000000000001e-05, "loss": 0.0128, "step": 3850 }, { "epoch": 40.63157894736842, "grad_norm": 0.13180460035800934, "learning_rate": 7.718e-05, "loss": 0.0106, "step": 3860 }, { "epoch": 40.73684210526316, "grad_norm": 0.1870930939912796, "learning_rate": 7.738000000000001e-05, "loss": 0.0121, "step": 3870 }, { "epoch": 40.8421052631579, "grad_norm": 0.24127325415611267, "learning_rate": 7.758000000000001e-05, "loss": 0.0115, "step": 3880 }, { "epoch": 40.94736842105263, "grad_norm": 0.18601706624031067, "learning_rate": 7.778e-05, "loss": 0.012, "step": 3890 }, { "epoch": 41.05263157894737, "grad_norm": 0.1703333705663681, "learning_rate": 7.798000000000001e-05, "loss": 0.0107, "step": 3900 }, { "epoch": 41.1578947368421, "grad_norm": 0.21299122273921967, "learning_rate": 7.818000000000001e-05, "loss": 0.0127, "step": 3910 }, { "epoch": 41.26315789473684, "grad_norm": 0.23282162845134735, "learning_rate": 7.838e-05, "loss": 0.0125, "step": 3920 }, { "epoch": 41.36842105263158, "grad_norm": 0.21252363920211792, "learning_rate": 7.858000000000001e-05, "loss": 0.0112, "step": 3930 }, { "epoch": 41.473684210526315, "grad_norm": 0.23398536443710327, "learning_rate": 7.878e-05, "loss": 0.0108, "step": 3940 }, { "epoch": 41.578947368421055, "grad_norm": 0.19570627808570862, "learning_rate": 7.897999999999999e-05, "loss": 0.0108, "step": 3950 }, { "epoch": 41.68421052631579, "grad_norm": 0.17268039286136627, "learning_rate": 7.918e-05, "loss": 0.0113, "step": 3960 }, { "epoch": 41.78947368421053, "grad_norm": 0.15861164033412933, "learning_rate": 7.938e-05, "loss": 0.0106, "step": 3970 }, { "epoch": 41.89473684210526, "grad_norm": 0.20067721605300903, "learning_rate": 7.958e-05, "loss": 0.0121, "step": 3980 }, { "epoch": 42.0, "grad_norm": 0.21180330216884613, "learning_rate": 7.978e-05, "loss": 0.011, "step": 3990 }, { "epoch": 42.10526315789474, "grad_norm": 0.21822966635227203, "learning_rate": 7.998e-05, "loss": 0.0121, "step": 4000 }, { "epoch": 42.21052631578947, "grad_norm": 0.18661390244960785, "learning_rate": 8.018e-05, "loss": 0.0117, "step": 4010 }, { "epoch": 42.31578947368421, "grad_norm": 0.18872861564159393, "learning_rate": 8.038e-05, "loss": 0.0115, "step": 4020 }, { "epoch": 42.421052631578945, "grad_norm": 0.20157185196876526, "learning_rate": 8.058e-05, "loss": 0.0116, "step": 4030 }, { "epoch": 42.526315789473685, "grad_norm": 0.22545406222343445, "learning_rate": 8.078e-05, "loss": 0.0107, "step": 4040 }, { "epoch": 42.63157894736842, "grad_norm": 0.19019801914691925, "learning_rate": 8.098e-05, "loss": 0.0115, "step": 4050 }, { "epoch": 42.73684210526316, "grad_norm": 0.17407260835170746, "learning_rate": 8.118e-05, "loss": 0.0107, "step": 4060 }, { "epoch": 42.8421052631579, "grad_norm": 0.1919659823179245, "learning_rate": 8.138e-05, "loss": 0.0106, "step": 4070 }, { "epoch": 42.94736842105263, "grad_norm": 0.19652576744556427, "learning_rate": 8.158e-05, "loss": 0.012, "step": 4080 }, { "epoch": 43.05263157894737, "grad_norm": 0.1990840584039688, "learning_rate": 8.178e-05, "loss": 0.0118, "step": 4090 }, { "epoch": 43.1578947368421, "grad_norm": 0.22769849002361298, "learning_rate": 8.198e-05, "loss": 0.0122, "step": 4100 }, { "epoch": 43.26315789473684, "grad_norm": 0.1592024564743042, "learning_rate": 8.218e-05, "loss": 0.0121, "step": 4110 }, { "epoch": 43.36842105263158, "grad_norm": 0.16610203683376312, "learning_rate": 8.238000000000001e-05, "loss": 0.0114, "step": 4120 }, { "epoch": 43.473684210526315, "grad_norm": 0.19175952672958374, "learning_rate": 8.258e-05, "loss": 0.0108, "step": 4130 }, { "epoch": 43.578947368421055, "grad_norm": 0.1870580017566681, "learning_rate": 8.278e-05, "loss": 0.011, "step": 4140 }, { "epoch": 43.68421052631579, "grad_norm": 0.18033215403556824, "learning_rate": 8.298000000000001e-05, "loss": 0.011, "step": 4150 }, { "epoch": 43.78947368421053, "grad_norm": 0.1625319868326187, "learning_rate": 8.318e-05, "loss": 0.011, "step": 4160 }, { "epoch": 43.89473684210526, "grad_norm": 0.20105122029781342, "learning_rate": 8.338e-05, "loss": 0.0105, "step": 4170 }, { "epoch": 44.0, "grad_norm": 0.300924152135849, "learning_rate": 8.358e-05, "loss": 0.0119, "step": 4180 }, { "epoch": 44.10526315789474, "grad_norm": 0.27038487792015076, "learning_rate": 8.378e-05, "loss": 0.012, "step": 4190 }, { "epoch": 44.21052631578947, "grad_norm": 0.19253966212272644, "learning_rate": 8.398e-05, "loss": 0.0122, "step": 4200 }, { "epoch": 44.31578947368421, "grad_norm": 0.2305767834186554, "learning_rate": 8.418e-05, "loss": 0.0123, "step": 4210 }, { "epoch": 44.421052631578945, "grad_norm": 0.19717970490455627, "learning_rate": 8.438e-05, "loss": 0.0116, "step": 4220 }, { "epoch": 44.526315789473685, "grad_norm": 0.29840514063835144, "learning_rate": 8.458e-05, "loss": 0.0114, "step": 4230 }, { "epoch": 44.63157894736842, "grad_norm": 0.20693133771419525, "learning_rate": 8.478e-05, "loss": 0.0124, "step": 4240 }, { "epoch": 44.73684210526316, "grad_norm": 0.1990007758140564, "learning_rate": 8.498e-05, "loss": 0.0107, "step": 4250 }, { "epoch": 44.8421052631579, "grad_norm": 0.1647491753101349, "learning_rate": 8.518000000000001e-05, "loss": 0.011, "step": 4260 }, { "epoch": 44.94736842105263, "grad_norm": 0.2084619402885437, "learning_rate": 8.538e-05, "loss": 0.01, "step": 4270 }, { "epoch": 45.05263157894737, "grad_norm": 0.17961111664772034, "learning_rate": 8.558e-05, "loss": 0.0106, "step": 4280 }, { "epoch": 45.1578947368421, "grad_norm": 0.1851383000612259, "learning_rate": 8.578000000000001e-05, "loss": 0.0118, "step": 4290 }, { "epoch": 45.26315789473684, "grad_norm": 0.22622188925743103, "learning_rate": 8.598e-05, "loss": 0.011, "step": 4300 }, { "epoch": 45.36842105263158, "grad_norm": 0.19176995754241943, "learning_rate": 8.618e-05, "loss": 0.0118, "step": 4310 }, { "epoch": 45.473684210526315, "grad_norm": 0.22198165953159332, "learning_rate": 8.638000000000001e-05, "loss": 0.0107, "step": 4320 }, { "epoch": 45.578947368421055, "grad_norm": 0.22405892610549927, "learning_rate": 8.658e-05, "loss": 0.0113, "step": 4330 }, { "epoch": 45.68421052631579, "grad_norm": 0.18319398164749146, "learning_rate": 8.678e-05, "loss": 0.0118, "step": 4340 }, { "epoch": 45.78947368421053, "grad_norm": 0.22047214210033417, "learning_rate": 8.698000000000001e-05, "loss": 0.0107, "step": 4350 }, { "epoch": 45.89473684210526, "grad_norm": 0.21890342235565186, "learning_rate": 8.718e-05, "loss": 0.0119, "step": 4360 }, { "epoch": 46.0, "grad_norm": 0.1678524911403656, "learning_rate": 8.738000000000001e-05, "loss": 0.0099, "step": 4370 }, { "epoch": 46.10526315789474, "grad_norm": 0.15417273342609406, "learning_rate": 8.758000000000001e-05, "loss": 0.0108, "step": 4380 }, { "epoch": 46.21052631578947, "grad_norm": 0.18818746507167816, "learning_rate": 8.778e-05, "loss": 0.0116, "step": 4390 }, { "epoch": 46.31578947368421, "grad_norm": 0.1861685961484909, "learning_rate": 8.798000000000001e-05, "loss": 0.0102, "step": 4400 }, { "epoch": 46.421052631578945, "grad_norm": 0.22351974248886108, "learning_rate": 8.818000000000001e-05, "loss": 0.0101, "step": 4410 }, { "epoch": 46.526315789473685, "grad_norm": 0.15710335969924927, "learning_rate": 8.838e-05, "loss": 0.0117, "step": 4420 }, { "epoch": 46.63157894736842, "grad_norm": 0.17801593244075775, "learning_rate": 8.858000000000001e-05, "loss": 0.0121, "step": 4430 }, { "epoch": 46.73684210526316, "grad_norm": 0.18110333383083344, "learning_rate": 8.878000000000001e-05, "loss": 0.0106, "step": 4440 }, { "epoch": 46.8421052631579, "grad_norm": 0.1773470789194107, "learning_rate": 8.898e-05, "loss": 0.0115, "step": 4450 }, { "epoch": 46.94736842105263, "grad_norm": 0.2346692979335785, "learning_rate": 8.918000000000001e-05, "loss": 0.0111, "step": 4460 }, { "epoch": 47.05263157894737, "grad_norm": 0.22668524086475372, "learning_rate": 8.938e-05, "loss": 0.0102, "step": 4470 }, { "epoch": 47.1578947368421, "grad_norm": 0.2413189560174942, "learning_rate": 8.958e-05, "loss": 0.0105, "step": 4480 }, { "epoch": 47.26315789473684, "grad_norm": 0.18773376941680908, "learning_rate": 8.978000000000001e-05, "loss": 0.0113, "step": 4490 }, { "epoch": 47.36842105263158, "grad_norm": 0.171018585562706, "learning_rate": 8.998e-05, "loss": 0.0109, "step": 4500 }, { "epoch": 47.473684210526315, "grad_norm": 0.17170758545398712, "learning_rate": 9.018000000000001e-05, "loss": 0.01, "step": 4510 }, { "epoch": 47.578947368421055, "grad_norm": 0.17056024074554443, "learning_rate": 9.038000000000001e-05, "loss": 0.0106, "step": 4520 }, { "epoch": 47.68421052631579, "grad_norm": 0.19392932951450348, "learning_rate": 9.058e-05, "loss": 0.0102, "step": 4530 }, { "epoch": 47.78947368421053, "grad_norm": 0.1558980792760849, "learning_rate": 9.078000000000001e-05, "loss": 0.0105, "step": 4540 }, { "epoch": 47.89473684210526, "grad_norm": 0.11771657317876816, "learning_rate": 9.098000000000001e-05, "loss": 0.0108, "step": 4550 }, { "epoch": 48.0, "grad_norm": 0.16250286996364594, "learning_rate": 9.118e-05, "loss": 0.0102, "step": 4560 }, { "epoch": 48.10526315789474, "grad_norm": 0.1942671686410904, "learning_rate": 9.138e-05, "loss": 0.0114, "step": 4570 }, { "epoch": 48.21052631578947, "grad_norm": 0.1566770225763321, "learning_rate": 9.158e-05, "loss": 0.0114, "step": 4580 }, { "epoch": 48.31578947368421, "grad_norm": 0.17785854637622833, "learning_rate": 9.178e-05, "loss": 0.0111, "step": 4590 }, { "epoch": 48.421052631578945, "grad_norm": 0.16568462550640106, "learning_rate": 9.198e-05, "loss": 0.0103, "step": 4600 }, { "epoch": 48.526315789473685, "grad_norm": 0.16783058643341064, "learning_rate": 9.218e-05, "loss": 0.011, "step": 4610 }, { "epoch": 48.63157894736842, "grad_norm": 0.17443189024925232, "learning_rate": 9.238e-05, "loss": 0.0114, "step": 4620 }, { "epoch": 48.73684210526316, "grad_norm": 0.21440953016281128, "learning_rate": 9.258e-05, "loss": 0.0113, "step": 4630 }, { "epoch": 48.8421052631579, "grad_norm": 0.23224467039108276, "learning_rate": 9.278e-05, "loss": 0.0119, "step": 4640 }, { "epoch": 48.94736842105263, "grad_norm": 0.16428428888320923, "learning_rate": 9.298e-05, "loss": 0.0118, "step": 4650 }, { "epoch": 49.05263157894737, "grad_norm": 0.21710319817066193, "learning_rate": 9.318e-05, "loss": 0.0101, "step": 4660 }, { "epoch": 49.1578947368421, "grad_norm": 0.17088653147220612, "learning_rate": 9.338e-05, "loss": 0.0117, "step": 4670 }, { "epoch": 49.26315789473684, "grad_norm": 0.16908703744411469, "learning_rate": 9.358e-05, "loss": 0.011, "step": 4680 }, { "epoch": 49.36842105263158, "grad_norm": 0.2050996720790863, "learning_rate": 9.378e-05, "loss": 0.0106, "step": 4690 }, { "epoch": 49.473684210526315, "grad_norm": 0.18306152522563934, "learning_rate": 9.398e-05, "loss": 0.0098, "step": 4700 }, { "epoch": 49.578947368421055, "grad_norm": 0.17459528148174286, "learning_rate": 9.418e-05, "loss": 0.0103, "step": 4710 }, { "epoch": 49.68421052631579, "grad_norm": 0.1954093724489212, "learning_rate": 9.438e-05, "loss": 0.0107, "step": 4720 }, { "epoch": 49.78947368421053, "grad_norm": 0.20250558853149414, "learning_rate": 9.458e-05, "loss": 0.0104, "step": 4730 }, { "epoch": 49.89473684210526, "grad_norm": 0.21961790323257446, "learning_rate": 9.478e-05, "loss": 0.0105, "step": 4740 }, { "epoch": 50.0, "grad_norm": 0.17740000784397125, "learning_rate": 9.498e-05, "loss": 0.012, "step": 4750 }, { "epoch": 50.10526315789474, "grad_norm": 0.22163300216197968, "learning_rate": 9.518000000000001e-05, "loss": 0.0104, "step": 4760 }, { "epoch": 50.21052631578947, "grad_norm": 0.147724911570549, "learning_rate": 9.538e-05, "loss": 0.0106, "step": 4770 }, { "epoch": 50.31578947368421, "grad_norm": 0.18383976817131042, "learning_rate": 9.558e-05, "loss": 0.0115, "step": 4780 }, { "epoch": 50.421052631578945, "grad_norm": 0.1635538786649704, "learning_rate": 9.578000000000001e-05, "loss": 0.0101, "step": 4790 }, { "epoch": 50.526315789473685, "grad_norm": 0.18582719564437866, "learning_rate": 9.598e-05, "loss": 0.0125, "step": 4800 }, { "epoch": 50.63157894736842, "grad_norm": 0.2038213312625885, "learning_rate": 9.618e-05, "loss": 0.0122, "step": 4810 }, { "epoch": 50.73684210526316, "grad_norm": 0.1846201866865158, "learning_rate": 9.638000000000001e-05, "loss": 0.0121, "step": 4820 }, { "epoch": 50.8421052631579, "grad_norm": 0.21574844419956207, "learning_rate": 9.658e-05, "loss": 0.0112, "step": 4830 }, { "epoch": 50.94736842105263, "grad_norm": 0.21967636048793793, "learning_rate": 9.678e-05, "loss": 0.0115, "step": 4840 }, { "epoch": 51.05263157894737, "grad_norm": 0.22442564368247986, "learning_rate": 9.698000000000001e-05, "loss": 0.0121, "step": 4850 }, { "epoch": 51.1578947368421, "grad_norm": 0.15954144299030304, "learning_rate": 9.718e-05, "loss": 0.0116, "step": 4860 }, { "epoch": 51.26315789473684, "grad_norm": 0.2003287672996521, "learning_rate": 9.738e-05, "loss": 0.0117, "step": 4870 }, { "epoch": 51.36842105263158, "grad_norm": 0.17973533272743225, "learning_rate": 9.758000000000001e-05, "loss": 0.0109, "step": 4880 }, { "epoch": 51.473684210526315, "grad_norm": 0.15353059768676758, "learning_rate": 9.778e-05, "loss": 0.0113, "step": 4890 }, { "epoch": 51.578947368421055, "grad_norm": 0.137799933552742, "learning_rate": 9.798000000000001e-05, "loss": 0.0102, "step": 4900 }, { "epoch": 51.68421052631579, "grad_norm": 0.16002607345581055, "learning_rate": 9.818000000000001e-05, "loss": 0.0094, "step": 4910 }, { "epoch": 51.78947368421053, "grad_norm": 0.2005079984664917, "learning_rate": 9.838e-05, "loss": 0.0111, "step": 4920 }, { "epoch": 51.89473684210526, "grad_norm": 0.19363661110401154, "learning_rate": 9.858000000000001e-05, "loss": 0.0108, "step": 4930 }, { "epoch": 52.0, "grad_norm": 0.23911581933498383, "learning_rate": 9.878e-05, "loss": 0.01, "step": 4940 }, { "epoch": 52.10526315789474, "grad_norm": 0.1481848657131195, "learning_rate": 9.898e-05, "loss": 0.0101, "step": 4950 }, { "epoch": 52.21052631578947, "grad_norm": 0.18552443385124207, "learning_rate": 9.918000000000001e-05, "loss": 0.0114, "step": 4960 }, { "epoch": 52.31578947368421, "grad_norm": 0.16553276777267456, "learning_rate": 9.938e-05, "loss": 0.0096, "step": 4970 }, { "epoch": 52.421052631578945, "grad_norm": 0.16358128190040588, "learning_rate": 9.958e-05, "loss": 0.0119, "step": 4980 }, { "epoch": 52.526315789473685, "grad_norm": 0.15967893600463867, "learning_rate": 9.978000000000001e-05, "loss": 0.0114, "step": 4990 }, { "epoch": 52.63157894736842, "grad_norm": 0.18860729038715363, "learning_rate": 9.998e-05, "loss": 0.0107, "step": 5000 }, { "epoch": 52.73684210526316, "grad_norm": 0.20670965313911438, "learning_rate": 9.999999778549045e-05, "loss": 0.011, "step": 5010 }, { "epoch": 52.8421052631579, "grad_norm": 0.1918308287858963, "learning_rate": 9.999999013039593e-05, "loss": 0.0117, "step": 5020 }, { "epoch": 52.94736842105263, "grad_norm": 0.18079149723052979, "learning_rate": 9.999997700737766e-05, "loss": 0.011, "step": 5030 }, { "epoch": 53.05263157894737, "grad_norm": 0.14320296049118042, "learning_rate": 9.999995841643709e-05, "loss": 0.0094, "step": 5040 }, { "epoch": 53.1578947368421, "grad_norm": 0.19478894770145416, "learning_rate": 9.999993435757623e-05, "loss": 0.0098, "step": 5050 }, { "epoch": 53.26315789473684, "grad_norm": 0.19877856969833374, "learning_rate": 9.999990483079773e-05, "loss": 0.0102, "step": 5060 }, { "epoch": 53.36842105263158, "grad_norm": 0.22134117782115936, "learning_rate": 9.999986983610481e-05, "loss": 0.0104, "step": 5070 }, { "epoch": 53.473684210526315, "grad_norm": 0.23054654896259308, "learning_rate": 9.99998293735013e-05, "loss": 0.011, "step": 5080 }, { "epoch": 53.578947368421055, "grad_norm": 0.21688352525234222, "learning_rate": 9.999978344299161e-05, "loss": 0.0104, "step": 5090 }, { "epoch": 53.68421052631579, "grad_norm": 0.1742480844259262, "learning_rate": 9.99997320445808e-05, "loss": 0.0108, "step": 5100 }, { "epoch": 53.78947368421053, "grad_norm": 0.12125600874423981, "learning_rate": 9.999967517827444e-05, "loss": 0.0103, "step": 5110 }, { "epoch": 53.89473684210526, "grad_norm": 0.13377055525779724, "learning_rate": 9.999961284407879e-05, "loss": 0.0104, "step": 5120 }, { "epoch": 54.0, "grad_norm": 0.14926579594612122, "learning_rate": 9.999954504200067e-05, "loss": 0.0106, "step": 5130 }, { "epoch": 54.10526315789474, "grad_norm": 0.13876181840896606, "learning_rate": 9.999947177204744e-05, "loss": 0.0107, "step": 5140 }, { "epoch": 54.21052631578947, "grad_norm": 0.17124806344509125, "learning_rate": 9.999939303422718e-05, "loss": 0.0097, "step": 5150 }, { "epoch": 54.31578947368421, "grad_norm": 0.1493268758058548, "learning_rate": 9.999930882854847e-05, "loss": 0.0109, "step": 5160 }, { "epoch": 54.421052631578945, "grad_norm": 0.1756788194179535, "learning_rate": 9.999921915502051e-05, "loss": 0.0094, "step": 5170 }, { "epoch": 54.526315789473685, "grad_norm": 0.1820589154958725, "learning_rate": 9.99991240136531e-05, "loss": 0.0101, "step": 5180 }, { "epoch": 54.63157894736842, "grad_norm": 0.16425181925296783, "learning_rate": 9.999902340445668e-05, "loss": 0.0105, "step": 5190 }, { "epoch": 54.73684210526316, "grad_norm": 0.1366964727640152, "learning_rate": 9.999891732744224e-05, "loss": 0.0093, "step": 5200 }, { "epoch": 54.8421052631579, "grad_norm": 0.19328008592128754, "learning_rate": 9.999880578262135e-05, "loss": 0.0096, "step": 5210 }, { "epoch": 54.94736842105263, "grad_norm": 0.17198260128498077, "learning_rate": 9.999868877000624e-05, "loss": 0.0116, "step": 5220 }, { "epoch": 55.05263157894737, "grad_norm": 0.1936512142419815, "learning_rate": 9.99985662896097e-05, "loss": 0.0095, "step": 5230 }, { "epoch": 55.1578947368421, "grad_norm": 0.2091033011674881, "learning_rate": 9.999843834144513e-05, "loss": 0.0094, "step": 5240 }, { "epoch": 55.26315789473684, "grad_norm": 0.13628341257572174, "learning_rate": 9.99983049255265e-05, "loss": 0.0109, "step": 5250 }, { "epoch": 55.36842105263158, "grad_norm": 0.1364404708147049, "learning_rate": 9.999816604186843e-05, "loss": 0.0098, "step": 5260 }, { "epoch": 55.473684210526315, "grad_norm": 0.13993944227695465, "learning_rate": 9.999802169048609e-05, "loss": 0.0106, "step": 5270 }, { "epoch": 55.578947368421055, "grad_norm": 0.19288496673107147, "learning_rate": 9.999787187139527e-05, "loss": 0.0107, "step": 5280 }, { "epoch": 55.68421052631579, "grad_norm": 0.1657884269952774, "learning_rate": 9.999771658461234e-05, "loss": 0.01, "step": 5290 }, { "epoch": 55.78947368421053, "grad_norm": 0.14469844102859497, "learning_rate": 9.999755583015431e-05, "loss": 0.0104, "step": 5300 }, { "epoch": 55.89473684210526, "grad_norm": 0.1709185391664505, "learning_rate": 9.999738960803874e-05, "loss": 0.0095, "step": 5310 }, { "epoch": 56.0, "grad_norm": 0.1474951207637787, "learning_rate": 9.99972179182838e-05, "loss": 0.0092, "step": 5320 }, { "epoch": 56.10526315789474, "grad_norm": 0.181667298078537, "learning_rate": 9.99970407609083e-05, "loss": 0.01, "step": 5330 }, { "epoch": 56.21052631578947, "grad_norm": 0.16350439190864563, "learning_rate": 9.999685813593159e-05, "loss": 0.0097, "step": 5340 }, { "epoch": 56.31578947368421, "grad_norm": 0.19336770474910736, "learning_rate": 9.999667004337362e-05, "loss": 0.0102, "step": 5350 }, { "epoch": 56.421052631578945, "grad_norm": 0.1777733415365219, "learning_rate": 9.9996476483255e-05, "loss": 0.0087, "step": 5360 }, { "epoch": 56.526315789473685, "grad_norm": 0.20297273993492126, "learning_rate": 9.999627745559688e-05, "loss": 0.0102, "step": 5370 }, { "epoch": 56.63157894736842, "grad_norm": 0.1623271107673645, "learning_rate": 9.999607296042101e-05, "loss": 0.0099, "step": 5380 }, { "epoch": 56.73684210526316, "grad_norm": 0.17753125727176666, "learning_rate": 9.99958629977498e-05, "loss": 0.0093, "step": 5390 }, { "epoch": 56.8421052631579, "grad_norm": 0.14664316177368164, "learning_rate": 9.999564756760615e-05, "loss": 0.0101, "step": 5400 }, { "epoch": 56.94736842105263, "grad_norm": 0.1283586174249649, "learning_rate": 9.999542667001366e-05, "loss": 0.0097, "step": 5410 }, { "epoch": 57.05263157894737, "grad_norm": 0.18282289803028107, "learning_rate": 9.999520030499647e-05, "loss": 0.0093, "step": 5420 }, { "epoch": 57.1578947368421, "grad_norm": 0.21146658062934875, "learning_rate": 9.999496847257936e-05, "loss": 0.0096, "step": 5430 }, { "epoch": 57.26315789473684, "grad_norm": 0.16015808284282684, "learning_rate": 9.999473117278764e-05, "loss": 0.0093, "step": 5440 }, { "epoch": 57.36842105263158, "grad_norm": 0.15822754800319672, "learning_rate": 9.999448840564731e-05, "loss": 0.0095, "step": 5450 }, { "epoch": 57.473684210526315, "grad_norm": 0.15499241650104523, "learning_rate": 9.999424017118488e-05, "loss": 0.0091, "step": 5460 }, { "epoch": 57.578947368421055, "grad_norm": 0.1432248055934906, "learning_rate": 9.999398646942751e-05, "loss": 0.0092, "step": 5470 }, { "epoch": 57.68421052631579, "grad_norm": 0.15302738547325134, "learning_rate": 9.999372730040296e-05, "loss": 0.01, "step": 5480 }, { "epoch": 57.78947368421053, "grad_norm": 0.12965810298919678, "learning_rate": 9.999346266413953e-05, "loss": 0.009, "step": 5490 }, { "epoch": 57.89473684210526, "grad_norm": 0.16545115411281586, "learning_rate": 9.99931925606662e-05, "loss": 0.0093, "step": 5500 }, { "epoch": 58.0, "grad_norm": 0.2322085201740265, "learning_rate": 9.99929169900125e-05, "loss": 0.0103, "step": 5510 }, { "epoch": 58.10526315789474, "grad_norm": 0.18458271026611328, "learning_rate": 9.999263595220855e-05, "loss": 0.0096, "step": 5520 }, { "epoch": 58.21052631578947, "grad_norm": 0.12357588857412338, "learning_rate": 9.99923494472851e-05, "loss": 0.0099, "step": 5530 }, { "epoch": 58.31578947368421, "grad_norm": 0.1631576120853424, "learning_rate": 9.999205747527348e-05, "loss": 0.0104, "step": 5540 }, { "epoch": 58.421052631578945, "grad_norm": 0.16488900780677795, "learning_rate": 9.999176003620561e-05, "loss": 0.0094, "step": 5550 }, { "epoch": 58.526315789473685, "grad_norm": 0.17670392990112305, "learning_rate": 9.999145713011405e-05, "loss": 0.0089, "step": 5560 }, { "epoch": 58.63157894736842, "grad_norm": 0.13911907374858856, "learning_rate": 9.999114875703186e-05, "loss": 0.0091, "step": 5570 }, { "epoch": 58.73684210526316, "grad_norm": 0.17339004576206207, "learning_rate": 9.999083491699281e-05, "loss": 0.0096, "step": 5580 }, { "epoch": 58.8421052631579, "grad_norm": 0.16414259374141693, "learning_rate": 9.999051561003123e-05, "loss": 0.009, "step": 5590 }, { "epoch": 58.94736842105263, "grad_norm": 0.15142418444156647, "learning_rate": 9.999019083618202e-05, "loss": 0.0097, "step": 5600 }, { "epoch": 59.05263157894737, "grad_norm": 0.16554191708564758, "learning_rate": 9.99898605954807e-05, "loss": 0.0087, "step": 5610 }, { "epoch": 59.1578947368421, "grad_norm": 0.16175711154937744, "learning_rate": 9.998952488796338e-05, "loss": 0.0104, "step": 5620 }, { "epoch": 59.26315789473684, "grad_norm": 0.14773352444171906, "learning_rate": 9.998918371366676e-05, "loss": 0.0107, "step": 5630 }, { "epoch": 59.36842105263158, "grad_norm": 0.1364739090204239, "learning_rate": 9.99888370726282e-05, "loss": 0.009, "step": 5640 }, { "epoch": 59.473684210526315, "grad_norm": 0.12696947157382965, "learning_rate": 9.998848496488556e-05, "loss": 0.0101, "step": 5650 }, { "epoch": 59.578947368421055, "grad_norm": 0.13747288286685944, "learning_rate": 9.998812739047736e-05, "loss": 0.0104, "step": 5660 }, { "epoch": 59.68421052631579, "grad_norm": 0.1762026846408844, "learning_rate": 9.99877643494427e-05, "loss": 0.0089, "step": 5670 }, { "epoch": 59.78947368421053, "grad_norm": 0.11868304014205933, "learning_rate": 9.998739584182128e-05, "loss": 0.0101, "step": 5680 }, { "epoch": 59.89473684210526, "grad_norm": 0.19812919199466705, "learning_rate": 9.998702186765342e-05, "loss": 0.0096, "step": 5690 }, { "epoch": 60.0, "grad_norm": 0.180701345205307, "learning_rate": 9.998664242698e-05, "loss": 0.0101, "step": 5700 }, { "epoch": 60.10526315789474, "grad_norm": 0.15576530992984772, "learning_rate": 9.998625751984251e-05, "loss": 0.0102, "step": 5710 }, { "epoch": 60.21052631578947, "grad_norm": 0.1506369560956955, "learning_rate": 9.998586714628307e-05, "loss": 0.0106, "step": 5720 }, { "epoch": 60.31578947368421, "grad_norm": 0.11458553373813629, "learning_rate": 9.998547130634432e-05, "loss": 0.0094, "step": 5730 }, { "epoch": 60.421052631578945, "grad_norm": 0.14538589119911194, "learning_rate": 9.99850700000696e-05, "loss": 0.01, "step": 5740 }, { "epoch": 60.526315789473685, "grad_norm": 0.15139704942703247, "learning_rate": 9.998466322750278e-05, "loss": 0.0078, "step": 5750 }, { "epoch": 60.63157894736842, "grad_norm": 0.12363656610250473, "learning_rate": 9.998425098868834e-05, "loss": 0.0085, "step": 5760 }, { "epoch": 60.73684210526316, "grad_norm": 0.14009208977222443, "learning_rate": 9.998383328367136e-05, "loss": 0.0094, "step": 5770 }, { "epoch": 60.8421052631579, "grad_norm": 0.1501094102859497, "learning_rate": 9.99834101124975e-05, "loss": 0.0093, "step": 5780 }, { "epoch": 60.94736842105263, "grad_norm": 0.1508973389863968, "learning_rate": 9.998298147521309e-05, "loss": 0.0089, "step": 5790 }, { "epoch": 61.05263157894737, "grad_norm": 0.16601379215717316, "learning_rate": 9.998254737186496e-05, "loss": 0.0087, "step": 5800 }, { "epoch": 61.1578947368421, "grad_norm": 0.1826418787240982, "learning_rate": 9.99821078025006e-05, "loss": 0.0094, "step": 5810 }, { "epoch": 61.26315789473684, "grad_norm": 0.1457967609167099, "learning_rate": 9.998166276716807e-05, "loss": 0.01, "step": 5820 }, { "epoch": 61.36842105263158, "grad_norm": 0.13877476751804352, "learning_rate": 9.998121226591606e-05, "loss": 0.0093, "step": 5830 }, { "epoch": 61.473684210526315, "grad_norm": 0.13871927559375763, "learning_rate": 9.998075629879382e-05, "loss": 0.0096, "step": 5840 }, { "epoch": 61.578947368421055, "grad_norm": 0.12461774051189423, "learning_rate": 9.99802948658512e-05, "loss": 0.0081, "step": 5850 }, { "epoch": 61.68421052631579, "grad_norm": 0.16352084279060364, "learning_rate": 9.99798279671387e-05, "loss": 0.0093, "step": 5860 }, { "epoch": 61.78947368421053, "grad_norm": 0.19579023122787476, "learning_rate": 9.997935560270734e-05, "loss": 0.0094, "step": 5870 }, { "epoch": 61.89473684210526, "grad_norm": 0.1514274775981903, "learning_rate": 9.997887777260879e-05, "loss": 0.0087, "step": 5880 }, { "epoch": 62.0, "grad_norm": 0.14203481376171112, "learning_rate": 9.997839447689532e-05, "loss": 0.0083, "step": 5890 }, { "epoch": 62.10526315789474, "grad_norm": 0.11290877312421799, "learning_rate": 9.997790571561978e-05, "loss": 0.0096, "step": 5900 }, { "epoch": 62.21052631578947, "grad_norm": 0.15248949825763702, "learning_rate": 9.99774114888356e-05, "loss": 0.0098, "step": 5910 }, { "epoch": 62.31578947368421, "grad_norm": 0.18024367094039917, "learning_rate": 9.997691179659684e-05, "loss": 0.0096, "step": 5920 }, { "epoch": 62.421052631578945, "grad_norm": 0.1857658177614212, "learning_rate": 9.997640663895815e-05, "loss": 0.0102, "step": 5930 }, { "epoch": 62.526315789473685, "grad_norm": 0.18365083634853363, "learning_rate": 9.997589601597477e-05, "loss": 0.0104, "step": 5940 }, { "epoch": 62.63157894736842, "grad_norm": 0.1843414008617401, "learning_rate": 9.997537992770252e-05, "loss": 0.0098, "step": 5950 }, { "epoch": 62.73684210526316, "grad_norm": 0.19069445133209229, "learning_rate": 9.997485837419788e-05, "loss": 0.0107, "step": 5960 }, { "epoch": 62.8421052631579, "grad_norm": 0.19017553329467773, "learning_rate": 9.997433135551786e-05, "loss": 0.0088, "step": 5970 }, { "epoch": 62.94736842105263, "grad_norm": 0.1876637488603592, "learning_rate": 9.997379887172009e-05, "loss": 0.0105, "step": 5980 }, { "epoch": 63.05263157894737, "grad_norm": 0.12983855605125427, "learning_rate": 9.997326092286281e-05, "loss": 0.0098, "step": 5990 }, { "epoch": 63.1578947368421, "grad_norm": 0.1388033926486969, "learning_rate": 9.997271750900486e-05, "loss": 0.01, "step": 6000 }, { "epoch": 63.26315789473684, "grad_norm": 0.14018623530864716, "learning_rate": 9.997216863020565e-05, "loss": 0.0101, "step": 6010 }, { "epoch": 63.36842105263158, "grad_norm": 0.17431820929050446, "learning_rate": 9.99716142865252e-05, "loss": 0.0089, "step": 6020 }, { "epoch": 63.473684210526315, "grad_norm": 0.2002701610326767, "learning_rate": 9.997105447802415e-05, "loss": 0.0088, "step": 6030 }, { "epoch": 63.578947368421055, "grad_norm": 0.14597174525260925, "learning_rate": 9.997048920476373e-05, "loss": 0.01, "step": 6040 }, { "epoch": 63.68421052631579, "grad_norm": 0.15541471540927887, "learning_rate": 9.996991846680572e-05, "loss": 0.0089, "step": 6050 }, { "epoch": 63.78947368421053, "grad_norm": 0.15486662089824677, "learning_rate": 9.996934226421257e-05, "loss": 0.0084, "step": 6060 }, { "epoch": 63.89473684210526, "grad_norm": 0.15883676707744598, "learning_rate": 9.996876059704726e-05, "loss": 0.0093, "step": 6070 }, { "epoch": 64.0, "grad_norm": 0.2113405019044876, "learning_rate": 9.996817346537343e-05, "loss": 0.0093, "step": 6080 }, { "epoch": 64.10526315789474, "grad_norm": 0.14586658775806427, "learning_rate": 9.996758086925526e-05, "loss": 0.0098, "step": 6090 }, { "epoch": 64.21052631578948, "grad_norm": 0.12627056241035461, "learning_rate": 9.996698280875759e-05, "loss": 0.0092, "step": 6100 }, { "epoch": 64.3157894736842, "grad_norm": 0.11351701617240906, "learning_rate": 9.99663792839458e-05, "loss": 0.0086, "step": 6110 }, { "epoch": 64.42105263157895, "grad_norm": 0.16290274262428284, "learning_rate": 9.99657702948859e-05, "loss": 0.0094, "step": 6120 }, { "epoch": 64.52631578947368, "grad_norm": 0.14398665726184845, "learning_rate": 9.996515584164448e-05, "loss": 0.0096, "step": 6130 }, { "epoch": 64.63157894736842, "grad_norm": 0.1672496646642685, "learning_rate": 9.996453592428873e-05, "loss": 0.0091, "step": 6140 }, { "epoch": 64.73684210526316, "grad_norm": 0.14771363139152527, "learning_rate": 9.996391054288646e-05, "loss": 0.0098, "step": 6150 }, { "epoch": 64.84210526315789, "grad_norm": 0.1438017040491104, "learning_rate": 9.996327969750605e-05, "loss": 0.01, "step": 6160 }, { "epoch": 64.94736842105263, "grad_norm": 0.16346044838428497, "learning_rate": 9.996264338821649e-05, "loss": 0.0085, "step": 6170 }, { "epoch": 65.05263157894737, "grad_norm": 0.17775575816631317, "learning_rate": 9.996200161508735e-05, "loss": 0.0085, "step": 6180 }, { "epoch": 65.15789473684211, "grad_norm": 0.13238279521465302, "learning_rate": 9.996135437818885e-05, "loss": 0.0084, "step": 6190 }, { "epoch": 65.26315789473684, "grad_norm": 0.09737861156463623, "learning_rate": 9.996070167759175e-05, "loss": 0.0086, "step": 6200 }, { "epoch": 65.36842105263158, "grad_norm": 0.11257822811603546, "learning_rate": 9.996004351336743e-05, "loss": 0.0079, "step": 6210 }, { "epoch": 65.47368421052632, "grad_norm": 0.11516859382390976, "learning_rate": 9.995937988558785e-05, "loss": 0.0078, "step": 6220 }, { "epoch": 65.57894736842105, "grad_norm": 0.13981062173843384, "learning_rate": 9.995871079432561e-05, "loss": 0.0083, "step": 6230 }, { "epoch": 65.6842105263158, "grad_norm": 0.13986778259277344, "learning_rate": 9.995803623965389e-05, "loss": 0.0087, "step": 6240 }, { "epoch": 65.78947368421052, "grad_norm": 0.14539757370948792, "learning_rate": 9.995735622164641e-05, "loss": 0.0094, "step": 6250 }, { "epoch": 65.89473684210526, "grad_norm": 0.13641588389873505, "learning_rate": 9.995667074037758e-05, "loss": 0.0087, "step": 6260 }, { "epoch": 66.0, "grad_norm": 0.16893237829208374, "learning_rate": 9.995597979592232e-05, "loss": 0.0098, "step": 6270 }, { "epoch": 66.10526315789474, "grad_norm": 0.1553732007741928, "learning_rate": 9.995528338835625e-05, "loss": 0.0097, "step": 6280 }, { "epoch": 66.21052631578948, "grad_norm": 0.12815383076667786, "learning_rate": 9.995458151775547e-05, "loss": 0.01, "step": 6290 }, { "epoch": 66.3157894736842, "grad_norm": 0.13050927221775055, "learning_rate": 9.995387418419677e-05, "loss": 0.0094, "step": 6300 }, { "epoch": 66.42105263157895, "grad_norm": 0.1411401331424713, "learning_rate": 9.99531613877575e-05, "loss": 0.0095, "step": 6310 }, { "epoch": 66.52631578947368, "grad_norm": 0.15626223385334015, "learning_rate": 9.995244312851559e-05, "loss": 0.0098, "step": 6320 }, { "epoch": 66.63157894736842, "grad_norm": 0.17229348421096802, "learning_rate": 9.995171940654961e-05, "loss": 0.0091, "step": 6330 }, { "epoch": 66.73684210526316, "grad_norm": 0.13778650760650635, "learning_rate": 9.995099022193871e-05, "loss": 0.0097, "step": 6340 }, { "epoch": 66.84210526315789, "grad_norm": 0.12857700884342194, "learning_rate": 9.995025557476261e-05, "loss": 0.0092, "step": 6350 }, { "epoch": 66.94736842105263, "grad_norm": 0.15290206670761108, "learning_rate": 9.994951546510165e-05, "loss": 0.0094, "step": 6360 }, { "epoch": 67.05263157894737, "grad_norm": 0.15689069032669067, "learning_rate": 9.994876989303679e-05, "loss": 0.0089, "step": 6370 }, { "epoch": 67.15789473684211, "grad_norm": 0.1968839019536972, "learning_rate": 9.994801885864955e-05, "loss": 0.009, "step": 6380 }, { "epoch": 67.26315789473684, "grad_norm": 0.21617542207241058, "learning_rate": 9.994726236202205e-05, "loss": 0.0096, "step": 6390 }, { "epoch": 67.36842105263158, "grad_norm": 0.18209591507911682, "learning_rate": 9.994650040323704e-05, "loss": 0.0092, "step": 6400 }, { "epoch": 67.47368421052632, "grad_norm": 0.16333486139774323, "learning_rate": 9.994573298237784e-05, "loss": 0.01, "step": 6410 }, { "epoch": 67.57894736842105, "grad_norm": 0.17723891139030457, "learning_rate": 9.994496009952837e-05, "loss": 0.0094, "step": 6420 }, { "epoch": 67.6842105263158, "grad_norm": 0.13339084386825562, "learning_rate": 9.994418175477316e-05, "loss": 0.0093, "step": 6430 }, { "epoch": 67.78947368421052, "grad_norm": 0.13959872722625732, "learning_rate": 9.994339794819733e-05, "loss": 0.0095, "step": 6440 }, { "epoch": 67.89473684210526, "grad_norm": 0.14133432507514954, "learning_rate": 9.994260867988658e-05, "loss": 0.0092, "step": 6450 }, { "epoch": 68.0, "grad_norm": 0.21912789344787598, "learning_rate": 9.994181394992723e-05, "loss": 0.0082, "step": 6460 }, { "epoch": 68.10526315789474, "grad_norm": 0.1554618626832962, "learning_rate": 9.994101375840618e-05, "loss": 0.0091, "step": 6470 }, { "epoch": 68.21052631578948, "grad_norm": 0.14466311037540436, "learning_rate": 9.994020810541098e-05, "loss": 0.0092, "step": 6480 }, { "epoch": 68.3157894736842, "grad_norm": 0.1327737271785736, "learning_rate": 9.99393969910297e-05, "loss": 0.0085, "step": 6490 }, { "epoch": 68.42105263157895, "grad_norm": 0.1864118129014969, "learning_rate": 9.993858041535104e-05, "loss": 0.0097, "step": 6500 }, { "epoch": 68.52631578947368, "grad_norm": 0.1674613058567047, "learning_rate": 9.99377583784643e-05, "loss": 0.0083, "step": 6510 }, { "epoch": 68.63157894736842, "grad_norm": 0.148840993642807, "learning_rate": 9.993693088045939e-05, "loss": 0.009, "step": 6520 }, { "epoch": 68.73684210526316, "grad_norm": 0.17198745906352997, "learning_rate": 9.99360979214268e-05, "loss": 0.0091, "step": 6530 }, { "epoch": 68.84210526315789, "grad_norm": 0.14425309002399445, "learning_rate": 9.99352595014576e-05, "loss": 0.0089, "step": 6540 }, { "epoch": 68.94736842105263, "grad_norm": 0.17350517213344574, "learning_rate": 9.993441562064354e-05, "loss": 0.0097, "step": 6550 }, { "epoch": 69.05263157894737, "grad_norm": 0.1362496018409729, "learning_rate": 9.993356627907685e-05, "loss": 0.0094, "step": 6560 }, { "epoch": 69.15789473684211, "grad_norm": 0.12359984964132309, "learning_rate": 9.99327114768504e-05, "loss": 0.0093, "step": 6570 }, { "epoch": 69.26315789473684, "grad_norm": 0.19654642045497894, "learning_rate": 9.99318512140577e-05, "loss": 0.0092, "step": 6580 }, { "epoch": 69.36842105263158, "grad_norm": 0.12445419281721115, "learning_rate": 9.993098549079284e-05, "loss": 0.0081, "step": 6590 }, { "epoch": 69.47368421052632, "grad_norm": 0.14248943328857422, "learning_rate": 9.993011430715047e-05, "loss": 0.0089, "step": 6600 }, { "epoch": 69.57894736842105, "grad_norm": 0.12037394940853119, "learning_rate": 9.992923766322586e-05, "loss": 0.0088, "step": 6610 }, { "epoch": 69.6842105263158, "grad_norm": 0.1510329395532608, "learning_rate": 9.99283555591149e-05, "loss": 0.0096, "step": 6620 }, { "epoch": 69.78947368421052, "grad_norm": 0.13216611742973328, "learning_rate": 9.992746799491404e-05, "loss": 0.0089, "step": 6630 }, { "epoch": 69.89473684210526, "grad_norm": 0.1367773413658142, "learning_rate": 9.992657497072033e-05, "loss": 0.0092, "step": 6640 }, { "epoch": 70.0, "grad_norm": 0.24007968604564667, "learning_rate": 9.992567648663147e-05, "loss": 0.0101, "step": 6650 }, { "epoch": 70.10526315789474, "grad_norm": 0.15646283328533173, "learning_rate": 9.992477254274568e-05, "loss": 0.0092, "step": 6660 }, { "epoch": 70.21052631578948, "grad_norm": 0.1784546971321106, "learning_rate": 9.992386313916183e-05, "loss": 0.0091, "step": 6670 }, { "epoch": 70.3157894736842, "grad_norm": 0.12028051167726517, "learning_rate": 9.992294827597934e-05, "loss": 0.0097, "step": 6680 }, { "epoch": 70.42105263157895, "grad_norm": 0.15164831280708313, "learning_rate": 9.992202795329831e-05, "loss": 0.0093, "step": 6690 }, { "epoch": 70.52631578947368, "grad_norm": 0.17381545901298523, "learning_rate": 9.992110217121936e-05, "loss": 0.0106, "step": 6700 }, { "epoch": 70.63157894736842, "grad_norm": 0.1539708971977234, "learning_rate": 9.992017092984372e-05, "loss": 0.0094, "step": 6710 }, { "epoch": 70.73684210526316, "grad_norm": 0.16470472514629364, "learning_rate": 9.991923422927326e-05, "loss": 0.009, "step": 6720 }, { "epoch": 70.84210526315789, "grad_norm": 0.13268184661865234, "learning_rate": 9.991829206961037e-05, "loss": 0.0087, "step": 6730 }, { "epoch": 70.94736842105263, "grad_norm": 0.12377230823040009, "learning_rate": 9.991734445095813e-05, "loss": 0.0097, "step": 6740 }, { "epoch": 71.05263157894737, "grad_norm": 0.12022443115711212, "learning_rate": 9.991639137342015e-05, "loss": 0.0099, "step": 6750 }, { "epoch": 71.15789473684211, "grad_norm": 0.1583162397146225, "learning_rate": 9.991543283710064e-05, "loss": 0.0088, "step": 6760 }, { "epoch": 71.26315789473684, "grad_norm": 0.14013615250587463, "learning_rate": 9.991446884210445e-05, "loss": 0.0088, "step": 6770 }, { "epoch": 71.36842105263158, "grad_norm": 0.1160735934972763, "learning_rate": 9.9913499388537e-05, "loss": 0.0086, "step": 6780 }, { "epoch": 71.47368421052632, "grad_norm": 0.09773584455251694, "learning_rate": 9.99125244765043e-05, "loss": 0.0077, "step": 6790 }, { "epoch": 71.57894736842105, "grad_norm": 0.13230980932712555, "learning_rate": 9.991154410611296e-05, "loss": 0.0085, "step": 6800 }, { "epoch": 71.6842105263158, "grad_norm": 0.13829822838306427, "learning_rate": 9.99105582774702e-05, "loss": 0.0088, "step": 6810 }, { "epoch": 71.78947368421052, "grad_norm": 0.14681865274906158, "learning_rate": 9.990956699068384e-05, "loss": 0.0095, "step": 6820 }, { "epoch": 71.89473684210526, "grad_norm": 0.13248483836650848, "learning_rate": 9.990857024586224e-05, "loss": 0.0081, "step": 6830 }, { "epoch": 72.0, "grad_norm": 0.10956913232803345, "learning_rate": 9.990756804311446e-05, "loss": 0.0073, "step": 6840 }, { "epoch": 72.10526315789474, "grad_norm": 0.13126641511917114, "learning_rate": 9.990656038255006e-05, "loss": 0.008, "step": 6850 }, { "epoch": 72.21052631578948, "grad_norm": 0.14451555907726288, "learning_rate": 9.990554726427926e-05, "loss": 0.0087, "step": 6860 }, { "epoch": 72.3157894736842, "grad_norm": 0.13306394219398499, "learning_rate": 9.990452868841284e-05, "loss": 0.0083, "step": 6870 }, { "epoch": 72.42105263157895, "grad_norm": 0.12162573635578156, "learning_rate": 9.99035046550622e-05, "loss": 0.0085, "step": 6880 }, { "epoch": 72.52631578947368, "grad_norm": 0.10861755907535553, "learning_rate": 9.99024751643393e-05, "loss": 0.0083, "step": 6890 }, { "epoch": 72.63157894736842, "grad_norm": 0.08701284229755402, "learning_rate": 9.990144021635677e-05, "loss": 0.0071, "step": 6900 }, { "epoch": 72.73684210526316, "grad_norm": 0.16344591975212097, "learning_rate": 9.990039981122775e-05, "loss": 0.0085, "step": 6910 }, { "epoch": 72.84210526315789, "grad_norm": 0.1957966834306717, "learning_rate": 9.989935394906602e-05, "loss": 0.0093, "step": 6920 }, { "epoch": 72.94736842105263, "grad_norm": 0.15214237570762634, "learning_rate": 9.989830262998598e-05, "loss": 0.0085, "step": 6930 }, { "epoch": 73.05263157894737, "grad_norm": 0.1269829273223877, "learning_rate": 9.989724585410259e-05, "loss": 0.0084, "step": 6940 }, { "epoch": 73.15789473684211, "grad_norm": 0.14643998444080353, "learning_rate": 9.989618362153139e-05, "loss": 0.0087, "step": 6950 }, { "epoch": 73.26315789473684, "grad_norm": 0.15061350166797638, "learning_rate": 9.989511593238859e-05, "loss": 0.0096, "step": 6960 }, { "epoch": 73.36842105263158, "grad_norm": 0.17344336211681366, "learning_rate": 9.98940427867909e-05, "loss": 0.0093, "step": 6970 }, { "epoch": 73.47368421052632, "grad_norm": 0.14135925471782684, "learning_rate": 9.989296418485573e-05, "loss": 0.0095, "step": 6980 }, { "epoch": 73.57894736842105, "grad_norm": 0.14211228489875793, "learning_rate": 9.989188012670101e-05, "loss": 0.0096, "step": 6990 }, { "epoch": 73.6842105263158, "grad_norm": 0.12507472932338715, "learning_rate": 9.989079061244528e-05, "loss": 0.0089, "step": 7000 }, { "epoch": 73.78947368421052, "grad_norm": 0.1371171772480011, "learning_rate": 9.988969564220769e-05, "loss": 0.0088, "step": 7010 }, { "epoch": 73.89473684210526, "grad_norm": 0.1568659394979477, "learning_rate": 9.988859521610801e-05, "loss": 0.0072, "step": 7020 }, { "epoch": 74.0, "grad_norm": 0.2104596197605133, "learning_rate": 9.988748933426656e-05, "loss": 0.0105, "step": 7030 }, { "epoch": 74.10526315789474, "grad_norm": 0.13843034207820892, "learning_rate": 9.988637799680428e-05, "loss": 0.0093, "step": 7040 }, { "epoch": 74.21052631578948, "grad_norm": 0.11731021851301193, "learning_rate": 9.98852612038427e-05, "loss": 0.009, "step": 7050 }, { "epoch": 74.3157894736842, "grad_norm": 0.1433178186416626, "learning_rate": 9.988413895550397e-05, "loss": 0.0095, "step": 7060 }, { "epoch": 74.42105263157895, "grad_norm": 0.10954952239990234, "learning_rate": 9.98830112519108e-05, "loss": 0.0093, "step": 7070 }, { "epoch": 74.52631578947368, "grad_norm": 0.10990846902132034, "learning_rate": 9.98818780931865e-05, "loss": 0.0095, "step": 7080 }, { "epoch": 74.63157894736842, "grad_norm": 0.17556770145893097, "learning_rate": 9.988073947945502e-05, "loss": 0.0093, "step": 7090 }, { "epoch": 74.73684210526316, "grad_norm": 0.14783655107021332, "learning_rate": 9.987959541084087e-05, "loss": 0.009, "step": 7100 }, { "epoch": 74.84210526315789, "grad_norm": 0.10949458181858063, "learning_rate": 9.987844588746915e-05, "loss": 0.0072, "step": 7110 }, { "epoch": 74.94736842105263, "grad_norm": 0.12496960908174515, "learning_rate": 9.987729090946558e-05, "loss": 0.0096, "step": 7120 }, { "epoch": 75.05263157894737, "grad_norm": 0.15738287568092346, "learning_rate": 9.987613047695647e-05, "loss": 0.0079, "step": 7130 }, { "epoch": 75.15789473684211, "grad_norm": 0.10422225296497345, "learning_rate": 9.987496459006871e-05, "loss": 0.0086, "step": 7140 }, { "epoch": 75.26315789473684, "grad_norm": 0.16067740321159363, "learning_rate": 9.987379324892982e-05, "loss": 0.0084, "step": 7150 }, { "epoch": 75.36842105263158, "grad_norm": 0.14161574840545654, "learning_rate": 9.987261645366788e-05, "loss": 0.0078, "step": 7160 }, { "epoch": 75.47368421052632, "grad_norm": 0.1681891232728958, "learning_rate": 9.987143420441158e-05, "loss": 0.008, "step": 7170 }, { "epoch": 75.57894736842105, "grad_norm": 0.16630128026008606, "learning_rate": 9.987024650129022e-05, "loss": 0.0086, "step": 7180 }, { "epoch": 75.6842105263158, "grad_norm": 0.13361284136772156, "learning_rate": 9.986905334443368e-05, "loss": 0.0075, "step": 7190 }, { "epoch": 75.78947368421052, "grad_norm": 0.11533629894256592, "learning_rate": 9.986785473397245e-05, "loss": 0.0087, "step": 7200 }, { "epoch": 75.89473684210526, "grad_norm": 0.11364570260047913, "learning_rate": 9.98666506700376e-05, "loss": 0.0081, "step": 7210 }, { "epoch": 76.0, "grad_norm": 0.178251713514328, "learning_rate": 9.986544115276081e-05, "loss": 0.0077, "step": 7220 }, { "epoch": 76.10526315789474, "grad_norm": 0.12085533887147903, "learning_rate": 9.986422618227433e-05, "loss": 0.0087, "step": 7230 }, { "epoch": 76.21052631578948, "grad_norm": 0.13704465329647064, "learning_rate": 9.986300575871106e-05, "loss": 0.008, "step": 7240 }, { "epoch": 76.3157894736842, "grad_norm": 0.13807828724384308, "learning_rate": 9.986177988220444e-05, "loss": 0.0096, "step": 7250 }, { "epoch": 76.42105263157895, "grad_norm": 0.16173753142356873, "learning_rate": 9.986054855288856e-05, "loss": 0.0089, "step": 7260 }, { "epoch": 76.52631578947368, "grad_norm": 0.11265914142131805, "learning_rate": 9.985931177089802e-05, "loss": 0.0084, "step": 7270 }, { "epoch": 76.63157894736842, "grad_norm": 0.13597887754440308, "learning_rate": 9.985806953636814e-05, "loss": 0.0094, "step": 7280 }, { "epoch": 76.73684210526316, "grad_norm": 0.1352846473455429, "learning_rate": 9.985682184943471e-05, "loss": 0.0089, "step": 7290 }, { "epoch": 76.84210526315789, "grad_norm": 0.13723209500312805, "learning_rate": 9.98555687102342e-05, "loss": 0.0096, "step": 7300 }, { "epoch": 76.94736842105263, "grad_norm": 0.1286899596452713, "learning_rate": 9.985431011890367e-05, "loss": 0.009, "step": 7310 }, { "epoch": 77.05263157894737, "grad_norm": 0.14199331402778625, "learning_rate": 9.985304607558075e-05, "loss": 0.0097, "step": 7320 }, { "epoch": 77.15789473684211, "grad_norm": 0.17284376919269562, "learning_rate": 9.985177658040364e-05, "loss": 0.009, "step": 7330 }, { "epoch": 77.26315789473684, "grad_norm": 0.16063302755355835, "learning_rate": 9.985050163351119e-05, "loss": 0.0095, "step": 7340 }, { "epoch": 77.36842105263158, "grad_norm": 0.19101262092590332, "learning_rate": 9.984922123504286e-05, "loss": 0.0092, "step": 7350 }, { "epoch": 77.47368421052632, "grad_norm": 0.12913529574871063, "learning_rate": 9.984793538513862e-05, "loss": 0.0085, "step": 7360 }, { "epoch": 77.57894736842105, "grad_norm": 0.10988644510507584, "learning_rate": 9.984664408393912e-05, "loss": 0.0084, "step": 7370 }, { "epoch": 77.6842105263158, "grad_norm": 0.11190415173768997, "learning_rate": 9.984534733158556e-05, "loss": 0.0087, "step": 7380 }, { "epoch": 77.78947368421052, "grad_norm": 0.12887075543403625, "learning_rate": 9.984404512821977e-05, "loss": 0.0082, "step": 7390 }, { "epoch": 77.89473684210526, "grad_norm": 0.14344915747642517, "learning_rate": 9.984273747398411e-05, "loss": 0.0085, "step": 7400 }, { "epoch": 78.0, "grad_norm": 0.16560892760753632, "learning_rate": 9.984142436902165e-05, "loss": 0.0081, "step": 7410 }, { "epoch": 78.10526315789474, "grad_norm": 0.15489619970321655, "learning_rate": 9.984010581347596e-05, "loss": 0.008, "step": 7420 }, { "epoch": 78.21052631578948, "grad_norm": 0.11268732696771622, "learning_rate": 9.983878180749121e-05, "loss": 0.0093, "step": 7430 }, { "epoch": 78.3157894736842, "grad_norm": 0.12938103079795837, "learning_rate": 9.983745235121222e-05, "loss": 0.0083, "step": 7440 }, { "epoch": 78.42105263157895, "grad_norm": 0.16002489626407623, "learning_rate": 9.983611744478438e-05, "loss": 0.009, "step": 7450 }, { "epoch": 78.52631578947368, "grad_norm": 0.14251355826854706, "learning_rate": 9.983477708835365e-05, "loss": 0.0083, "step": 7460 }, { "epoch": 78.63157894736842, "grad_norm": 0.14445307850837708, "learning_rate": 9.983343128206664e-05, "loss": 0.0092, "step": 7470 }, { "epoch": 78.73684210526316, "grad_norm": 0.12826554477214813, "learning_rate": 9.983208002607049e-05, "loss": 0.0086, "step": 7480 }, { "epoch": 78.84210526315789, "grad_norm": 0.11728822439908981, "learning_rate": 9.9830723320513e-05, "loss": 0.0078, "step": 7490 }, { "epoch": 78.94736842105263, "grad_norm": 0.11619746685028076, "learning_rate": 9.982936116554254e-05, "loss": 0.0085, "step": 7500 }, { "epoch": 79.05263157894737, "grad_norm": 0.21058784425258636, "learning_rate": 9.982799356130803e-05, "loss": 0.01, "step": 7510 }, { "epoch": 79.15789473684211, "grad_norm": 0.18139421939849854, "learning_rate": 9.982662050795908e-05, "loss": 0.0088, "step": 7520 }, { "epoch": 79.26315789473684, "grad_norm": 0.15634939074516296, "learning_rate": 9.982524200564583e-05, "loss": 0.0091, "step": 7530 }, { "epoch": 79.36842105263158, "grad_norm": 0.14992967247962952, "learning_rate": 9.982385805451901e-05, "loss": 0.009, "step": 7540 }, { "epoch": 79.47368421052632, "grad_norm": 0.1571742594242096, "learning_rate": 9.982246865472998e-05, "loss": 0.0086, "step": 7550 }, { "epoch": 79.57894736842105, "grad_norm": 0.11760490387678146, "learning_rate": 9.982107380643069e-05, "loss": 0.0078, "step": 7560 }, { "epoch": 79.6842105263158, "grad_norm": 0.12865082919597626, "learning_rate": 9.981967350977368e-05, "loss": 0.0084, "step": 7570 }, { "epoch": 79.78947368421052, "grad_norm": 0.11204690486192703, "learning_rate": 9.981826776491208e-05, "loss": 0.009, "step": 7580 }, { "epoch": 79.89473684210526, "grad_norm": 0.12357516586780548, "learning_rate": 9.98168565719996e-05, "loss": 0.009, "step": 7590 }, { "epoch": 80.0, "grad_norm": 0.16107064485549927, "learning_rate": 9.98154399311906e-05, "loss": 0.0081, "step": 7600 }, { "epoch": 80.10526315789474, "grad_norm": 0.10917758196592331, "learning_rate": 9.981401784263997e-05, "loss": 0.0084, "step": 7610 }, { "epoch": 80.21052631578948, "grad_norm": 0.13071192800998688, "learning_rate": 9.981259030650326e-05, "loss": 0.0086, "step": 7620 }, { "epoch": 80.3157894736842, "grad_norm": 0.1876363456249237, "learning_rate": 9.981115732293655e-05, "loss": 0.0084, "step": 7630 }, { "epoch": 80.42105263157895, "grad_norm": 0.18205945193767548, "learning_rate": 9.980971889209659e-05, "loss": 0.0087, "step": 7640 }, { "epoch": 80.52631578947368, "grad_norm": 0.14961877465248108, "learning_rate": 9.980827501414064e-05, "loss": 0.0086, "step": 7650 }, { "epoch": 80.63157894736842, "grad_norm": 0.14722533524036407, "learning_rate": 9.980682568922663e-05, "loss": 0.0098, "step": 7660 }, { "epoch": 80.73684210526316, "grad_norm": 0.12436176091432571, "learning_rate": 9.980537091751304e-05, "loss": 0.008, "step": 7670 }, { "epoch": 80.84210526315789, "grad_norm": 0.09325896948575974, "learning_rate": 9.980391069915897e-05, "loss": 0.0074, "step": 7680 }, { "epoch": 80.94736842105263, "grad_norm": 0.14335604012012482, "learning_rate": 9.98024450343241e-05, "loss": 0.0074, "step": 7690 }, { "epoch": 81.05263157894737, "grad_norm": 0.12250857055187225, "learning_rate": 9.980097392316872e-05, "loss": 0.0081, "step": 7700 }, { "epoch": 81.15789473684211, "grad_norm": 0.13732917606830597, "learning_rate": 9.97994973658537e-05, "loss": 0.0083, "step": 7710 }, { "epoch": 81.26315789473684, "grad_norm": 0.1199154183268547, "learning_rate": 9.979801536254054e-05, "loss": 0.0068, "step": 7720 }, { "epoch": 81.36842105263158, "grad_norm": 0.1330307126045227, "learning_rate": 9.979652791339127e-05, "loss": 0.0081, "step": 7730 }, { "epoch": 81.47368421052632, "grad_norm": 0.15670964121818542, "learning_rate": 9.97950350185686e-05, "loss": 0.0086, "step": 7740 }, { "epoch": 81.57894736842105, "grad_norm": 0.1559881716966629, "learning_rate": 9.979353667823574e-05, "loss": 0.0078, "step": 7750 }, { "epoch": 81.6842105263158, "grad_norm": 0.13924287259578705, "learning_rate": 9.979203289255658e-05, "loss": 0.0072, "step": 7760 }, { "epoch": 81.78947368421052, "grad_norm": 0.11922169476747513, "learning_rate": 9.979052366169557e-05, "loss": 0.0078, "step": 7770 }, { "epoch": 81.89473684210526, "grad_norm": 0.15588219463825226, "learning_rate": 9.978900898581775e-05, "loss": 0.0076, "step": 7780 }, { "epoch": 82.0, "grad_norm": 0.17651787400245667, "learning_rate": 9.978748886508875e-05, "loss": 0.008, "step": 7790 }, { "epoch": 82.10526315789474, "grad_norm": 0.1503012329339981, "learning_rate": 9.978596329967484e-05, "loss": 0.0087, "step": 7800 }, { "epoch": 82.21052631578948, "grad_norm": 0.11930978298187256, "learning_rate": 9.978443228974284e-05, "loss": 0.0081, "step": 7810 }, { "epoch": 82.3157894736842, "grad_norm": 0.12284386903047562, "learning_rate": 9.978289583546015e-05, "loss": 0.0081, "step": 7820 }, { "epoch": 82.42105263157895, "grad_norm": 0.15306170284748077, "learning_rate": 9.978135393699484e-05, "loss": 0.0082, "step": 7830 }, { "epoch": 82.52631578947368, "grad_norm": 0.11916281282901764, "learning_rate": 9.977980659451548e-05, "loss": 0.0084, "step": 7840 }, { "epoch": 82.63157894736842, "grad_norm": 0.13465888798236847, "learning_rate": 9.977825380819135e-05, "loss": 0.0077, "step": 7850 }, { "epoch": 82.73684210526316, "grad_norm": 0.14859779179096222, "learning_rate": 9.97766955781922e-05, "loss": 0.0084, "step": 7860 }, { "epoch": 82.84210526315789, "grad_norm": 0.12754911184310913, "learning_rate": 9.977513190468848e-05, "loss": 0.0072, "step": 7870 }, { "epoch": 82.94736842105263, "grad_norm": 0.14732849597930908, "learning_rate": 9.977356278785116e-05, "loss": 0.0085, "step": 7880 }, { "epoch": 83.05263157894737, "grad_norm": 0.15942969918251038, "learning_rate": 9.977198822785184e-05, "loss": 0.0077, "step": 7890 }, { "epoch": 83.15789473684211, "grad_norm": 0.14577117562294006, "learning_rate": 9.977040822486273e-05, "loss": 0.0082, "step": 7900 }, { "epoch": 83.26315789473684, "grad_norm": 0.15686805546283722, "learning_rate": 9.97688227790566e-05, "loss": 0.0083, "step": 7910 }, { "epoch": 83.36842105263158, "grad_norm": 0.14981985092163086, "learning_rate": 9.976723189060684e-05, "loss": 0.0082, "step": 7920 }, { "epoch": 83.47368421052632, "grad_norm": 0.11101488023996353, "learning_rate": 9.976563555968742e-05, "loss": 0.0078, "step": 7930 }, { "epoch": 83.57894736842105, "grad_norm": 0.14810404181480408, "learning_rate": 9.976403378647292e-05, "loss": 0.0078, "step": 7940 }, { "epoch": 83.6842105263158, "grad_norm": 0.1269906461238861, "learning_rate": 9.97624265711385e-05, "loss": 0.008, "step": 7950 }, { "epoch": 83.78947368421052, "grad_norm": 0.09984484314918518, "learning_rate": 9.976081391385993e-05, "loss": 0.0075, "step": 7960 }, { "epoch": 83.89473684210526, "grad_norm": 0.15317031741142273, "learning_rate": 9.975919581481356e-05, "loss": 0.0082, "step": 7970 }, { "epoch": 84.0, "grad_norm": 0.13129134476184845, "learning_rate": 9.975757227417634e-05, "loss": 0.0071, "step": 7980 }, { "epoch": 84.10526315789474, "grad_norm": 0.14871802926063538, "learning_rate": 9.975594329212586e-05, "loss": 0.0077, "step": 7990 }, { "epoch": 84.21052631578948, "grad_norm": 0.1346968114376068, "learning_rate": 9.97543088688402e-05, "loss": 0.008, "step": 8000 }, { "epoch": 84.3157894736842, "grad_norm": 0.11748415976762772, "learning_rate": 9.975266900449814e-05, "loss": 0.0087, "step": 8010 }, { "epoch": 84.42105263157895, "grad_norm": 0.11455702781677246, "learning_rate": 9.975102369927898e-05, "loss": 0.0086, "step": 8020 }, { "epoch": 84.52631578947368, "grad_norm": 0.15419349074363708, "learning_rate": 9.974937295336269e-05, "loss": 0.0085, "step": 8030 }, { "epoch": 84.63157894736842, "grad_norm": 0.13644053041934967, "learning_rate": 9.974771676692975e-05, "loss": 0.0081, "step": 8040 }, { "epoch": 84.73684210526316, "grad_norm": 0.1347365528345108, "learning_rate": 9.974605514016131e-05, "loss": 0.0087, "step": 8050 }, { "epoch": 84.84210526315789, "grad_norm": 0.13503490388393402, "learning_rate": 9.974438807323907e-05, "loss": 0.0078, "step": 8060 }, { "epoch": 84.94736842105263, "grad_norm": 0.11858990043401718, "learning_rate": 9.974271556634535e-05, "loss": 0.0091, "step": 8070 }, { "epoch": 85.05263157894737, "grad_norm": 0.14185172319412231, "learning_rate": 9.974103761966302e-05, "loss": 0.0075, "step": 8080 }, { "epoch": 85.15789473684211, "grad_norm": 0.1270749419927597, "learning_rate": 9.973935423337563e-05, "loss": 0.0085, "step": 8090 }, { "epoch": 85.26315789473684, "grad_norm": 0.11450864374637604, "learning_rate": 9.973766540766722e-05, "loss": 0.008, "step": 8100 }, { "epoch": 85.36842105263158, "grad_norm": 0.11955317109823227, "learning_rate": 9.97359711427225e-05, "loss": 0.008, "step": 8110 }, { "epoch": 85.47368421052632, "grad_norm": 0.1286165416240692, "learning_rate": 9.973427143872677e-05, "loss": 0.0078, "step": 8120 }, { "epoch": 85.57894736842105, "grad_norm": 0.13018307089805603, "learning_rate": 9.973256629586589e-05, "loss": 0.0076, "step": 8130 }, { "epoch": 85.6842105263158, "grad_norm": 0.11989453434944153, "learning_rate": 9.973085571432632e-05, "loss": 0.0089, "step": 8140 }, { "epoch": 85.78947368421052, "grad_norm": 0.11375569552183151, "learning_rate": 9.972913969429513e-05, "loss": 0.0076, "step": 8150 }, { "epoch": 85.89473684210526, "grad_norm": 0.15104949474334717, "learning_rate": 9.972741823596e-05, "loss": 0.0084, "step": 8160 }, { "epoch": 86.0, "grad_norm": 0.1864442080259323, "learning_rate": 9.972569133950917e-05, "loss": 0.0093, "step": 8170 }, { "epoch": 86.10526315789474, "grad_norm": 0.13496601581573486, "learning_rate": 9.972395900513151e-05, "loss": 0.0079, "step": 8180 }, { "epoch": 86.21052631578948, "grad_norm": 0.13919402658939362, "learning_rate": 9.972222123301645e-05, "loss": 0.009, "step": 8190 }, { "epoch": 86.3157894736842, "grad_norm": 0.1281871795654297, "learning_rate": 9.972047802335403e-05, "loss": 0.0083, "step": 8200 }, { "epoch": 86.42105263157895, "grad_norm": 0.11563489586114883, "learning_rate": 9.971872937633488e-05, "loss": 0.0085, "step": 8210 }, { "epoch": 86.52631578947368, "grad_norm": 0.14409105479717255, "learning_rate": 9.971697529215024e-05, "loss": 0.0086, "step": 8220 }, { "epoch": 86.63157894736842, "grad_norm": 0.126327782869339, "learning_rate": 9.971521577099192e-05, "loss": 0.0079, "step": 8230 }, { "epoch": 86.73684210526316, "grad_norm": 0.16198782622814178, "learning_rate": 9.971345081305236e-05, "loss": 0.0081, "step": 8240 }, { "epoch": 86.84210526315789, "grad_norm": 0.12818509340286255, "learning_rate": 9.971168041852456e-05, "loss": 0.0077, "step": 8250 }, { "epoch": 86.94736842105263, "grad_norm": 0.1363496631383896, "learning_rate": 9.970990458760215e-05, "loss": 0.009, "step": 8260 }, { "epoch": 87.05263157894737, "grad_norm": 0.15016278624534607, "learning_rate": 9.970812332047929e-05, "loss": 0.0071, "step": 8270 }, { "epoch": 87.15789473684211, "grad_norm": 0.13040857017040253, "learning_rate": 9.97063366173508e-05, "loss": 0.0088, "step": 8280 }, { "epoch": 87.26315789473684, "grad_norm": 0.13095791637897491, "learning_rate": 9.970454447841207e-05, "loss": 0.0083, "step": 8290 }, { "epoch": 87.36842105263158, "grad_norm": 0.12901216745376587, "learning_rate": 9.970274690385909e-05, "loss": 0.0071, "step": 8300 }, { "epoch": 87.47368421052632, "grad_norm": 0.10863874107599258, "learning_rate": 9.970094389388844e-05, "loss": 0.008, "step": 8310 }, { "epoch": 87.57894736842105, "grad_norm": 0.14334727823734283, "learning_rate": 9.969913544869728e-05, "loss": 0.0074, "step": 8320 }, { "epoch": 87.6842105263158, "grad_norm": 0.12013314664363861, "learning_rate": 9.96973215684834e-05, "loss": 0.0068, "step": 8330 }, { "epoch": 87.78947368421052, "grad_norm": 0.13201214373111725, "learning_rate": 9.969550225344513e-05, "loss": 0.0078, "step": 8340 }, { "epoch": 87.89473684210526, "grad_norm": 0.12228573113679886, "learning_rate": 9.969367750378147e-05, "loss": 0.0076, "step": 8350 }, { "epoch": 88.0, "grad_norm": 0.20834249258041382, "learning_rate": 9.969184731969194e-05, "loss": 0.0075, "step": 8360 }, { "epoch": 88.10526315789474, "grad_norm": 0.135776087641716, "learning_rate": 9.96900117013767e-05, "loss": 0.0086, "step": 8370 }, { "epoch": 88.21052631578948, "grad_norm": 0.1218498945236206, "learning_rate": 9.96881706490365e-05, "loss": 0.0087, "step": 8380 }, { "epoch": 88.3157894736842, "grad_norm": 0.10732058435678482, "learning_rate": 9.968632416287265e-05, "loss": 0.0078, "step": 8390 }, { "epoch": 88.42105263157895, "grad_norm": 0.18059632182121277, "learning_rate": 9.96844722430871e-05, "loss": 0.0084, "step": 8400 }, { "epoch": 88.52631578947368, "grad_norm": 0.12333998829126358, "learning_rate": 9.968261488988235e-05, "loss": 0.0072, "step": 8410 }, { "epoch": 88.63157894736842, "grad_norm": 0.10359501093626022, "learning_rate": 9.968075210346155e-05, "loss": 0.0079, "step": 8420 }, { "epoch": 88.73684210526316, "grad_norm": 0.15271081030368805, "learning_rate": 9.967888388402839e-05, "loss": 0.009, "step": 8430 }, { "epoch": 88.84210526315789, "grad_norm": 0.10777752101421356, "learning_rate": 9.967701023178717e-05, "loss": 0.0081, "step": 8440 }, { "epoch": 88.94736842105263, "grad_norm": 0.13953076303005219, "learning_rate": 9.967513114694282e-05, "loss": 0.008, "step": 8450 }, { "epoch": 89.05263157894737, "grad_norm": 0.14363528788089752, "learning_rate": 9.967324662970079e-05, "loss": 0.008, "step": 8460 }, { "epoch": 89.15789473684211, "grad_norm": 0.11896394938230515, "learning_rate": 9.96713566802672e-05, "loss": 0.0079, "step": 8470 }, { "epoch": 89.26315789473684, "grad_norm": 0.17671330273151398, "learning_rate": 9.966946129884873e-05, "loss": 0.008, "step": 8480 }, { "epoch": 89.36842105263158, "grad_norm": 0.12953107059001923, "learning_rate": 9.966756048565265e-05, "loss": 0.0087, "step": 8490 }, { "epoch": 89.47368421052632, "grad_norm": 0.1410985291004181, "learning_rate": 9.966565424088681e-05, "loss": 0.008, "step": 8500 }, { "epoch": 89.57894736842105, "grad_norm": 0.10633615404367447, "learning_rate": 9.96637425647597e-05, "loss": 0.0083, "step": 8510 }, { "epoch": 89.6842105263158, "grad_norm": 0.13133376836776733, "learning_rate": 9.966182545748038e-05, "loss": 0.0078, "step": 8520 }, { "epoch": 89.78947368421052, "grad_norm": 0.15206265449523926, "learning_rate": 9.96599029192585e-05, "loss": 0.008, "step": 8530 }, { "epoch": 89.89473684210526, "grad_norm": 0.15031924843788147, "learning_rate": 9.965797495030428e-05, "loss": 0.0088, "step": 8540 }, { "epoch": 90.0, "grad_norm": 0.19861438870429993, "learning_rate": 9.96560415508286e-05, "loss": 0.0075, "step": 8550 }, { "epoch": 90.10526315789474, "grad_norm": 0.14780353009700775, "learning_rate": 9.965410272104286e-05, "loss": 0.0081, "step": 8560 }, { "epoch": 90.21052631578948, "grad_norm": 0.10752233117818832, "learning_rate": 9.96521584611591e-05, "loss": 0.0076, "step": 8570 }, { "epoch": 90.3157894736842, "grad_norm": 0.10518690943717957, "learning_rate": 9.965020877138994e-05, "loss": 0.0089, "step": 8580 }, { "epoch": 90.42105263157895, "grad_norm": 0.10953518748283386, "learning_rate": 9.964825365194861e-05, "loss": 0.0072, "step": 8590 }, { "epoch": 90.52631578947368, "grad_norm": 0.15428170561790466, "learning_rate": 9.96462931030489e-05, "loss": 0.008, "step": 8600 }, { "epoch": 90.63157894736842, "grad_norm": 0.14839759469032288, "learning_rate": 9.96443271249052e-05, "loss": 0.0074, "step": 8610 }, { "epoch": 90.73684210526316, "grad_norm": 0.11179295182228088, "learning_rate": 9.964235571773255e-05, "loss": 0.008, "step": 8620 }, { "epoch": 90.84210526315789, "grad_norm": 0.11329787224531174, "learning_rate": 9.96403788817465e-05, "loss": 0.0074, "step": 8630 }, { "epoch": 90.94736842105263, "grad_norm": 0.1407729536294937, "learning_rate": 9.963839661716325e-05, "loss": 0.0083, "step": 8640 }, { "epoch": 91.05263157894737, "grad_norm": 0.14744658768177032, "learning_rate": 9.963640892419958e-05, "loss": 0.0099, "step": 8650 }, { "epoch": 91.15789473684211, "grad_norm": 0.13620629906654358, "learning_rate": 9.963441580307286e-05, "loss": 0.0088, "step": 8660 }, { "epoch": 91.26315789473684, "grad_norm": 0.15325023233890533, "learning_rate": 9.963241725400104e-05, "loss": 0.0085, "step": 8670 }, { "epoch": 91.36842105263158, "grad_norm": 0.12003758549690247, "learning_rate": 9.963041327720271e-05, "loss": 0.0073, "step": 8680 }, { "epoch": 91.47368421052632, "grad_norm": 0.11186318844556808, "learning_rate": 9.962840387289697e-05, "loss": 0.0082, "step": 8690 }, { "epoch": 91.57894736842105, "grad_norm": 0.1471473127603531, "learning_rate": 9.962638904130363e-05, "loss": 0.008, "step": 8700 }, { "epoch": 91.6842105263158, "grad_norm": 0.11337539553642273, "learning_rate": 9.962436878264298e-05, "loss": 0.0083, "step": 8710 }, { "epoch": 91.78947368421052, "grad_norm": 0.17804725468158722, "learning_rate": 9.962234309713598e-05, "loss": 0.0088, "step": 8720 }, { "epoch": 91.89473684210526, "grad_norm": 0.1514543890953064, "learning_rate": 9.962031198500414e-05, "loss": 0.0079, "step": 8730 }, { "epoch": 92.0, "grad_norm": 0.12637931108474731, "learning_rate": 9.961827544646958e-05, "loss": 0.0067, "step": 8740 }, { "epoch": 92.10526315789474, "grad_norm": 0.13061395287513733, "learning_rate": 9.961623348175501e-05, "loss": 0.0081, "step": 8750 }, { "epoch": 92.21052631578948, "grad_norm": 0.15067316591739655, "learning_rate": 9.961418609108377e-05, "loss": 0.0091, "step": 8760 }, { "epoch": 92.3157894736842, "grad_norm": 0.12806618213653564, "learning_rate": 9.961213327467971e-05, "loss": 0.0082, "step": 8770 }, { "epoch": 92.42105263157895, "grad_norm": 0.12306497246026993, "learning_rate": 9.961007503276736e-05, "loss": 0.0079, "step": 8780 }, { "epoch": 92.52631578947368, "grad_norm": 0.1412457376718521, "learning_rate": 9.960801136557179e-05, "loss": 0.0075, "step": 8790 }, { "epoch": 92.63157894736842, "grad_norm": 0.14187520742416382, "learning_rate": 9.960594227331866e-05, "loss": 0.008, "step": 8800 }, { "epoch": 92.73684210526316, "grad_norm": 0.1349034160375595, "learning_rate": 9.960386775623429e-05, "loss": 0.008, "step": 8810 }, { "epoch": 92.84210526315789, "grad_norm": 0.16166633367538452, "learning_rate": 9.96017878145455e-05, "loss": 0.0075, "step": 8820 }, { "epoch": 92.94736842105263, "grad_norm": 0.12446432560682297, "learning_rate": 9.959970244847977e-05, "loss": 0.0083, "step": 8830 }, { "epoch": 93.05263157894737, "grad_norm": 0.11329825967550278, "learning_rate": 9.959761165826518e-05, "loss": 0.0075, "step": 8840 }, { "epoch": 93.15789473684211, "grad_norm": 0.07822027802467346, "learning_rate": 9.959551544413033e-05, "loss": 0.0068, "step": 8850 }, { "epoch": 93.26315789473684, "grad_norm": 0.12507620453834534, "learning_rate": 9.959341380630448e-05, "loss": 0.0069, "step": 8860 }, { "epoch": 93.36842105263158, "grad_norm": 0.12959857285022736, "learning_rate": 9.959130674501746e-05, "loss": 0.0072, "step": 8870 }, { "epoch": 93.47368421052632, "grad_norm": 0.10857640206813812, "learning_rate": 9.958919426049968e-05, "loss": 0.007, "step": 8880 }, { "epoch": 93.57894736842105, "grad_norm": 0.10622160881757736, "learning_rate": 9.958707635298219e-05, "loss": 0.0081, "step": 8890 }, { "epoch": 93.6842105263158, "grad_norm": 0.14386439323425293, "learning_rate": 9.958495302269657e-05, "loss": 0.0075, "step": 8900 }, { "epoch": 93.78947368421052, "grad_norm": 0.09523274004459381, "learning_rate": 9.958282426987503e-05, "loss": 0.0078, "step": 8910 }, { "epoch": 93.89473684210526, "grad_norm": 0.0994139313697815, "learning_rate": 9.95806900947504e-05, "loss": 0.0081, "step": 8920 }, { "epoch": 94.0, "grad_norm": 0.23948563635349274, "learning_rate": 9.957855049755604e-05, "loss": 0.0086, "step": 8930 }, { "epoch": 94.10526315789474, "grad_norm": 0.1514819711446762, "learning_rate": 9.957640547852593e-05, "loss": 0.0076, "step": 8940 }, { "epoch": 94.21052631578948, "grad_norm": 0.13165780901908875, "learning_rate": 9.957425503789466e-05, "loss": 0.0085, "step": 8950 }, { "epoch": 94.3157894736842, "grad_norm": 0.12934722006320953, "learning_rate": 9.957209917589738e-05, "loss": 0.008, "step": 8960 }, { "epoch": 94.42105263157895, "grad_norm": 0.11375588178634644, "learning_rate": 9.956993789276987e-05, "loss": 0.0085, "step": 8970 }, { "epoch": 94.52631578947368, "grad_norm": 0.13487444818019867, "learning_rate": 9.956777118874847e-05, "loss": 0.0074, "step": 8980 }, { "epoch": 94.63157894736842, "grad_norm": 0.11492281407117844, "learning_rate": 9.956559906407016e-05, "loss": 0.0079, "step": 8990 }, { "epoch": 94.73684210526316, "grad_norm": 0.10866376757621765, "learning_rate": 9.956342151897245e-05, "loss": 0.007, "step": 9000 }, { "epoch": 94.84210526315789, "grad_norm": 0.10281024873256683, "learning_rate": 9.956123855369346e-05, "loss": 0.0072, "step": 9010 }, { "epoch": 94.94736842105263, "grad_norm": 0.11325273662805557, "learning_rate": 9.955905016847196e-05, "loss": 0.0074, "step": 9020 }, { "epoch": 95.05263157894737, "grad_norm": 0.1340373307466507, "learning_rate": 9.955685636354723e-05, "loss": 0.0072, "step": 9030 }, { "epoch": 95.15789473684211, "grad_norm": 0.13346648216247559, "learning_rate": 9.95546571391592e-05, "loss": 0.0073, "step": 9040 }, { "epoch": 95.26315789473684, "grad_norm": 0.11400440335273743, "learning_rate": 9.955245249554837e-05, "loss": 0.0083, "step": 9050 }, { "epoch": 95.36842105263158, "grad_norm": 0.11283431947231293, "learning_rate": 9.955024243295582e-05, "loss": 0.0078, "step": 9060 }, { "epoch": 95.47368421052632, "grad_norm": 0.13030850887298584, "learning_rate": 9.954802695162328e-05, "loss": 0.0086, "step": 9070 }, { "epoch": 95.57894736842105, "grad_norm": 0.10786012560129166, "learning_rate": 9.954580605179302e-05, "loss": 0.0073, "step": 9080 }, { "epoch": 95.6842105263158, "grad_norm": 0.13767044246196747, "learning_rate": 9.954357973370788e-05, "loss": 0.0083, "step": 9090 }, { "epoch": 95.78947368421052, "grad_norm": 0.12894177436828613, "learning_rate": 9.954134799761135e-05, "loss": 0.0071, "step": 9100 }, { "epoch": 95.89473684210526, "grad_norm": 0.11473215371370316, "learning_rate": 9.953911084374748e-05, "loss": 0.0066, "step": 9110 }, { "epoch": 96.0, "grad_norm": 0.18452368676662445, "learning_rate": 9.953686827236093e-05, "loss": 0.0076, "step": 9120 }, { "epoch": 96.10526315789474, "grad_norm": 0.1604478508234024, "learning_rate": 9.953462028369695e-05, "loss": 0.008, "step": 9130 }, { "epoch": 96.21052631578948, "grad_norm": 0.15613768994808197, "learning_rate": 9.953236687800136e-05, "loss": 0.0083, "step": 9140 }, { "epoch": 96.3157894736842, "grad_norm": 0.12644821405410767, "learning_rate": 9.95301080555206e-05, "loss": 0.0076, "step": 9150 }, { "epoch": 96.42105263157895, "grad_norm": 0.10778552293777466, "learning_rate": 9.952784381650171e-05, "loss": 0.0073, "step": 9160 }, { "epoch": 96.52631578947368, "grad_norm": 0.13216543197631836, "learning_rate": 9.952557416119226e-05, "loss": 0.0097, "step": 9170 }, { "epoch": 96.63157894736842, "grad_norm": 0.16271711885929108, "learning_rate": 9.95232990898405e-05, "loss": 0.0078, "step": 9180 }, { "epoch": 96.73684210526316, "grad_norm": 0.1228848472237587, "learning_rate": 9.95210186026952e-05, "loss": 0.0081, "step": 9190 }, { "epoch": 96.84210526315789, "grad_norm": 0.15679624676704407, "learning_rate": 9.951873270000576e-05, "loss": 0.0079, "step": 9200 }, { "epoch": 96.94736842105263, "grad_norm": 0.12882697582244873, "learning_rate": 9.951644138202216e-05, "loss": 0.0079, "step": 9210 }, { "epoch": 97.05263157894737, "grad_norm": 0.097397081553936, "learning_rate": 9.951414464899498e-05, "loss": 0.0077, "step": 9220 }, { "epoch": 97.15789473684211, "grad_norm": 0.11240321397781372, "learning_rate": 9.951184250117538e-05, "loss": 0.0092, "step": 9230 }, { "epoch": 97.26315789473684, "grad_norm": 0.09620057791471481, "learning_rate": 9.950953493881513e-05, "loss": 0.0073, "step": 9240 }, { "epoch": 97.36842105263158, "grad_norm": 0.13269703090190887, "learning_rate": 9.950722196216658e-05, "loss": 0.0073, "step": 9250 }, { "epoch": 97.47368421052632, "grad_norm": 0.09583780169487, "learning_rate": 9.950490357148265e-05, "loss": 0.0074, "step": 9260 }, { "epoch": 97.57894736842105, "grad_norm": 0.12728653848171234, "learning_rate": 9.950257976701692e-05, "loss": 0.0072, "step": 9270 }, { "epoch": 97.6842105263158, "grad_norm": 0.11766964942216873, "learning_rate": 9.950025054902348e-05, "loss": 0.0077, "step": 9280 }, { "epoch": 97.78947368421052, "grad_norm": 0.09859863668680191, "learning_rate": 9.949791591775706e-05, "loss": 0.008, "step": 9290 }, { "epoch": 97.89473684210526, "grad_norm": 0.141700878739357, "learning_rate": 9.949557587347298e-05, "loss": 0.0077, "step": 9300 }, { "epoch": 98.0, "grad_norm": 0.1933646947145462, "learning_rate": 9.949323041642713e-05, "loss": 0.0078, "step": 9310 }, { "epoch": 98.10526315789474, "grad_norm": 0.1343238353729248, "learning_rate": 9.949087954687602e-05, "loss": 0.0072, "step": 9320 }, { "epoch": 98.21052631578948, "grad_norm": 0.10723429918289185, "learning_rate": 9.948852326507672e-05, "loss": 0.0083, "step": 9330 }, { "epoch": 98.3157894736842, "grad_norm": 0.10216823220252991, "learning_rate": 9.948616157128694e-05, "loss": 0.0079, "step": 9340 }, { "epoch": 98.42105263157895, "grad_norm": 0.1374451071023941, "learning_rate": 9.948379446576493e-05, "loss": 0.0068, "step": 9350 }, { "epoch": 98.52631578947368, "grad_norm": 0.12377262860536575, "learning_rate": 9.948142194876952e-05, "loss": 0.007, "step": 9360 }, { "epoch": 98.63157894736842, "grad_norm": 0.09821200370788574, "learning_rate": 9.947904402056024e-05, "loss": 0.0082, "step": 9370 }, { "epoch": 98.73684210526316, "grad_norm": 0.14092206954956055, "learning_rate": 9.947666068139708e-05, "loss": 0.0071, "step": 9380 }, { "epoch": 98.84210526315789, "grad_norm": 0.16417473554611206, "learning_rate": 9.947427193154071e-05, "loss": 0.0078, "step": 9390 }, { "epoch": 98.94736842105263, "grad_norm": 0.12779264152050018, "learning_rate": 9.947187777125233e-05, "loss": 0.0081, "step": 9400 }, { "epoch": 99.05263157894737, "grad_norm": 0.12239699810743332, "learning_rate": 9.946947820079377e-05, "loss": 0.0075, "step": 9410 }, { "epoch": 99.15789473684211, "grad_norm": 0.13785751163959503, "learning_rate": 9.946707322042747e-05, "loss": 0.0076, "step": 9420 }, { "epoch": 99.26315789473684, "grad_norm": 0.10429288446903229, "learning_rate": 9.94646628304164e-05, "loss": 0.0074, "step": 9430 }, { "epoch": 99.36842105263158, "grad_norm": 0.11825589835643768, "learning_rate": 9.946224703102418e-05, "loss": 0.0082, "step": 9440 }, { "epoch": 99.47368421052632, "grad_norm": 0.10573185980319977, "learning_rate": 9.945982582251498e-05, "loss": 0.0084, "step": 9450 }, { "epoch": 99.57894736842105, "grad_norm": 0.10730963945388794, "learning_rate": 9.94573992051536e-05, "loss": 0.0078, "step": 9460 }, { "epoch": 99.6842105263158, "grad_norm": 0.1566423922777176, "learning_rate": 9.94549671792054e-05, "loss": 0.0086, "step": 9470 }, { "epoch": 99.78947368421052, "grad_norm": 0.13876153528690338, "learning_rate": 9.945252974493635e-05, "loss": 0.008, "step": 9480 }, { "epoch": 99.89473684210526, "grad_norm": 0.15409305691719055, "learning_rate": 9.9450086902613e-05, "loss": 0.0073, "step": 9490 }, { "epoch": 100.0, "grad_norm": 0.23904435336589813, "learning_rate": 9.944763865250248e-05, "loss": 0.0077, "step": 9500 }, { "epoch": 100.10526315789474, "grad_norm": 0.1453717201948166, "learning_rate": 9.944518499487254e-05, "loss": 0.0081, "step": 9510 }, { "epoch": 100.21052631578948, "grad_norm": 0.15188424289226532, "learning_rate": 9.944272592999151e-05, "loss": 0.0076, "step": 9520 }, { "epoch": 100.3157894736842, "grad_norm": 0.1291910707950592, "learning_rate": 9.94402614581283e-05, "loss": 0.0081, "step": 9530 }, { "epoch": 100.42105263157895, "grad_norm": 0.12359101325273514, "learning_rate": 9.943779157955244e-05, "loss": 0.0082, "step": 9540 }, { "epoch": 100.52631578947368, "grad_norm": 0.11946360766887665, "learning_rate": 9.943531629453403e-05, "loss": 0.0074, "step": 9550 }, { "epoch": 100.63157894736842, "grad_norm": 0.12296666949987411, "learning_rate": 9.943283560334375e-05, "loss": 0.0082, "step": 9560 }, { "epoch": 100.73684210526316, "grad_norm": 0.09563307464122772, "learning_rate": 9.943034950625288e-05, "loss": 0.0069, "step": 9570 }, { "epoch": 100.84210526315789, "grad_norm": 0.1433829367160797, "learning_rate": 9.942785800353332e-05, "loss": 0.0069, "step": 9580 }, { "epoch": 100.94736842105263, "grad_norm": 0.12942394614219666, "learning_rate": 9.942536109545751e-05, "loss": 0.0075, "step": 9590 }, { "epoch": 101.05263157894737, "grad_norm": 0.11944699287414551, "learning_rate": 9.942285878229853e-05, "loss": 0.0071, "step": 9600 }, { "epoch": 101.15789473684211, "grad_norm": 0.10882293432950974, "learning_rate": 9.942035106433001e-05, "loss": 0.0074, "step": 9610 }, { "epoch": 101.26315789473684, "grad_norm": 0.1583690196275711, "learning_rate": 9.94178379418262e-05, "loss": 0.0084, "step": 9620 }, { "epoch": 101.36842105263158, "grad_norm": 0.12637443840503693, "learning_rate": 9.941531941506194e-05, "loss": 0.0073, "step": 9630 }, { "epoch": 101.47368421052632, "grad_norm": 0.12616808712482452, "learning_rate": 9.941279548431263e-05, "loss": 0.0067, "step": 9640 }, { "epoch": 101.57894736842105, "grad_norm": 0.12600423395633698, "learning_rate": 9.941026614985431e-05, "loss": 0.0075, "step": 9650 }, { "epoch": 101.6842105263158, "grad_norm": 0.12598419189453125, "learning_rate": 9.940773141196357e-05, "loss": 0.0081, "step": 9660 }, { "epoch": 101.78947368421052, "grad_norm": 0.1447250247001648, "learning_rate": 9.94051912709176e-05, "loss": 0.0067, "step": 9670 }, { "epoch": 101.89473684210526, "grad_norm": 0.09377957880496979, "learning_rate": 9.940264572699421e-05, "loss": 0.007, "step": 9680 }, { "epoch": 102.0, "grad_norm": 0.14670619368553162, "learning_rate": 9.940009478047174e-05, "loss": 0.0078, "step": 9690 }, { "epoch": 102.10526315789474, "grad_norm": 0.11505894362926483, "learning_rate": 9.939753843162918e-05, "loss": 0.0073, "step": 9700 }, { "epoch": 102.21052631578948, "grad_norm": 0.14945942163467407, "learning_rate": 9.939497668074609e-05, "loss": 0.0072, "step": 9710 }, { "epoch": 102.3157894736842, "grad_norm": 0.10579783469438553, "learning_rate": 9.93924095281026e-05, "loss": 0.0085, "step": 9720 }, { "epoch": 102.42105263157895, "grad_norm": 0.10222389549016953, "learning_rate": 9.938983697397948e-05, "loss": 0.0073, "step": 9730 }, { "epoch": 102.52631578947368, "grad_norm": 0.10191529989242554, "learning_rate": 9.938725901865805e-05, "loss": 0.0067, "step": 9740 }, { "epoch": 102.63157894736842, "grad_norm": 0.1269174963235855, "learning_rate": 9.93846756624202e-05, "loss": 0.0072, "step": 9750 }, { "epoch": 102.73684210526316, "grad_norm": 0.10815367847681046, "learning_rate": 9.938208690554849e-05, "loss": 0.0073, "step": 9760 }, { "epoch": 102.84210526315789, "grad_norm": 0.1085016056895256, "learning_rate": 9.9379492748326e-05, "loss": 0.007, "step": 9770 }, { "epoch": 102.94736842105263, "grad_norm": 0.12447935342788696, "learning_rate": 9.937689319103641e-05, "loss": 0.0072, "step": 9780 }, { "epoch": 103.05263157894737, "grad_norm": 0.101992167532444, "learning_rate": 9.937428823396404e-05, "loss": 0.0073, "step": 9790 }, { "epoch": 103.15789473684211, "grad_norm": 0.11905419081449509, "learning_rate": 9.937167787739372e-05, "loss": 0.008, "step": 9800 }, { "epoch": 103.26315789473684, "grad_norm": 0.10430438071489334, "learning_rate": 9.936906212161095e-05, "loss": 0.0073, "step": 9810 }, { "epoch": 103.36842105263158, "grad_norm": 0.12381041795015335, "learning_rate": 9.936644096690176e-05, "loss": 0.0061, "step": 9820 }, { "epoch": 103.47368421052632, "grad_norm": 0.09457768499851227, "learning_rate": 9.936381441355282e-05, "loss": 0.0076, "step": 9830 }, { "epoch": 103.57894736842105, "grad_norm": 0.11449840664863586, "learning_rate": 9.936118246185136e-05, "loss": 0.0071, "step": 9840 }, { "epoch": 103.6842105263158, "grad_norm": 0.10861928761005402, "learning_rate": 9.935854511208518e-05, "loss": 0.0062, "step": 9850 }, { "epoch": 103.78947368421052, "grad_norm": 0.10320445895195007, "learning_rate": 9.935590236454272e-05, "loss": 0.0068, "step": 9860 }, { "epoch": 103.89473684210526, "grad_norm": 0.11636320501565933, "learning_rate": 9.935325421951298e-05, "loss": 0.0074, "step": 9870 }, { "epoch": 104.0, "grad_norm": 0.12424568831920624, "learning_rate": 9.935060067728557e-05, "loss": 0.0068, "step": 9880 }, { "epoch": 104.10526315789474, "grad_norm": 0.17984867095947266, "learning_rate": 9.934794173815067e-05, "loss": 0.0072, "step": 9890 }, { "epoch": 104.21052631578948, "grad_norm": 0.11422181874513626, "learning_rate": 9.934527740239906e-05, "loss": 0.0065, "step": 9900 }, { "epoch": 104.3157894736842, "grad_norm": 0.0972343236207962, "learning_rate": 9.934260767032209e-05, "loss": 0.0087, "step": 9910 }, { "epoch": 104.42105263157895, "grad_norm": 0.12529635429382324, "learning_rate": 9.933993254221172e-05, "loss": 0.0083, "step": 9920 }, { "epoch": 104.52631578947368, "grad_norm": 0.14177510142326355, "learning_rate": 9.933725201836053e-05, "loss": 0.0079, "step": 9930 }, { "epoch": 104.63157894736842, "grad_norm": 0.10650669038295746, "learning_rate": 9.933456609906162e-05, "loss": 0.0074, "step": 9940 }, { "epoch": 104.73684210526316, "grad_norm": 0.13711029291152954, "learning_rate": 9.933187478460875e-05, "loss": 0.0058, "step": 9950 }, { "epoch": 104.84210526315789, "grad_norm": 0.13468720018863678, "learning_rate": 9.93291780752962e-05, "loss": 0.0073, "step": 9960 }, { "epoch": 104.94736842105263, "grad_norm": 0.12175214290618896, "learning_rate": 9.932647597141893e-05, "loss": 0.0078, "step": 9970 }, { "epoch": 105.05263157894737, "grad_norm": 0.12953054904937744, "learning_rate": 9.932376847327239e-05, "loss": 0.0076, "step": 9980 }, { "epoch": 105.15789473684211, "grad_norm": 0.13486401736736298, "learning_rate": 9.932105558115268e-05, "loss": 0.0071, "step": 9990 }, { "epoch": 105.26315789473684, "grad_norm": 0.10279963910579681, "learning_rate": 9.931833729535651e-05, "loss": 0.0085, "step": 10000 }, { "epoch": 105.36842105263158, "grad_norm": 0.12698274850845337, "learning_rate": 9.931561361618111e-05, "loss": 0.007, "step": 10010 }, { "epoch": 105.47368421052632, "grad_norm": 0.12331541627645493, "learning_rate": 9.931288454392435e-05, "loss": 0.0075, "step": 10020 }, { "epoch": 105.57894736842105, "grad_norm": 0.13951101899147034, "learning_rate": 9.931015007888467e-05, "loss": 0.0084, "step": 10030 }, { "epoch": 105.6842105263158, "grad_norm": 0.13261210918426514, "learning_rate": 9.930741022136112e-05, "loss": 0.007, "step": 10040 }, { "epoch": 105.78947368421052, "grad_norm": 0.1327696442604065, "learning_rate": 9.930466497165333e-05, "loss": 0.0071, "step": 10050 }, { "epoch": 105.89473684210526, "grad_norm": 0.08862864971160889, "learning_rate": 9.93019143300615e-05, "loss": 0.0061, "step": 10060 }, { "epoch": 106.0, "grad_norm": 0.1092723160982132, "learning_rate": 9.929915829688644e-05, "loss": 0.007, "step": 10070 }, { "epoch": 106.10526315789474, "grad_norm": 0.15010863542556763, "learning_rate": 9.929639687242955e-05, "loss": 0.008, "step": 10080 }, { "epoch": 106.21052631578948, "grad_norm": 0.1453062742948532, "learning_rate": 9.929363005699281e-05, "loss": 0.0079, "step": 10090 }, { "epoch": 106.3157894736842, "grad_norm": 0.13476888835430145, "learning_rate": 9.92908578508788e-05, "loss": 0.0083, "step": 10100 }, { "epoch": 106.42105263157895, "grad_norm": 0.12175904214382172, "learning_rate": 9.928808025439069e-05, "loss": 0.0069, "step": 10110 }, { "epoch": 106.52631578947368, "grad_norm": 0.12407585233449936, "learning_rate": 9.928529726783223e-05, "loss": 0.0078, "step": 10120 }, { "epoch": 106.63157894736842, "grad_norm": 0.12477261573076248, "learning_rate": 9.928250889150774e-05, "loss": 0.0069, "step": 10130 }, { "epoch": 106.73684210526316, "grad_norm": 0.08643396198749542, "learning_rate": 9.92797151257222e-05, "loss": 0.008, "step": 10140 }, { "epoch": 106.84210526315789, "grad_norm": 0.13328570127487183, "learning_rate": 9.927691597078108e-05, "loss": 0.0075, "step": 10150 }, { "epoch": 106.94736842105263, "grad_norm": 0.11889760196208954, "learning_rate": 9.927411142699053e-05, "loss": 0.0076, "step": 10160 }, { "epoch": 107.05263157894737, "grad_norm": 0.130068838596344, "learning_rate": 9.927130149465725e-05, "loss": 0.0078, "step": 10170 }, { "epoch": 107.15789473684211, "grad_norm": 0.1565490961074829, "learning_rate": 9.92684861740885e-05, "loss": 0.0075, "step": 10180 }, { "epoch": 107.26315789473684, "grad_norm": 0.1399718075990677, "learning_rate": 9.926566546559217e-05, "loss": 0.0075, "step": 10190 }, { "epoch": 107.36842105263158, "grad_norm": 0.13362926244735718, "learning_rate": 9.926283936947673e-05, "loss": 0.007, "step": 10200 }, { "epoch": 107.47368421052632, "grad_norm": 0.1525583118200302, "learning_rate": 9.926000788605126e-05, "loss": 0.0067, "step": 10210 }, { "epoch": 107.57894736842105, "grad_norm": 0.1055540218949318, "learning_rate": 9.92571710156254e-05, "loss": 0.0071, "step": 10220 }, { "epoch": 107.6842105263158, "grad_norm": 0.09650088846683502, "learning_rate": 9.925432875850936e-05, "loss": 0.007, "step": 10230 }, { "epoch": 107.78947368421052, "grad_norm": 0.10767079889774323, "learning_rate": 9.925148111501396e-05, "loss": 0.0067, "step": 10240 }, { "epoch": 107.89473684210526, "grad_norm": 0.09888990223407745, "learning_rate": 9.924862808545066e-05, "loss": 0.0068, "step": 10250 }, { "epoch": 108.0, "grad_norm": 0.1485253870487213, "learning_rate": 9.924576967013141e-05, "loss": 0.0079, "step": 10260 }, { "epoch": 108.10526315789474, "grad_norm": 0.11624163389205933, "learning_rate": 9.924290586936887e-05, "loss": 0.0075, "step": 10270 }, { "epoch": 108.21052631578948, "grad_norm": 0.12795805931091309, "learning_rate": 9.924003668347614e-05, "loss": 0.0068, "step": 10280 }, { "epoch": 108.3157894736842, "grad_norm": 0.13842640817165375, "learning_rate": 9.923716211276704e-05, "loss": 0.0075, "step": 10290 }, { "epoch": 108.42105263157895, "grad_norm": 0.19393950700759888, "learning_rate": 9.923428215755594e-05, "loss": 0.0078, "step": 10300 }, { "epoch": 108.52631578947368, "grad_norm": 0.11785121262073517, "learning_rate": 9.923139681815775e-05, "loss": 0.0068, "step": 10310 }, { "epoch": 108.63157894736842, "grad_norm": 0.11315825581550598, "learning_rate": 9.922850609488801e-05, "loss": 0.0081, "step": 10320 }, { "epoch": 108.73684210526316, "grad_norm": 0.12167701125144958, "learning_rate": 9.922560998806287e-05, "loss": 0.0076, "step": 10330 }, { "epoch": 108.84210526315789, "grad_norm": 0.10714850574731827, "learning_rate": 9.922270849799905e-05, "loss": 0.0078, "step": 10340 }, { "epoch": 108.94736842105263, "grad_norm": 0.13180918991565704, "learning_rate": 9.92198016250138e-05, "loss": 0.007, "step": 10350 }, { "epoch": 109.05263157894737, "grad_norm": 0.13352611660957336, "learning_rate": 9.921688936942506e-05, "loss": 0.0084, "step": 10360 }, { "epoch": 109.15789473684211, "grad_norm": 0.11284000426530838, "learning_rate": 9.921397173155129e-05, "loss": 0.007, "step": 10370 }, { "epoch": 109.26315789473684, "grad_norm": 0.10394179075956345, "learning_rate": 9.921104871171157e-05, "loss": 0.008, "step": 10380 }, { "epoch": 109.36842105263158, "grad_norm": 0.12979534268379211, "learning_rate": 9.920812031022554e-05, "loss": 0.0069, "step": 10390 }, { "epoch": 109.47368421052632, "grad_norm": 0.12432803958654404, "learning_rate": 9.920518652741348e-05, "loss": 0.0074, "step": 10400 }, { "epoch": 109.57894736842105, "grad_norm": 0.12890595197677612, "learning_rate": 9.920224736359618e-05, "loss": 0.008, "step": 10410 }, { "epoch": 109.6842105263158, "grad_norm": 0.12917010486125946, "learning_rate": 9.91993028190951e-05, "loss": 0.0071, "step": 10420 }, { "epoch": 109.78947368421052, "grad_norm": 0.1357768028974533, "learning_rate": 9.919635289423222e-05, "loss": 0.0081, "step": 10430 }, { "epoch": 109.89473684210526, "grad_norm": 0.14413800835609436, "learning_rate": 9.919339758933015e-05, "loss": 0.0066, "step": 10440 }, { "epoch": 110.0, "grad_norm": 0.20640653371810913, "learning_rate": 9.919043690471209e-05, "loss": 0.0078, "step": 10450 }, { "epoch": 110.10526315789474, "grad_norm": 0.13709606230258942, "learning_rate": 9.91874708407018e-05, "loss": 0.007, "step": 10460 }, { "epoch": 110.21052631578948, "grad_norm": 0.147393137216568, "learning_rate": 9.918449939762367e-05, "loss": 0.0077, "step": 10470 }, { "epoch": 110.3157894736842, "grad_norm": 0.12653180956840515, "learning_rate": 9.91815225758026e-05, "loss": 0.0069, "step": 10480 }, { "epoch": 110.42105263157895, "grad_norm": 0.12302938848733902, "learning_rate": 9.917854037556419e-05, "loss": 0.0072, "step": 10490 }, { "epoch": 110.52631578947368, "grad_norm": 0.10854997485876083, "learning_rate": 9.917555279723454e-05, "loss": 0.0067, "step": 10500 }, { "epoch": 110.63157894736842, "grad_norm": 0.11087555438280106, "learning_rate": 9.917255984114036e-05, "loss": 0.0069, "step": 10510 }, { "epoch": 110.73684210526316, "grad_norm": 0.11278069764375687, "learning_rate": 9.916956150760896e-05, "loss": 0.0071, "step": 10520 }, { "epoch": 110.84210526315789, "grad_norm": 0.09612365812063217, "learning_rate": 9.916655779696826e-05, "loss": 0.0072, "step": 10530 }, { "epoch": 110.94736842105263, "grad_norm": 0.14426292479038239, "learning_rate": 9.916354870954671e-05, "loss": 0.0075, "step": 10540 }, { "epoch": 111.05263157894737, "grad_norm": 0.1961567997932434, "learning_rate": 9.91605342456734e-05, "loss": 0.0075, "step": 10550 }, { "epoch": 111.15789473684211, "grad_norm": 0.10281945765018463, "learning_rate": 9.915751440567795e-05, "loss": 0.0071, "step": 10560 }, { "epoch": 111.26315789473684, "grad_norm": 0.1153169795870781, "learning_rate": 9.915448918989066e-05, "loss": 0.0078, "step": 10570 }, { "epoch": 111.36842105263158, "grad_norm": 0.10629049688577652, "learning_rate": 9.915145859864232e-05, "loss": 0.0067, "step": 10580 }, { "epoch": 111.47368421052632, "grad_norm": 0.1179846003651619, "learning_rate": 9.914842263226437e-05, "loss": 0.0066, "step": 10590 }, { "epoch": 111.57894736842105, "grad_norm": 0.12678496539592743, "learning_rate": 9.914538129108882e-05, "loss": 0.0088, "step": 10600 }, { "epoch": 111.6842105263158, "grad_norm": 0.10384052246809006, "learning_rate": 9.914233457544825e-05, "loss": 0.0065, "step": 10610 }, { "epoch": 111.78947368421052, "grad_norm": 0.12358850985765457, "learning_rate": 9.913928248567586e-05, "loss": 0.0074, "step": 10620 }, { "epoch": 111.89473684210526, "grad_norm": 0.10518169403076172, "learning_rate": 9.913622502210542e-05, "loss": 0.0073, "step": 10630 }, { "epoch": 112.0, "grad_norm": 0.10739091783761978, "learning_rate": 9.913316218507128e-05, "loss": 0.0064, "step": 10640 }, { "epoch": 112.10526315789474, "grad_norm": 0.14395783841609955, "learning_rate": 9.91300939749084e-05, "loss": 0.0074, "step": 10650 }, { "epoch": 112.21052631578948, "grad_norm": 0.14957135915756226, "learning_rate": 9.91270203919523e-05, "loss": 0.0073, "step": 10660 }, { "epoch": 112.3157894736842, "grad_norm": 0.1173473447561264, "learning_rate": 9.912394143653912e-05, "loss": 0.0069, "step": 10670 }, { "epoch": 112.42105263157895, "grad_norm": 0.13610222935676575, "learning_rate": 9.912085710900555e-05, "loss": 0.0077, "step": 10680 }, { "epoch": 112.52631578947368, "grad_norm": 0.1500745564699173, "learning_rate": 9.911776740968892e-05, "loss": 0.0082, "step": 10690 }, { "epoch": 112.63157894736842, "grad_norm": 0.1237211525440216, "learning_rate": 9.911467233892709e-05, "loss": 0.0073, "step": 10700 }, { "epoch": 112.73684210526316, "grad_norm": 0.09862775355577469, "learning_rate": 9.911157189705853e-05, "loss": 0.0076, "step": 10710 }, { "epoch": 112.84210526315789, "grad_norm": 0.11750315129756927, "learning_rate": 9.910846608442229e-05, "loss": 0.0076, "step": 10720 }, { "epoch": 112.94736842105263, "grad_norm": 0.09682919830083847, "learning_rate": 9.910535490135805e-05, "loss": 0.0073, "step": 10730 }, { "epoch": 113.05263157894737, "grad_norm": 0.14944495260715485, "learning_rate": 9.910223834820603e-05, "loss": 0.0074, "step": 10740 }, { "epoch": 113.15789473684211, "grad_norm": 0.12755371630191803, "learning_rate": 9.909911642530703e-05, "loss": 0.0073, "step": 10750 }, { "epoch": 113.26315789473684, "grad_norm": 0.15766991674900055, "learning_rate": 9.909598913300249e-05, "loss": 0.0076, "step": 10760 }, { "epoch": 113.36842105263158, "grad_norm": 0.11905724555253983, "learning_rate": 9.909285647163438e-05, "loss": 0.0073, "step": 10770 }, { "epoch": 113.47368421052632, "grad_norm": 0.12471886724233627, "learning_rate": 9.908971844154531e-05, "loss": 0.0073, "step": 10780 }, { "epoch": 113.57894736842105, "grad_norm": 0.14526134729385376, "learning_rate": 9.908657504307843e-05, "loss": 0.0065, "step": 10790 }, { "epoch": 113.6842105263158, "grad_norm": 0.088313989341259, "learning_rate": 9.908342627657751e-05, "loss": 0.007, "step": 10800 }, { "epoch": 113.78947368421052, "grad_norm": 0.09619778394699097, "learning_rate": 9.908027214238689e-05, "loss": 0.0072, "step": 10810 }, { "epoch": 113.89473684210526, "grad_norm": 0.12210401147603989, "learning_rate": 9.90771126408515e-05, "loss": 0.0058, "step": 10820 }, { "epoch": 114.0, "grad_norm": 0.17567585408687592, "learning_rate": 9.907394777231685e-05, "loss": 0.0075, "step": 10830 }, { "epoch": 114.10526315789474, "grad_norm": 0.13673630356788635, "learning_rate": 9.907077753712905e-05, "loss": 0.0078, "step": 10840 }, { "epoch": 114.21052631578948, "grad_norm": 0.10644416511058807, "learning_rate": 9.906760193563482e-05, "loss": 0.0074, "step": 10850 }, { "epoch": 114.3157894736842, "grad_norm": 0.09621451795101166, "learning_rate": 9.906442096818139e-05, "loss": 0.0072, "step": 10860 }, { "epoch": 114.42105263157895, "grad_norm": 0.0913161188364029, "learning_rate": 9.906123463511665e-05, "loss": 0.0075, "step": 10870 }, { "epoch": 114.52631578947368, "grad_norm": 0.16844592988491058, "learning_rate": 9.905804293678907e-05, "loss": 0.0074, "step": 10880 }, { "epoch": 114.63157894736842, "grad_norm": 0.17523697018623352, "learning_rate": 9.905484587354766e-05, "loss": 0.0081, "step": 10890 }, { "epoch": 114.73684210526316, "grad_norm": 0.12193790823221207, "learning_rate": 9.905164344574205e-05, "loss": 0.0074, "step": 10900 }, { "epoch": 114.84210526315789, "grad_norm": 0.10345201194286346, "learning_rate": 9.904843565372248e-05, "loss": 0.0069, "step": 10910 }, { "epoch": 114.94736842105263, "grad_norm": 0.10950011014938354, "learning_rate": 9.904522249783972e-05, "loss": 0.0075, "step": 10920 }, { "epoch": 115.05263157894737, "grad_norm": 0.13672535121440887, "learning_rate": 9.904200397844517e-05, "loss": 0.0071, "step": 10930 }, { "epoch": 115.15789473684211, "grad_norm": 0.12776020169258118, "learning_rate": 9.903878009589078e-05, "loss": 0.0064, "step": 10940 }, { "epoch": 115.26315789473684, "grad_norm": 0.1116962656378746, "learning_rate": 9.903555085052915e-05, "loss": 0.0065, "step": 10950 }, { "epoch": 115.36842105263158, "grad_norm": 0.12602457404136658, "learning_rate": 9.903231624271338e-05, "loss": 0.0073, "step": 10960 }, { "epoch": 115.47368421052632, "grad_norm": 0.11069713532924652, "learning_rate": 9.902907627279724e-05, "loss": 0.0074, "step": 10970 }, { "epoch": 115.57894736842105, "grad_norm": 0.11077066510915756, "learning_rate": 9.902583094113504e-05, "loss": 0.0081, "step": 10980 }, { "epoch": 115.6842105263158, "grad_norm": 0.1128193810582161, "learning_rate": 9.902258024808168e-05, "loss": 0.0078, "step": 10990 }, { "epoch": 115.78947368421052, "grad_norm": 0.08592390269041061, "learning_rate": 9.901932419399264e-05, "loss": 0.0068, "step": 11000 }, { "epoch": 115.89473684210526, "grad_norm": 0.09894435852766037, "learning_rate": 9.9016062779224e-05, "loss": 0.0073, "step": 11010 }, { "epoch": 116.0, "grad_norm": 0.16473501920700073, "learning_rate": 9.901279600413242e-05, "loss": 0.0066, "step": 11020 }, { "epoch": 116.10526315789474, "grad_norm": 0.09421315044164658, "learning_rate": 9.900952386907518e-05, "loss": 0.0076, "step": 11030 }, { "epoch": 116.21052631578948, "grad_norm": 0.11828994750976562, "learning_rate": 9.90062463744101e-05, "loss": 0.0076, "step": 11040 }, { "epoch": 116.3157894736842, "grad_norm": 0.12555275857448578, "learning_rate": 9.900296352049558e-05, "loss": 0.0076, "step": 11050 }, { "epoch": 116.42105263157895, "grad_norm": 0.12229546159505844, "learning_rate": 9.899967530769065e-05, "loss": 0.008, "step": 11060 }, { "epoch": 116.52631578947368, "grad_norm": 0.12758874893188477, "learning_rate": 9.899638173635489e-05, "loss": 0.007, "step": 11070 }, { "epoch": 116.63157894736842, "grad_norm": 0.09363964945077896, "learning_rate": 9.899308280684849e-05, "loss": 0.0069, "step": 11080 }, { "epoch": 116.73684210526316, "grad_norm": 0.1160290464758873, "learning_rate": 9.898977851953222e-05, "loss": 0.0065, "step": 11090 }, { "epoch": 116.84210526315789, "grad_norm": 0.10735781490802765, "learning_rate": 9.898646887476741e-05, "loss": 0.0065, "step": 11100 }, { "epoch": 116.94736842105263, "grad_norm": 0.12930306792259216, "learning_rate": 9.898315387291603e-05, "loss": 0.0074, "step": 11110 }, { "epoch": 117.05263157894737, "grad_norm": 0.1727401465177536, "learning_rate": 9.89798335143406e-05, "loss": 0.0074, "step": 11120 }, { "epoch": 117.15789473684211, "grad_norm": 0.15105003118515015, "learning_rate": 9.897650779940419e-05, "loss": 0.0086, "step": 11130 }, { "epoch": 117.26315789473684, "grad_norm": 0.094263955950737, "learning_rate": 9.897317672847054e-05, "loss": 0.0072, "step": 11140 }, { "epoch": 117.36842105263158, "grad_norm": 0.17060479521751404, "learning_rate": 9.89698403019039e-05, "loss": 0.0092, "step": 11150 }, { "epoch": 117.47368421052632, "grad_norm": 0.10869809985160828, "learning_rate": 9.896649852006917e-05, "loss": 0.0071, "step": 11160 }, { "epoch": 117.57894736842105, "grad_norm": 0.15887823700904846, "learning_rate": 9.896315138333177e-05, "loss": 0.0071, "step": 11170 }, { "epoch": 117.6842105263158, "grad_norm": 0.1258375644683838, "learning_rate": 9.895979889205774e-05, "loss": 0.0071, "step": 11180 }, { "epoch": 117.78947368421052, "grad_norm": 0.10151493549346924, "learning_rate": 9.895644104661372e-05, "loss": 0.0068, "step": 11190 }, { "epoch": 117.89473684210526, "grad_norm": 0.10809575766324997, "learning_rate": 9.895307784736691e-05, "loss": 0.0069, "step": 11200 }, { "epoch": 118.0, "grad_norm": 0.1597270369529724, "learning_rate": 9.894970929468512e-05, "loss": 0.0067, "step": 11210 }, { "epoch": 118.10526315789474, "grad_norm": 0.13503943383693695, "learning_rate": 9.89463353889367e-05, "loss": 0.0073, "step": 11220 }, { "epoch": 118.21052631578948, "grad_norm": 0.12572774291038513, "learning_rate": 9.894295613049065e-05, "loss": 0.0084, "step": 11230 }, { "epoch": 118.3157894736842, "grad_norm": 0.13532376289367676, "learning_rate": 9.893957151971649e-05, "loss": 0.0065, "step": 11240 }, { "epoch": 118.42105263157895, "grad_norm": 0.08871018141508102, "learning_rate": 9.893618155698436e-05, "loss": 0.0066, "step": 11250 }, { "epoch": 118.52631578947368, "grad_norm": 0.13673138618469238, "learning_rate": 9.8932786242665e-05, "loss": 0.0084, "step": 11260 }, { "epoch": 118.63157894736842, "grad_norm": 0.11279681324958801, "learning_rate": 9.89293855771297e-05, "loss": 0.0074, "step": 11270 }, { "epoch": 118.73684210526316, "grad_norm": 0.1142764762043953, "learning_rate": 9.892597956075036e-05, "loss": 0.0064, "step": 11280 }, { "epoch": 118.84210526315789, "grad_norm": 0.12559561431407928, "learning_rate": 9.892256819389947e-05, "loss": 0.0071, "step": 11290 }, { "epoch": 118.94736842105263, "grad_norm": 0.11252374202013016, "learning_rate": 9.891915147695006e-05, "loss": 0.0069, "step": 11300 }, { "epoch": 119.05263157894737, "grad_norm": 0.1353713423013687, "learning_rate": 9.891572941027577e-05, "loss": 0.0074, "step": 11310 }, { "epoch": 119.15789473684211, "grad_norm": 0.13129208981990814, "learning_rate": 9.89123019942509e-05, "loss": 0.0079, "step": 11320 }, { "epoch": 119.26315789473684, "grad_norm": 0.12106342613697052, "learning_rate": 9.89088692292502e-05, "loss": 0.0061, "step": 11330 }, { "epoch": 119.36842105263158, "grad_norm": 0.13408184051513672, "learning_rate": 9.89054311156491e-05, "loss": 0.0081, "step": 11340 }, { "epoch": 119.47368421052632, "grad_norm": 0.11541828513145447, "learning_rate": 9.890198765382357e-05, "loss": 0.0071, "step": 11350 }, { "epoch": 119.57894736842105, "grad_norm": 0.1192685067653656, "learning_rate": 9.889853884415021e-05, "loss": 0.007, "step": 11360 }, { "epoch": 119.6842105263158, "grad_norm": 0.10648507624864578, "learning_rate": 9.889508468700614e-05, "loss": 0.0076, "step": 11370 }, { "epoch": 119.78947368421052, "grad_norm": 0.09546975791454315, "learning_rate": 9.889162518276915e-05, "loss": 0.006, "step": 11380 }, { "epoch": 119.89473684210526, "grad_norm": 0.11123473942279816, "learning_rate": 9.888816033181752e-05, "loss": 0.0071, "step": 11390 }, { "epoch": 120.0, "grad_norm": 0.15651078522205353, "learning_rate": 9.888469013453018e-05, "loss": 0.0068, "step": 11400 }, { "epoch": 120.10526315789474, "grad_norm": 0.10550056397914886, "learning_rate": 9.888121459128663e-05, "loss": 0.0077, "step": 11410 }, { "epoch": 120.21052631578948, "grad_norm": 0.10074958950281143, "learning_rate": 9.887773370246693e-05, "loss": 0.0068, "step": 11420 }, { "epoch": 120.3157894736842, "grad_norm": 0.09014368802309036, "learning_rate": 9.887424746845177e-05, "loss": 0.0062, "step": 11430 }, { "epoch": 120.42105263157895, "grad_norm": 0.11773741990327835, "learning_rate": 9.887075588962239e-05, "loss": 0.0063, "step": 11440 }, { "epoch": 120.52631578947368, "grad_norm": 0.14808928966522217, "learning_rate": 9.88672589663606e-05, "loss": 0.0071, "step": 11450 }, { "epoch": 120.63157894736842, "grad_norm": 0.11231149733066559, "learning_rate": 9.886375669904886e-05, "loss": 0.0074, "step": 11460 }, { "epoch": 120.73684210526316, "grad_norm": 0.12749329209327698, "learning_rate": 9.886024908807014e-05, "loss": 0.0075, "step": 11470 }, { "epoch": 120.84210526315789, "grad_norm": 0.11668363213539124, "learning_rate": 9.885673613380806e-05, "loss": 0.0074, "step": 11480 }, { "epoch": 120.94736842105263, "grad_norm": 0.11069215834140778, "learning_rate": 9.885321783664676e-05, "loss": 0.006, "step": 11490 }, { "epoch": 121.05263157894737, "grad_norm": 0.11578954756259918, "learning_rate": 9.884969419697101e-05, "loss": 0.0068, "step": 11500 }, { "epoch": 121.15789473684211, "grad_norm": 0.10960417985916138, "learning_rate": 9.884616521516614e-05, "loss": 0.0067, "step": 11510 }, { "epoch": 121.26315789473684, "grad_norm": 0.08508331328630447, "learning_rate": 9.88426308916181e-05, "loss": 0.0063, "step": 11520 }, { "epoch": 121.36842105263158, "grad_norm": 0.11892615258693695, "learning_rate": 9.883909122671335e-05, "loss": 0.0072, "step": 11530 }, { "epoch": 121.47368421052632, "grad_norm": 0.11277160048484802, "learning_rate": 9.883554622083904e-05, "loss": 0.0062, "step": 11540 }, { "epoch": 121.57894736842105, "grad_norm": 0.1324908584356308, "learning_rate": 9.88319958743828e-05, "loss": 0.0062, "step": 11550 }, { "epoch": 121.6842105263158, "grad_norm": 0.13087598979473114, "learning_rate": 9.882844018773291e-05, "loss": 0.0062, "step": 11560 }, { "epoch": 121.78947368421052, "grad_norm": 0.13328325748443604, "learning_rate": 9.882487916127823e-05, "loss": 0.007, "step": 11570 }, { "epoch": 121.89473684210526, "grad_norm": 0.11278009414672852, "learning_rate": 9.882131279540815e-05, "loss": 0.0071, "step": 11580 }, { "epoch": 122.0, "grad_norm": 0.15364998579025269, "learning_rate": 9.881774109051271e-05, "loss": 0.0072, "step": 11590 }, { "epoch": 122.10526315789474, "grad_norm": 0.12676186859607697, "learning_rate": 9.881416404698252e-05, "loss": 0.0076, "step": 11600 }, { "epoch": 122.21052631578948, "grad_norm": 0.1295202225446701, "learning_rate": 9.881058166520873e-05, "loss": 0.0064, "step": 11610 }, { "epoch": 122.3157894736842, "grad_norm": 0.13035742938518524, "learning_rate": 9.880699394558311e-05, "loss": 0.0063, "step": 11620 }, { "epoch": 122.42105263157895, "grad_norm": 0.12471061944961548, "learning_rate": 9.880340088849801e-05, "loss": 0.0076, "step": 11630 }, { "epoch": 122.52631578947368, "grad_norm": 0.11835433542728424, "learning_rate": 9.879980249434637e-05, "loss": 0.0061, "step": 11640 }, { "epoch": 122.63157894736842, "grad_norm": 0.1100078746676445, "learning_rate": 9.879619876352168e-05, "loss": 0.0065, "step": 11650 }, { "epoch": 122.73684210526316, "grad_norm": 0.11116872727870941, "learning_rate": 9.879258969641809e-05, "loss": 0.0064, "step": 11660 }, { "epoch": 122.84210526315789, "grad_norm": 0.1251794844865799, "learning_rate": 9.878897529343023e-05, "loss": 0.0069, "step": 11670 }, { "epoch": 122.94736842105263, "grad_norm": 0.1054270938038826, "learning_rate": 9.878535555495338e-05, "loss": 0.0065, "step": 11680 }, { "epoch": 123.05263157894737, "grad_norm": 0.10480573028326035, "learning_rate": 9.87817304813834e-05, "loss": 0.0072, "step": 11690 }, { "epoch": 123.15789473684211, "grad_norm": 0.13451333343982697, "learning_rate": 9.877810007311671e-05, "loss": 0.007, "step": 11700 }, { "epoch": 123.26315789473684, "grad_norm": 0.12862323224544525, "learning_rate": 9.877446433055035e-05, "loss": 0.0082, "step": 11710 }, { "epoch": 123.36842105263158, "grad_norm": 0.09321793168783188, "learning_rate": 9.877082325408191e-05, "loss": 0.0059, "step": 11720 }, { "epoch": 123.47368421052632, "grad_norm": 0.1310686469078064, "learning_rate": 9.876717684410954e-05, "loss": 0.0075, "step": 11730 }, { "epoch": 123.57894736842105, "grad_norm": 0.12234800308942795, "learning_rate": 9.876352510103204e-05, "loss": 0.0066, "step": 11740 }, { "epoch": 123.6842105263158, "grad_norm": 0.09801743179559708, "learning_rate": 9.875986802524875e-05, "loss": 0.0066, "step": 11750 }, { "epoch": 123.78947368421052, "grad_norm": 0.1249106377363205, "learning_rate": 9.87562056171596e-05, "loss": 0.0069, "step": 11760 }, { "epoch": 123.89473684210526, "grad_norm": 0.14254862070083618, "learning_rate": 9.875253787716511e-05, "loss": 0.0066, "step": 11770 }, { "epoch": 124.0, "grad_norm": 0.15278418362140656, "learning_rate": 9.874886480566637e-05, "loss": 0.0065, "step": 11780 }, { "epoch": 124.10526315789474, "grad_norm": 0.10681987553834915, "learning_rate": 9.874518640306507e-05, "loss": 0.0077, "step": 11790 }, { "epoch": 124.21052631578948, "grad_norm": 0.1528470516204834, "learning_rate": 9.874150266976347e-05, "loss": 0.0071, "step": 11800 }, { "epoch": 124.3157894736842, "grad_norm": 0.12221971154212952, "learning_rate": 9.873781360616443e-05, "loss": 0.0064, "step": 11810 }, { "epoch": 124.42105263157895, "grad_norm": 0.09387538582086563, "learning_rate": 9.873411921267137e-05, "loss": 0.0075, "step": 11820 }, { "epoch": 124.52631578947368, "grad_norm": 0.11478770524263382, "learning_rate": 9.873041948968829e-05, "loss": 0.0069, "step": 11830 }, { "epoch": 124.63157894736842, "grad_norm": 0.1397726982831955, "learning_rate": 9.872671443761981e-05, "loss": 0.0067, "step": 11840 }, { "epoch": 124.73684210526316, "grad_norm": 0.10683067888021469, "learning_rate": 9.872300405687109e-05, "loss": 0.0058, "step": 11850 }, { "epoch": 124.84210526315789, "grad_norm": 0.12475340813398361, "learning_rate": 9.871928834784792e-05, "loss": 0.0067, "step": 11860 }, { "epoch": 124.94736842105263, "grad_norm": 0.10899939388036728, "learning_rate": 9.871556731095661e-05, "loss": 0.0059, "step": 11870 }, { "epoch": 125.05263157894737, "grad_norm": 0.10366328805685043, "learning_rate": 9.871184094660411e-05, "loss": 0.0066, "step": 11880 }, { "epoch": 125.15789473684211, "grad_norm": 0.08016665279865265, "learning_rate": 9.870810925519791e-05, "loss": 0.007, "step": 11890 }, { "epoch": 125.26315789473684, "grad_norm": 0.1154344379901886, "learning_rate": 9.870437223714612e-05, "loss": 0.0068, "step": 11900 }, { "epoch": 125.36842105263158, "grad_norm": 0.10640610754489899, "learning_rate": 9.87006298928574e-05, "loss": 0.0067, "step": 11910 }, { "epoch": 125.47368421052632, "grad_norm": 0.09692821651697159, "learning_rate": 9.869688222274103e-05, "loss": 0.0059, "step": 11920 }, { "epoch": 125.57894736842105, "grad_norm": 0.16041874885559082, "learning_rate": 9.869312922720681e-05, "loss": 0.0064, "step": 11930 }, { "epoch": 125.6842105263158, "grad_norm": 0.09132453799247742, "learning_rate": 9.868937090666521e-05, "loss": 0.0073, "step": 11940 }, { "epoch": 125.78947368421052, "grad_norm": 0.13571293652057648, "learning_rate": 9.86856072615272e-05, "loss": 0.0069, "step": 11950 }, { "epoch": 125.89473684210526, "grad_norm": 0.0971255823969841, "learning_rate": 9.868183829220438e-05, "loss": 0.0062, "step": 11960 }, { "epoch": 126.0, "grad_norm": 0.11008875072002411, "learning_rate": 9.867806399910893e-05, "loss": 0.0071, "step": 11970 }, { "epoch": 126.10526315789474, "grad_norm": 0.13370102643966675, "learning_rate": 9.867428438265356e-05, "loss": 0.0071, "step": 11980 }, { "epoch": 126.21052631578948, "grad_norm": 0.09139281511306763, "learning_rate": 9.867049944325165e-05, "loss": 0.0073, "step": 11990 }, { "epoch": 126.3157894736842, "grad_norm": 0.10204178839921951, "learning_rate": 9.86667091813171e-05, "loss": 0.0069, "step": 12000 }, { "epoch": 126.42105263157895, "grad_norm": 0.10118898004293442, "learning_rate": 9.866291359726438e-05, "loss": 0.0065, "step": 12010 }, { "epoch": 126.52631578947368, "grad_norm": 0.10271890461444855, "learning_rate": 9.865911269150861e-05, "loss": 0.0062, "step": 12020 }, { "epoch": 126.63157894736842, "grad_norm": 0.10409326106309891, "learning_rate": 9.865530646446544e-05, "loss": 0.0063, "step": 12030 }, { "epoch": 126.73684210526316, "grad_norm": 0.11347350478172302, "learning_rate": 9.86514949165511e-05, "loss": 0.0067, "step": 12040 }, { "epoch": 126.84210526315789, "grad_norm": 0.15390285849571228, "learning_rate": 9.864767804818243e-05, "loss": 0.0066, "step": 12050 }, { "epoch": 126.94736842105263, "grad_norm": 0.1444290578365326, "learning_rate": 9.86438558597768e-05, "loss": 0.006, "step": 12060 }, { "epoch": 127.05263157894737, "grad_norm": 0.12926363945007324, "learning_rate": 9.864002835175225e-05, "loss": 0.008, "step": 12070 }, { "epoch": 127.15789473684211, "grad_norm": 0.10224888473749161, "learning_rate": 9.863619552452734e-05, "loss": 0.0071, "step": 12080 }, { "epoch": 127.26315789473684, "grad_norm": 0.08802203088998795, "learning_rate": 9.863235737852119e-05, "loss": 0.0068, "step": 12090 }, { "epoch": 127.36842105263158, "grad_norm": 0.14852212369441986, "learning_rate": 9.862851391415356e-05, "loss": 0.0062, "step": 12100 }, { "epoch": 127.47368421052632, "grad_norm": 0.11683733016252518, "learning_rate": 9.862466513184477e-05, "loss": 0.0059, "step": 12110 }, { "epoch": 127.57894736842105, "grad_norm": 0.12413021177053452, "learning_rate": 9.86208110320157e-05, "loss": 0.0068, "step": 12120 }, { "epoch": 127.6842105263158, "grad_norm": 0.14567716419696808, "learning_rate": 9.861695161508784e-05, "loss": 0.0066, "step": 12130 }, { "epoch": 127.78947368421052, "grad_norm": 0.10097550600767136, "learning_rate": 9.861308688148324e-05, "loss": 0.0067, "step": 12140 }, { "epoch": 127.89473684210526, "grad_norm": 0.11372367292642593, "learning_rate": 9.860921683162455e-05, "loss": 0.0066, "step": 12150 }, { "epoch": 128.0, "grad_norm": 0.12749327719211578, "learning_rate": 9.860534146593499e-05, "loss": 0.0066, "step": 12160 }, { "epoch": 128.10526315789474, "grad_norm": 0.11151307076215744, "learning_rate": 9.860146078483836e-05, "loss": 0.0074, "step": 12170 }, { "epoch": 128.21052631578948, "grad_norm": 0.1077490970492363, "learning_rate": 9.859757478875905e-05, "loss": 0.0061, "step": 12180 }, { "epoch": 128.31578947368422, "grad_norm": 0.09475722163915634, "learning_rate": 9.859368347812204e-05, "loss": 0.006, "step": 12190 }, { "epoch": 128.42105263157896, "grad_norm": 0.12202521413564682, "learning_rate": 9.858978685335285e-05, "loss": 0.0062, "step": 12200 }, { "epoch": 128.52631578947367, "grad_norm": 0.12211856245994568, "learning_rate": 9.858588491487763e-05, "loss": 0.0064, "step": 12210 }, { "epoch": 128.6315789473684, "grad_norm": 0.12197234481573105, "learning_rate": 9.858197766312308e-05, "loss": 0.007, "step": 12220 }, { "epoch": 128.73684210526315, "grad_norm": 0.11558926105499268, "learning_rate": 9.857806509851649e-05, "loss": 0.0065, "step": 12230 }, { "epoch": 128.8421052631579, "grad_norm": 0.09523840993642807, "learning_rate": 9.857414722148574e-05, "loss": 0.0074, "step": 12240 }, { "epoch": 128.94736842105263, "grad_norm": 0.12549778819084167, "learning_rate": 9.857022403245928e-05, "loss": 0.0075, "step": 12250 }, { "epoch": 129.05263157894737, "grad_norm": 0.12443146854639053, "learning_rate": 9.856629553186615e-05, "loss": 0.0067, "step": 12260 }, { "epoch": 129.1578947368421, "grad_norm": 0.10476195812225342, "learning_rate": 9.856236172013595e-05, "loss": 0.0067, "step": 12270 }, { "epoch": 129.26315789473685, "grad_norm": 0.11675841361284256, "learning_rate": 9.85584225976989e-05, "loss": 0.0061, "step": 12280 }, { "epoch": 129.3684210526316, "grad_norm": 0.0872441753745079, "learning_rate": 9.855447816498575e-05, "loss": 0.0079, "step": 12290 }, { "epoch": 129.47368421052633, "grad_norm": 0.12599436938762665, "learning_rate": 9.855052842242787e-05, "loss": 0.0072, "step": 12300 }, { "epoch": 129.57894736842104, "grad_norm": 0.09482535719871521, "learning_rate": 9.85465733704572e-05, "loss": 0.0061, "step": 12310 }, { "epoch": 129.68421052631578, "grad_norm": 0.12087057530879974, "learning_rate": 9.854261300950624e-05, "loss": 0.0061, "step": 12320 }, { "epoch": 129.78947368421052, "grad_norm": 0.10061466693878174, "learning_rate": 9.853864734000813e-05, "loss": 0.0065, "step": 12330 }, { "epoch": 129.89473684210526, "grad_norm": 0.1139116883277893, "learning_rate": 9.85346763623965e-05, "loss": 0.0064, "step": 12340 }, { "epoch": 130.0, "grad_norm": 0.1269531100988388, "learning_rate": 9.853070007710564e-05, "loss": 0.0073, "step": 12350 }, { "epoch": 130.10526315789474, "grad_norm": 0.13900335133075714, "learning_rate": 9.85267184845704e-05, "loss": 0.0071, "step": 12360 }, { "epoch": 130.21052631578948, "grad_norm": 0.11418324708938599, "learning_rate": 9.852273158522616e-05, "loss": 0.007, "step": 12370 }, { "epoch": 130.31578947368422, "grad_norm": 0.11903784424066544, "learning_rate": 9.851873937950896e-05, "loss": 0.0064, "step": 12380 }, { "epoch": 130.42105263157896, "grad_norm": 0.11671419441699982, "learning_rate": 9.851474186785537e-05, "loss": 0.0067, "step": 12390 }, { "epoch": 130.52631578947367, "grad_norm": 0.12152779847383499, "learning_rate": 9.851073905070254e-05, "loss": 0.0072, "step": 12400 }, { "epoch": 130.6315789473684, "grad_norm": 0.10030547529459, "learning_rate": 9.850673092848824e-05, "loss": 0.0069, "step": 12410 }, { "epoch": 130.73684210526315, "grad_norm": 0.10689917206764221, "learning_rate": 9.850271750165077e-05, "loss": 0.007, "step": 12420 }, { "epoch": 130.8421052631579, "grad_norm": 0.11027085036039352, "learning_rate": 9.849869877062902e-05, "loss": 0.0062, "step": 12430 }, { "epoch": 130.94736842105263, "grad_norm": 0.1292971968650818, "learning_rate": 9.849467473586252e-05, "loss": 0.0063, "step": 12440 }, { "epoch": 131.05263157894737, "grad_norm": 0.12588882446289062, "learning_rate": 9.849064539779127e-05, "loss": 0.0068, "step": 12450 }, { "epoch": 131.1578947368421, "grad_norm": 0.11754495650529861, "learning_rate": 9.848661075685594e-05, "loss": 0.0066, "step": 12460 }, { "epoch": 131.26315789473685, "grad_norm": 0.09805987775325775, "learning_rate": 9.848257081349778e-05, "loss": 0.0067, "step": 12470 }, { "epoch": 131.3684210526316, "grad_norm": 0.11798666417598724, "learning_rate": 9.847852556815856e-05, "loss": 0.006, "step": 12480 }, { "epoch": 131.47368421052633, "grad_norm": 0.11072800308465958, "learning_rate": 9.847447502128067e-05, "loss": 0.0068, "step": 12490 }, { "epoch": 131.57894736842104, "grad_norm": 0.10365515202283859, "learning_rate": 9.847041917330708e-05, "loss": 0.0068, "step": 12500 }, { "epoch": 131.68421052631578, "grad_norm": 0.0873807966709137, "learning_rate": 9.846635802468132e-05, "loss": 0.0069, "step": 12510 }, { "epoch": 131.78947368421052, "grad_norm": 0.09488113969564438, "learning_rate": 9.84622915758475e-05, "loss": 0.006, "step": 12520 }, { "epoch": 131.89473684210526, "grad_norm": 0.09462578594684601, "learning_rate": 9.845821982725034e-05, "loss": 0.0067, "step": 12530 }, { "epoch": 132.0, "grad_norm": 0.11480776965618134, "learning_rate": 9.845414277933514e-05, "loss": 0.0056, "step": 12540 }, { "epoch": 132.10526315789474, "grad_norm": 0.10447590798139572, "learning_rate": 9.845006043254771e-05, "loss": 0.0071, "step": 12550 }, { "epoch": 132.21052631578948, "grad_norm": 0.10198118537664413, "learning_rate": 9.844597278733451e-05, "loss": 0.0071, "step": 12560 }, { "epoch": 132.31578947368422, "grad_norm": 0.10850925743579865, "learning_rate": 9.844187984414259e-05, "loss": 0.0065, "step": 12570 }, { "epoch": 132.42105263157896, "grad_norm": 0.11711620539426804, "learning_rate": 9.84377816034195e-05, "loss": 0.0069, "step": 12580 }, { "epoch": 132.52631578947367, "grad_norm": 0.11352997273206711, "learning_rate": 9.843367806561345e-05, "loss": 0.0061, "step": 12590 }, { "epoch": 132.6315789473684, "grad_norm": 0.09832363575696945, "learning_rate": 9.842956923117317e-05, "loss": 0.0066, "step": 12600 }, { "epoch": 132.73684210526315, "grad_norm": 0.10421469807624817, "learning_rate": 9.842545510054802e-05, "loss": 0.0071, "step": 12610 }, { "epoch": 132.8421052631579, "grad_norm": 0.12157443165779114, "learning_rate": 9.842133567418792e-05, "loss": 0.0068, "step": 12620 }, { "epoch": 132.94736842105263, "grad_norm": 0.12756484746932983, "learning_rate": 9.841721095254333e-05, "loss": 0.0064, "step": 12630 }, { "epoch": 133.05263157894737, "grad_norm": 0.10045566409826279, "learning_rate": 9.841308093606537e-05, "loss": 0.0065, "step": 12640 }, { "epoch": 133.1578947368421, "grad_norm": 0.11803105473518372, "learning_rate": 9.840894562520565e-05, "loss": 0.0073, "step": 12650 }, { "epoch": 133.26315789473685, "grad_norm": 0.14279302954673767, "learning_rate": 9.840480502041642e-05, "loss": 0.0058, "step": 12660 }, { "epoch": 133.3684210526316, "grad_norm": 0.1263798028230667, "learning_rate": 9.840065912215049e-05, "loss": 0.0068, "step": 12670 }, { "epoch": 133.47368421052633, "grad_norm": 0.12159503251314163, "learning_rate": 9.839650793086124e-05, "loss": 0.006, "step": 12680 }, { "epoch": 133.57894736842104, "grad_norm": 0.09399081766605377, "learning_rate": 9.839235144700265e-05, "loss": 0.0062, "step": 12690 }, { "epoch": 133.68421052631578, "grad_norm": 0.11986082047224045, "learning_rate": 9.838818967102926e-05, "loss": 0.0064, "step": 12700 }, { "epoch": 133.78947368421052, "grad_norm": 0.10354716330766678, "learning_rate": 9.83840226033962e-05, "loss": 0.0058, "step": 12710 }, { "epoch": 133.89473684210526, "grad_norm": 0.08768772333860397, "learning_rate": 9.837985024455918e-05, "loss": 0.0064, "step": 12720 }, { "epoch": 134.0, "grad_norm": 0.09483372420072556, "learning_rate": 9.837567259497447e-05, "loss": 0.0063, "step": 12730 }, { "epoch": 134.10526315789474, "grad_norm": 0.09156618267297745, "learning_rate": 9.837148965509894e-05, "loss": 0.0063, "step": 12740 }, { "epoch": 134.21052631578948, "grad_norm": 0.11922232806682587, "learning_rate": 9.836730142539001e-05, "loss": 0.0069, "step": 12750 }, { "epoch": 134.31578947368422, "grad_norm": 0.14137721061706543, "learning_rate": 9.836310790630574e-05, "loss": 0.007, "step": 12760 }, { "epoch": 134.42105263157896, "grad_norm": 0.10143239796161652, "learning_rate": 9.83589090983047e-05, "loss": 0.0059, "step": 12770 }, { "epoch": 134.52631578947367, "grad_norm": 0.08806347846984863, "learning_rate": 9.835470500184605e-05, "loss": 0.0068, "step": 12780 }, { "epoch": 134.6315789473684, "grad_norm": 0.10948656499385834, "learning_rate": 9.835049561738957e-05, "loss": 0.0071, "step": 12790 }, { "epoch": 134.73684210526315, "grad_norm": 0.12984275817871094, "learning_rate": 9.834628094539558e-05, "loss": 0.0073, "step": 12800 }, { "epoch": 134.8421052631579, "grad_norm": 0.11989716440439224, "learning_rate": 9.834206098632499e-05, "loss": 0.0067, "step": 12810 }, { "epoch": 134.94736842105263, "grad_norm": 0.13511574268341064, "learning_rate": 9.833783574063931e-05, "loss": 0.0062, "step": 12820 }, { "epoch": 135.05263157894737, "grad_norm": 0.12243366986513138, "learning_rate": 9.833360520880058e-05, "loss": 0.007, "step": 12830 }, { "epoch": 135.1578947368421, "grad_norm": 0.09331075102090836, "learning_rate": 9.832936939127144e-05, "loss": 0.0062, "step": 12840 }, { "epoch": 135.26315789473685, "grad_norm": 0.08237658441066742, "learning_rate": 9.832512828851515e-05, "loss": 0.0063, "step": 12850 }, { "epoch": 135.3684210526316, "grad_norm": 0.10546699166297913, "learning_rate": 9.832088190099546e-05, "loss": 0.006, "step": 12860 }, { "epoch": 135.47368421052633, "grad_norm": 0.11553333699703217, "learning_rate": 9.831663022917679e-05, "loss": 0.0062, "step": 12870 }, { "epoch": 135.57894736842104, "grad_norm": 0.1340731978416443, "learning_rate": 9.831237327352407e-05, "loss": 0.0064, "step": 12880 }, { "epoch": 135.68421052631578, "grad_norm": 0.13000522553920746, "learning_rate": 9.830811103450286e-05, "loss": 0.007, "step": 12890 }, { "epoch": 135.78947368421052, "grad_norm": 0.12860654294490814, "learning_rate": 9.830384351257924e-05, "loss": 0.0063, "step": 12900 }, { "epoch": 135.89473684210526, "grad_norm": 0.11113713681697845, "learning_rate": 9.829957070821993e-05, "loss": 0.007, "step": 12910 }, { "epoch": 136.0, "grad_norm": 0.14426125586032867, "learning_rate": 9.829529262189218e-05, "loss": 0.0062, "step": 12920 }, { "epoch": 136.10526315789474, "grad_norm": 0.12325100600719452, "learning_rate": 9.829100925406385e-05, "loss": 0.006, "step": 12930 }, { "epoch": 136.21052631578948, "grad_norm": 0.13209716975688934, "learning_rate": 9.828672060520333e-05, "loss": 0.0062, "step": 12940 }, { "epoch": 136.31578947368422, "grad_norm": 0.11776181310415268, "learning_rate": 9.828242667577966e-05, "loss": 0.0063, "step": 12950 }, { "epoch": 136.42105263157896, "grad_norm": 0.14828786253929138, "learning_rate": 9.82781274662624e-05, "loss": 0.0071, "step": 12960 }, { "epoch": 136.52631578947367, "grad_norm": 0.1777542531490326, "learning_rate": 9.82738229771217e-05, "loss": 0.0067, "step": 12970 }, { "epoch": 136.6315789473684, "grad_norm": 0.11011964827775955, "learning_rate": 9.826951320882829e-05, "loss": 0.0061, "step": 12980 }, { "epoch": 136.73684210526315, "grad_norm": 0.09157565236091614, "learning_rate": 9.826519816185351e-05, "loss": 0.0073, "step": 12990 }, { "epoch": 136.8421052631579, "grad_norm": 0.10782536119222641, "learning_rate": 9.826087783666921e-05, "loss": 0.0067, "step": 13000 }, { "epoch": 136.94736842105263, "grad_norm": 0.10878381133079529, "learning_rate": 9.825655223374787e-05, "loss": 0.0085, "step": 13010 }, { "epoch": 137.05263157894737, "grad_norm": 0.09362064301967621, "learning_rate": 9.825222135356253e-05, "loss": 0.0073, "step": 13020 }, { "epoch": 137.1578947368421, "grad_norm": 0.1361800879240036, "learning_rate": 9.82478851965868e-05, "loss": 0.0073, "step": 13030 }, { "epoch": 137.26315789473685, "grad_norm": 0.10328274965286255, "learning_rate": 9.82435437632949e-05, "loss": 0.0062, "step": 13040 }, { "epoch": 137.3684210526316, "grad_norm": 0.1095409095287323, "learning_rate": 9.823919705416158e-05, "loss": 0.0062, "step": 13050 }, { "epoch": 137.47368421052633, "grad_norm": 0.15068548917770386, "learning_rate": 9.82348450696622e-05, "loss": 0.0062, "step": 13060 }, { "epoch": 137.57894736842104, "grad_norm": 0.10054443776607513, "learning_rate": 9.823048781027268e-05, "loss": 0.0069, "step": 13070 }, { "epoch": 137.68421052631578, "grad_norm": 0.1199939027428627, "learning_rate": 9.822612527646953e-05, "loss": 0.0059, "step": 13080 }, { "epoch": 137.78947368421052, "grad_norm": 0.09776423871517181, "learning_rate": 9.822175746872984e-05, "loss": 0.0064, "step": 13090 }, { "epoch": 137.89473684210526, "grad_norm": 0.10806853324174881, "learning_rate": 9.821738438753123e-05, "loss": 0.0067, "step": 13100 }, { "epoch": 138.0, "grad_norm": 0.1354711353778839, "learning_rate": 9.821300603335196e-05, "loss": 0.0058, "step": 13110 }, { "epoch": 138.10526315789474, "grad_norm": 0.10798846185207367, "learning_rate": 9.820862240667085e-05, "loss": 0.0059, "step": 13120 }, { "epoch": 138.21052631578948, "grad_norm": 0.10563022643327713, "learning_rate": 9.820423350796726e-05, "loss": 0.0062, "step": 13130 }, { "epoch": 138.31578947368422, "grad_norm": 0.09248574823141098, "learning_rate": 9.819983933772118e-05, "loss": 0.0055, "step": 13140 }, { "epoch": 138.42105263157896, "grad_norm": 0.10190785676240921, "learning_rate": 9.819543989641314e-05, "loss": 0.0062, "step": 13150 }, { "epoch": 138.52631578947367, "grad_norm": 0.13356290757656097, "learning_rate": 9.819103518452423e-05, "loss": 0.0064, "step": 13160 }, { "epoch": 138.6315789473684, "grad_norm": 0.09210803359746933, "learning_rate": 9.818662520253618e-05, "loss": 0.0054, "step": 13170 }, { "epoch": 138.73684210526315, "grad_norm": 0.08953816443681717, "learning_rate": 9.818220995093126e-05, "loss": 0.0065, "step": 13180 }, { "epoch": 138.8421052631579, "grad_norm": 0.10350607335567474, "learning_rate": 9.817778943019228e-05, "loss": 0.0054, "step": 13190 }, { "epoch": 138.94736842105263, "grad_norm": 0.10156502574682236, "learning_rate": 9.81733636408027e-05, "loss": 0.0068, "step": 13200 }, { "epoch": 139.05263157894737, "grad_norm": 0.10849910974502563, "learning_rate": 9.816893258324649e-05, "loss": 0.0066, "step": 13210 }, { "epoch": 139.1578947368421, "grad_norm": 0.10574297606945038, "learning_rate": 9.816449625800823e-05, "loss": 0.0062, "step": 13220 }, { "epoch": 139.26315789473685, "grad_norm": 0.07238058745861053, "learning_rate": 9.816005466557308e-05, "loss": 0.0068, "step": 13230 }, { "epoch": 139.3684210526316, "grad_norm": 0.11837809532880783, "learning_rate": 9.815560780642674e-05, "loss": 0.0059, "step": 13240 }, { "epoch": 139.47368421052633, "grad_norm": 0.11038076877593994, "learning_rate": 9.815115568105555e-05, "loss": 0.007, "step": 13250 }, { "epoch": 139.57894736842104, "grad_norm": 0.08973398059606552, "learning_rate": 9.814669828994638e-05, "loss": 0.0064, "step": 13260 }, { "epoch": 139.68421052631578, "grad_norm": 0.11171893030405045, "learning_rate": 9.814223563358665e-05, "loss": 0.0078, "step": 13270 }, { "epoch": 139.78947368421052, "grad_norm": 0.0979628935456276, "learning_rate": 9.813776771246443e-05, "loss": 0.0064, "step": 13280 }, { "epoch": 139.89473684210526, "grad_norm": 0.12237384170293808, "learning_rate": 9.813329452706829e-05, "loss": 0.0061, "step": 13290 }, { "epoch": 140.0, "grad_norm": 0.14808569848537445, "learning_rate": 9.812881607788744e-05, "loss": 0.0065, "step": 13300 }, { "epoch": 140.10526315789474, "grad_norm": 0.13136471807956696, "learning_rate": 9.812433236541163e-05, "loss": 0.0064, "step": 13310 }, { "epoch": 140.21052631578948, "grad_norm": 0.12184444814920425, "learning_rate": 9.811984339013116e-05, "loss": 0.006, "step": 13320 }, { "epoch": 140.31578947368422, "grad_norm": 0.10059401392936707, "learning_rate": 9.811534915253698e-05, "loss": 0.0058, "step": 13330 }, { "epoch": 140.42105263157896, "grad_norm": 0.10126055032014847, "learning_rate": 9.811084965312056e-05, "loss": 0.0063, "step": 13340 }, { "epoch": 140.52631578947367, "grad_norm": 0.0794675201177597, "learning_rate": 9.810634489237396e-05, "loss": 0.0056, "step": 13350 }, { "epoch": 140.6315789473684, "grad_norm": 0.08039086312055588, "learning_rate": 9.81018348707898e-05, "loss": 0.0063, "step": 13360 }, { "epoch": 140.73684210526315, "grad_norm": 0.118436299264431, "learning_rate": 9.809731958886131e-05, "loss": 0.007, "step": 13370 }, { "epoch": 140.8421052631579, "grad_norm": 0.09553225338459015, "learning_rate": 9.809279904708224e-05, "loss": 0.0061, "step": 13380 }, { "epoch": 140.94736842105263, "grad_norm": 0.1399151235818863, "learning_rate": 9.808827324594699e-05, "loss": 0.0062, "step": 13390 }, { "epoch": 141.05263157894737, "grad_norm": 0.11053816229104996, "learning_rate": 9.808374218595046e-05, "loss": 0.0061, "step": 13400 }, { "epoch": 141.1578947368421, "grad_norm": 0.12744121253490448, "learning_rate": 9.80792058675882e-05, "loss": 0.0069, "step": 13410 }, { "epoch": 141.26315789473685, "grad_norm": 0.14364251494407654, "learning_rate": 9.807466429135627e-05, "loss": 0.007, "step": 13420 }, { "epoch": 141.3684210526316, "grad_norm": 0.12149327248334885, "learning_rate": 9.807011745775132e-05, "loss": 0.0073, "step": 13430 }, { "epoch": 141.47368421052633, "grad_norm": 0.11288629472255707, "learning_rate": 9.806556536727061e-05, "loss": 0.0064, "step": 13440 }, { "epoch": 141.57894736842104, "grad_norm": 0.12204630672931671, "learning_rate": 9.806100802041193e-05, "loss": 0.0062, "step": 13450 }, { "epoch": 141.68421052631578, "grad_norm": 0.12733839452266693, "learning_rate": 9.805644541767368e-05, "loss": 0.0061, "step": 13460 }, { "epoch": 141.78947368421052, "grad_norm": 0.09546263515949249, "learning_rate": 9.805187755955478e-05, "loss": 0.0056, "step": 13470 }, { "epoch": 141.89473684210526, "grad_norm": 0.08623010665178299, "learning_rate": 9.804730444655483e-05, "loss": 0.0061, "step": 13480 }, { "epoch": 142.0, "grad_norm": 0.1535148173570633, "learning_rate": 9.804272607917388e-05, "loss": 0.0065, "step": 13490 }, { "epoch": 142.10526315789474, "grad_norm": 0.11703010648488998, "learning_rate": 9.803814245791265e-05, "loss": 0.0065, "step": 13500 }, { "epoch": 142.21052631578948, "grad_norm": 0.09568590670824051, "learning_rate": 9.803355358327239e-05, "loss": 0.0064, "step": 13510 }, { "epoch": 142.31578947368422, "grad_norm": 0.11711076647043228, "learning_rate": 9.802895945575492e-05, "loss": 0.0063, "step": 13520 }, { "epoch": 142.42105263157896, "grad_norm": 0.14236724376678467, "learning_rate": 9.802436007586266e-05, "loss": 0.0071, "step": 13530 }, { "epoch": 142.52631578947367, "grad_norm": 0.09575182944536209, "learning_rate": 9.801975544409858e-05, "loss": 0.0062, "step": 13540 }, { "epoch": 142.6315789473684, "grad_norm": 0.11315842717885971, "learning_rate": 9.801514556096625e-05, "loss": 0.0065, "step": 13550 }, { "epoch": 142.73684210526315, "grad_norm": 0.10693608224391937, "learning_rate": 9.801053042696977e-05, "loss": 0.0063, "step": 13560 }, { "epoch": 142.8421052631579, "grad_norm": 0.1034511923789978, "learning_rate": 9.800591004261388e-05, "loss": 0.0065, "step": 13570 }, { "epoch": 142.94736842105263, "grad_norm": 0.13950347900390625, "learning_rate": 9.800128440840385e-05, "loss": 0.0066, "step": 13580 }, { "epoch": 143.05263157894737, "grad_norm": 0.14724069833755493, "learning_rate": 9.799665352484552e-05, "loss": 0.0068, "step": 13590 }, { "epoch": 143.1578947368421, "grad_norm": 0.10338518023490906, "learning_rate": 9.799201739244532e-05, "loss": 0.0075, "step": 13600 }, { "epoch": 143.26315789473685, "grad_norm": 0.1115468442440033, "learning_rate": 9.798737601171025e-05, "loss": 0.0067, "step": 13610 }, { "epoch": 143.3684210526316, "grad_norm": 0.08588545024394989, "learning_rate": 9.79827293831479e-05, "loss": 0.006, "step": 13620 }, { "epoch": 143.47368421052633, "grad_norm": 0.1223704144358635, "learning_rate": 9.797807750726638e-05, "loss": 0.0066, "step": 13630 }, { "epoch": 143.57894736842104, "grad_norm": 0.11704360693693161, "learning_rate": 9.797342038457446e-05, "loss": 0.0065, "step": 13640 }, { "epoch": 143.68421052631578, "grad_norm": 0.11207153648138046, "learning_rate": 9.796875801558141e-05, "loss": 0.0055, "step": 13650 }, { "epoch": 143.78947368421052, "grad_norm": 0.12368658185005188, "learning_rate": 9.79640904007971e-05, "loss": 0.0059, "step": 13660 }, { "epoch": 143.89473684210526, "grad_norm": 0.1331542730331421, "learning_rate": 9.795941754073199e-05, "loss": 0.0065, "step": 13670 }, { "epoch": 144.0, "grad_norm": 0.14980727434158325, "learning_rate": 9.795473943589705e-05, "loss": 0.0057, "step": 13680 }, { "epoch": 144.10526315789474, "grad_norm": 0.11533532291650772, "learning_rate": 9.795005608680394e-05, "loss": 0.0069, "step": 13690 }, { "epoch": 144.21052631578948, "grad_norm": 0.13101254403591156, "learning_rate": 9.794536749396477e-05, "loss": 0.0059, "step": 13700 }, { "epoch": 144.31578947368422, "grad_norm": 0.10358690470457077, "learning_rate": 9.79406736578923e-05, "loss": 0.0064, "step": 13710 }, { "epoch": 144.42105263157896, "grad_norm": 0.11898922175168991, "learning_rate": 9.793597457909984e-05, "loss": 0.0069, "step": 13720 }, { "epoch": 144.52631578947367, "grad_norm": 0.10591776669025421, "learning_rate": 9.793127025810127e-05, "loss": 0.0065, "step": 13730 }, { "epoch": 144.6315789473684, "grad_norm": 0.10343646258115768, "learning_rate": 9.792656069541104e-05, "loss": 0.005, "step": 13740 }, { "epoch": 144.73684210526315, "grad_norm": 0.13424411416053772, "learning_rate": 9.79218458915442e-05, "loss": 0.0063, "step": 13750 }, { "epoch": 144.8421052631579, "grad_norm": 0.09104565531015396, "learning_rate": 9.791712584701634e-05, "loss": 0.0064, "step": 13760 }, { "epoch": 144.94736842105263, "grad_norm": 0.0903177484869957, "learning_rate": 9.791240056234364e-05, "loss": 0.0061, "step": 13770 }, { "epoch": 145.05263157894737, "grad_norm": 0.14818336069583893, "learning_rate": 9.790767003804283e-05, "loss": 0.0062, "step": 13780 }, { "epoch": 145.1578947368421, "grad_norm": 0.10535188764333725, "learning_rate": 9.790293427463126e-05, "loss": 0.0065, "step": 13790 }, { "epoch": 145.26315789473685, "grad_norm": 0.11707215011119843, "learning_rate": 9.789819327262684e-05, "loss": 0.0065, "step": 13800 }, { "epoch": 145.3684210526316, "grad_norm": 0.09701425582170486, "learning_rate": 9.7893447032548e-05, "loss": 0.0058, "step": 13810 }, { "epoch": 145.47368421052633, "grad_norm": 0.09237522631883621, "learning_rate": 9.78886955549138e-05, "loss": 0.006, "step": 13820 }, { "epoch": 145.57894736842104, "grad_norm": 0.10126321017742157, "learning_rate": 9.788393884024387e-05, "loss": 0.0066, "step": 13830 }, { "epoch": 145.68421052631578, "grad_norm": 0.11800383031368256, "learning_rate": 9.787917688905836e-05, "loss": 0.0063, "step": 13840 }, { "epoch": 145.78947368421052, "grad_norm": 0.08877461403608322, "learning_rate": 9.787440970187807e-05, "loss": 0.0059, "step": 13850 }, { "epoch": 145.89473684210526, "grad_norm": 0.12939152121543884, "learning_rate": 9.786963727922429e-05, "loss": 0.0067, "step": 13860 }, { "epoch": 146.0, "grad_norm": 0.10592437535524368, "learning_rate": 9.786485962161897e-05, "loss": 0.006, "step": 13870 }, { "epoch": 146.10526315789474, "grad_norm": 0.1659889668226242, "learning_rate": 9.786007672958455e-05, "loss": 0.0063, "step": 13880 }, { "epoch": 146.21052631578948, "grad_norm": 0.11093369871377945, "learning_rate": 9.78552886036441e-05, "loss": 0.0065, "step": 13890 }, { "epoch": 146.31578947368422, "grad_norm": 0.12322403490543365, "learning_rate": 9.785049524432124e-05, "loss": 0.0068, "step": 13900 }, { "epoch": 146.42105263157896, "grad_norm": 0.1667276918888092, "learning_rate": 9.784569665214016e-05, "loss": 0.0063, "step": 13910 }, { "epoch": 146.52631578947367, "grad_norm": 0.1304740011692047, "learning_rate": 9.784089282762563e-05, "loss": 0.0063, "step": 13920 }, { "epoch": 146.6315789473684, "grad_norm": 0.1327562928199768, "learning_rate": 9.7836083771303e-05, "loss": 0.0067, "step": 13930 }, { "epoch": 146.73684210526315, "grad_norm": 0.1096552312374115, "learning_rate": 9.783126948369817e-05, "loss": 0.0063, "step": 13940 }, { "epoch": 146.8421052631579, "grad_norm": 0.1293153166770935, "learning_rate": 9.78264499653376e-05, "loss": 0.0063, "step": 13950 }, { "epoch": 146.94736842105263, "grad_norm": 0.11907489597797394, "learning_rate": 9.782162521674838e-05, "loss": 0.0073, "step": 13960 }, { "epoch": 147.05263157894737, "grad_norm": 0.13044562935829163, "learning_rate": 9.781679523845812e-05, "loss": 0.0052, "step": 13970 }, { "epoch": 147.1578947368421, "grad_norm": 0.09786625951528549, "learning_rate": 9.781196003099502e-05, "loss": 0.0063, "step": 13980 }, { "epoch": 147.26315789473685, "grad_norm": 0.11696356534957886, "learning_rate": 9.780711959488786e-05, "loss": 0.0054, "step": 13990 }, { "epoch": 147.3684210526316, "grad_norm": 0.07295782119035721, "learning_rate": 9.780227393066599e-05, "loss": 0.0062, "step": 14000 }, { "epoch": 147.47368421052633, "grad_norm": 0.10334135591983795, "learning_rate": 9.77974230388593e-05, "loss": 0.007, "step": 14010 }, { "epoch": 147.57894736842104, "grad_norm": 0.09259862452745438, "learning_rate": 9.779256691999829e-05, "loss": 0.0061, "step": 14020 }, { "epoch": 147.68421052631578, "grad_norm": 0.12586738169193268, "learning_rate": 9.778770557461403e-05, "loss": 0.0068, "step": 14030 }, { "epoch": 147.78947368421052, "grad_norm": 0.08902116864919662, "learning_rate": 9.778283900323812e-05, "loss": 0.0065, "step": 14040 }, { "epoch": 147.89473684210526, "grad_norm": 0.12306950241327286, "learning_rate": 9.777796720640277e-05, "loss": 0.0064, "step": 14050 }, { "epoch": 148.0, "grad_norm": 0.10884469002485275, "learning_rate": 9.777309018464078e-05, "loss": 0.0059, "step": 14060 }, { "epoch": 148.10526315789474, "grad_norm": 0.10513205826282501, "learning_rate": 9.776820793848547e-05, "loss": 0.0064, "step": 14070 }, { "epoch": 148.21052631578948, "grad_norm": 0.12333497405052185, "learning_rate": 9.776332046847075e-05, "loss": 0.0058, "step": 14080 }, { "epoch": 148.31578947368422, "grad_norm": 0.09083478152751923, "learning_rate": 9.775842777513111e-05, "loss": 0.0058, "step": 14090 }, { "epoch": 148.42105263157896, "grad_norm": 0.10982291400432587, "learning_rate": 9.775352985900163e-05, "loss": 0.006, "step": 14100 }, { "epoch": 148.52631578947367, "grad_norm": 0.11592737585306168, "learning_rate": 9.774862672061791e-05, "loss": 0.0062, "step": 14110 }, { "epoch": 148.6315789473684, "grad_norm": 0.08607106655836105, "learning_rate": 9.774371836051616e-05, "loss": 0.0072, "step": 14120 }, { "epoch": 148.73684210526315, "grad_norm": 0.1089625284075737, "learning_rate": 9.773880477923315e-05, "loss": 0.0066, "step": 14130 }, { "epoch": 148.8421052631579, "grad_norm": 0.10424116998910904, "learning_rate": 9.773388597730623e-05, "loss": 0.0066, "step": 14140 }, { "epoch": 148.94736842105263, "grad_norm": 0.11732890456914902, "learning_rate": 9.77289619552733e-05, "loss": 0.0057, "step": 14150 }, { "epoch": 149.05263157894737, "grad_norm": 0.13491962850093842, "learning_rate": 9.772403271367285e-05, "loss": 0.006, "step": 14160 }, { "epoch": 149.1578947368421, "grad_norm": 0.09297728538513184, "learning_rate": 9.771909825304396e-05, "loss": 0.0061, "step": 14170 }, { "epoch": 149.26315789473685, "grad_norm": 0.08795227855443954, "learning_rate": 9.771415857392619e-05, "loss": 0.0068, "step": 14180 }, { "epoch": 149.3684210526316, "grad_norm": 0.08932247757911682, "learning_rate": 9.770921367685978e-05, "loss": 0.0069, "step": 14190 }, { "epoch": 149.47368421052633, "grad_norm": 0.09641904383897781, "learning_rate": 9.770426356238551e-05, "loss": 0.0068, "step": 14200 }, { "epoch": 149.57894736842104, "grad_norm": 0.0879126638174057, "learning_rate": 9.769930823104469e-05, "loss": 0.0066, "step": 14210 }, { "epoch": 149.68421052631578, "grad_norm": 0.13338081538677216, "learning_rate": 9.769434768337926e-05, "loss": 0.0071, "step": 14220 }, { "epoch": 149.78947368421052, "grad_norm": 0.09619131684303284, "learning_rate": 9.768938191993164e-05, "loss": 0.0059, "step": 14230 }, { "epoch": 149.89473684210526, "grad_norm": 0.11410927772521973, "learning_rate": 9.768441094124494e-05, "loss": 0.0064, "step": 14240 }, { "epoch": 150.0, "grad_norm": 0.10342440754175186, "learning_rate": 9.767943474786275e-05, "loss": 0.0061, "step": 14250 }, { "epoch": 150.10526315789474, "grad_norm": 0.14583691954612732, "learning_rate": 9.767445334032923e-05, "loss": 0.0061, "step": 14260 }, { "epoch": 150.21052631578948, "grad_norm": 0.14853619039058685, "learning_rate": 9.766946671918919e-05, "loss": 0.0072, "step": 14270 }, { "epoch": 150.31578947368422, "grad_norm": 0.11536464840173721, "learning_rate": 9.766447488498796e-05, "loss": 0.0064, "step": 14280 }, { "epoch": 150.42105263157896, "grad_norm": 0.08865881711244583, "learning_rate": 9.765947783827139e-05, "loss": 0.006, "step": 14290 }, { "epoch": 150.52631578947367, "grad_norm": 0.11143821477890015, "learning_rate": 9.765447557958599e-05, "loss": 0.0064, "step": 14300 }, { "epoch": 150.6315789473684, "grad_norm": 0.11787416785955429, "learning_rate": 9.764946810947879e-05, "loss": 0.0059, "step": 14310 }, { "epoch": 150.73684210526315, "grad_norm": 0.11179876327514648, "learning_rate": 9.764445542849738e-05, "loss": 0.006, "step": 14320 }, { "epoch": 150.8421052631579, "grad_norm": 0.08205567300319672, "learning_rate": 9.763943753718998e-05, "loss": 0.0063, "step": 14330 }, { "epoch": 150.94736842105263, "grad_norm": 0.10034269839525223, "learning_rate": 9.76344144361053e-05, "loss": 0.0058, "step": 14340 }, { "epoch": 151.05263157894737, "grad_norm": 0.10845828056335449, "learning_rate": 9.762938612579269e-05, "loss": 0.0063, "step": 14350 }, { "epoch": 151.1578947368421, "grad_norm": 0.12671522796154022, "learning_rate": 9.762435260680202e-05, "loss": 0.0062, "step": 14360 }, { "epoch": 151.26315789473685, "grad_norm": 0.09339872002601624, "learning_rate": 9.761931387968373e-05, "loss": 0.0061, "step": 14370 }, { "epoch": 151.3684210526316, "grad_norm": 0.10098486393690109, "learning_rate": 9.76142699449889e-05, "loss": 0.0059, "step": 14380 }, { "epoch": 151.47368421052633, "grad_norm": 0.136272132396698, "learning_rate": 9.760922080326908e-05, "loss": 0.0064, "step": 14390 }, { "epoch": 151.57894736842104, "grad_norm": 0.09693536907434464, "learning_rate": 9.760416645507644e-05, "loss": 0.0063, "step": 14400 }, { "epoch": 151.68421052631578, "grad_norm": 0.0985080823302269, "learning_rate": 9.759910690096375e-05, "loss": 0.0061, "step": 14410 }, { "epoch": 151.78947368421052, "grad_norm": 0.10717394202947617, "learning_rate": 9.759404214148429e-05, "loss": 0.0063, "step": 14420 }, { "epoch": 151.89473684210526, "grad_norm": 0.10205375403165817, "learning_rate": 9.758897217719191e-05, "loss": 0.0061, "step": 14430 }, { "epoch": 152.0, "grad_norm": 0.1410951465368271, "learning_rate": 9.758389700864113e-05, "loss": 0.0056, "step": 14440 }, { "epoch": 152.10526315789474, "grad_norm": 0.13772372901439667, "learning_rate": 9.757881663638688e-05, "loss": 0.0068, "step": 14450 }, { "epoch": 152.21052631578948, "grad_norm": 0.11870266497135162, "learning_rate": 9.757373106098478e-05, "loss": 0.0069, "step": 14460 }, { "epoch": 152.31578947368422, "grad_norm": 0.12423451989889145, "learning_rate": 9.756864028299097e-05, "loss": 0.007, "step": 14470 }, { "epoch": 152.42105263157896, "grad_norm": 0.07461009174585342, "learning_rate": 9.75635443029622e-05, "loss": 0.0059, "step": 14480 }, { "epoch": 152.52631578947367, "grad_norm": 0.08727014809846878, "learning_rate": 9.755844312145572e-05, "loss": 0.0053, "step": 14490 }, { "epoch": 152.6315789473684, "grad_norm": 0.10781914740800858, "learning_rate": 9.755333673902941e-05, "loss": 0.0059, "step": 14500 }, { "epoch": 152.73684210526315, "grad_norm": 0.09212779253721237, "learning_rate": 9.75482251562417e-05, "loss": 0.0052, "step": 14510 }, { "epoch": 152.8421052631579, "grad_norm": 0.09171384572982788, "learning_rate": 9.754310837365155e-05, "loss": 0.0057, "step": 14520 }, { "epoch": 152.94736842105263, "grad_norm": 0.07735206186771393, "learning_rate": 9.753798639181856e-05, "loss": 0.0054, "step": 14530 }, { "epoch": 153.05263157894737, "grad_norm": 0.12397102266550064, "learning_rate": 9.753285921130286e-05, "loss": 0.0059, "step": 14540 }, { "epoch": 153.1578947368421, "grad_norm": 0.0944150909781456, "learning_rate": 9.752772683266512e-05, "loss": 0.0063, "step": 14550 }, { "epoch": 153.26315789473685, "grad_norm": 0.09865710139274597, "learning_rate": 9.752258925646665e-05, "loss": 0.0061, "step": 14560 }, { "epoch": 153.3684210526316, "grad_norm": 0.0964842364192009, "learning_rate": 9.751744648326926e-05, "loss": 0.0054, "step": 14570 }, { "epoch": 153.47368421052633, "grad_norm": 0.08011260628700256, "learning_rate": 9.751229851363536e-05, "loss": 0.0057, "step": 14580 }, { "epoch": 153.57894736842104, "grad_norm": 0.09720030426979065, "learning_rate": 9.750714534812793e-05, "loss": 0.0063, "step": 14590 }, { "epoch": 153.68421052631578, "grad_norm": 0.1020357757806778, "learning_rate": 9.750198698731053e-05, "loss": 0.0055, "step": 14600 }, { "epoch": 153.78947368421052, "grad_norm": 0.10864708572626114, "learning_rate": 9.749682343174722e-05, "loss": 0.0057, "step": 14610 }, { "epoch": 153.89473684210526, "grad_norm": 0.07444767653942108, "learning_rate": 9.749165468200272e-05, "loss": 0.0053, "step": 14620 }, { "epoch": 154.0, "grad_norm": 0.13548637926578522, "learning_rate": 9.748648073864229e-05, "loss": 0.007, "step": 14630 }, { "epoch": 154.10526315789474, "grad_norm": 0.08843310177326202, "learning_rate": 9.748130160223168e-05, "loss": 0.0059, "step": 14640 }, { "epoch": 154.21052631578948, "grad_norm": 0.09739404916763306, "learning_rate": 9.747611727333734e-05, "loss": 0.0059, "step": 14650 }, { "epoch": 154.31578947368422, "grad_norm": 0.10842596739530563, "learning_rate": 9.74709277525262e-05, "loss": 0.0066, "step": 14660 }, { "epoch": 154.42105263157896, "grad_norm": 0.08895198255777359, "learning_rate": 9.746573304036576e-05, "loss": 0.0066, "step": 14670 }, { "epoch": 154.52631578947367, "grad_norm": 0.10492676496505737, "learning_rate": 9.746053313742412e-05, "loss": 0.0056, "step": 14680 }, { "epoch": 154.6315789473684, "grad_norm": 0.1041608601808548, "learning_rate": 9.745532804426994e-05, "loss": 0.0066, "step": 14690 }, { "epoch": 154.73684210526315, "grad_norm": 0.07015719264745712, "learning_rate": 9.745011776147242e-05, "loss": 0.0059, "step": 14700 }, { "epoch": 154.8421052631579, "grad_norm": 0.09705916792154312, "learning_rate": 9.744490228960138e-05, "loss": 0.0055, "step": 14710 }, { "epoch": 154.94736842105263, "grad_norm": 0.07972677797079086, "learning_rate": 9.743968162922713e-05, "loss": 0.006, "step": 14720 }, { "epoch": 155.05263157894737, "grad_norm": 0.12203462421894073, "learning_rate": 9.743445578092064e-05, "loss": 0.0065, "step": 14730 }, { "epoch": 155.1578947368421, "grad_norm": 0.12093003839254379, "learning_rate": 9.742922474525338e-05, "loss": 0.0057, "step": 14740 }, { "epoch": 155.26315789473685, "grad_norm": 0.10076072067022324, "learning_rate": 9.742398852279741e-05, "loss": 0.0068, "step": 14750 }, { "epoch": 155.3684210526316, "grad_norm": 0.08312252908945084, "learning_rate": 9.741874711412535e-05, "loss": 0.0061, "step": 14760 }, { "epoch": 155.47368421052633, "grad_norm": 0.11630146205425262, "learning_rate": 9.741350051981042e-05, "loss": 0.006, "step": 14770 }, { "epoch": 155.57894736842104, "grad_norm": 0.11526572704315186, "learning_rate": 9.740824874042633e-05, "loss": 0.0063, "step": 14780 }, { "epoch": 155.68421052631578, "grad_norm": 0.1659160852432251, "learning_rate": 9.740299177654746e-05, "loss": 0.0061, "step": 14790 }, { "epoch": 155.78947368421052, "grad_norm": 0.11650712788105011, "learning_rate": 9.739772962874867e-05, "loss": 0.0057, "step": 14800 }, { "epoch": 155.89473684210526, "grad_norm": 0.13062746822834015, "learning_rate": 9.739246229760541e-05, "loss": 0.0068, "step": 14810 }, { "epoch": 156.0, "grad_norm": 0.17034994065761566, "learning_rate": 9.738718978369376e-05, "loss": 0.0067, "step": 14820 }, { "epoch": 156.10526315789474, "grad_norm": 0.12568596005439758, "learning_rate": 9.738191208759025e-05, "loss": 0.0058, "step": 14830 }, { "epoch": 156.21052631578948, "grad_norm": 0.13174381852149963, "learning_rate": 9.73766292098721e-05, "loss": 0.007, "step": 14840 }, { "epoch": 156.31578947368422, "grad_norm": 0.10607745498418808, "learning_rate": 9.737134115111699e-05, "loss": 0.0069, "step": 14850 }, { "epoch": 156.42105263157896, "grad_norm": 0.11391407251358032, "learning_rate": 9.736604791190323e-05, "loss": 0.0052, "step": 14860 }, { "epoch": 156.52631578947367, "grad_norm": 0.09663858264684677, "learning_rate": 9.73607494928097e-05, "loss": 0.006, "step": 14870 }, { "epoch": 156.6315789473684, "grad_norm": 0.11927337199449539, "learning_rate": 9.735544589441581e-05, "loss": 0.0067, "step": 14880 }, { "epoch": 156.73684210526315, "grad_norm": 0.09161271899938583, "learning_rate": 9.735013711730154e-05, "loss": 0.0054, "step": 14890 }, { "epoch": 156.8421052631579, "grad_norm": 0.09159369766712189, "learning_rate": 9.734482316204747e-05, "loss": 0.0053, "step": 14900 }, { "epoch": 156.94736842105263, "grad_norm": 0.10474693775177002, "learning_rate": 9.733950402923473e-05, "loss": 0.0057, "step": 14910 }, { "epoch": 157.05263157894737, "grad_norm": 0.11188424378633499, "learning_rate": 9.7334179719445e-05, "loss": 0.0069, "step": 14920 }, { "epoch": 157.1578947368421, "grad_norm": 0.09962841123342514, "learning_rate": 9.732885023326053e-05, "loss": 0.0053, "step": 14930 }, { "epoch": 157.26315789473685, "grad_norm": 0.12412749975919724, "learning_rate": 9.732351557126418e-05, "loss": 0.0058, "step": 14940 }, { "epoch": 157.3684210526316, "grad_norm": 0.09029325097799301, "learning_rate": 9.731817573403929e-05, "loss": 0.0062, "step": 14950 }, { "epoch": 157.47368421052633, "grad_norm": 0.11267896741628647, "learning_rate": 9.731283072216985e-05, "loss": 0.0056, "step": 14960 }, { "epoch": 157.57894736842104, "grad_norm": 0.0821489617228508, "learning_rate": 9.730748053624039e-05, "loss": 0.0064, "step": 14970 }, { "epoch": 157.68421052631578, "grad_norm": 0.09562251716852188, "learning_rate": 9.730212517683598e-05, "loss": 0.0054, "step": 14980 }, { "epoch": 157.78947368421052, "grad_norm": 0.09023728221654892, "learning_rate": 9.729676464454228e-05, "loss": 0.0063, "step": 14990 }, { "epoch": 157.89473684210526, "grad_norm": 0.09799344837665558, "learning_rate": 9.72913989399455e-05, "loss": 0.0065, "step": 15000 }, { "epoch": 158.0, "grad_norm": 0.11649885028600693, "learning_rate": 9.728602806363242e-05, "loss": 0.0062, "step": 15010 }, { "epoch": 158.10526315789474, "grad_norm": 0.07598704099655151, "learning_rate": 9.728065201619043e-05, "loss": 0.006, "step": 15020 }, { "epoch": 158.21052631578948, "grad_norm": 0.11433202773332596, "learning_rate": 9.727527079820742e-05, "loss": 0.0054, "step": 15030 }, { "epoch": 158.31578947368422, "grad_norm": 0.09383378177881241, "learning_rate": 9.726988441027186e-05, "loss": 0.0054, "step": 15040 }, { "epoch": 158.42105263157896, "grad_norm": 0.12668243050575256, "learning_rate": 9.726449285297281e-05, "loss": 0.0056, "step": 15050 }, { "epoch": 158.52631578947367, "grad_norm": 0.0879107117652893, "learning_rate": 9.72590961268999e-05, "loss": 0.0059, "step": 15060 }, { "epoch": 158.6315789473684, "grad_norm": 0.11746475100517273, "learning_rate": 9.725369423264328e-05, "loss": 0.0063, "step": 15070 }, { "epoch": 158.73684210526315, "grad_norm": 0.11951445043087006, "learning_rate": 9.72482871707937e-05, "loss": 0.0058, "step": 15080 }, { "epoch": 158.8421052631579, "grad_norm": 0.11609849333763123, "learning_rate": 9.724287494194247e-05, "loss": 0.0065, "step": 15090 }, { "epoch": 158.94736842105263, "grad_norm": 0.11217866837978363, "learning_rate": 9.723745754668147e-05, "loss": 0.006, "step": 15100 }, { "epoch": 159.05263157894737, "grad_norm": 0.1589696705341339, "learning_rate": 9.723203498560313e-05, "loss": 0.006, "step": 15110 }, { "epoch": 159.1578947368421, "grad_norm": 0.09902767091989517, "learning_rate": 9.722660725930046e-05, "loss": 0.0067, "step": 15120 }, { "epoch": 159.26315789473685, "grad_norm": 0.12807710468769073, "learning_rate": 9.722117436836702e-05, "loss": 0.0055, "step": 15130 }, { "epoch": 159.3684210526316, "grad_norm": 0.10024532675743103, "learning_rate": 9.721573631339696e-05, "loss": 0.0062, "step": 15140 }, { "epoch": 159.47368421052633, "grad_norm": 0.09242168068885803, "learning_rate": 9.721029309498494e-05, "loss": 0.006, "step": 15150 }, { "epoch": 159.57894736842104, "grad_norm": 0.08846838772296906, "learning_rate": 9.720484471372627e-05, "loss": 0.0059, "step": 15160 }, { "epoch": 159.68421052631578, "grad_norm": 0.1001964658498764, "learning_rate": 9.719939117021673e-05, "loss": 0.006, "step": 15170 }, { "epoch": 159.78947368421052, "grad_norm": 0.10001203417778015, "learning_rate": 9.719393246505275e-05, "loss": 0.0062, "step": 15180 }, { "epoch": 159.89473684210526, "grad_norm": 0.12100008130073547, "learning_rate": 9.718846859883128e-05, "loss": 0.0056, "step": 15190 }, { "epoch": 160.0, "grad_norm": 0.13505269587039948, "learning_rate": 9.718299957214982e-05, "loss": 0.0053, "step": 15200 }, { "epoch": 160.10526315789474, "grad_norm": 0.10815172642469406, "learning_rate": 9.717752538560646e-05, "loss": 0.0069, "step": 15210 }, { "epoch": 160.21052631578948, "grad_norm": 0.11616551131010056, "learning_rate": 9.717204603979986e-05, "loss": 0.0059, "step": 15220 }, { "epoch": 160.31578947368422, "grad_norm": 0.1351521611213684, "learning_rate": 9.716656153532922e-05, "loss": 0.0066, "step": 15230 }, { "epoch": 160.42105263157896, "grad_norm": 0.11647641658782959, "learning_rate": 9.716107187279434e-05, "loss": 0.006, "step": 15240 }, { "epoch": 160.52631578947367, "grad_norm": 0.1148165613412857, "learning_rate": 9.715557705279555e-05, "loss": 0.0063, "step": 15250 }, { "epoch": 160.6315789473684, "grad_norm": 0.07822070270776749, "learning_rate": 9.715007707593372e-05, "loss": 0.0059, "step": 15260 }, { "epoch": 160.73684210526315, "grad_norm": 0.08203118294477463, "learning_rate": 9.714457194281036e-05, "loss": 0.0058, "step": 15270 }, { "epoch": 160.8421052631579, "grad_norm": 0.11027787625789642, "learning_rate": 9.713906165402751e-05, "loss": 0.0063, "step": 15280 }, { "epoch": 160.94736842105263, "grad_norm": 0.10436559468507767, "learning_rate": 9.713354621018774e-05, "loss": 0.0055, "step": 15290 }, { "epoch": 161.05263157894737, "grad_norm": 0.08048702776432037, "learning_rate": 9.712802561189422e-05, "loss": 0.0062, "step": 15300 }, { "epoch": 161.1578947368421, "grad_norm": 0.11251739412546158, "learning_rate": 9.712249985975069e-05, "loss": 0.0063, "step": 15310 }, { "epoch": 161.26315789473685, "grad_norm": 0.10447315871715546, "learning_rate": 9.71169689543614e-05, "loss": 0.0068, "step": 15320 }, { "epoch": 161.3684210526316, "grad_norm": 0.09289635717868805, "learning_rate": 9.711143289633123e-05, "loss": 0.0062, "step": 15330 }, { "epoch": 161.47368421052633, "grad_norm": 0.07397113740444183, "learning_rate": 9.710589168626561e-05, "loss": 0.0065, "step": 15340 }, { "epoch": 161.57894736842104, "grad_norm": 0.08559083938598633, "learning_rate": 9.710034532477048e-05, "loss": 0.0056, "step": 15350 }, { "epoch": 161.68421052631578, "grad_norm": 0.09893789142370224, "learning_rate": 9.709479381245239e-05, "loss": 0.0065, "step": 15360 }, { "epoch": 161.78947368421052, "grad_norm": 0.09096041321754456, "learning_rate": 9.708923714991847e-05, "loss": 0.0057, "step": 15370 }, { "epoch": 161.89473684210526, "grad_norm": 0.0913262665271759, "learning_rate": 9.708367533777638e-05, "loss": 0.0067, "step": 15380 }, { "epoch": 162.0, "grad_norm": 0.09572716057300568, "learning_rate": 9.707810837663431e-05, "loss": 0.0052, "step": 15390 }, { "epoch": 162.10526315789474, "grad_norm": 0.10408979654312134, "learning_rate": 9.707253626710113e-05, "loss": 0.0056, "step": 15400 }, { "epoch": 162.21052631578948, "grad_norm": 0.07520575076341629, "learning_rate": 9.706695900978613e-05, "loss": 0.0057, "step": 15410 }, { "epoch": 162.31578947368422, "grad_norm": 0.07208319753408432, "learning_rate": 9.706137660529926e-05, "loss": 0.0067, "step": 15420 }, { "epoch": 162.42105263157896, "grad_norm": 0.07458256185054779, "learning_rate": 9.705578905425101e-05, "loss": 0.0061, "step": 15430 }, { "epoch": 162.52631578947367, "grad_norm": 0.08696509152650833, "learning_rate": 9.705019635725241e-05, "loss": 0.006, "step": 15440 }, { "epoch": 162.6315789473684, "grad_norm": 0.0813346654176712, "learning_rate": 9.704459851491508e-05, "loss": 0.0057, "step": 15450 }, { "epoch": 162.73684210526315, "grad_norm": 0.09850164502859116, "learning_rate": 9.703899552785118e-05, "loss": 0.006, "step": 15460 }, { "epoch": 162.8421052631579, "grad_norm": 0.08877905458211899, "learning_rate": 9.703338739667346e-05, "loss": 0.0056, "step": 15470 }, { "epoch": 162.94736842105263, "grad_norm": 0.10328131169080734, "learning_rate": 9.70277741219952e-05, "loss": 0.0059, "step": 15480 }, { "epoch": 163.05263157894737, "grad_norm": 0.14689330756664276, "learning_rate": 9.702215570443027e-05, "loss": 0.0069, "step": 15490 }, { "epoch": 163.1578947368421, "grad_norm": 0.15254616737365723, "learning_rate": 9.701653214459309e-05, "loss": 0.0061, "step": 15500 }, { "epoch": 163.26315789473685, "grad_norm": 0.11372317373752594, "learning_rate": 9.701090344309865e-05, "loss": 0.0056, "step": 15510 }, { "epoch": 163.3684210526316, "grad_norm": 0.12281787395477295, "learning_rate": 9.700526960056247e-05, "loss": 0.0056, "step": 15520 }, { "epoch": 163.47368421052633, "grad_norm": 0.08897586166858673, "learning_rate": 9.699963061760068e-05, "loss": 0.0051, "step": 15530 }, { "epoch": 163.57894736842104, "grad_norm": 0.0993124321103096, "learning_rate": 9.699398649482997e-05, "loss": 0.006, "step": 15540 }, { "epoch": 163.68421052631578, "grad_norm": 0.09951714426279068, "learning_rate": 9.698833723286753e-05, "loss": 0.0058, "step": 15550 }, { "epoch": 163.78947368421052, "grad_norm": 0.10883090645074844, "learning_rate": 9.698268283233118e-05, "loss": 0.0062, "step": 15560 }, { "epoch": 163.89473684210526, "grad_norm": 0.131513312458992, "learning_rate": 9.697702329383929e-05, "loss": 0.006, "step": 15570 }, { "epoch": 164.0, "grad_norm": 0.14266039431095123, "learning_rate": 9.697135861801074e-05, "loss": 0.007, "step": 15580 }, { "epoch": 164.10526315789474, "grad_norm": 0.11636766046285629, "learning_rate": 9.696568880546505e-05, "loss": 0.0063, "step": 15590 }, { "epoch": 164.21052631578948, "grad_norm": 0.09705577045679092, "learning_rate": 9.696001385682223e-05, "loss": 0.0055, "step": 15600 }, { "epoch": 164.31578947368422, "grad_norm": 0.10131474584341049, "learning_rate": 9.695433377270291e-05, "loss": 0.0063, "step": 15610 }, { "epoch": 164.42105263157896, "grad_norm": 0.12224453687667847, "learning_rate": 9.694864855372824e-05, "loss": 0.0056, "step": 15620 }, { "epoch": 164.52631578947367, "grad_norm": 0.10917645692825317, "learning_rate": 9.694295820051995e-05, "loss": 0.0058, "step": 15630 }, { "epoch": 164.6315789473684, "grad_norm": 0.09978047758340836, "learning_rate": 9.693726271370032e-05, "loss": 0.006, "step": 15640 }, { "epoch": 164.73684210526315, "grad_norm": 0.09088002890348434, "learning_rate": 9.693156209389221e-05, "loss": 0.0053, "step": 15650 }, { "epoch": 164.8421052631579, "grad_norm": 0.10270854830741882, "learning_rate": 9.692585634171905e-05, "loss": 0.0053, "step": 15660 }, { "epoch": 164.94736842105263, "grad_norm": 0.0995563417673111, "learning_rate": 9.692014545780476e-05, "loss": 0.0064, "step": 15670 }, { "epoch": 165.05263157894737, "grad_norm": 0.14027832448482513, "learning_rate": 9.691442944277393e-05, "loss": 0.0064, "step": 15680 }, { "epoch": 165.1578947368421, "grad_norm": 0.11869968473911285, "learning_rate": 9.690870829725162e-05, "loss": 0.0065, "step": 15690 }, { "epoch": 165.26315789473685, "grad_norm": 0.1107565239071846, "learning_rate": 9.69029820218635e-05, "loss": 0.0062, "step": 15700 }, { "epoch": 165.3684210526316, "grad_norm": 0.11681481450796127, "learning_rate": 9.689725061723579e-05, "loss": 0.0068, "step": 15710 }, { "epoch": 165.47368421052633, "grad_norm": 0.08099032938480377, "learning_rate": 9.689151408399527e-05, "loss": 0.0064, "step": 15720 }, { "epoch": 165.57894736842104, "grad_norm": 0.09294518828392029, "learning_rate": 9.688577242276924e-05, "loss": 0.0073, "step": 15730 }, { "epoch": 165.68421052631578, "grad_norm": 0.10443460941314697, "learning_rate": 9.688002563418566e-05, "loss": 0.0059, "step": 15740 }, { "epoch": 165.78947368421052, "grad_norm": 0.08506909757852554, "learning_rate": 9.687427371887293e-05, "loss": 0.006, "step": 15750 }, { "epoch": 165.89473684210526, "grad_norm": 0.10517652332782745, "learning_rate": 9.686851667746012e-05, "loss": 0.0063, "step": 15760 }, { "epoch": 166.0, "grad_norm": 0.13594461977481842, "learning_rate": 9.686275451057677e-05, "loss": 0.0062, "step": 15770 }, { "epoch": 166.10526315789474, "grad_norm": 0.0907692238688469, "learning_rate": 9.685698721885308e-05, "loss": 0.0068, "step": 15780 }, { "epoch": 166.21052631578948, "grad_norm": 0.09986280649900436, "learning_rate": 9.68512148029197e-05, "loss": 0.0061, "step": 15790 }, { "epoch": 166.31578947368422, "grad_norm": 0.10269057005643845, "learning_rate": 9.684543726340791e-05, "loss": 0.0067, "step": 15800 }, { "epoch": 166.42105263157896, "grad_norm": 0.10779926925897598, "learning_rate": 9.683965460094952e-05, "loss": 0.0071, "step": 15810 }, { "epoch": 166.52631578947367, "grad_norm": 0.09947656095027924, "learning_rate": 9.683386681617694e-05, "loss": 0.0063, "step": 15820 }, { "epoch": 166.6315789473684, "grad_norm": 0.0981997698545456, "learning_rate": 9.68280739097231e-05, "loss": 0.0055, "step": 15830 }, { "epoch": 166.73684210526315, "grad_norm": 0.10985822975635529, "learning_rate": 9.682227588222148e-05, "loss": 0.0059, "step": 15840 }, { "epoch": 166.8421052631579, "grad_norm": 0.08261359483003616, "learning_rate": 9.681647273430618e-05, "loss": 0.0058, "step": 15850 }, { "epoch": 166.94736842105263, "grad_norm": 0.11951493471860886, "learning_rate": 9.681066446661182e-05, "loss": 0.0058, "step": 15860 }, { "epoch": 167.05263157894737, "grad_norm": 0.1082393229007721, "learning_rate": 9.680485107977357e-05, "loss": 0.0069, "step": 15870 }, { "epoch": 167.1578947368421, "grad_norm": 0.12003245949745178, "learning_rate": 9.679903257442716e-05, "loss": 0.0072, "step": 15880 }, { "epoch": 167.26315789473685, "grad_norm": 0.10545317828655243, "learning_rate": 9.679320895120891e-05, "loss": 0.0054, "step": 15890 }, { "epoch": 167.3684210526316, "grad_norm": 0.10432327538728714, "learning_rate": 9.67873802107557e-05, "loss": 0.0058, "step": 15900 }, { "epoch": 167.47368421052633, "grad_norm": 0.09775014221668243, "learning_rate": 9.67815463537049e-05, "loss": 0.0067, "step": 15910 }, { "epoch": 167.57894736842104, "grad_norm": 0.10219010710716248, "learning_rate": 9.677570738069457e-05, "loss": 0.0061, "step": 15920 }, { "epoch": 167.68421052631578, "grad_norm": 0.08753865212202072, "learning_rate": 9.676986329236318e-05, "loss": 0.0061, "step": 15930 }, { "epoch": 167.78947368421052, "grad_norm": 0.10096964240074158, "learning_rate": 9.676401408934987e-05, "loss": 0.0058, "step": 15940 }, { "epoch": 167.89473684210526, "grad_norm": 0.14665286242961884, "learning_rate": 9.675815977229428e-05, "loss": 0.0057, "step": 15950 }, { "epoch": 168.0, "grad_norm": 0.11779864132404327, "learning_rate": 9.675230034183664e-05, "loss": 0.006, "step": 15960 }, { "epoch": 168.10526315789474, "grad_norm": 0.1164892390370369, "learning_rate": 9.674643579861773e-05, "loss": 0.0067, "step": 15970 }, { "epoch": 168.21052631578948, "grad_norm": 0.08913131058216095, "learning_rate": 9.674056614327886e-05, "loss": 0.0057, "step": 15980 }, { "epoch": 168.31578947368422, "grad_norm": 0.11129714548587799, "learning_rate": 9.673469137646198e-05, "loss": 0.0057, "step": 15990 }, { "epoch": 168.42105263157896, "grad_norm": 0.10419096052646637, "learning_rate": 9.67288114988095e-05, "loss": 0.0051, "step": 16000 }, { "epoch": 168.52631578947367, "grad_norm": 0.09184510260820389, "learning_rate": 9.672292651096447e-05, "loss": 0.0062, "step": 16010 }, { "epoch": 168.6315789473684, "grad_norm": 0.11464101076126099, "learning_rate": 9.671703641357042e-05, "loss": 0.0061, "step": 16020 }, { "epoch": 168.73684210526315, "grad_norm": 0.10781626403331757, "learning_rate": 9.67111412072715e-05, "loss": 0.0054, "step": 16030 }, { "epoch": 168.8421052631579, "grad_norm": 0.08849211782217026, "learning_rate": 9.670524089271242e-05, "loss": 0.0064, "step": 16040 }, { "epoch": 168.94736842105263, "grad_norm": 0.10730911046266556, "learning_rate": 9.669933547053842e-05, "loss": 0.0067, "step": 16050 }, { "epoch": 169.05263157894737, "grad_norm": 0.12665890157222748, "learning_rate": 9.669342494139531e-05, "loss": 0.0065, "step": 16060 }, { "epoch": 169.1578947368421, "grad_norm": 0.10394436120986938, "learning_rate": 9.668750930592943e-05, "loss": 0.0061, "step": 16070 }, { "epoch": 169.26315789473685, "grad_norm": 0.11444704234600067, "learning_rate": 9.668158856478775e-05, "loss": 0.0061, "step": 16080 }, { "epoch": 169.3684210526316, "grad_norm": 0.1064368188381195, "learning_rate": 9.66756627186177e-05, "loss": 0.006, "step": 16090 }, { "epoch": 169.47368421052633, "grad_norm": 0.10948581993579865, "learning_rate": 9.666973176806737e-05, "loss": 0.0063, "step": 16100 }, { "epoch": 169.57894736842104, "grad_norm": 0.10355694591999054, "learning_rate": 9.666379571378534e-05, "loss": 0.0057, "step": 16110 }, { "epoch": 169.68421052631578, "grad_norm": 0.12559114396572113, "learning_rate": 9.665785455642076e-05, "loss": 0.0061, "step": 16120 }, { "epoch": 169.78947368421052, "grad_norm": 0.10479604452848434, "learning_rate": 9.665190829662337e-05, "loss": 0.0061, "step": 16130 }, { "epoch": 169.89473684210526, "grad_norm": 0.14055009186267853, "learning_rate": 9.664595693504342e-05, "loss": 0.006, "step": 16140 }, { "epoch": 170.0, "grad_norm": 0.12050525099039078, "learning_rate": 9.664000047233175e-05, "loss": 0.0064, "step": 16150 }, { "epoch": 170.10526315789474, "grad_norm": 0.08702632784843445, "learning_rate": 9.663403890913976e-05, "loss": 0.0057, "step": 16160 }, { "epoch": 170.21052631578948, "grad_norm": 0.12009923160076141, "learning_rate": 9.662807224611938e-05, "loss": 0.0067, "step": 16170 }, { "epoch": 170.31578947368422, "grad_norm": 0.08324727416038513, "learning_rate": 9.662210048392311e-05, "loss": 0.0054, "step": 16180 }, { "epoch": 170.42105263157896, "grad_norm": 0.10812536627054214, "learning_rate": 9.661612362320405e-05, "loss": 0.0055, "step": 16190 }, { "epoch": 170.52631578947367, "grad_norm": 0.09282735735177994, "learning_rate": 9.661014166461579e-05, "loss": 0.0058, "step": 16200 }, { "epoch": 170.6315789473684, "grad_norm": 0.09891632944345474, "learning_rate": 9.66041546088125e-05, "loss": 0.0053, "step": 16210 }, { "epoch": 170.73684210526315, "grad_norm": 0.08778538554906845, "learning_rate": 9.659816245644895e-05, "loss": 0.0065, "step": 16220 }, { "epoch": 170.8421052631579, "grad_norm": 0.09970757365226746, "learning_rate": 9.65921652081804e-05, "loss": 0.006, "step": 16230 }, { "epoch": 170.94736842105263, "grad_norm": 0.07879586517810822, "learning_rate": 9.658616286466271e-05, "loss": 0.0055, "step": 16240 }, { "epoch": 171.05263157894737, "grad_norm": 0.09365987777709961, "learning_rate": 9.65801554265523e-05, "loss": 0.0052, "step": 16250 }, { "epoch": 171.1578947368421, "grad_norm": 0.09969311952590942, "learning_rate": 9.657414289450612e-05, "loss": 0.0061, "step": 16260 }, { "epoch": 171.26315789473685, "grad_norm": 0.08913831412792206, "learning_rate": 9.656812526918171e-05, "loss": 0.0051, "step": 16270 }, { "epoch": 171.3684210526316, "grad_norm": 0.10673962533473969, "learning_rate": 9.656210255123712e-05, "loss": 0.0052, "step": 16280 }, { "epoch": 171.47368421052633, "grad_norm": 0.10695390403270721, "learning_rate": 9.6556074741331e-05, "loss": 0.0058, "step": 16290 }, { "epoch": 171.57894736842104, "grad_norm": 0.12365935742855072, "learning_rate": 9.655004184012256e-05, "loss": 0.0061, "step": 16300 }, { "epoch": 171.68421052631578, "grad_norm": 0.12617820501327515, "learning_rate": 9.654400384827152e-05, "loss": 0.0052, "step": 16310 }, { "epoch": 171.78947368421052, "grad_norm": 0.10612557083368301, "learning_rate": 9.653796076643818e-05, "loss": 0.0054, "step": 16320 }, { "epoch": 171.89473684210526, "grad_norm": 0.0950983315706253, "learning_rate": 9.653191259528344e-05, "loss": 0.006, "step": 16330 }, { "epoch": 172.0, "grad_norm": 0.1300671100616455, "learning_rate": 9.65258593354687e-05, "loss": 0.006, "step": 16340 }, { "epoch": 172.10526315789474, "grad_norm": 0.0921817198395729, "learning_rate": 9.651980098765591e-05, "loss": 0.0052, "step": 16350 }, { "epoch": 172.21052631578948, "grad_norm": 0.08947551250457764, "learning_rate": 9.651373755250765e-05, "loss": 0.0052, "step": 16360 }, { "epoch": 172.31578947368422, "grad_norm": 0.09370794147253036, "learning_rate": 9.650766903068697e-05, "loss": 0.0064, "step": 16370 }, { "epoch": 172.42105263157896, "grad_norm": 0.09645667672157288, "learning_rate": 9.650159542285753e-05, "loss": 0.0061, "step": 16380 }, { "epoch": 172.52631578947367, "grad_norm": 0.0871637836098671, "learning_rate": 9.649551672968353e-05, "loss": 0.0058, "step": 16390 }, { "epoch": 172.6315789473684, "grad_norm": 0.09386560320854187, "learning_rate": 9.648943295182973e-05, "loss": 0.0053, "step": 16400 }, { "epoch": 172.73684210526315, "grad_norm": 0.09511848539113998, "learning_rate": 9.648334408996144e-05, "loss": 0.0054, "step": 16410 }, { "epoch": 172.8421052631579, "grad_norm": 0.08465959876775742, "learning_rate": 9.647725014474452e-05, "loss": 0.0053, "step": 16420 }, { "epoch": 172.94736842105263, "grad_norm": 0.07781356573104858, "learning_rate": 9.64711511168454e-05, "loss": 0.0061, "step": 16430 }, { "epoch": 173.05263157894737, "grad_norm": 0.1124839261174202, "learning_rate": 9.646504700693108e-05, "loss": 0.0064, "step": 16440 }, { "epoch": 173.1578947368421, "grad_norm": 0.1186879426240921, "learning_rate": 9.645893781566907e-05, "loss": 0.0062, "step": 16450 }, { "epoch": 173.26315789473685, "grad_norm": 0.10134558379650116, "learning_rate": 9.645282354372744e-05, "loss": 0.0069, "step": 16460 }, { "epoch": 173.3684210526316, "grad_norm": 0.12409409135580063, "learning_rate": 9.644670419177491e-05, "loss": 0.0058, "step": 16470 }, { "epoch": 173.47368421052633, "grad_norm": 0.11293862015008926, "learning_rate": 9.644057976048062e-05, "loss": 0.0066, "step": 16480 }, { "epoch": 173.57894736842104, "grad_norm": 0.10609892755746841, "learning_rate": 9.643445025051435e-05, "loss": 0.0063, "step": 16490 }, { "epoch": 173.68421052631578, "grad_norm": 0.1380523145198822, "learning_rate": 9.642831566254641e-05, "loss": 0.0057, "step": 16500 }, { "epoch": 173.78947368421052, "grad_norm": 0.10687968134880066, "learning_rate": 9.642217599724769e-05, "loss": 0.0055, "step": 16510 }, { "epoch": 173.89473684210526, "grad_norm": 0.09793029725551605, "learning_rate": 9.64160312552896e-05, "loss": 0.0048, "step": 16520 }, { "epoch": 174.0, "grad_norm": 0.11989811807870865, "learning_rate": 9.64098814373441e-05, "loss": 0.0054, "step": 16530 }, { "epoch": 174.10526315789474, "grad_norm": 0.09762968868017197, "learning_rate": 9.640372654408374e-05, "loss": 0.0062, "step": 16540 }, { "epoch": 174.21052631578948, "grad_norm": 0.08668243885040283, "learning_rate": 9.639756657618162e-05, "loss": 0.0056, "step": 16550 }, { "epoch": 174.31578947368422, "grad_norm": 0.0855623334646225, "learning_rate": 9.639140153431138e-05, "loss": 0.0056, "step": 16560 }, { "epoch": 174.42105263157896, "grad_norm": 0.11512044817209244, "learning_rate": 9.638523141914721e-05, "loss": 0.0067, "step": 16570 }, { "epoch": 174.52631578947367, "grad_norm": 0.10706936568021774, "learning_rate": 9.637905623136388e-05, "loss": 0.0057, "step": 16580 }, { "epoch": 174.6315789473684, "grad_norm": 0.12963342666625977, "learning_rate": 9.637287597163669e-05, "loss": 0.0051, "step": 16590 }, { "epoch": 174.73684210526315, "grad_norm": 0.11037561297416687, "learning_rate": 9.63666906406415e-05, "loss": 0.0063, "step": 16600 }, { "epoch": 174.8421052631579, "grad_norm": 0.0941309854388237, "learning_rate": 9.636050023905473e-05, "loss": 0.0056, "step": 16610 }, { "epoch": 174.94736842105263, "grad_norm": 0.11097069084644318, "learning_rate": 9.635430476755336e-05, "loss": 0.0059, "step": 16620 }, { "epoch": 175.05263157894737, "grad_norm": 0.1143418624997139, "learning_rate": 9.63481042268149e-05, "loss": 0.006, "step": 16630 }, { "epoch": 175.1578947368421, "grad_norm": 0.09016439318656921, "learning_rate": 9.634189861751745e-05, "loss": 0.0058, "step": 16640 }, { "epoch": 175.26315789473685, "grad_norm": 0.11548951268196106, "learning_rate": 9.633568794033967e-05, "loss": 0.0057, "step": 16650 }, { "epoch": 175.3684210526316, "grad_norm": 0.09563266485929489, "learning_rate": 9.63294721959607e-05, "loss": 0.0057, "step": 16660 }, { "epoch": 175.47368421052633, "grad_norm": 0.09660106152296066, "learning_rate": 9.63232513850603e-05, "loss": 0.0058, "step": 16670 }, { "epoch": 175.57894736842104, "grad_norm": 0.09068552404642105, "learning_rate": 9.631702550831878e-05, "loss": 0.0054, "step": 16680 }, { "epoch": 175.68421052631578, "grad_norm": 0.0733691081404686, "learning_rate": 9.631079456641698e-05, "loss": 0.0057, "step": 16690 }, { "epoch": 175.78947368421052, "grad_norm": 0.09196390956640244, "learning_rate": 9.630455856003632e-05, "loss": 0.0054, "step": 16700 }, { "epoch": 175.89473684210526, "grad_norm": 0.10067440569400787, "learning_rate": 9.629831748985876e-05, "loss": 0.0064, "step": 16710 }, { "epoch": 176.0, "grad_norm": 0.1315017193555832, "learning_rate": 9.629207135656679e-05, "loss": 0.0052, "step": 16720 }, { "epoch": 176.10526315789474, "grad_norm": 0.094994455575943, "learning_rate": 9.628582016084353e-05, "loss": 0.006, "step": 16730 }, { "epoch": 176.21052631578948, "grad_norm": 0.09334778040647507, "learning_rate": 9.627956390337254e-05, "loss": 0.0051, "step": 16740 }, { "epoch": 176.31578947368422, "grad_norm": 0.0861540287733078, "learning_rate": 9.627330258483802e-05, "loss": 0.0053, "step": 16750 }, { "epoch": 176.42105263157896, "grad_norm": 0.11145459860563278, "learning_rate": 9.62670362059247e-05, "loss": 0.0056, "step": 16760 }, { "epoch": 176.52631578947367, "grad_norm": 0.11011065542697906, "learning_rate": 9.626076476731786e-05, "loss": 0.0057, "step": 16770 }, { "epoch": 176.6315789473684, "grad_norm": 0.09117409586906433, "learning_rate": 9.625448826970336e-05, "loss": 0.0052, "step": 16780 }, { "epoch": 176.73684210526315, "grad_norm": 0.08705424517393112, "learning_rate": 9.624820671376755e-05, "loss": 0.0063, "step": 16790 }, { "epoch": 176.8421052631579, "grad_norm": 0.07140037417411804, "learning_rate": 9.62419201001974e-05, "loss": 0.0051, "step": 16800 }, { "epoch": 176.94736842105263, "grad_norm": 0.0981048047542572, "learning_rate": 9.623562842968037e-05, "loss": 0.0056, "step": 16810 }, { "epoch": 177.05263157894737, "grad_norm": 0.08558149635791779, "learning_rate": 9.622933170290454e-05, "loss": 0.005, "step": 16820 }, { "epoch": 177.1578947368421, "grad_norm": 0.08178994804620743, "learning_rate": 9.622302992055849e-05, "loss": 0.0054, "step": 16830 }, { "epoch": 177.26315789473685, "grad_norm": 0.12404053658246994, "learning_rate": 9.62167230833314e-05, "loss": 0.0059, "step": 16840 }, { "epoch": 177.3684210526316, "grad_norm": 0.1298750340938568, "learning_rate": 9.621041119191295e-05, "loss": 0.0056, "step": 16850 }, { "epoch": 177.47368421052633, "grad_norm": 0.12551365792751312, "learning_rate": 9.620409424699342e-05, "loss": 0.0064, "step": 16860 }, { "epoch": 177.57894736842104, "grad_norm": 0.14423397183418274, "learning_rate": 9.619777224926359e-05, "loss": 0.0055, "step": 16870 }, { "epoch": 177.68421052631578, "grad_norm": 0.11666254699230194, "learning_rate": 9.619144519941485e-05, "loss": 0.006, "step": 16880 }, { "epoch": 177.78947368421052, "grad_norm": 0.1046954095363617, "learning_rate": 9.618511309813912e-05, "loss": 0.0045, "step": 16890 }, { "epoch": 177.89473684210526, "grad_norm": 0.11897090077400208, "learning_rate": 9.617877594612886e-05, "loss": 0.0051, "step": 16900 }, { "epoch": 178.0, "grad_norm": 0.08425489813089371, "learning_rate": 9.617243374407707e-05, "loss": 0.0047, "step": 16910 }, { "epoch": 178.10526315789474, "grad_norm": 0.10590685158967972, "learning_rate": 9.616608649267736e-05, "loss": 0.0065, "step": 16920 }, { "epoch": 178.21052631578948, "grad_norm": 0.09558332711458206, "learning_rate": 9.615973419262385e-05, "loss": 0.0053, "step": 16930 }, { "epoch": 178.31578947368422, "grad_norm": 0.09094730764627457, "learning_rate": 9.615337684461119e-05, "loss": 0.0053, "step": 16940 }, { "epoch": 178.42105263157896, "grad_norm": 0.08769617974758148, "learning_rate": 9.614701444933465e-05, "loss": 0.0052, "step": 16950 }, { "epoch": 178.52631578947367, "grad_norm": 0.0785774439573288, "learning_rate": 9.614064700748997e-05, "loss": 0.0061, "step": 16960 }, { "epoch": 178.6315789473684, "grad_norm": 0.09572357684373856, "learning_rate": 9.613427451977352e-05, "loss": 0.005, "step": 16970 }, { "epoch": 178.73684210526315, "grad_norm": 0.0910499170422554, "learning_rate": 9.612789698688216e-05, "loss": 0.0053, "step": 16980 }, { "epoch": 178.8421052631579, "grad_norm": 0.1148422583937645, "learning_rate": 9.612151440951334e-05, "loss": 0.0063, "step": 16990 }, { "epoch": 178.94736842105263, "grad_norm": 0.12671008706092834, "learning_rate": 9.611512678836506e-05, "loss": 0.0052, "step": 17000 }, { "epoch": 179.05263157894737, "grad_norm": 0.13144293427467346, "learning_rate": 9.610873412413584e-05, "loss": 0.0057, "step": 17010 }, { "epoch": 179.1578947368421, "grad_norm": 0.09473998099565506, "learning_rate": 9.610233641752476e-05, "loss": 0.0054, "step": 17020 }, { "epoch": 179.26315789473685, "grad_norm": 0.09532013535499573, "learning_rate": 9.609593366923151e-05, "loss": 0.006, "step": 17030 }, { "epoch": 179.3684210526316, "grad_norm": 0.11495666205883026, "learning_rate": 9.608952587995625e-05, "loss": 0.0057, "step": 17040 }, { "epoch": 179.47368421052633, "grad_norm": 0.09097711741924286, "learning_rate": 9.608311305039972e-05, "loss": 0.0055, "step": 17050 }, { "epoch": 179.57894736842104, "grad_norm": 0.09741006046533585, "learning_rate": 9.607669518126326e-05, "loss": 0.0053, "step": 17060 }, { "epoch": 179.68421052631578, "grad_norm": 0.10862749069929123, "learning_rate": 9.607027227324866e-05, "loss": 0.0063, "step": 17070 }, { "epoch": 179.78947368421052, "grad_norm": 0.10541144758462906, "learning_rate": 9.606384432705837e-05, "loss": 0.0058, "step": 17080 }, { "epoch": 179.89473684210526, "grad_norm": 0.11453728377819061, "learning_rate": 9.60574113433953e-05, "loss": 0.0058, "step": 17090 }, { "epoch": 180.0, "grad_norm": 0.13595616817474365, "learning_rate": 9.6050973322963e-05, "loss": 0.0052, "step": 17100 }, { "epoch": 180.10526315789474, "grad_norm": 0.08164487034082413, "learning_rate": 9.604453026646547e-05, "loss": 0.0065, "step": 17110 }, { "epoch": 180.21052631578948, "grad_norm": 0.11023581773042679, "learning_rate": 9.603808217460735e-05, "loss": 0.0069, "step": 17120 }, { "epoch": 180.31578947368422, "grad_norm": 0.10723347216844559, "learning_rate": 9.603162904809377e-05, "loss": 0.0056, "step": 17130 }, { "epoch": 180.42105263157896, "grad_norm": 0.08650550246238708, "learning_rate": 9.602517088763045e-05, "loss": 0.0053, "step": 17140 }, { "epoch": 180.52631578947367, "grad_norm": 0.0902615636587143, "learning_rate": 9.601870769392365e-05, "loss": 0.0058, "step": 17150 }, { "epoch": 180.6315789473684, "grad_norm": 0.07320526987314224, "learning_rate": 9.601223946768017e-05, "loss": 0.0064, "step": 17160 }, { "epoch": 180.73684210526315, "grad_norm": 0.09108800441026688, "learning_rate": 9.600576620960734e-05, "loss": 0.0054, "step": 17170 }, { "epoch": 180.8421052631579, "grad_norm": 0.09846489131450653, "learning_rate": 9.599928792041308e-05, "loss": 0.007, "step": 17180 }, { "epoch": 180.94736842105263, "grad_norm": 0.10450387001037598, "learning_rate": 9.599280460080587e-05, "loss": 0.0054, "step": 17190 }, { "epoch": 181.05263157894737, "grad_norm": 0.10475200414657593, "learning_rate": 9.59863162514947e-05, "loss": 0.0049, "step": 17200 }, { "epoch": 181.1578947368421, "grad_norm": 0.11362241208553314, "learning_rate": 9.597982287318911e-05, "loss": 0.0061, "step": 17210 }, { "epoch": 181.26315789473685, "grad_norm": 0.08892772346735, "learning_rate": 9.597332446659923e-05, "loss": 0.006, "step": 17220 }, { "epoch": 181.3684210526316, "grad_norm": 0.10979536175727844, "learning_rate": 9.59668210324357e-05, "loss": 0.0055, "step": 17230 }, { "epoch": 181.47368421052633, "grad_norm": 0.11613641679286957, "learning_rate": 9.596031257140974e-05, "loss": 0.0068, "step": 17240 }, { "epoch": 181.57894736842104, "grad_norm": 0.1088760569691658, "learning_rate": 9.59537990842331e-05, "loss": 0.0054, "step": 17250 }, { "epoch": 181.68421052631578, "grad_norm": 0.09740595519542694, "learning_rate": 9.594728057161806e-05, "loss": 0.0055, "step": 17260 }, { "epoch": 181.78947368421052, "grad_norm": 0.13076171278953552, "learning_rate": 9.594075703427752e-05, "loss": 0.0054, "step": 17270 }, { "epoch": 181.89473684210526, "grad_norm": 0.08908698707818985, "learning_rate": 9.593422847292486e-05, "loss": 0.0052, "step": 17280 }, { "epoch": 182.0, "grad_norm": 0.1390591710805893, "learning_rate": 9.592769488827402e-05, "loss": 0.006, "step": 17290 }, { "epoch": 182.10526315789474, "grad_norm": 0.07564643770456314, "learning_rate": 9.592115628103952e-05, "loss": 0.0055, "step": 17300 }, { "epoch": 182.21052631578948, "grad_norm": 0.08598346263170242, "learning_rate": 9.591461265193643e-05, "loss": 0.0056, "step": 17310 }, { "epoch": 182.31578947368422, "grad_norm": 0.1079203188419342, "learning_rate": 9.590806400168032e-05, "loss": 0.0055, "step": 17320 }, { "epoch": 182.42105263157896, "grad_norm": 0.08149214088916779, "learning_rate": 9.590151033098735e-05, "loss": 0.0059, "step": 17330 }, { "epoch": 182.52631578947367, "grad_norm": 0.08353514969348907, "learning_rate": 9.589495164057423e-05, "loss": 0.0056, "step": 17340 }, { "epoch": 182.6315789473684, "grad_norm": 0.09648147970438004, "learning_rate": 9.58883879311582e-05, "loss": 0.0061, "step": 17350 }, { "epoch": 182.73684210526315, "grad_norm": 0.11217759549617767, "learning_rate": 9.588181920345705e-05, "loss": 0.0052, "step": 17360 }, { "epoch": 182.8421052631579, "grad_norm": 0.11430205404758453, "learning_rate": 9.587524545818913e-05, "loss": 0.0056, "step": 17370 }, { "epoch": 182.94736842105263, "grad_norm": 0.1082398071885109, "learning_rate": 9.586866669607335e-05, "loss": 0.0058, "step": 17380 }, { "epoch": 183.05263157894737, "grad_norm": 0.08259918540716171, "learning_rate": 9.586208291782915e-05, "loss": 0.0057, "step": 17390 }, { "epoch": 183.1578947368421, "grad_norm": 0.10179761797189713, "learning_rate": 9.58554941241765e-05, "loss": 0.0066, "step": 17400 }, { "epoch": 183.26315789473685, "grad_norm": 0.07107558846473694, "learning_rate": 9.584890031583596e-05, "loss": 0.0055, "step": 17410 }, { "epoch": 183.3684210526316, "grad_norm": 0.09005969762802124, "learning_rate": 9.584230149352861e-05, "loss": 0.0067, "step": 17420 }, { "epoch": 183.47368421052633, "grad_norm": 0.09055021405220032, "learning_rate": 9.58356976579761e-05, "loss": 0.0063, "step": 17430 }, { "epoch": 183.57894736842104, "grad_norm": 0.10779004544019699, "learning_rate": 9.58290888099006e-05, "loss": 0.0057, "step": 17440 }, { "epoch": 183.68421052631578, "grad_norm": 0.08696645498275757, "learning_rate": 9.582247495002486e-05, "loss": 0.0055, "step": 17450 }, { "epoch": 183.78947368421052, "grad_norm": 0.09732159227132797, "learning_rate": 9.581585607907214e-05, "loss": 0.0053, "step": 17460 }, { "epoch": 183.89473684210526, "grad_norm": 0.14474932849407196, "learning_rate": 9.580923219776628e-05, "loss": 0.0063, "step": 17470 }, { "epoch": 184.0, "grad_norm": 0.1269233226776123, "learning_rate": 9.580260330683167e-05, "loss": 0.0059, "step": 17480 }, { "epoch": 184.10526315789474, "grad_norm": 0.09480349719524384, "learning_rate": 9.579596940699322e-05, "loss": 0.0064, "step": 17490 }, { "epoch": 184.21052631578948, "grad_norm": 0.08177075535058975, "learning_rate": 9.578933049897643e-05, "loss": 0.0061, "step": 17500 }, { "epoch": 184.31578947368422, "grad_norm": 0.06671208143234253, "learning_rate": 9.578268658350728e-05, "loss": 0.006, "step": 17510 }, { "epoch": 184.42105263157896, "grad_norm": 0.09853783994913101, "learning_rate": 9.577603766131235e-05, "loss": 0.0064, "step": 17520 }, { "epoch": 184.52631578947367, "grad_norm": 0.12233302742242813, "learning_rate": 9.576938373311878e-05, "loss": 0.0059, "step": 17530 }, { "epoch": 184.6315789473684, "grad_norm": 0.08507564663887024, "learning_rate": 9.576272479965421e-05, "loss": 0.0053, "step": 17540 }, { "epoch": 184.73684210526315, "grad_norm": 0.1256791353225708, "learning_rate": 9.575606086164687e-05, "loss": 0.0052, "step": 17550 }, { "epoch": 184.8421052631579, "grad_norm": 0.08507006615400314, "learning_rate": 9.57493919198255e-05, "loss": 0.005, "step": 17560 }, { "epoch": 184.94736842105263, "grad_norm": 0.09072346985340118, "learning_rate": 9.57427179749194e-05, "loss": 0.0063, "step": 17570 }, { "epoch": 185.05263157894737, "grad_norm": 0.07639488577842712, "learning_rate": 9.573603902765846e-05, "loss": 0.0049, "step": 17580 }, { "epoch": 185.1578947368421, "grad_norm": 0.07356898486614227, "learning_rate": 9.572935507877304e-05, "loss": 0.0054, "step": 17590 }, { "epoch": 185.26315789473685, "grad_norm": 0.10636232048273087, "learning_rate": 9.57226661289941e-05, "loss": 0.006, "step": 17600 }, { "epoch": 185.3684210526316, "grad_norm": 0.08115315437316895, "learning_rate": 9.571597217905315e-05, "loss": 0.0057, "step": 17610 }, { "epoch": 185.47368421052633, "grad_norm": 0.09415759146213531, "learning_rate": 9.57092732296822e-05, "loss": 0.0055, "step": 17620 }, { "epoch": 185.57894736842104, "grad_norm": 0.10145479440689087, "learning_rate": 9.570256928161385e-05, "loss": 0.0056, "step": 17630 }, { "epoch": 185.68421052631578, "grad_norm": 0.07230809330940247, "learning_rate": 9.569586033558126e-05, "loss": 0.0052, "step": 17640 }, { "epoch": 185.78947368421052, "grad_norm": 0.08651606738567352, "learning_rate": 9.568914639231807e-05, "loss": 0.0052, "step": 17650 }, { "epoch": 185.89473684210526, "grad_norm": 0.1229088082909584, "learning_rate": 9.568242745255852e-05, "loss": 0.0049, "step": 17660 }, { "epoch": 186.0, "grad_norm": 0.11716331541538239, "learning_rate": 9.567570351703739e-05, "loss": 0.0056, "step": 17670 }, { "epoch": 186.10526315789474, "grad_norm": 0.09402994066476822, "learning_rate": 9.566897458649001e-05, "loss": 0.006, "step": 17680 }, { "epoch": 186.21052631578948, "grad_norm": 0.08382748812437057, "learning_rate": 9.566224066165221e-05, "loss": 0.0059, "step": 17690 }, { "epoch": 186.31578947368422, "grad_norm": 0.09807534515857697, "learning_rate": 9.565550174326043e-05, "loss": 0.0058, "step": 17700 }, { "epoch": 186.42105263157896, "grad_norm": 0.12501707673072815, "learning_rate": 9.564875783205162e-05, "loss": 0.0064, "step": 17710 }, { "epoch": 186.52631578947367, "grad_norm": 0.0937684029340744, "learning_rate": 9.564200892876328e-05, "loss": 0.0053, "step": 17720 }, { "epoch": 186.6315789473684, "grad_norm": 0.09393413364887238, "learning_rate": 9.563525503413348e-05, "loss": 0.005, "step": 17730 }, { "epoch": 186.73684210526315, "grad_norm": 0.1059558168053627, "learning_rate": 9.562849614890079e-05, "loss": 0.0056, "step": 17740 }, { "epoch": 186.8421052631579, "grad_norm": 0.07172396034002304, "learning_rate": 9.562173227380436e-05, "loss": 0.0068, "step": 17750 }, { "epoch": 186.94736842105263, "grad_norm": 0.07133765518665314, "learning_rate": 9.561496340958389e-05, "loss": 0.005, "step": 17760 }, { "epoch": 187.05263157894737, "grad_norm": 0.10553847253322601, "learning_rate": 9.560818955697959e-05, "loss": 0.0054, "step": 17770 }, { "epoch": 187.1578947368421, "grad_norm": 0.08046780526638031, "learning_rate": 9.560141071673228e-05, "loss": 0.0058, "step": 17780 }, { "epoch": 187.26315789473685, "grad_norm": 0.11511830240488052, "learning_rate": 9.559462688958323e-05, "loss": 0.0068, "step": 17790 }, { "epoch": 187.3684210526316, "grad_norm": 0.07903485745191574, "learning_rate": 9.558783807627434e-05, "loss": 0.0055, "step": 17800 }, { "epoch": 187.47368421052633, "grad_norm": 0.08397987484931946, "learning_rate": 9.558104427754801e-05, "loss": 0.0055, "step": 17810 }, { "epoch": 187.57894736842104, "grad_norm": 0.12288576364517212, "learning_rate": 9.557424549414722e-05, "loss": 0.0061, "step": 17820 }, { "epoch": 187.68421052631578, "grad_norm": 0.0763506218791008, "learning_rate": 9.556744172681546e-05, "loss": 0.0055, "step": 17830 }, { "epoch": 187.78947368421052, "grad_norm": 0.09145005792379379, "learning_rate": 9.556063297629677e-05, "loss": 0.0057, "step": 17840 }, { "epoch": 187.89473684210526, "grad_norm": 0.08467885851860046, "learning_rate": 9.555381924333578e-05, "loss": 0.0051, "step": 17850 }, { "epoch": 188.0, "grad_norm": 0.12354733794927597, "learning_rate": 9.554700052867758e-05, "loss": 0.0051, "step": 17860 }, { "epoch": 188.10526315789474, "grad_norm": 0.09389479458332062, "learning_rate": 9.554017683306789e-05, "loss": 0.0055, "step": 17870 }, { "epoch": 188.21052631578948, "grad_norm": 0.1133095994591713, "learning_rate": 9.553334815725294e-05, "loss": 0.0058, "step": 17880 }, { "epoch": 188.31578947368422, "grad_norm": 0.08226996660232544, "learning_rate": 9.552651450197949e-05, "loss": 0.005, "step": 17890 }, { "epoch": 188.42105263157896, "grad_norm": 0.09442288428544998, "learning_rate": 9.551967586799486e-05, "loss": 0.0052, "step": 17900 }, { "epoch": 188.52631578947367, "grad_norm": 0.10282671451568604, "learning_rate": 9.551283225604692e-05, "loss": 0.0061, "step": 17910 }, { "epoch": 188.6315789473684, "grad_norm": 0.10782934725284576, "learning_rate": 9.550598366688406e-05, "loss": 0.0057, "step": 17920 }, { "epoch": 188.73684210526315, "grad_norm": 0.10147484391927719, "learning_rate": 9.549913010125526e-05, "loss": 0.0052, "step": 17930 }, { "epoch": 188.8421052631579, "grad_norm": 0.09786023944616318, "learning_rate": 9.549227155990999e-05, "loss": 0.0055, "step": 17940 }, { "epoch": 188.94736842105263, "grad_norm": 0.12183920294046402, "learning_rate": 9.548540804359828e-05, "loss": 0.0059, "step": 17950 }, { "epoch": 189.05263157894737, "grad_norm": 0.09452280402183533, "learning_rate": 9.547853955307077e-05, "loss": 0.0048, "step": 17960 }, { "epoch": 189.1578947368421, "grad_norm": 0.10775373131036758, "learning_rate": 9.547166608907853e-05, "loss": 0.0052, "step": 17970 }, { "epoch": 189.26315789473685, "grad_norm": 0.11184996366500854, "learning_rate": 9.546478765237326e-05, "loss": 0.0059, "step": 17980 }, { "epoch": 189.3684210526316, "grad_norm": 0.0863642692565918, "learning_rate": 9.545790424370715e-05, "loss": 0.0056, "step": 17990 }, { "epoch": 189.47368421052633, "grad_norm": 0.11795289814472198, "learning_rate": 9.5451015863833e-05, "loss": 0.0055, "step": 18000 }, { "epoch": 189.57894736842104, "grad_norm": 0.1018434390425682, "learning_rate": 9.544412251350408e-05, "loss": 0.0052, "step": 18010 }, { "epoch": 189.68421052631578, "grad_norm": 0.0992245003581047, "learning_rate": 9.543722419347422e-05, "loss": 0.0055, "step": 18020 }, { "epoch": 189.78947368421052, "grad_norm": 0.1062994971871376, "learning_rate": 9.543032090449788e-05, "loss": 0.0049, "step": 18030 }, { "epoch": 189.89473684210526, "grad_norm": 0.07991468906402588, "learning_rate": 9.542341264732992e-05, "loss": 0.0047, "step": 18040 }, { "epoch": 190.0, "grad_norm": 0.10099149495363235, "learning_rate": 9.541649942272585e-05, "loss": 0.0046, "step": 18050 }, { "epoch": 190.10526315789474, "grad_norm": 0.11005444079637527, "learning_rate": 9.54095812314417e-05, "loss": 0.0054, "step": 18060 }, { "epoch": 190.21052631578948, "grad_norm": 0.11735799163579941, "learning_rate": 9.540265807423401e-05, "loss": 0.0058, "step": 18070 }, { "epoch": 190.31578947368422, "grad_norm": 0.11450332403182983, "learning_rate": 9.53957299518599e-05, "loss": 0.0046, "step": 18080 }, { "epoch": 190.42105263157896, "grad_norm": 0.09020315110683441, "learning_rate": 9.5388796865077e-05, "loss": 0.0052, "step": 18090 }, { "epoch": 190.52631578947367, "grad_norm": 0.08833105117082596, "learning_rate": 9.538185881464353e-05, "loss": 0.006, "step": 18100 }, { "epoch": 190.6315789473684, "grad_norm": 0.08688335865736008, "learning_rate": 9.537491580131821e-05, "loss": 0.0048, "step": 18110 }, { "epoch": 190.73684210526315, "grad_norm": 0.09913775324821472, "learning_rate": 9.53679678258603e-05, "loss": 0.0046, "step": 18120 }, { "epoch": 190.8421052631579, "grad_norm": 0.07524233311414719, "learning_rate": 9.536101488902966e-05, "loss": 0.0063, "step": 18130 }, { "epoch": 190.94736842105263, "grad_norm": 0.06980802863836288, "learning_rate": 9.535405699158663e-05, "loss": 0.0057, "step": 18140 }, { "epoch": 191.05263157894737, "grad_norm": 0.09734215587377548, "learning_rate": 9.53470941342921e-05, "loss": 0.0049, "step": 18150 }, { "epoch": 191.1578947368421, "grad_norm": 0.08383326232433319, "learning_rate": 9.534012631790756e-05, "loss": 0.0057, "step": 18160 }, { "epoch": 191.26315789473685, "grad_norm": 0.08664380759000778, "learning_rate": 9.533315354319494e-05, "loss": 0.0053, "step": 18170 }, { "epoch": 191.3684210526316, "grad_norm": 0.087955541908741, "learning_rate": 9.532617581091682e-05, "loss": 0.005, "step": 18180 }, { "epoch": 191.47368421052633, "grad_norm": 0.10498349368572235, "learning_rate": 9.531919312183629e-05, "loss": 0.0051, "step": 18190 }, { "epoch": 191.57894736842104, "grad_norm": 0.1314578801393509, "learning_rate": 9.531220547671688e-05, "loss": 0.0062, "step": 18200 }, { "epoch": 191.68421052631578, "grad_norm": 0.10031463950872421, "learning_rate": 9.530521287632285e-05, "loss": 0.0052, "step": 18210 }, { "epoch": 191.78947368421052, "grad_norm": 0.07232140004634857, "learning_rate": 9.529821532141884e-05, "loss": 0.0058, "step": 18220 }, { "epoch": 191.89473684210526, "grad_norm": 0.09564090520143509, "learning_rate": 9.52912128127701e-05, "loss": 0.0047, "step": 18230 }, { "epoch": 192.0, "grad_norm": 0.08566493541002274, "learning_rate": 9.528420535114244e-05, "loss": 0.0054, "step": 18240 }, { "epoch": 192.10526315789474, "grad_norm": 0.10594399273395538, "learning_rate": 9.527719293730215e-05, "loss": 0.006, "step": 18250 }, { "epoch": 192.21052631578948, "grad_norm": 0.09844136238098145, "learning_rate": 9.527017557201611e-05, "loss": 0.0054, "step": 18260 }, { "epoch": 192.31578947368422, "grad_norm": 0.09018511325120926, "learning_rate": 9.526315325605176e-05, "loss": 0.0053, "step": 18270 }, { "epoch": 192.42105263157896, "grad_norm": 0.11464013904333115, "learning_rate": 9.525612599017699e-05, "loss": 0.0053, "step": 18280 }, { "epoch": 192.52631578947367, "grad_norm": 0.09259311109781265, "learning_rate": 9.524909377516033e-05, "loss": 0.006, "step": 18290 }, { "epoch": 192.6315789473684, "grad_norm": 0.09918615221977234, "learning_rate": 9.524205661177081e-05, "loss": 0.0057, "step": 18300 }, { "epoch": 192.73684210526315, "grad_norm": 0.09562665969133377, "learning_rate": 9.523501450077801e-05, "loss": 0.006, "step": 18310 }, { "epoch": 192.8421052631579, "grad_norm": 0.10229291021823883, "learning_rate": 9.522796744295202e-05, "loss": 0.0054, "step": 18320 }, { "epoch": 192.94736842105263, "grad_norm": 0.1014680340886116, "learning_rate": 9.522091543906352e-05, "loss": 0.0058, "step": 18330 }, { "epoch": 193.05263157894737, "grad_norm": 0.1223004162311554, "learning_rate": 9.521385848988369e-05, "loss": 0.0064, "step": 18340 }, { "epoch": 193.1578947368421, "grad_norm": 0.09115135669708252, "learning_rate": 9.520679659618428e-05, "loss": 0.0052, "step": 18350 }, { "epoch": 193.26315789473685, "grad_norm": 0.10748081654310226, "learning_rate": 9.519972975873754e-05, "loss": 0.0056, "step": 18360 }, { "epoch": 193.3684210526316, "grad_norm": 0.09065068513154984, "learning_rate": 9.519265797831633e-05, "loss": 0.006, "step": 18370 }, { "epoch": 193.47368421052633, "grad_norm": 0.09564093500375748, "learning_rate": 9.518558125569399e-05, "loss": 0.0055, "step": 18380 }, { "epoch": 193.57894736842104, "grad_norm": 0.10992763936519623, "learning_rate": 9.517849959164442e-05, "loss": 0.0054, "step": 18390 }, { "epoch": 193.68421052631578, "grad_norm": 0.09234621375799179, "learning_rate": 9.517141298694205e-05, "loss": 0.0054, "step": 18400 }, { "epoch": 193.78947368421052, "grad_norm": 0.12144038826227188, "learning_rate": 9.516432144236188e-05, "loss": 0.0056, "step": 18410 }, { "epoch": 193.89473684210526, "grad_norm": 0.07223552465438843, "learning_rate": 9.515722495867941e-05, "loss": 0.0068, "step": 18420 }, { "epoch": 194.0, "grad_norm": 0.10048957169055939, "learning_rate": 9.515012353667072e-05, "loss": 0.0053, "step": 18430 }, { "epoch": 194.10526315789474, "grad_norm": 0.08738689124584198, "learning_rate": 9.51430171771124e-05, "loss": 0.0059, "step": 18440 }, { "epoch": 194.21052631578948, "grad_norm": 0.10997642576694489, "learning_rate": 9.513590588078159e-05, "loss": 0.006, "step": 18450 }, { "epoch": 194.31578947368422, "grad_norm": 0.10459021478891373, "learning_rate": 9.512878964845597e-05, "loss": 0.0054, "step": 18460 }, { "epoch": 194.42105263157896, "grad_norm": 0.08523332327604294, "learning_rate": 9.512166848091377e-05, "loss": 0.0052, "step": 18470 }, { "epoch": 194.52631578947367, "grad_norm": 0.10950907319784164, "learning_rate": 9.511454237893376e-05, "loss": 0.0056, "step": 18480 }, { "epoch": 194.6315789473684, "grad_norm": 0.10121098160743713, "learning_rate": 9.51074113432952e-05, "loss": 0.0058, "step": 18490 }, { "epoch": 194.73684210526315, "grad_norm": 0.10156667977571487, "learning_rate": 9.510027537477797e-05, "loss": 0.0053, "step": 18500 }, { "epoch": 194.8421052631579, "grad_norm": 0.09932121634483337, "learning_rate": 9.509313447416242e-05, "loss": 0.0053, "step": 18510 }, { "epoch": 194.94736842105263, "grad_norm": 0.10064784437417984, "learning_rate": 9.508598864222949e-05, "loss": 0.006, "step": 18520 }, { "epoch": 195.05263157894737, "grad_norm": 0.09651435166597366, "learning_rate": 9.507883787976062e-05, "loss": 0.0047, "step": 18530 }, { "epoch": 195.1578947368421, "grad_norm": 0.09404672682285309, "learning_rate": 9.507168218753781e-05, "loss": 0.0062, "step": 18540 }, { "epoch": 195.26315789473685, "grad_norm": 0.09428716450929642, "learning_rate": 9.506452156634362e-05, "loss": 0.0061, "step": 18550 }, { "epoch": 195.3684210526316, "grad_norm": 0.09175308793783188, "learning_rate": 9.505735601696109e-05, "loss": 0.0059, "step": 18560 }, { "epoch": 195.47368421052633, "grad_norm": 0.10800617933273315, "learning_rate": 9.505018554017385e-05, "loss": 0.005, "step": 18570 }, { "epoch": 195.57894736842104, "grad_norm": 0.08208729326725006, "learning_rate": 9.504301013676604e-05, "loss": 0.0052, "step": 18580 }, { "epoch": 195.68421052631578, "grad_norm": 0.09814596176147461, "learning_rate": 9.503582980752238e-05, "loss": 0.0057, "step": 18590 }, { "epoch": 195.78947368421052, "grad_norm": 0.09879244118928909, "learning_rate": 9.502864455322809e-05, "loss": 0.0049, "step": 18600 }, { "epoch": 195.89473684210526, "grad_norm": 0.08306244015693665, "learning_rate": 9.502145437466891e-05, "loss": 0.0044, "step": 18610 }, { "epoch": 196.0, "grad_norm": 0.09770520776510239, "learning_rate": 9.501425927263116e-05, "loss": 0.0055, "step": 18620 }, { "epoch": 196.10526315789474, "grad_norm": 0.09300059825181961, "learning_rate": 9.500705924790172e-05, "loss": 0.0046, "step": 18630 }, { "epoch": 196.21052631578948, "grad_norm": 0.0988309383392334, "learning_rate": 9.499985430126794e-05, "loss": 0.006, "step": 18640 }, { "epoch": 196.31578947368422, "grad_norm": 0.09161316603422165, "learning_rate": 9.499264443351775e-05, "loss": 0.0053, "step": 18650 }, { "epoch": 196.42105263157896, "grad_norm": 0.0695241168141365, "learning_rate": 9.498542964543961e-05, "loss": 0.0061, "step": 18660 }, { "epoch": 196.52631578947367, "grad_norm": 0.08304384350776672, "learning_rate": 9.497820993782252e-05, "loss": 0.0053, "step": 18670 }, { "epoch": 196.6315789473684, "grad_norm": 0.08585744351148605, "learning_rate": 9.497098531145601e-05, "loss": 0.0054, "step": 18680 }, { "epoch": 196.73684210526315, "grad_norm": 0.08903979510068893, "learning_rate": 9.496375576713017e-05, "loss": 0.0051, "step": 18690 }, { "epoch": 196.8421052631579, "grad_norm": 0.10806342214345932, "learning_rate": 9.49565213056356e-05, "loss": 0.0058, "step": 18700 }, { "epoch": 196.94736842105263, "grad_norm": 0.08129089325666428, "learning_rate": 9.494928192776342e-05, "loss": 0.005, "step": 18710 }, { "epoch": 197.05263157894737, "grad_norm": 0.10140802711248398, "learning_rate": 9.494203763430538e-05, "loss": 0.0052, "step": 18720 }, { "epoch": 197.1578947368421, "grad_norm": 0.10660256445407867, "learning_rate": 9.493478842605366e-05, "loss": 0.0055, "step": 18730 }, { "epoch": 197.26315789473685, "grad_norm": 0.09290044754743576, "learning_rate": 9.492753430380105e-05, "loss": 0.0057, "step": 18740 }, { "epoch": 197.3684210526316, "grad_norm": 0.09320078045129776, "learning_rate": 9.492027526834083e-05, "loss": 0.0053, "step": 18750 }, { "epoch": 197.47368421052633, "grad_norm": 0.10205413401126862, "learning_rate": 9.491301132046684e-05, "loss": 0.0052, "step": 18760 }, { "epoch": 197.57894736842104, "grad_norm": 0.10857052356004715, "learning_rate": 9.490574246097345e-05, "loss": 0.0054, "step": 18770 }, { "epoch": 197.68421052631578, "grad_norm": 0.1136835366487503, "learning_rate": 9.48984686906556e-05, "loss": 0.0054, "step": 18780 }, { "epoch": 197.78947368421052, "grad_norm": 0.11511198431253433, "learning_rate": 9.489119001030871e-05, "loss": 0.0057, "step": 18790 }, { "epoch": 197.89473684210526, "grad_norm": 0.08594632893800735, "learning_rate": 9.488390642072878e-05, "loss": 0.0056, "step": 18800 }, { "epoch": 198.0, "grad_norm": 0.1125936433672905, "learning_rate": 9.48766179227123e-05, "loss": 0.0058, "step": 18810 }, { "epoch": 198.10526315789474, "grad_norm": 0.12131507694721222, "learning_rate": 9.486932451705636e-05, "loss": 0.0056, "step": 18820 }, { "epoch": 198.21052631578948, "grad_norm": 0.10835666209459305, "learning_rate": 9.486202620455857e-05, "loss": 0.0061, "step": 18830 }, { "epoch": 198.31578947368422, "grad_norm": 0.11050567030906677, "learning_rate": 9.485472298601704e-05, "loss": 0.0052, "step": 18840 }, { "epoch": 198.42105263157896, "grad_norm": 0.11421769112348557, "learning_rate": 9.484741486223043e-05, "loss": 0.0051, "step": 18850 }, { "epoch": 198.52631578947367, "grad_norm": 0.09856433421373367, "learning_rate": 9.484010183399797e-05, "loss": 0.0048, "step": 18860 }, { "epoch": 198.6315789473684, "grad_norm": 0.10910903662443161, "learning_rate": 9.483278390211938e-05, "loss": 0.0055, "step": 18870 }, { "epoch": 198.73684210526315, "grad_norm": 0.06728316843509674, "learning_rate": 9.482546106739496e-05, "loss": 0.0052, "step": 18880 }, { "epoch": 198.8421052631579, "grad_norm": 0.07540174573659897, "learning_rate": 9.48181333306255e-05, "loss": 0.0048, "step": 18890 }, { "epoch": 198.94736842105263, "grad_norm": 0.07704653590917587, "learning_rate": 9.481080069261237e-05, "loss": 0.0053, "step": 18900 }, { "epoch": 199.05263157894737, "grad_norm": 0.09004538506269455, "learning_rate": 9.480346315415745e-05, "loss": 0.0047, "step": 18910 }, { "epoch": 199.1578947368421, "grad_norm": 0.11200690269470215, "learning_rate": 9.479612071606314e-05, "loss": 0.0063, "step": 18920 }, { "epoch": 199.26315789473685, "grad_norm": 0.08753960579633713, "learning_rate": 9.478877337913244e-05, "loss": 0.0053, "step": 18930 }, { "epoch": 199.3684210526316, "grad_norm": 0.09322004020214081, "learning_rate": 9.478142114416881e-05, "loss": 0.005, "step": 18940 }, { "epoch": 199.47368421052633, "grad_norm": 0.09975776076316833, "learning_rate": 9.47740640119763e-05, "loss": 0.0049, "step": 18950 }, { "epoch": 199.57894736842104, "grad_norm": 0.07746563106775284, "learning_rate": 9.476670198335947e-05, "loss": 0.0053, "step": 18960 }, { "epoch": 199.68421052631578, "grad_norm": 0.08851278573274612, "learning_rate": 9.47593350591234e-05, "loss": 0.0051, "step": 18970 }, { "epoch": 199.78947368421052, "grad_norm": 0.10217593610286713, "learning_rate": 9.475196324007376e-05, "loss": 0.0045, "step": 18980 }, { "epoch": 199.89473684210526, "grad_norm": 0.07446050643920898, "learning_rate": 9.474458652701669e-05, "loss": 0.0047, "step": 18990 }, { "epoch": 200.0, "grad_norm": 0.11288639158010483, "learning_rate": 9.473720492075892e-05, "loss": 0.005, "step": 19000 }, { "epoch": 200.10526315789474, "grad_norm": 0.0826745331287384, "learning_rate": 9.472981842210768e-05, "loss": 0.0048, "step": 19010 }, { "epoch": 200.21052631578948, "grad_norm": 0.07897073030471802, "learning_rate": 9.472242703187074e-05, "loss": 0.006, "step": 19020 }, { "epoch": 200.31578947368422, "grad_norm": 0.08525894582271576, "learning_rate": 9.471503075085643e-05, "loss": 0.0052, "step": 19030 }, { "epoch": 200.42105263157896, "grad_norm": 0.06440197676420212, "learning_rate": 9.470762957987359e-05, "loss": 0.0045, "step": 19040 }, { "epoch": 200.52631578947367, "grad_norm": 0.09628661721944809, "learning_rate": 9.470022351973158e-05, "loss": 0.0055, "step": 19050 }, { "epoch": 200.6315789473684, "grad_norm": 0.08259394764900208, "learning_rate": 9.469281257124034e-05, "loss": 0.0059, "step": 19060 }, { "epoch": 200.73684210526315, "grad_norm": 0.09494814276695251, "learning_rate": 9.46853967352103e-05, "loss": 0.0058, "step": 19070 }, { "epoch": 200.8421052631579, "grad_norm": 0.09577471017837524, "learning_rate": 9.467797601245246e-05, "loss": 0.0047, "step": 19080 }, { "epoch": 200.94736842105263, "grad_norm": 0.08920130133628845, "learning_rate": 9.467055040377834e-05, "loss": 0.0055, "step": 19090 }, { "epoch": 201.05263157894737, "grad_norm": 0.08132235705852509, "learning_rate": 9.466311990999999e-05, "loss": 0.0056, "step": 19100 }, { "epoch": 201.1578947368421, "grad_norm": 0.08733727037906647, "learning_rate": 9.465568453193e-05, "loss": 0.0054, "step": 19110 }, { "epoch": 201.26315789473685, "grad_norm": 0.11187614500522614, "learning_rate": 9.464824427038148e-05, "loss": 0.0049, "step": 19120 }, { "epoch": 201.3684210526316, "grad_norm": 0.07102920860052109, "learning_rate": 9.46407991261681e-05, "loss": 0.0053, "step": 19130 }, { "epoch": 201.47368421052633, "grad_norm": 0.0912785530090332, "learning_rate": 9.463334910010404e-05, "loss": 0.006, "step": 19140 }, { "epoch": 201.57894736842104, "grad_norm": 0.08878538757562637, "learning_rate": 9.462589419300403e-05, "loss": 0.006, "step": 19150 }, { "epoch": 201.68421052631578, "grad_norm": 0.09036549925804138, "learning_rate": 9.461843440568333e-05, "loss": 0.0058, "step": 19160 }, { "epoch": 201.78947368421052, "grad_norm": 0.10864312946796417, "learning_rate": 9.461096973895773e-05, "loss": 0.0061, "step": 19170 }, { "epoch": 201.89473684210526, "grad_norm": 0.10819237679243088, "learning_rate": 9.460350019364355e-05, "loss": 0.0064, "step": 19180 }, { "epoch": 202.0, "grad_norm": 0.14416790008544922, "learning_rate": 9.459602577055764e-05, "loss": 0.0055, "step": 19190 }, { "epoch": 202.10526315789474, "grad_norm": 0.12200731039047241, "learning_rate": 9.45885464705174e-05, "loss": 0.006, "step": 19200 }, { "epoch": 202.21052631578948, "grad_norm": 0.09328319877386093, "learning_rate": 9.458106229434076e-05, "loss": 0.0059, "step": 19210 }, { "epoch": 202.31578947368422, "grad_norm": 0.10859799385070801, "learning_rate": 9.457357324284617e-05, "loss": 0.0055, "step": 19220 }, { "epoch": 202.42105263157896, "grad_norm": 0.08148252218961716, "learning_rate": 9.456607931685262e-05, "loss": 0.0048, "step": 19230 }, { "epoch": 202.52631578947367, "grad_norm": 0.11715289950370789, "learning_rate": 9.455858051717965e-05, "loss": 0.0049, "step": 19240 }, { "epoch": 202.6315789473684, "grad_norm": 0.09834036231040955, "learning_rate": 9.45510768446473e-05, "loss": 0.0054, "step": 19250 }, { "epoch": 202.73684210526315, "grad_norm": 0.08165912330150604, "learning_rate": 9.454356830007618e-05, "loss": 0.005, "step": 19260 }, { "epoch": 202.8421052631579, "grad_norm": 0.12786118686199188, "learning_rate": 9.45360548842874e-05, "loss": 0.006, "step": 19270 }, { "epoch": 202.94736842105263, "grad_norm": 0.07939919084310532, "learning_rate": 9.452853659810261e-05, "loss": 0.0064, "step": 19280 }, { "epoch": 203.05263157894737, "grad_norm": 0.07339566200971603, "learning_rate": 9.452101344234401e-05, "loss": 0.0049, "step": 19290 }, { "epoch": 203.1578947368421, "grad_norm": 0.07751643657684326, "learning_rate": 9.451348541783431e-05, "loss": 0.0055, "step": 19300 }, { "epoch": 203.26315789473685, "grad_norm": 0.09149680286645889, "learning_rate": 9.450595252539678e-05, "loss": 0.0051, "step": 19310 }, { "epoch": 203.3684210526316, "grad_norm": 0.09669169783592224, "learning_rate": 9.449841476585518e-05, "loss": 0.005, "step": 19320 }, { "epoch": 203.47368421052633, "grad_norm": 0.129618838429451, "learning_rate": 9.449087214003384e-05, "loss": 0.0059, "step": 19330 }, { "epoch": 203.57894736842104, "grad_norm": 0.08649791777133942, "learning_rate": 9.448332464875765e-05, "loss": 0.0047, "step": 19340 }, { "epoch": 203.68421052631578, "grad_norm": 0.07093901187181473, "learning_rate": 9.447577229285192e-05, "loss": 0.0053, "step": 19350 }, { "epoch": 203.78947368421052, "grad_norm": 0.0960673838853836, "learning_rate": 9.446821507314261e-05, "loss": 0.0046, "step": 19360 }, { "epoch": 203.89473684210526, "grad_norm": 0.11478099972009659, "learning_rate": 9.446065299045617e-05, "loss": 0.0048, "step": 19370 }, { "epoch": 204.0, "grad_norm": 0.13477498292922974, "learning_rate": 9.445308604561955e-05, "loss": 0.0057, "step": 19380 }, { "epoch": 204.10526315789474, "grad_norm": 0.12364468723535538, "learning_rate": 9.444551423946028e-05, "loss": 0.0051, "step": 19390 }, { "epoch": 204.21052631578948, "grad_norm": 0.11762908846139908, "learning_rate": 9.443793757280638e-05, "loss": 0.0065, "step": 19400 }, { "epoch": 204.31578947368422, "grad_norm": 0.09544281661510468, "learning_rate": 9.443035604648646e-05, "loss": 0.0054, "step": 19410 }, { "epoch": 204.42105263157896, "grad_norm": 0.12891481816768646, "learning_rate": 9.44227696613296e-05, "loss": 0.0065, "step": 19420 }, { "epoch": 204.52631578947367, "grad_norm": 0.10694330930709839, "learning_rate": 9.441517841816542e-05, "loss": 0.0054, "step": 19430 }, { "epoch": 204.6315789473684, "grad_norm": 0.1092243641614914, "learning_rate": 9.440758231782413e-05, "loss": 0.0053, "step": 19440 }, { "epoch": 204.73684210526315, "grad_norm": 0.11552444100379944, "learning_rate": 9.439998136113639e-05, "loss": 0.0057, "step": 19450 }, { "epoch": 204.8421052631579, "grad_norm": 0.07901804149150848, "learning_rate": 9.439237554893344e-05, "loss": 0.0059, "step": 19460 }, { "epoch": 204.94736842105263, "grad_norm": 0.0873388797044754, "learning_rate": 9.438476488204705e-05, "loss": 0.0054, "step": 19470 }, { "epoch": 205.05263157894737, "grad_norm": 0.07380732893943787, "learning_rate": 9.43771493613095e-05, "loss": 0.005, "step": 19480 }, { "epoch": 205.1578947368421, "grad_norm": 0.07498955726623535, "learning_rate": 9.436952898755362e-05, "loss": 0.005, "step": 19490 }, { "epoch": 205.26315789473685, "grad_norm": 0.0927809625864029, "learning_rate": 9.436190376161276e-05, "loss": 0.0053, "step": 19500 }, { "epoch": 205.3684210526316, "grad_norm": 0.07343907654285431, "learning_rate": 9.43542736843208e-05, "loss": 0.0047, "step": 19510 }, { "epoch": 205.47368421052633, "grad_norm": 0.1056886538863182, "learning_rate": 9.434663875651216e-05, "loss": 0.0045, "step": 19520 }, { "epoch": 205.57894736842104, "grad_norm": 0.06505971401929855, "learning_rate": 9.433899897902177e-05, "loss": 0.0052, "step": 19530 }, { "epoch": 205.68421052631578, "grad_norm": 0.09466693550348282, "learning_rate": 9.433135435268511e-05, "loss": 0.0061, "step": 19540 }, { "epoch": 205.78947368421052, "grad_norm": 0.08101682364940643, "learning_rate": 9.432370487833819e-05, "loss": 0.0048, "step": 19550 }, { "epoch": 205.89473684210526, "grad_norm": 0.08043239265680313, "learning_rate": 9.431605055681756e-05, "loss": 0.0048, "step": 19560 }, { "epoch": 206.0, "grad_norm": 0.07636161148548126, "learning_rate": 9.430839138896026e-05, "loss": 0.0055, "step": 19570 }, { "epoch": 206.10526315789474, "grad_norm": 0.07368014752864838, "learning_rate": 9.43007273756039e-05, "loss": 0.0055, "step": 19580 }, { "epoch": 206.21052631578948, "grad_norm": 0.11481732130050659, "learning_rate": 9.429305851758658e-05, "loss": 0.0048, "step": 19590 }, { "epoch": 206.31578947368422, "grad_norm": 0.08316943794488907, "learning_rate": 9.428538481574699e-05, "loss": 0.0054, "step": 19600 }, { "epoch": 206.42105263157896, "grad_norm": 0.10448942333459854, "learning_rate": 9.42777062709243e-05, "loss": 0.005, "step": 19610 }, { "epoch": 206.52631578947367, "grad_norm": 0.08123279362916946, "learning_rate": 9.427002288395821e-05, "loss": 0.0047, "step": 19620 }, { "epoch": 206.6315789473684, "grad_norm": 0.08698562532663345, "learning_rate": 9.426233465568898e-05, "loss": 0.0046, "step": 19630 }, { "epoch": 206.73684210526315, "grad_norm": 0.09983783215284348, "learning_rate": 9.42546415869574e-05, "loss": 0.0056, "step": 19640 }, { "epoch": 206.8421052631579, "grad_norm": 0.08898084610700607, "learning_rate": 9.424694367860473e-05, "loss": 0.0062, "step": 19650 }, { "epoch": 206.94736842105263, "grad_norm": 0.0835774838924408, "learning_rate": 9.423924093147284e-05, "loss": 0.0051, "step": 19660 }, { "epoch": 207.05263157894737, "grad_norm": 0.07460326701402664, "learning_rate": 9.423153334640407e-05, "loss": 0.0049, "step": 19670 }, { "epoch": 207.1578947368421, "grad_norm": 0.11057882755994797, "learning_rate": 9.42238209242413e-05, "loss": 0.0053, "step": 19680 }, { "epoch": 207.26315789473685, "grad_norm": 0.09957701712846756, "learning_rate": 9.421610366582798e-05, "loss": 0.0063, "step": 19690 }, { "epoch": 207.3684210526316, "grad_norm": 0.11137846112251282, "learning_rate": 9.420838157200803e-05, "loss": 0.0056, "step": 19700 }, { "epoch": 207.47368421052633, "grad_norm": 0.06810256838798523, "learning_rate": 9.420065464362594e-05, "loss": 0.0059, "step": 19710 }, { "epoch": 207.57894736842104, "grad_norm": 0.110895536839962, "learning_rate": 9.419292288152673e-05, "loss": 0.0053, "step": 19720 }, { "epoch": 207.68421052631578, "grad_norm": 0.1265445202589035, "learning_rate": 9.418518628655588e-05, "loss": 0.005, "step": 19730 }, { "epoch": 207.78947368421052, "grad_norm": 0.11128674447536469, "learning_rate": 9.417744485955951e-05, "loss": 0.0051, "step": 19740 }, { "epoch": 207.89473684210526, "grad_norm": 0.10175863653421402, "learning_rate": 9.41696986013842e-05, "loss": 0.0048, "step": 19750 }, { "epoch": 208.0, "grad_norm": 0.1113918125629425, "learning_rate": 9.416194751287705e-05, "loss": 0.0047, "step": 19760 }, { "epoch": 208.10526315789474, "grad_norm": 0.10981211811304092, "learning_rate": 9.415419159488572e-05, "loss": 0.0048, "step": 19770 }, { "epoch": 208.21052631578948, "grad_norm": 0.10332616418600082, "learning_rate": 9.414643084825837e-05, "loss": 0.005, "step": 19780 }, { "epoch": 208.31578947368422, "grad_norm": 0.07308477908372879, "learning_rate": 9.413866527384372e-05, "loss": 0.0057, "step": 19790 }, { "epoch": 208.42105263157896, "grad_norm": 0.0942094698548317, "learning_rate": 9.4130894872491e-05, "loss": 0.0062, "step": 19800 }, { "epoch": 208.52631578947367, "grad_norm": 0.085349440574646, "learning_rate": 9.412311964504998e-05, "loss": 0.0051, "step": 19810 }, { "epoch": 208.6315789473684, "grad_norm": 0.09796016663312912, "learning_rate": 9.411533959237091e-05, "loss": 0.0056, "step": 19820 }, { "epoch": 208.73684210526315, "grad_norm": 0.1025775671005249, "learning_rate": 9.410755471530464e-05, "loss": 0.0049, "step": 19830 }, { "epoch": 208.8421052631579, "grad_norm": 0.09978478401899338, "learning_rate": 9.40997650147025e-05, "loss": 0.0055, "step": 19840 }, { "epoch": 208.94736842105263, "grad_norm": 0.08836795389652252, "learning_rate": 9.409197049141637e-05, "loss": 0.005, "step": 19850 }, { "epoch": 209.05263157894737, "grad_norm": 0.08556957542896271, "learning_rate": 9.408417114629863e-05, "loss": 0.0054, "step": 19860 }, { "epoch": 209.1578947368421, "grad_norm": 0.07666759192943573, "learning_rate": 9.40763669802022e-05, "loss": 0.0059, "step": 19870 }, { "epoch": 209.26315789473685, "grad_norm": 0.09552009403705597, "learning_rate": 9.406855799398056e-05, "loss": 0.0048, "step": 19880 }, { "epoch": 209.3684210526316, "grad_norm": 0.06767871230840683, "learning_rate": 9.406074418848767e-05, "loss": 0.0047, "step": 19890 }, { "epoch": 209.47368421052633, "grad_norm": 0.10250048339366913, "learning_rate": 9.405292556457805e-05, "loss": 0.0065, "step": 19900 }, { "epoch": 209.57894736842104, "grad_norm": 0.10386500507593155, "learning_rate": 9.404510212310671e-05, "loss": 0.0045, "step": 19910 }, { "epoch": 209.68421052631578, "grad_norm": 0.07723086327314377, "learning_rate": 9.403727386492924e-05, "loss": 0.0049, "step": 19920 }, { "epoch": 209.78947368421052, "grad_norm": 0.07325049489736557, "learning_rate": 9.40294407909017e-05, "loss": 0.0045, "step": 19930 }, { "epoch": 209.89473684210526, "grad_norm": 0.09100504219532013, "learning_rate": 9.40216029018807e-05, "loss": 0.0046, "step": 19940 }, { "epoch": 210.0, "grad_norm": 0.12742513418197632, "learning_rate": 9.401376019872338e-05, "loss": 0.0056, "step": 19950 }, { "epoch": 210.10526315789474, "grad_norm": 0.10910771042108536, "learning_rate": 9.400591268228746e-05, "loss": 0.0043, "step": 19960 }, { "epoch": 210.21052631578948, "grad_norm": 0.11435840278863907, "learning_rate": 9.399806035343106e-05, "loss": 0.0056, "step": 19970 }, { "epoch": 210.31578947368422, "grad_norm": 0.10450158268213272, "learning_rate": 9.399020321301294e-05, "loss": 0.0053, "step": 19980 }, { "epoch": 210.42105263157896, "grad_norm": 0.08184730261564255, "learning_rate": 9.398234126189234e-05, "loss": 0.0055, "step": 19990 }, { "epoch": 210.52631578947367, "grad_norm": 0.10813973098993301, "learning_rate": 9.397447450092902e-05, "loss": 0.0056, "step": 20000 }, { "epoch": 210.6315789473684, "grad_norm": 0.10270272195339203, "learning_rate": 9.39666029309833e-05, "loss": 0.0051, "step": 20010 }, { "epoch": 210.73684210526315, "grad_norm": 0.11873231828212738, "learning_rate": 9.395872655291596e-05, "loss": 0.005, "step": 20020 }, { "epoch": 210.8421052631579, "grad_norm": 0.09353121370077133, "learning_rate": 9.395084536758838e-05, "loss": 0.0058, "step": 20030 }, { "epoch": 210.94736842105263, "grad_norm": 0.12627990543842316, "learning_rate": 9.394295937586243e-05, "loss": 0.0061, "step": 20040 }, { "epoch": 211.05263157894737, "grad_norm": 0.1304464191198349, "learning_rate": 9.393506857860052e-05, "loss": 0.0061, "step": 20050 }, { "epoch": 211.1578947368421, "grad_norm": 0.11639498919248581, "learning_rate": 9.392717297666555e-05, "loss": 0.0065, "step": 20060 }, { "epoch": 211.26315789473685, "grad_norm": 0.12942714989185333, "learning_rate": 9.391927257092101e-05, "loss": 0.0057, "step": 20070 }, { "epoch": 211.3684210526316, "grad_norm": 0.10831288993358612, "learning_rate": 9.391136736223085e-05, "loss": 0.0054, "step": 20080 }, { "epoch": 211.47368421052633, "grad_norm": 0.08768301457166672, "learning_rate": 9.390345735145956e-05, "loss": 0.0058, "step": 20090 }, { "epoch": 211.57894736842104, "grad_norm": 0.11305667459964752, "learning_rate": 9.389554253947219e-05, "loss": 0.0051, "step": 20100 }, { "epoch": 211.68421052631578, "grad_norm": 0.0882592499256134, "learning_rate": 9.388762292713428e-05, "loss": 0.005, "step": 20110 }, { "epoch": 211.78947368421052, "grad_norm": 0.09246308356523514, "learning_rate": 9.38796985153119e-05, "loss": 0.0064, "step": 20120 }, { "epoch": 211.89473684210526, "grad_norm": 0.09377451241016388, "learning_rate": 9.387176930487169e-05, "loss": 0.0056, "step": 20130 }, { "epoch": 212.0, "grad_norm": 0.145900160074234, "learning_rate": 9.386383529668072e-05, "loss": 0.0054, "step": 20140 }, { "epoch": 212.10526315789474, "grad_norm": 0.09627629071474075, "learning_rate": 9.385589649160669e-05, "loss": 0.006, "step": 20150 }, { "epoch": 212.21052631578948, "grad_norm": 0.11023987829685211, "learning_rate": 9.384795289051775e-05, "loss": 0.0058, "step": 20160 }, { "epoch": 212.31578947368422, "grad_norm": 0.10249980539083481, "learning_rate": 9.384000449428261e-05, "loss": 0.0051, "step": 20170 }, { "epoch": 212.42105263157896, "grad_norm": 0.08233065158128738, "learning_rate": 9.383205130377048e-05, "loss": 0.005, "step": 20180 }, { "epoch": 212.52631578947367, "grad_norm": 0.0968620628118515, "learning_rate": 9.382409331985114e-05, "loss": 0.0055, "step": 20190 }, { "epoch": 212.6315789473684, "grad_norm": 0.07458353787660599, "learning_rate": 9.381613054339482e-05, "loss": 0.0047, "step": 20200 }, { "epoch": 212.73684210526315, "grad_norm": 0.07995102554559708, "learning_rate": 9.380816297527235e-05, "loss": 0.0047, "step": 20210 }, { "epoch": 212.8421052631579, "grad_norm": 0.08121421933174133, "learning_rate": 9.380019061635506e-05, "loss": 0.005, "step": 20220 }, { "epoch": 212.94736842105263, "grad_norm": 0.1074252799153328, "learning_rate": 9.379221346751474e-05, "loss": 0.005, "step": 20230 }, { "epoch": 213.05263157894737, "grad_norm": 0.07512903213500977, "learning_rate": 9.378423152962382e-05, "loss": 0.0049, "step": 20240 }, { "epoch": 213.1578947368421, "grad_norm": 0.08385896682739258, "learning_rate": 9.377624480355517e-05, "loss": 0.0049, "step": 20250 }, { "epoch": 213.26315789473685, "grad_norm": 0.09065373241901398, "learning_rate": 9.376825329018219e-05, "loss": 0.006, "step": 20260 }, { "epoch": 213.3684210526316, "grad_norm": 0.10730097442865372, "learning_rate": 9.376025699037884e-05, "loss": 0.0053, "step": 20270 }, { "epoch": 213.47368421052633, "grad_norm": 0.09522482007741928, "learning_rate": 9.37522559050196e-05, "loss": 0.0047, "step": 20280 }, { "epoch": 213.57894736842104, "grad_norm": 0.11447836458683014, "learning_rate": 9.37442500349794e-05, "loss": 0.0041, "step": 20290 }, { "epoch": 213.68421052631578, "grad_norm": 0.10918283462524414, "learning_rate": 9.373623938113381e-05, "loss": 0.005, "step": 20300 }, { "epoch": 213.78947368421052, "grad_norm": 0.11353535205125809, "learning_rate": 9.372822394435883e-05, "loss": 0.0051, "step": 20310 }, { "epoch": 213.89473684210526, "grad_norm": 0.09594425559043884, "learning_rate": 9.372020372553102e-05, "loss": 0.005, "step": 20320 }, { "epoch": 214.0, "grad_norm": 0.10285834968090057, "learning_rate": 9.371217872552746e-05, "loss": 0.0051, "step": 20330 }, { "epoch": 214.10526315789474, "grad_norm": 0.09370587021112442, "learning_rate": 9.370414894522576e-05, "loss": 0.0051, "step": 20340 }, { "epoch": 214.21052631578948, "grad_norm": 0.10037998110055923, "learning_rate": 9.369611438550406e-05, "loss": 0.0049, "step": 20350 }, { "epoch": 214.31578947368422, "grad_norm": 0.07900876551866531, "learning_rate": 9.368807504724095e-05, "loss": 0.0054, "step": 20360 }, { "epoch": 214.42105263157896, "grad_norm": 0.08325327932834625, "learning_rate": 9.368003093131565e-05, "loss": 0.0057, "step": 20370 }, { "epoch": 214.52631578947367, "grad_norm": 0.08997838199138641, "learning_rate": 9.367198203860785e-05, "loss": 0.0055, "step": 20380 }, { "epoch": 214.6315789473684, "grad_norm": 0.09500209242105484, "learning_rate": 9.366392836999774e-05, "loss": 0.0054, "step": 20390 }, { "epoch": 214.73684210526315, "grad_norm": 0.09257620573043823, "learning_rate": 9.365586992636607e-05, "loss": 0.0048, "step": 20400 }, { "epoch": 214.8421052631579, "grad_norm": 0.08723931759595871, "learning_rate": 9.364780670859412e-05, "loss": 0.0044, "step": 20410 }, { "epoch": 214.94736842105263, "grad_norm": 0.10628555715084076, "learning_rate": 9.363973871756364e-05, "loss": 0.0053, "step": 20420 }, { "epoch": 215.05263157894737, "grad_norm": 0.10462025552988052, "learning_rate": 9.363166595415696e-05, "loss": 0.0054, "step": 20430 }, { "epoch": 215.1578947368421, "grad_norm": 0.10597800463438034, "learning_rate": 9.362358841925686e-05, "loss": 0.006, "step": 20440 }, { "epoch": 215.26315789473685, "grad_norm": 0.0940181165933609, "learning_rate": 9.361550611374674e-05, "loss": 0.0053, "step": 20450 }, { "epoch": 215.3684210526316, "grad_norm": 0.11530766636133194, "learning_rate": 9.360741903851043e-05, "loss": 0.0053, "step": 20460 }, { "epoch": 215.47368421052633, "grad_norm": 0.08594803512096405, "learning_rate": 9.359932719443236e-05, "loss": 0.0061, "step": 20470 }, { "epoch": 215.57894736842104, "grad_norm": 0.11585386842489243, "learning_rate": 9.35912305823974e-05, "loss": 0.0049, "step": 20480 }, { "epoch": 215.68421052631578, "grad_norm": 0.08266984671354294, "learning_rate": 9.358312920329101e-05, "loss": 0.0044, "step": 20490 }, { "epoch": 215.78947368421052, "grad_norm": 0.08733273297548294, "learning_rate": 9.357502305799914e-05, "loss": 0.0043, "step": 20500 }, { "epoch": 215.89473684210526, "grad_norm": 0.09496360272169113, "learning_rate": 9.356691214740824e-05, "loss": 0.0053, "step": 20510 }, { "epoch": 216.0, "grad_norm": 0.11547680199146271, "learning_rate": 9.355879647240535e-05, "loss": 0.0048, "step": 20520 }, { "epoch": 216.10526315789474, "grad_norm": 0.0879686251282692, "learning_rate": 9.355067603387798e-05, "loss": 0.0054, "step": 20530 }, { "epoch": 216.21052631578948, "grad_norm": 0.0913296490907669, "learning_rate": 9.354255083271412e-05, "loss": 0.0048, "step": 20540 }, { "epoch": 216.31578947368422, "grad_norm": 0.0906638354063034, "learning_rate": 9.353442086980239e-05, "loss": 0.005, "step": 20550 }, { "epoch": 216.42105263157896, "grad_norm": 0.12339568138122559, "learning_rate": 9.352628614603185e-05, "loss": 0.0052, "step": 20560 }, { "epoch": 216.52631578947367, "grad_norm": 0.1298765391111374, "learning_rate": 9.351814666229209e-05, "loss": 0.0047, "step": 20570 }, { "epoch": 216.6315789473684, "grad_norm": 0.099777951836586, "learning_rate": 9.351000241947324e-05, "loss": 0.0051, "step": 20580 }, { "epoch": 216.73684210526315, "grad_norm": 0.10363582521677017, "learning_rate": 9.350185341846594e-05, "loss": 0.0058, "step": 20590 }, { "epoch": 216.8421052631579, "grad_norm": 0.10156618058681488, "learning_rate": 9.349369966016134e-05, "loss": 0.0054, "step": 20600 }, { "epoch": 216.94736842105263, "grad_norm": 0.0836644321680069, "learning_rate": 9.348554114545117e-05, "loss": 0.0055, "step": 20610 }, { "epoch": 217.05263157894737, "grad_norm": 0.11154348403215408, "learning_rate": 9.347737787522758e-05, "loss": 0.0053, "step": 20620 }, { "epoch": 217.1578947368421, "grad_norm": 0.12974970042705536, "learning_rate": 9.346920985038332e-05, "loss": 0.0046, "step": 20630 }, { "epoch": 217.26315789473685, "grad_norm": 0.08840488642454147, "learning_rate": 9.346103707181162e-05, "loss": 0.0053, "step": 20640 }, { "epoch": 217.3684210526316, "grad_norm": 0.1021587923169136, "learning_rate": 9.345285954040626e-05, "loss": 0.0051, "step": 20650 }, { "epoch": 217.47368421052633, "grad_norm": 0.0767628625035286, "learning_rate": 9.34446772570615e-05, "loss": 0.0063, "step": 20660 }, { "epoch": 217.57894736842104, "grad_norm": 0.08087176084518433, "learning_rate": 9.343649022267214e-05, "loss": 0.006, "step": 20670 }, { "epoch": 217.68421052631578, "grad_norm": 0.0890268012881279, "learning_rate": 9.342829843813353e-05, "loss": 0.0049, "step": 20680 }, { "epoch": 217.78947368421052, "grad_norm": 0.11250331997871399, "learning_rate": 9.342010190434149e-05, "loss": 0.0059, "step": 20690 }, { "epoch": 217.89473684210526, "grad_norm": 0.0832214206457138, "learning_rate": 9.34119006221924e-05, "loss": 0.0047, "step": 20700 }, { "epoch": 218.0, "grad_norm": 0.11983665078878403, "learning_rate": 9.340369459258313e-05, "loss": 0.0048, "step": 20710 }, { "epoch": 218.10526315789474, "grad_norm": 0.09841437637805939, "learning_rate": 9.339548381641106e-05, "loss": 0.0053, "step": 20720 }, { "epoch": 218.21052631578948, "grad_norm": 0.09643573313951492, "learning_rate": 9.338726829457413e-05, "loss": 0.0056, "step": 20730 }, { "epoch": 218.31578947368422, "grad_norm": 0.10438665747642517, "learning_rate": 9.337904802797078e-05, "loss": 0.0055, "step": 20740 }, { "epoch": 218.42105263157896, "grad_norm": 0.10271177440881729, "learning_rate": 9.337082301749993e-05, "loss": 0.0057, "step": 20750 }, { "epoch": 218.52631578947367, "grad_norm": 0.09939359873533249, "learning_rate": 9.336259326406109e-05, "loss": 0.0057, "step": 20760 }, { "epoch": 218.6315789473684, "grad_norm": 0.1307879537343979, "learning_rate": 9.335435876855427e-05, "loss": 0.0047, "step": 20770 }, { "epoch": 218.73684210526315, "grad_norm": 0.12502525746822357, "learning_rate": 9.334611953187994e-05, "loss": 0.0049, "step": 20780 }, { "epoch": 218.8421052631579, "grad_norm": 0.11180505156517029, "learning_rate": 9.333787555493914e-05, "loss": 0.0049, "step": 20790 }, { "epoch": 218.94736842105263, "grad_norm": 0.10302519053220749, "learning_rate": 9.332962683863345e-05, "loss": 0.0051, "step": 20800 }, { "epoch": 219.05263157894737, "grad_norm": 0.12259330600500107, "learning_rate": 9.332137338386489e-05, "loss": 0.0057, "step": 20810 }, { "epoch": 219.1578947368421, "grad_norm": 0.07369441539049149, "learning_rate": 9.33131151915361e-05, "loss": 0.0058, "step": 20820 }, { "epoch": 219.26315789473685, "grad_norm": 0.09656079858541489, "learning_rate": 9.330485226255012e-05, "loss": 0.0059, "step": 20830 }, { "epoch": 219.3684210526316, "grad_norm": 0.08977683633565903, "learning_rate": 9.329658459781061e-05, "loss": 0.005, "step": 20840 }, { "epoch": 219.47368421052633, "grad_norm": 0.08973968029022217, "learning_rate": 9.328831219822172e-05, "loss": 0.005, "step": 20850 }, { "epoch": 219.57894736842104, "grad_norm": 0.07894519716501236, "learning_rate": 9.328003506468808e-05, "loss": 0.0055, "step": 20860 }, { "epoch": 219.68421052631578, "grad_norm": 0.06458108127117157, "learning_rate": 9.327175319811488e-05, "loss": 0.0051, "step": 20870 }, { "epoch": 219.78947368421052, "grad_norm": 0.08920023590326309, "learning_rate": 9.326346659940781e-05, "loss": 0.0059, "step": 20880 }, { "epoch": 219.89473684210526, "grad_norm": 0.08337945491075516, "learning_rate": 9.325517526947308e-05, "loss": 0.0055, "step": 20890 }, { "epoch": 220.0, "grad_norm": 0.11776619404554367, "learning_rate": 9.32468792092174e-05, "loss": 0.0046, "step": 20900 }, { "epoch": 220.10526315789474, "grad_norm": 0.08759979903697968, "learning_rate": 9.323857841954803e-05, "loss": 0.0051, "step": 20910 }, { "epoch": 220.21052631578948, "grad_norm": 0.10001180320978165, "learning_rate": 9.323027290137276e-05, "loss": 0.0047, "step": 20920 }, { "epoch": 220.31578947368422, "grad_norm": 0.11602969467639923, "learning_rate": 9.322196265559981e-05, "loss": 0.0052, "step": 20930 }, { "epoch": 220.42105263157896, "grad_norm": 0.10229936242103577, "learning_rate": 9.321364768313803e-05, "loss": 0.0053, "step": 20940 }, { "epoch": 220.52631578947367, "grad_norm": 0.10819990187883377, "learning_rate": 9.32053279848967e-05, "loss": 0.0048, "step": 20950 }, { "epoch": 220.6315789473684, "grad_norm": 0.11822228878736496, "learning_rate": 9.319700356178567e-05, "loss": 0.0051, "step": 20960 }, { "epoch": 220.73684210526315, "grad_norm": 0.07568807154893875, "learning_rate": 9.318867441471527e-05, "loss": 0.0052, "step": 20970 }, { "epoch": 220.8421052631579, "grad_norm": 0.08289973437786102, "learning_rate": 9.318034054459637e-05, "loss": 0.0047, "step": 20980 }, { "epoch": 220.94736842105263, "grad_norm": 0.0894988402724266, "learning_rate": 9.317200195234034e-05, "loss": 0.006, "step": 20990 }, { "epoch": 221.05263157894737, "grad_norm": 0.0931633710861206, "learning_rate": 9.316365863885909e-05, "loss": 0.0051, "step": 21000 }, { "epoch": 221.1578947368421, "grad_norm": 0.09513578563928604, "learning_rate": 9.315531060506502e-05, "loss": 0.0055, "step": 21010 }, { "epoch": 221.26315789473685, "grad_norm": 0.08475489169359207, "learning_rate": 9.314695785187108e-05, "loss": 0.0043, "step": 21020 }, { "epoch": 221.3684210526316, "grad_norm": 0.09925791621208191, "learning_rate": 9.313860038019069e-05, "loss": 0.0045, "step": 21030 }, { "epoch": 221.47368421052633, "grad_norm": 0.08526481688022614, "learning_rate": 9.313023819093782e-05, "loss": 0.0054, "step": 21040 }, { "epoch": 221.57894736842104, "grad_norm": 0.08443191647529602, "learning_rate": 9.312187128502695e-05, "loss": 0.0048, "step": 21050 }, { "epoch": 221.68421052631578, "grad_norm": 0.07730673253536224, "learning_rate": 9.311349966337307e-05, "loss": 0.0042, "step": 21060 }, { "epoch": 221.78947368421052, "grad_norm": 0.07438924163579941, "learning_rate": 9.310512332689169e-05, "loss": 0.0041, "step": 21070 }, { "epoch": 221.89473684210526, "grad_norm": 0.10123375058174133, "learning_rate": 9.309674227649883e-05, "loss": 0.0049, "step": 21080 }, { "epoch": 222.0, "grad_norm": 0.09969462454319, "learning_rate": 9.308835651311103e-05, "loss": 0.0052, "step": 21090 }, { "epoch": 222.10526315789474, "grad_norm": 0.08910614997148514, "learning_rate": 9.307996603764533e-05, "loss": 0.0041, "step": 21100 }, { "epoch": 222.21052631578948, "grad_norm": 0.10580173879861832, "learning_rate": 9.307157085101932e-05, "loss": 0.0053, "step": 21110 }, { "epoch": 222.31578947368422, "grad_norm": 0.1016528457403183, "learning_rate": 9.306317095415109e-05, "loss": 0.0045, "step": 21120 }, { "epoch": 222.42105263157896, "grad_norm": 0.08136841654777527, "learning_rate": 9.305476634795922e-05, "loss": 0.0055, "step": 21130 }, { "epoch": 222.52631578947367, "grad_norm": 0.08356888592243195, "learning_rate": 9.304635703336284e-05, "loss": 0.0046, "step": 21140 }, { "epoch": 222.6315789473684, "grad_norm": 0.11041110754013062, "learning_rate": 9.303794301128157e-05, "loss": 0.0053, "step": 21150 }, { "epoch": 222.73684210526315, "grad_norm": 0.1185794323682785, "learning_rate": 9.302952428263555e-05, "loss": 0.0052, "step": 21160 }, { "epoch": 222.8421052631579, "grad_norm": 0.09445054829120636, "learning_rate": 9.302110084834545e-05, "loss": 0.0051, "step": 21170 }, { "epoch": 222.94736842105263, "grad_norm": 0.10462455451488495, "learning_rate": 9.301267270933245e-05, "loss": 0.0051, "step": 21180 }, { "epoch": 223.05263157894737, "grad_norm": 0.11008958518505096, "learning_rate": 9.300423986651823e-05, "loss": 0.0048, "step": 21190 }, { "epoch": 223.1578947368421, "grad_norm": 0.08979330211877823, "learning_rate": 9.299580232082501e-05, "loss": 0.0055, "step": 21200 }, { "epoch": 223.26315789473685, "grad_norm": 0.09021782875061035, "learning_rate": 9.298736007317547e-05, "loss": 0.006, "step": 21210 }, { "epoch": 223.3684210526316, "grad_norm": 0.09357183426618576, "learning_rate": 9.297891312449288e-05, "loss": 0.005, "step": 21220 }, { "epoch": 223.47368421052633, "grad_norm": 0.09305062144994736, "learning_rate": 9.297046147570094e-05, "loss": 0.0047, "step": 21230 }, { "epoch": 223.57894736842104, "grad_norm": 0.08703404664993286, "learning_rate": 9.296200512772396e-05, "loss": 0.0054, "step": 21240 }, { "epoch": 223.68421052631578, "grad_norm": 0.08584420382976532, "learning_rate": 9.295354408148668e-05, "loss": 0.0045, "step": 21250 }, { "epoch": 223.78947368421052, "grad_norm": 0.08088276535272598, "learning_rate": 9.294507833791441e-05, "loss": 0.0045, "step": 21260 }, { "epoch": 223.89473684210526, "grad_norm": 0.09993323683738708, "learning_rate": 9.293660789793295e-05, "loss": 0.0046, "step": 21270 }, { "epoch": 224.0, "grad_norm": 0.1283658891916275, "learning_rate": 9.292813276246858e-05, "loss": 0.0049, "step": 21280 }, { "epoch": 224.10526315789474, "grad_norm": 0.11442077159881592, "learning_rate": 9.291965293244816e-05, "loss": 0.0047, "step": 21290 }, { "epoch": 224.21052631578948, "grad_norm": 0.11044860631227493, "learning_rate": 9.291116840879904e-05, "loss": 0.0053, "step": 21300 }, { "epoch": 224.31578947368422, "grad_norm": 0.09890852868556976, "learning_rate": 9.290267919244904e-05, "loss": 0.005, "step": 21310 }, { "epoch": 224.42105263157896, "grad_norm": 0.09613480418920517, "learning_rate": 9.289418528432655e-05, "loss": 0.0052, "step": 21320 }, { "epoch": 224.52631578947367, "grad_norm": 0.07443972676992416, "learning_rate": 9.288568668536045e-05, "loss": 0.0054, "step": 21330 }, { "epoch": 224.6315789473684, "grad_norm": 0.1126861572265625, "learning_rate": 9.287718339648013e-05, "loss": 0.0056, "step": 21340 }, { "epoch": 224.73684210526315, "grad_norm": 0.07478903979063034, "learning_rate": 9.28686754186155e-05, "loss": 0.0052, "step": 21350 }, { "epoch": 224.8421052631579, "grad_norm": 0.07806690037250519, "learning_rate": 9.286016275269698e-05, "loss": 0.0052, "step": 21360 }, { "epoch": 224.94736842105263, "grad_norm": 0.10250221192836761, "learning_rate": 9.285164539965551e-05, "loss": 0.0052, "step": 21370 }, { "epoch": 225.05263157894737, "grad_norm": 0.12439588457345963, "learning_rate": 9.284312336042251e-05, "loss": 0.0048, "step": 21380 }, { "epoch": 225.1578947368421, "grad_norm": 0.07597228139638901, "learning_rate": 9.283459663592996e-05, "loss": 0.0046, "step": 21390 }, { "epoch": 225.26315789473685, "grad_norm": 0.12736502289772034, "learning_rate": 9.282606522711033e-05, "loss": 0.0046, "step": 21400 }, { "epoch": 225.3684210526316, "grad_norm": 0.09100698679685593, "learning_rate": 9.281752913489657e-05, "loss": 0.0049, "step": 21410 }, { "epoch": 225.47368421052633, "grad_norm": 0.10369879007339478, "learning_rate": 9.280898836022222e-05, "loss": 0.0058, "step": 21420 }, { "epoch": 225.57894736842104, "grad_norm": 0.10605253279209137, "learning_rate": 9.280044290402126e-05, "loss": 0.0053, "step": 21430 }, { "epoch": 225.68421052631578, "grad_norm": 0.09239837527275085, "learning_rate": 9.279189276722821e-05, "loss": 0.0043, "step": 21440 }, { "epoch": 225.78947368421052, "grad_norm": 0.08773194998502731, "learning_rate": 9.278333795077812e-05, "loss": 0.0047, "step": 21450 }, { "epoch": 225.89473684210526, "grad_norm": 0.09348850697278976, "learning_rate": 9.27747784556065e-05, "loss": 0.0048, "step": 21460 }, { "epoch": 226.0, "grad_norm": 0.09699859470129013, "learning_rate": 9.276621428264942e-05, "loss": 0.0043, "step": 21470 }, { "epoch": 226.10526315789474, "grad_norm": 0.10818800330162048, "learning_rate": 9.275764543284345e-05, "loss": 0.0053, "step": 21480 }, { "epoch": 226.21052631578948, "grad_norm": 0.06589154899120331, "learning_rate": 9.274907190712566e-05, "loss": 0.0053, "step": 21490 }, { "epoch": 226.31578947368422, "grad_norm": 0.09075087308883667, "learning_rate": 9.274049370643363e-05, "loss": 0.0056, "step": 21500 }, { "epoch": 226.42105263157896, "grad_norm": 0.11430711299180984, "learning_rate": 9.273191083170547e-05, "loss": 0.0051, "step": 21510 }, { "epoch": 226.52631578947367, "grad_norm": 0.07675759494304657, "learning_rate": 9.27233232838798e-05, "loss": 0.0045, "step": 21520 }, { "epoch": 226.6315789473684, "grad_norm": 0.11721966415643692, "learning_rate": 9.27147310638957e-05, "loss": 0.0056, "step": 21530 }, { "epoch": 226.73684210526315, "grad_norm": 0.09672722220420837, "learning_rate": 9.270613417269286e-05, "loss": 0.0049, "step": 21540 }, { "epoch": 226.8421052631579, "grad_norm": 0.10501982271671295, "learning_rate": 9.269753261121138e-05, "loss": 0.0046, "step": 21550 }, { "epoch": 226.94736842105263, "grad_norm": 0.08955240249633789, "learning_rate": 9.268892638039194e-05, "loss": 0.0054, "step": 21560 }, { "epoch": 227.05263157894737, "grad_norm": 0.09980728477239609, "learning_rate": 9.268031548117569e-05, "loss": 0.0048, "step": 21570 }, { "epoch": 227.1578947368421, "grad_norm": 0.09216956049203873, "learning_rate": 9.26716999145043e-05, "loss": 0.0053, "step": 21580 }, { "epoch": 227.26315789473685, "grad_norm": 0.12352706491947174, "learning_rate": 9.266307968131998e-05, "loss": 0.0046, "step": 21590 }, { "epoch": 227.3684210526316, "grad_norm": 0.06046817824244499, "learning_rate": 9.26544547825654e-05, "loss": 0.0043, "step": 21600 }, { "epoch": 227.47368421052633, "grad_norm": 0.10073709487915039, "learning_rate": 9.264582521918376e-05, "loss": 0.0045, "step": 21610 }, { "epoch": 227.57894736842104, "grad_norm": 0.10522671043872833, "learning_rate": 9.263719099211881e-05, "loss": 0.0057, "step": 21620 }, { "epoch": 227.68421052631578, "grad_norm": 0.07593127340078354, "learning_rate": 9.262855210231476e-05, "loss": 0.0045, "step": 21630 }, { "epoch": 227.78947368421052, "grad_norm": 0.09037493169307709, "learning_rate": 9.261990855071633e-05, "loss": 0.0041, "step": 21640 }, { "epoch": 227.89473684210526, "grad_norm": 0.08297672867774963, "learning_rate": 9.261126033826878e-05, "loss": 0.0045, "step": 21650 }, { "epoch": 228.0, "grad_norm": 0.13249386847019196, "learning_rate": 9.260260746591786e-05, "loss": 0.005, "step": 21660 }, { "epoch": 228.10526315789474, "grad_norm": 0.09484949707984924, "learning_rate": 9.259394993460985e-05, "loss": 0.005, "step": 21670 }, { "epoch": 228.21052631578948, "grad_norm": 0.10077936947345734, "learning_rate": 9.258528774529151e-05, "loss": 0.0049, "step": 21680 }, { "epoch": 228.31578947368422, "grad_norm": 0.11134756356477737, "learning_rate": 9.257662089891013e-05, "loss": 0.0063, "step": 21690 }, { "epoch": 228.42105263157896, "grad_norm": 0.07473544031381607, "learning_rate": 9.25679493964135e-05, "loss": 0.005, "step": 21700 }, { "epoch": 228.52631578947367, "grad_norm": 0.11294106394052505, "learning_rate": 9.255927323874994e-05, "loss": 0.0049, "step": 21710 }, { "epoch": 228.6315789473684, "grad_norm": 0.0826835185289383, "learning_rate": 9.255059242686822e-05, "loss": 0.0043, "step": 21720 }, { "epoch": 228.73684210526315, "grad_norm": 0.06917119771242142, "learning_rate": 9.254190696171769e-05, "loss": 0.005, "step": 21730 }, { "epoch": 228.8421052631579, "grad_norm": 0.08161822706460953, "learning_rate": 9.25332168442482e-05, "loss": 0.0044, "step": 21740 }, { "epoch": 228.94736842105263, "grad_norm": 0.09031887352466583, "learning_rate": 9.252452207541004e-05, "loss": 0.0057, "step": 21750 }, { "epoch": 229.05263157894737, "grad_norm": 0.09225823730230331, "learning_rate": 9.251582265615409e-05, "loss": 0.0049, "step": 21760 }, { "epoch": 229.1578947368421, "grad_norm": 0.09902383387088776, "learning_rate": 9.250711858743169e-05, "loss": 0.0048, "step": 21770 }, { "epoch": 229.26315789473685, "grad_norm": 0.0862918421626091, "learning_rate": 9.24984098701947e-05, "loss": 0.0049, "step": 21780 }, { "epoch": 229.3684210526316, "grad_norm": 0.06053105369210243, "learning_rate": 9.248969650539552e-05, "loss": 0.0061, "step": 21790 }, { "epoch": 229.47368421052633, "grad_norm": 0.09889543801546097, "learning_rate": 9.2480978493987e-05, "loss": 0.0056, "step": 21800 }, { "epoch": 229.57894736842104, "grad_norm": 0.0948820635676384, "learning_rate": 9.247225583692256e-05, "loss": 0.0053, "step": 21810 }, { "epoch": 229.68421052631578, "grad_norm": 0.11278508603572845, "learning_rate": 9.246352853515607e-05, "loss": 0.0052, "step": 21820 }, { "epoch": 229.78947368421052, "grad_norm": 0.10766803473234177, "learning_rate": 9.245479658964194e-05, "loss": 0.0047, "step": 21830 }, { "epoch": 229.89473684210526, "grad_norm": 0.07301165908575058, "learning_rate": 9.244606000133507e-05, "loss": 0.0049, "step": 21840 }, { "epoch": 230.0, "grad_norm": 0.0802464708685875, "learning_rate": 9.24373187711909e-05, "loss": 0.0044, "step": 21850 }, { "epoch": 230.10526315789474, "grad_norm": 0.11225441098213196, "learning_rate": 9.242857290016537e-05, "loss": 0.005, "step": 21860 }, { "epoch": 230.21052631578948, "grad_norm": 0.06583172082901001, "learning_rate": 9.241982238921488e-05, "loss": 0.0047, "step": 21870 }, { "epoch": 230.31578947368422, "grad_norm": 0.08848714083433151, "learning_rate": 9.24110672392964e-05, "loss": 0.0056, "step": 21880 }, { "epoch": 230.42105263157896, "grad_norm": 0.07193911820650101, "learning_rate": 9.240230745136737e-05, "loss": 0.0045, "step": 21890 }, { "epoch": 230.52631578947367, "grad_norm": 0.09227006137371063, "learning_rate": 9.239354302638575e-05, "loss": 0.0048, "step": 21900 }, { "epoch": 230.6315789473684, "grad_norm": 0.08994708955287933, "learning_rate": 9.238477396531e-05, "loss": 0.0041, "step": 21910 }, { "epoch": 230.73684210526315, "grad_norm": 0.09489414095878601, "learning_rate": 9.23760002690991e-05, "loss": 0.005, "step": 21920 }, { "epoch": 230.8421052631579, "grad_norm": 0.11154148727655411, "learning_rate": 9.236722193871252e-05, "loss": 0.0049, "step": 21930 }, { "epoch": 230.94736842105263, "grad_norm": 0.09733685106039047, "learning_rate": 9.235843897511023e-05, "loss": 0.0051, "step": 21940 }, { "epoch": 231.05263157894737, "grad_norm": 0.09746429324150085, "learning_rate": 9.234965137925276e-05, "loss": 0.0046, "step": 21950 }, { "epoch": 231.1578947368421, "grad_norm": 0.10818134248256683, "learning_rate": 9.234085915210108e-05, "loss": 0.0049, "step": 21960 }, { "epoch": 231.26315789473685, "grad_norm": 0.09541594237089157, "learning_rate": 9.23320622946167e-05, "loss": 0.0052, "step": 21970 }, { "epoch": 231.3684210526316, "grad_norm": 0.10579414665699005, "learning_rate": 9.232326080776163e-05, "loss": 0.0048, "step": 21980 }, { "epoch": 231.47368421052633, "grad_norm": 0.09814228117465973, "learning_rate": 9.23144546924984e-05, "loss": 0.0053, "step": 21990 }, { "epoch": 231.57894736842104, "grad_norm": 0.09188726544380188, "learning_rate": 9.230564394979e-05, "loss": 0.0055, "step": 22000 }, { "epoch": 231.68421052631578, "grad_norm": 0.09712284058332443, "learning_rate": 9.22968285806e-05, "loss": 0.0053, "step": 22010 }, { "epoch": 231.78947368421052, "grad_norm": 0.10242398828268051, "learning_rate": 9.228800858589242e-05, "loss": 0.0055, "step": 22020 }, { "epoch": 231.89473684210526, "grad_norm": 0.11618710309267044, "learning_rate": 9.227918396663179e-05, "loss": 0.006, "step": 22030 }, { "epoch": 232.0, "grad_norm": 0.07782705128192902, "learning_rate": 9.227035472378319e-05, "loss": 0.0045, "step": 22040 }, { "epoch": 232.10526315789474, "grad_norm": 0.07661119848489761, "learning_rate": 9.226152085831213e-05, "loss": 0.0054, "step": 22050 }, { "epoch": 232.21052631578948, "grad_norm": 0.09495744109153748, "learning_rate": 9.22526823711847e-05, "loss": 0.0051, "step": 22060 }, { "epoch": 232.31578947368422, "grad_norm": 0.09135855734348297, "learning_rate": 9.224383926336745e-05, "loss": 0.0057, "step": 22070 }, { "epoch": 232.42105263157896, "grad_norm": 0.12937091290950775, "learning_rate": 9.223499153582744e-05, "loss": 0.0059, "step": 22080 }, { "epoch": 232.52631578947367, "grad_norm": 0.11951113492250443, "learning_rate": 9.222613918953226e-05, "loss": 0.0043, "step": 22090 }, { "epoch": 232.6315789473684, "grad_norm": 0.08559547364711761, "learning_rate": 9.221728222544999e-05, "loss": 0.0046, "step": 22100 }, { "epoch": 232.73684210526315, "grad_norm": 0.12571115791797638, "learning_rate": 9.22084206445492e-05, "loss": 0.0043, "step": 22110 }, { "epoch": 232.8421052631579, "grad_norm": 0.09690631181001663, "learning_rate": 9.2199554447799e-05, "loss": 0.0054, "step": 22120 }, { "epoch": 232.94736842105263, "grad_norm": 0.07842592895030975, "learning_rate": 9.219068363616897e-05, "loss": 0.0046, "step": 22130 }, { "epoch": 233.05263157894737, "grad_norm": 0.09841160476207733, "learning_rate": 9.218180821062919e-05, "loss": 0.0043, "step": 22140 }, { "epoch": 233.1578947368421, "grad_norm": 0.11560285091400146, "learning_rate": 9.21729281721503e-05, "loss": 0.0042, "step": 22150 }, { "epoch": 233.26315789473685, "grad_norm": 0.10504244267940521, "learning_rate": 9.216404352170339e-05, "loss": 0.0048, "step": 22160 }, { "epoch": 233.3684210526316, "grad_norm": 0.08539700508117676, "learning_rate": 9.215515426026007e-05, "loss": 0.0047, "step": 22170 }, { "epoch": 233.47368421052633, "grad_norm": 0.0800115093588829, "learning_rate": 9.214626038879246e-05, "loss": 0.0044, "step": 22180 }, { "epoch": 233.57894736842104, "grad_norm": 0.08743495494127274, "learning_rate": 9.21373619082732e-05, "loss": 0.0054, "step": 22190 }, { "epoch": 233.68421052631578, "grad_norm": 0.10172592103481293, "learning_rate": 9.212845881967535e-05, "loss": 0.0053, "step": 22200 }, { "epoch": 233.78947368421052, "grad_norm": 0.10154042392969131, "learning_rate": 9.211955112397262e-05, "loss": 0.0047, "step": 22210 }, { "epoch": 233.89473684210526, "grad_norm": 0.08866249024868011, "learning_rate": 9.211063882213909e-05, "loss": 0.0042, "step": 22220 }, { "epoch": 234.0, "grad_norm": 0.13246700167655945, "learning_rate": 9.210172191514942e-05, "loss": 0.0052, "step": 22230 }, { "epoch": 234.10526315789474, "grad_norm": 0.09192723035812378, "learning_rate": 9.209280040397874e-05, "loss": 0.0049, "step": 22240 }, { "epoch": 234.21052631578948, "grad_norm": 0.09955965727567673, "learning_rate": 9.208387428960268e-05, "loss": 0.0052, "step": 22250 }, { "epoch": 234.31578947368422, "grad_norm": 0.0928320437669754, "learning_rate": 9.20749435729974e-05, "loss": 0.0051, "step": 22260 }, { "epoch": 234.42105263157896, "grad_norm": 0.07064713537693024, "learning_rate": 9.206600825513957e-05, "loss": 0.0054, "step": 22270 }, { "epoch": 234.52631578947367, "grad_norm": 0.07049229741096497, "learning_rate": 9.20570683370063e-05, "loss": 0.0045, "step": 22280 }, { "epoch": 234.6315789473684, "grad_norm": 0.10702194273471832, "learning_rate": 9.204812381957528e-05, "loss": 0.005, "step": 22290 }, { "epoch": 234.73684210526315, "grad_norm": 0.08690179884433746, "learning_rate": 9.203917470382465e-05, "loss": 0.0047, "step": 22300 }, { "epoch": 234.8421052631579, "grad_norm": 0.09367760270833969, "learning_rate": 9.203022099073309e-05, "loss": 0.0059, "step": 22310 }, { "epoch": 234.94736842105263, "grad_norm": 0.07539884746074677, "learning_rate": 9.202126268127976e-05, "loss": 0.0042, "step": 22320 }, { "epoch": 235.05263157894737, "grad_norm": 0.09850426018238068, "learning_rate": 9.20122997764443e-05, "loss": 0.0051, "step": 22330 }, { "epoch": 235.1578947368421, "grad_norm": 0.12127961963415146, "learning_rate": 9.200333227720692e-05, "loss": 0.0049, "step": 22340 }, { "epoch": 235.26315789473685, "grad_norm": 0.09915795177221298, "learning_rate": 9.199436018454826e-05, "loss": 0.0055, "step": 22350 }, { "epoch": 235.3684210526316, "grad_norm": 0.08738639950752258, "learning_rate": 9.198538349944952e-05, "loss": 0.0053, "step": 22360 }, { "epoch": 235.47368421052633, "grad_norm": 0.08200608938932419, "learning_rate": 9.197640222289234e-05, "loss": 0.0052, "step": 22370 }, { "epoch": 235.57894736842104, "grad_norm": 0.12676459550857544, "learning_rate": 9.196741635585895e-05, "loss": 0.005, "step": 22380 }, { "epoch": 235.68421052631578, "grad_norm": 0.08172592520713806, "learning_rate": 9.195842589933199e-05, "loss": 0.0067, "step": 22390 }, { "epoch": 235.78947368421052, "grad_norm": 0.099916011095047, "learning_rate": 9.194943085429466e-05, "loss": 0.0053, "step": 22400 }, { "epoch": 235.89473684210526, "grad_norm": 0.10654307901859283, "learning_rate": 9.194043122173065e-05, "loss": 0.0058, "step": 22410 }, { "epoch": 236.0, "grad_norm": 0.12875771522521973, "learning_rate": 9.193142700262413e-05, "loss": 0.0041, "step": 22420 }, { "epoch": 236.10526315789474, "grad_norm": 0.08838526904582977, "learning_rate": 9.192241819795979e-05, "loss": 0.0057, "step": 22430 }, { "epoch": 236.21052631578948, "grad_norm": 0.0997476726770401, "learning_rate": 9.191340480872284e-05, "loss": 0.0054, "step": 22440 }, { "epoch": 236.31578947368422, "grad_norm": 0.08990754187107086, "learning_rate": 9.190438683589895e-05, "loss": 0.0062, "step": 22450 }, { "epoch": 236.42105263157896, "grad_norm": 0.09634021669626236, "learning_rate": 9.189536428047432e-05, "loss": 0.0044, "step": 22460 }, { "epoch": 236.52631578947367, "grad_norm": 0.09351784735918045, "learning_rate": 9.188633714343564e-05, "loss": 0.0052, "step": 22470 }, { "epoch": 236.6315789473684, "grad_norm": 0.08077388256788254, "learning_rate": 9.18773054257701e-05, "loss": 0.0049, "step": 22480 }, { "epoch": 236.73684210526315, "grad_norm": 0.07956048846244812, "learning_rate": 9.18682691284654e-05, "loss": 0.005, "step": 22490 }, { "epoch": 236.8421052631579, "grad_norm": 0.10041644424200058, "learning_rate": 9.185922825250974e-05, "loss": 0.0044, "step": 22500 }, { "epoch": 236.94736842105263, "grad_norm": 0.08218209445476532, "learning_rate": 9.185018279889181e-05, "loss": 0.0037, "step": 22510 }, { "epoch": 237.05263157894737, "grad_norm": 0.087824746966362, "learning_rate": 9.184113276860082e-05, "loss": 0.0048, "step": 22520 }, { "epoch": 237.1578947368421, "grad_norm": 0.0846189334988594, "learning_rate": 9.183207816262645e-05, "loss": 0.0053, "step": 22530 }, { "epoch": 237.26315789473685, "grad_norm": 0.08893334865570068, "learning_rate": 9.182301898195891e-05, "loss": 0.0039, "step": 22540 }, { "epoch": 237.3684210526316, "grad_norm": 0.10665535926818848, "learning_rate": 9.181395522758889e-05, "loss": 0.0045, "step": 22550 }, { "epoch": 237.47368421052633, "grad_norm": 0.0916961133480072, "learning_rate": 9.180488690050759e-05, "loss": 0.0044, "step": 22560 }, { "epoch": 237.57894736842104, "grad_norm": 0.08292534947395325, "learning_rate": 9.179581400170671e-05, "loss": 0.0049, "step": 22570 }, { "epoch": 237.68421052631578, "grad_norm": 0.06852658092975616, "learning_rate": 9.178673653217845e-05, "loss": 0.0048, "step": 22580 }, { "epoch": 237.78947368421052, "grad_norm": 0.08216863125562668, "learning_rate": 9.177765449291551e-05, "loss": 0.0044, "step": 22590 }, { "epoch": 237.89473684210526, "grad_norm": 0.09192163497209549, "learning_rate": 9.176856788491109e-05, "loss": 0.004, "step": 22600 }, { "epoch": 238.0, "grad_norm": 0.10377254337072372, "learning_rate": 9.175947670915887e-05, "loss": 0.0041, "step": 22610 }, { "epoch": 238.10526315789474, "grad_norm": 0.09770612418651581, "learning_rate": 9.175038096665309e-05, "loss": 0.0051, "step": 22620 }, { "epoch": 238.21052631578948, "grad_norm": 0.09249929338693619, "learning_rate": 9.17412806583884e-05, "loss": 0.0047, "step": 22630 }, { "epoch": 238.31578947368422, "grad_norm": 0.10935710370540619, "learning_rate": 9.173217578536002e-05, "loss": 0.0049, "step": 22640 }, { "epoch": 238.42105263157896, "grad_norm": 0.08307832479476929, "learning_rate": 9.172306634856362e-05, "loss": 0.0042, "step": 22650 }, { "epoch": 238.52631578947367, "grad_norm": 0.08755365759134293, "learning_rate": 9.171395234899545e-05, "loss": 0.0047, "step": 22660 }, { "epoch": 238.6315789473684, "grad_norm": 0.09510734677314758, "learning_rate": 9.170483378765214e-05, "loss": 0.0046, "step": 22670 }, { "epoch": 238.73684210526315, "grad_norm": 0.0914827361702919, "learning_rate": 9.169571066553091e-05, "loss": 0.0044, "step": 22680 }, { "epoch": 238.8421052631579, "grad_norm": 0.0945587232708931, "learning_rate": 9.168658298362946e-05, "loss": 0.0045, "step": 22690 }, { "epoch": 238.94736842105263, "grad_norm": 0.07207690924406052, "learning_rate": 9.167745074294598e-05, "loss": 0.0047, "step": 22700 }, { "epoch": 239.05263157894737, "grad_norm": 0.07603525370359421, "learning_rate": 9.166831394447913e-05, "loss": 0.0057, "step": 22710 }, { "epoch": 239.1578947368421, "grad_norm": 0.1414797604084015, "learning_rate": 9.165917258922812e-05, "loss": 0.005, "step": 22720 }, { "epoch": 239.26315789473685, "grad_norm": 0.11342702805995941, "learning_rate": 9.165002667819262e-05, "loss": 0.0054, "step": 22730 }, { "epoch": 239.3684210526316, "grad_norm": 0.09689325094223022, "learning_rate": 9.164087621237282e-05, "loss": 0.0053, "step": 22740 }, { "epoch": 239.47368421052633, "grad_norm": 0.09454178065061569, "learning_rate": 9.163172119276942e-05, "loss": 0.0051, "step": 22750 }, { "epoch": 239.57894736842104, "grad_norm": 0.07798517495393753, "learning_rate": 9.162256162038358e-05, "loss": 0.005, "step": 22760 }, { "epoch": 239.68421052631578, "grad_norm": 0.08280234038829803, "learning_rate": 9.161339749621698e-05, "loss": 0.005, "step": 22770 }, { "epoch": 239.78947368421052, "grad_norm": 0.09641771018505096, "learning_rate": 9.160422882127177e-05, "loss": 0.0053, "step": 22780 }, { "epoch": 239.89473684210526, "grad_norm": 0.11817659437656403, "learning_rate": 9.159505559655069e-05, "loss": 0.0057, "step": 22790 }, { "epoch": 240.0, "grad_norm": 0.10546832531690598, "learning_rate": 9.158587782305684e-05, "loss": 0.0041, "step": 22800 }, { "epoch": 240.10526315789474, "grad_norm": 0.06715750694274902, "learning_rate": 9.157669550179391e-05, "loss": 0.0052, "step": 22810 }, { "epoch": 240.21052631578948, "grad_norm": 0.11064883321523666, "learning_rate": 9.156750863376609e-05, "loss": 0.0047, "step": 22820 }, { "epoch": 240.31578947368422, "grad_norm": 0.08194030076265335, "learning_rate": 9.155831721997801e-05, "loss": 0.0049, "step": 22830 }, { "epoch": 240.42105263157896, "grad_norm": 0.10611561685800552, "learning_rate": 9.154912126143484e-05, "loss": 0.0052, "step": 22840 }, { "epoch": 240.52631578947367, "grad_norm": 0.10975619405508041, "learning_rate": 9.153992075914224e-05, "loss": 0.0046, "step": 22850 }, { "epoch": 240.6315789473684, "grad_norm": 0.07129723578691483, "learning_rate": 9.153071571410635e-05, "loss": 0.0048, "step": 22860 }, { "epoch": 240.73684210526315, "grad_norm": 0.07782033830881119, "learning_rate": 9.152150612733384e-05, "loss": 0.0046, "step": 22870 }, { "epoch": 240.8421052631579, "grad_norm": 0.09615527838468552, "learning_rate": 9.151229199983184e-05, "loss": 0.0046, "step": 22880 }, { "epoch": 240.94736842105263, "grad_norm": 0.07788117229938507, "learning_rate": 9.150307333260802e-05, "loss": 0.0042, "step": 22890 }, { "epoch": 241.05263157894737, "grad_norm": 0.08325617015361786, "learning_rate": 9.149385012667048e-05, "loss": 0.0041, "step": 22900 }, { "epoch": 241.1578947368421, "grad_norm": 0.10076621919870377, "learning_rate": 9.148462238302788e-05, "loss": 0.0061, "step": 22910 }, { "epoch": 241.26315789473685, "grad_norm": 0.10900676995515823, "learning_rate": 9.147539010268936e-05, "loss": 0.0051, "step": 22920 }, { "epoch": 241.3684210526316, "grad_norm": 0.09299885481595993, "learning_rate": 9.14661532866645e-05, "loss": 0.0056, "step": 22930 }, { "epoch": 241.47368421052633, "grad_norm": 0.07812844961881638, "learning_rate": 9.145691193596348e-05, "loss": 0.0048, "step": 22940 }, { "epoch": 241.57894736842104, "grad_norm": 0.07572343200445175, "learning_rate": 9.144766605159691e-05, "loss": 0.0055, "step": 22950 }, { "epoch": 241.68421052631578, "grad_norm": 0.08195517957210541, "learning_rate": 9.14384156345759e-05, "loss": 0.0044, "step": 22960 }, { "epoch": 241.78947368421052, "grad_norm": 0.08120331913232803, "learning_rate": 9.142916068591204e-05, "loss": 0.0051, "step": 22970 }, { "epoch": 241.89473684210526, "grad_norm": 0.06528232991695404, "learning_rate": 9.141990120661746e-05, "loss": 0.004, "step": 22980 }, { "epoch": 242.0, "grad_norm": 0.06140722334384918, "learning_rate": 9.141063719770475e-05, "loss": 0.0039, "step": 22990 }, { "epoch": 242.10526315789474, "grad_norm": 0.09246635437011719, "learning_rate": 9.140136866018704e-05, "loss": 0.005, "step": 23000 }, { "epoch": 242.21052631578948, "grad_norm": 0.09309699386358261, "learning_rate": 9.139209559507788e-05, "loss": 0.0049, "step": 23010 }, { "epoch": 242.31578947368422, "grad_norm": 0.08090782910585403, "learning_rate": 9.13828180033914e-05, "loss": 0.0048, "step": 23020 }, { "epoch": 242.42105263157896, "grad_norm": 0.10515516251325607, "learning_rate": 9.137353588614212e-05, "loss": 0.005, "step": 23030 }, { "epoch": 242.52631578947367, "grad_norm": 0.09139833599328995, "learning_rate": 9.136424924434519e-05, "loss": 0.0051, "step": 23040 }, { "epoch": 242.6315789473684, "grad_norm": 0.10170422494411469, "learning_rate": 9.135495807901615e-05, "loss": 0.0048, "step": 23050 }, { "epoch": 242.73684210526315, "grad_norm": 0.10197393596172333, "learning_rate": 9.134566239117108e-05, "loss": 0.0049, "step": 23060 }, { "epoch": 242.8421052631579, "grad_norm": 0.10707356780767441, "learning_rate": 9.13363621818265e-05, "loss": 0.0044, "step": 23070 }, { "epoch": 242.94736842105263, "grad_norm": 0.10135133564472198, "learning_rate": 9.132705745199953e-05, "loss": 0.0045, "step": 23080 }, { "epoch": 243.05263157894737, "grad_norm": 0.10281630605459213, "learning_rate": 9.131774820270768e-05, "loss": 0.0051, "step": 23090 }, { "epoch": 243.1578947368421, "grad_norm": 0.0787549838423729, "learning_rate": 9.130843443496901e-05, "loss": 0.0047, "step": 23100 }, { "epoch": 243.26315789473685, "grad_norm": 0.09946662932634354, "learning_rate": 9.129911614980206e-05, "loss": 0.0055, "step": 23110 }, { "epoch": 243.3684210526316, "grad_norm": 0.0893026739358902, "learning_rate": 9.128979334822584e-05, "loss": 0.0047, "step": 23120 }, { "epoch": 243.47368421052633, "grad_norm": 0.09145066887140274, "learning_rate": 9.128046603125992e-05, "loss": 0.0057, "step": 23130 }, { "epoch": 243.57894736842104, "grad_norm": 0.08795006573200226, "learning_rate": 9.12711341999243e-05, "loss": 0.0048, "step": 23140 }, { "epoch": 243.68421052631578, "grad_norm": 0.08502758294343948, "learning_rate": 9.12617978552395e-05, "loss": 0.0045, "step": 23150 }, { "epoch": 243.78947368421052, "grad_norm": 0.09842435270547867, "learning_rate": 9.12524569982265e-05, "loss": 0.0047, "step": 23160 }, { "epoch": 243.89473684210526, "grad_norm": 0.10225595533847809, "learning_rate": 9.124311162990684e-05, "loss": 0.0052, "step": 23170 }, { "epoch": 244.0, "grad_norm": 0.132956862449646, "learning_rate": 9.12337617513025e-05, "loss": 0.006, "step": 23180 }, { "epoch": 244.10526315789474, "grad_norm": 0.11275745183229446, "learning_rate": 9.122440736343596e-05, "loss": 0.0047, "step": 23190 }, { "epoch": 244.21052631578948, "grad_norm": 0.10249364376068115, "learning_rate": 9.12150484673302e-05, "loss": 0.0052, "step": 23200 }, { "epoch": 244.31578947368422, "grad_norm": 0.09423460811376572, "learning_rate": 9.120568506400873e-05, "loss": 0.0043, "step": 23210 }, { "epoch": 244.42105263157896, "grad_norm": 0.09284797310829163, "learning_rate": 9.119631715449548e-05, "loss": 0.0059, "step": 23220 }, { "epoch": 244.52631578947367, "grad_norm": 0.1103181391954422, "learning_rate": 9.118694473981493e-05, "loss": 0.0056, "step": 23230 }, { "epoch": 244.6315789473684, "grad_norm": 0.0860523208975792, "learning_rate": 9.117756782099203e-05, "loss": 0.0049, "step": 23240 }, { "epoch": 244.73684210526315, "grad_norm": 0.09586432576179504, "learning_rate": 9.11681863990522e-05, "loss": 0.0048, "step": 23250 }, { "epoch": 244.8421052631579, "grad_norm": 0.07418300956487656, "learning_rate": 9.115880047502142e-05, "loss": 0.0046, "step": 23260 }, { "epoch": 244.94736842105263, "grad_norm": 0.08322425931692123, "learning_rate": 9.114941004992609e-05, "loss": 0.0055, "step": 23270 }, { "epoch": 245.05263157894737, "grad_norm": 0.11338356137275696, "learning_rate": 9.114001512479317e-05, "loss": 0.0056, "step": 23280 }, { "epoch": 245.1578947368421, "grad_norm": 0.09779495745897293, "learning_rate": 9.113061570065003e-05, "loss": 0.0049, "step": 23290 }, { "epoch": 245.26315789473685, "grad_norm": 0.0775829404592514, "learning_rate": 9.112121177852459e-05, "loss": 0.0046, "step": 23300 }, { "epoch": 245.3684210526316, "grad_norm": 0.09964293986558914, "learning_rate": 9.111180335944527e-05, "loss": 0.0044, "step": 23310 }, { "epoch": 245.47368421052633, "grad_norm": 0.08731137961149216, "learning_rate": 9.110239044444093e-05, "loss": 0.0051, "step": 23320 }, { "epoch": 245.57894736842104, "grad_norm": 0.07495295256376266, "learning_rate": 9.109297303454099e-05, "loss": 0.0048, "step": 23330 }, { "epoch": 245.68421052631578, "grad_norm": 0.07745756208896637, "learning_rate": 9.108355113077526e-05, "loss": 0.0045, "step": 23340 }, { "epoch": 245.78947368421052, "grad_norm": 0.11412325501441956, "learning_rate": 9.107412473417419e-05, "loss": 0.0053, "step": 23350 }, { "epoch": 245.89473684210526, "grad_norm": 0.0942738950252533, "learning_rate": 9.106469384576858e-05, "loss": 0.0057, "step": 23360 }, { "epoch": 246.0, "grad_norm": 0.07528401166200638, "learning_rate": 9.105525846658978e-05, "loss": 0.005, "step": 23370 }, { "epoch": 246.10526315789474, "grad_norm": 0.09947749227285385, "learning_rate": 9.104581859766965e-05, "loss": 0.005, "step": 23380 }, { "epoch": 246.21052631578948, "grad_norm": 0.08642572909593582, "learning_rate": 9.10363742400405e-05, "loss": 0.0048, "step": 23390 }, { "epoch": 246.31578947368422, "grad_norm": 0.09067782759666443, "learning_rate": 9.102692539473518e-05, "loss": 0.0051, "step": 23400 }, { "epoch": 246.42105263157896, "grad_norm": 0.08783146739006042, "learning_rate": 9.101747206278697e-05, "loss": 0.0055, "step": 23410 }, { "epoch": 246.52631578947367, "grad_norm": 0.09521854668855667, "learning_rate": 9.100801424522968e-05, "loss": 0.005, "step": 23420 }, { "epoch": 246.6315789473684, "grad_norm": 0.07011570036411285, "learning_rate": 9.099855194309762e-05, "loss": 0.0049, "step": 23430 }, { "epoch": 246.73684210526315, "grad_norm": 0.08573336154222488, "learning_rate": 9.098908515742554e-05, "loss": 0.0048, "step": 23440 }, { "epoch": 246.8421052631579, "grad_norm": 0.07650633901357651, "learning_rate": 9.097961388924873e-05, "loss": 0.0045, "step": 23450 }, { "epoch": 246.94736842105263, "grad_norm": 0.09443485736846924, "learning_rate": 9.097013813960298e-05, "loss": 0.0052, "step": 23460 }, { "epoch": 247.05263157894737, "grad_norm": 0.0755956843495369, "learning_rate": 9.09606579095245e-05, "loss": 0.0046, "step": 23470 }, { "epoch": 247.1578947368421, "grad_norm": 0.09966499358415604, "learning_rate": 9.095117320005008e-05, "loss": 0.0041, "step": 23480 }, { "epoch": 247.26315789473685, "grad_norm": 0.0849524661898613, "learning_rate": 9.094168401221691e-05, "loss": 0.0048, "step": 23490 }, { "epoch": 247.3684210526316, "grad_norm": 0.09650569409132004, "learning_rate": 9.093219034706273e-05, "loss": 0.005, "step": 23500 }, { "epoch": 247.47368421052633, "grad_norm": 0.0809812843799591, "learning_rate": 9.092269220562577e-05, "loss": 0.0055, "step": 23510 }, { "epoch": 247.57894736842104, "grad_norm": 0.10821619629859924, "learning_rate": 9.09131895889447e-05, "loss": 0.0053, "step": 23520 }, { "epoch": 247.68421052631578, "grad_norm": 0.08655670285224915, "learning_rate": 9.090368249805873e-05, "loss": 0.0041, "step": 23530 }, { "epoch": 247.78947368421052, "grad_norm": 0.09134835749864578, "learning_rate": 9.089417093400754e-05, "loss": 0.0053, "step": 23540 }, { "epoch": 247.89473684210526, "grad_norm": 0.06592364609241486, "learning_rate": 9.088465489783131e-05, "loss": 0.0049, "step": 23550 }, { "epoch": 248.0, "grad_norm": 0.11600016802549362, "learning_rate": 9.087513439057068e-05, "loss": 0.0039, "step": 23560 }, { "epoch": 248.10526315789474, "grad_norm": 0.09561937302350998, "learning_rate": 9.08656094132668e-05, "loss": 0.0056, "step": 23570 }, { "epoch": 248.21052631578948, "grad_norm": 0.0898534432053566, "learning_rate": 9.085607996696134e-05, "loss": 0.0046, "step": 23580 }, { "epoch": 248.31578947368422, "grad_norm": 0.110638327896595, "learning_rate": 9.084654605269639e-05, "loss": 0.0054, "step": 23590 }, { "epoch": 248.42105263157896, "grad_norm": 0.09947089850902557, "learning_rate": 9.083700767151457e-05, "loss": 0.0052, "step": 23600 }, { "epoch": 248.52631578947367, "grad_norm": 0.10138768702745438, "learning_rate": 9.082746482445898e-05, "loss": 0.0048, "step": 23610 }, { "epoch": 248.6315789473684, "grad_norm": 0.08945556730031967, "learning_rate": 9.081791751257325e-05, "loss": 0.0047, "step": 23620 }, { "epoch": 248.73684210526315, "grad_norm": 0.07594466209411621, "learning_rate": 9.080836573690142e-05, "loss": 0.0038, "step": 23630 }, { "epoch": 248.8421052631579, "grad_norm": 0.10993480682373047, "learning_rate": 9.079880949848805e-05, "loss": 0.005, "step": 23640 }, { "epoch": 248.94736842105263, "grad_norm": 0.10146602988243103, "learning_rate": 9.078924879837822e-05, "loss": 0.0045, "step": 23650 }, { "epoch": 249.05263157894737, "grad_norm": 0.08585251867771149, "learning_rate": 9.077968363761747e-05, "loss": 0.0055, "step": 23660 }, { "epoch": 249.1578947368421, "grad_norm": 0.09343577921390533, "learning_rate": 9.077011401725182e-05, "loss": 0.0052, "step": 23670 }, { "epoch": 249.26315789473685, "grad_norm": 0.08282516151666641, "learning_rate": 9.07605399383278e-05, "loss": 0.0049, "step": 23680 }, { "epoch": 249.3684210526316, "grad_norm": 0.09948794543743134, "learning_rate": 9.075096140189243e-05, "loss": 0.0047, "step": 23690 }, { "epoch": 249.47368421052633, "grad_norm": 0.08121038973331451, "learning_rate": 9.074137840899318e-05, "loss": 0.0058, "step": 23700 }, { "epoch": 249.57894736842104, "grad_norm": 0.09523860365152359, "learning_rate": 9.073179096067804e-05, "loss": 0.0044, "step": 23710 }, { "epoch": 249.68421052631578, "grad_norm": 0.07148752361536026, "learning_rate": 9.072219905799549e-05, "loss": 0.0045, "step": 23720 }, { "epoch": 249.78947368421052, "grad_norm": 0.07432721555233002, "learning_rate": 9.071260270199447e-05, "loss": 0.0048, "step": 23730 }, { "epoch": 249.89473684210526, "grad_norm": 0.08927646279335022, "learning_rate": 9.070300189372441e-05, "loss": 0.0048, "step": 23740 }, { "epoch": 250.0, "grad_norm": 0.08876445144414902, "learning_rate": 9.069339663423528e-05, "loss": 0.0046, "step": 23750 }, { "epoch": 250.10526315789474, "grad_norm": 0.109311543405056, "learning_rate": 9.068378692457747e-05, "loss": 0.0052, "step": 23760 }, { "epoch": 250.21052631578948, "grad_norm": 0.06328414380550385, "learning_rate": 9.067417276580189e-05, "loss": 0.0045, "step": 23770 }, { "epoch": 250.31578947368422, "grad_norm": 0.06989211589097977, "learning_rate": 9.066455415895993e-05, "loss": 0.0048, "step": 23780 }, { "epoch": 250.42105263157896, "grad_norm": 0.09645038098096848, "learning_rate": 9.065493110510346e-05, "loss": 0.0046, "step": 23790 }, { "epoch": 250.52631578947367, "grad_norm": 0.08857565373182297, "learning_rate": 9.064530360528484e-05, "loss": 0.0056, "step": 23800 }, { "epoch": 250.6315789473684, "grad_norm": 0.09868523478507996, "learning_rate": 9.063567166055695e-05, "loss": 0.0053, "step": 23810 }, { "epoch": 250.73684210526315, "grad_norm": 0.08307787775993347, "learning_rate": 9.062603527197308e-05, "loss": 0.0048, "step": 23820 }, { "epoch": 250.8421052631579, "grad_norm": 0.05872514843940735, "learning_rate": 9.06163944405871e-05, "loss": 0.0045, "step": 23830 }, { "epoch": 250.94736842105263, "grad_norm": 0.08355265110731125, "learning_rate": 9.060674916745327e-05, "loss": 0.0051, "step": 23840 }, { "epoch": 251.05263157894737, "grad_norm": 0.1225322037935257, "learning_rate": 9.05970994536264e-05, "loss": 0.0043, "step": 23850 }, { "epoch": 251.1578947368421, "grad_norm": 0.08716987818479538, "learning_rate": 9.05874453001618e-05, "loss": 0.0053, "step": 23860 }, { "epoch": 251.26315789473685, "grad_norm": 0.08592429757118225, "learning_rate": 9.057778670811517e-05, "loss": 0.004, "step": 23870 }, { "epoch": 251.3684210526316, "grad_norm": 0.08821776509284973, "learning_rate": 9.056812367854281e-05, "loss": 0.0048, "step": 23880 }, { "epoch": 251.47368421052633, "grad_norm": 0.07231392711400986, "learning_rate": 9.055845621250143e-05, "loss": 0.0041, "step": 23890 }, { "epoch": 251.57894736842104, "grad_norm": 0.0770423486828804, "learning_rate": 9.054878431104825e-05, "loss": 0.0042, "step": 23900 }, { "epoch": 251.68421052631578, "grad_norm": 0.0868351086974144, "learning_rate": 9.0539107975241e-05, "loss": 0.0049, "step": 23910 }, { "epoch": 251.78947368421052, "grad_norm": 0.06868606805801392, "learning_rate": 9.052942720613784e-05, "loss": 0.0046, "step": 23920 }, { "epoch": 251.89473684210526, "grad_norm": 0.08195365965366364, "learning_rate": 9.051974200479745e-05, "loss": 0.0046, "step": 23930 }, { "epoch": 252.0, "grad_norm": 0.10459356755018234, "learning_rate": 9.051005237227901e-05, "loss": 0.0042, "step": 23940 }, { "epoch": 252.10526315789474, "grad_norm": 0.07770103216171265, "learning_rate": 9.050035830964215e-05, "loss": 0.0047, "step": 23950 }, { "epoch": 252.21052631578948, "grad_norm": 0.08181679993867874, "learning_rate": 9.049065981794698e-05, "loss": 0.0045, "step": 23960 }, { "epoch": 252.31578947368422, "grad_norm": 0.10882919281721115, "learning_rate": 9.048095689825414e-05, "loss": 0.005, "step": 23970 }, { "epoch": 252.42105263157896, "grad_norm": 0.11738716810941696, "learning_rate": 9.047124955162472e-05, "loss": 0.0047, "step": 23980 }, { "epoch": 252.52631578947367, "grad_norm": 0.09284193813800812, "learning_rate": 9.046153777912028e-05, "loss": 0.0048, "step": 23990 }, { "epoch": 252.6315789473684, "grad_norm": 0.08232200890779495, "learning_rate": 9.045182158180292e-05, "loss": 0.005, "step": 24000 }, { "epoch": 252.73684210526315, "grad_norm": 0.08188033103942871, "learning_rate": 9.044210096073516e-05, "loss": 0.005, "step": 24010 }, { "epoch": 252.8421052631579, "grad_norm": 0.09080926328897476, "learning_rate": 9.043237591698004e-05, "loss": 0.0042, "step": 24020 }, { "epoch": 252.94736842105263, "grad_norm": 0.0761692002415657, "learning_rate": 9.04226464516011e-05, "loss": 0.005, "step": 24030 }, { "epoch": 253.05263157894737, "grad_norm": 0.09699016064405441, "learning_rate": 9.041291256566229e-05, "loss": 0.004, "step": 24040 }, { "epoch": 253.1578947368421, "grad_norm": 0.0893397331237793, "learning_rate": 9.040317426022814e-05, "loss": 0.0045, "step": 24050 }, { "epoch": 253.26315789473685, "grad_norm": 0.10107294470071793, "learning_rate": 9.03934315363636e-05, "loss": 0.0049, "step": 24060 }, { "epoch": 253.3684210526316, "grad_norm": 0.08357974141836166, "learning_rate": 9.038368439513409e-05, "loss": 0.006, "step": 24070 }, { "epoch": 253.47368421052633, "grad_norm": 0.08103679865598679, "learning_rate": 9.03739328376056e-05, "loss": 0.0039, "step": 24080 }, { "epoch": 253.57894736842104, "grad_norm": 0.08546243607997894, "learning_rate": 9.036417686484451e-05, "loss": 0.005, "step": 24090 }, { "epoch": 253.68421052631578, "grad_norm": 0.11623452603816986, "learning_rate": 9.035441647791773e-05, "loss": 0.0058, "step": 24100 }, { "epoch": 253.78947368421052, "grad_norm": 0.10642939060926437, "learning_rate": 9.034465167789263e-05, "loss": 0.0044, "step": 24110 }, { "epoch": 253.89473684210526, "grad_norm": 0.10678545385599136, "learning_rate": 9.033488246583706e-05, "loss": 0.0048, "step": 24120 }, { "epoch": 254.0, "grad_norm": 0.09002236276865005, "learning_rate": 9.032510884281941e-05, "loss": 0.0055, "step": 24130 }, { "epoch": 254.10526315789474, "grad_norm": 0.09414684772491455, "learning_rate": 9.031533080990848e-05, "loss": 0.0044, "step": 24140 }, { "epoch": 254.21052631578948, "grad_norm": 0.09916674345731735, "learning_rate": 9.030554836817358e-05, "loss": 0.0056, "step": 24150 }, { "epoch": 254.31578947368422, "grad_norm": 0.06678690016269684, "learning_rate": 9.029576151868451e-05, "loss": 0.0043, "step": 24160 }, { "epoch": 254.42105263157896, "grad_norm": 0.10782409459352493, "learning_rate": 9.028597026251155e-05, "loss": 0.0047, "step": 24170 }, { "epoch": 254.52631578947367, "grad_norm": 0.08687081933021545, "learning_rate": 9.027617460072547e-05, "loss": 0.0061, "step": 24180 }, { "epoch": 254.6315789473684, "grad_norm": 0.10923073440790176, "learning_rate": 9.026637453439745e-05, "loss": 0.0043, "step": 24190 }, { "epoch": 254.73684210526315, "grad_norm": 0.09814601391553879, "learning_rate": 9.025657006459927e-05, "loss": 0.0039, "step": 24200 }, { "epoch": 254.8421052631579, "grad_norm": 0.08685661852359772, "learning_rate": 9.024676119240311e-05, "loss": 0.0041, "step": 24210 }, { "epoch": 254.94736842105263, "grad_norm": 0.08172531425952911, "learning_rate": 9.023694791888166e-05, "loss": 0.0034, "step": 24220 }, { "epoch": 255.05263157894737, "grad_norm": 0.0640685185790062, "learning_rate": 9.022713024510808e-05, "loss": 0.0042, "step": 24230 }, { "epoch": 255.1578947368421, "grad_norm": 0.10331367701292038, "learning_rate": 9.021730817215601e-05, "loss": 0.0042, "step": 24240 }, { "epoch": 255.26315789473685, "grad_norm": 0.09030447155237198, "learning_rate": 9.02074817010996e-05, "loss": 0.0044, "step": 24250 }, { "epoch": 255.3684210526316, "grad_norm": 0.08852477371692657, "learning_rate": 9.019765083301342e-05, "loss": 0.0044, "step": 24260 }, { "epoch": 255.47368421052633, "grad_norm": 0.11999541521072388, "learning_rate": 9.01878155689726e-05, "loss": 0.0043, "step": 24270 }, { "epoch": 255.57894736842104, "grad_norm": 0.12364986538887024, "learning_rate": 9.017797591005268e-05, "loss": 0.0045, "step": 24280 }, { "epoch": 255.68421052631578, "grad_norm": 0.0893399566411972, "learning_rate": 9.016813185732972e-05, "loss": 0.0049, "step": 24290 }, { "epoch": 255.78947368421052, "grad_norm": 0.09106048941612244, "learning_rate": 9.015828341188027e-05, "loss": 0.0054, "step": 24300 }, { "epoch": 255.89473684210526, "grad_norm": 0.1045936718583107, "learning_rate": 9.01484305747813e-05, "loss": 0.0044, "step": 24310 }, { "epoch": 256.0, "grad_norm": 0.10703691840171814, "learning_rate": 9.013857334711033e-05, "loss": 0.0047, "step": 24320 }, { "epoch": 256.10526315789474, "grad_norm": 0.10012687742710114, "learning_rate": 9.012871172994534e-05, "loss": 0.0049, "step": 24330 }, { "epoch": 256.2105263157895, "grad_norm": 0.11987627297639847, "learning_rate": 9.011884572436476e-05, "loss": 0.0055, "step": 24340 }, { "epoch": 256.3157894736842, "grad_norm": 0.07379309833049774, "learning_rate": 9.010897533144754e-05, "loss": 0.0046, "step": 24350 }, { "epoch": 256.42105263157896, "grad_norm": 0.062425266951322556, "learning_rate": 9.009910055227306e-05, "loss": 0.0041, "step": 24360 }, { "epoch": 256.5263157894737, "grad_norm": 0.0787573978304863, "learning_rate": 9.008922138792124e-05, "loss": 0.0055, "step": 24370 }, { "epoch": 256.63157894736844, "grad_norm": 0.08086098730564117, "learning_rate": 9.007933783947244e-05, "loss": 0.0044, "step": 24380 }, { "epoch": 256.7368421052632, "grad_norm": 0.08992540091276169, "learning_rate": 9.006944990800752e-05, "loss": 0.0055, "step": 24390 }, { "epoch": 256.8421052631579, "grad_norm": 0.06688720732927322, "learning_rate": 9.005955759460779e-05, "loss": 0.0053, "step": 24400 }, { "epoch": 256.94736842105266, "grad_norm": 0.07554023712873459, "learning_rate": 9.004966090035508e-05, "loss": 0.0047, "step": 24410 }, { "epoch": 257.05263157894734, "grad_norm": 0.08493076264858246, "learning_rate": 9.003975982633166e-05, "loss": 0.005, "step": 24420 }, { "epoch": 257.1578947368421, "grad_norm": 0.10376620292663574, "learning_rate": 9.00298543736203e-05, "loss": 0.0052, "step": 24430 }, { "epoch": 257.2631578947368, "grad_norm": 0.10275474190711975, "learning_rate": 9.001994454330427e-05, "loss": 0.0051, "step": 24440 }, { "epoch": 257.36842105263156, "grad_norm": 0.10507708787918091, "learning_rate": 9.001003033646727e-05, "loss": 0.0055, "step": 24450 }, { "epoch": 257.4736842105263, "grad_norm": 0.06345563381910324, "learning_rate": 9.00001117541935e-05, "loss": 0.0046, "step": 24460 }, { "epoch": 257.57894736842104, "grad_norm": 0.08516007661819458, "learning_rate": 8.999018879756764e-05, "loss": 0.0042, "step": 24470 }, { "epoch": 257.6842105263158, "grad_norm": 0.07549402117729187, "learning_rate": 8.998026146767487e-05, "loss": 0.0052, "step": 24480 }, { "epoch": 257.7894736842105, "grad_norm": 0.07157187163829803, "learning_rate": 8.99703297656008e-05, "loss": 0.0043, "step": 24490 }, { "epoch": 257.89473684210526, "grad_norm": 0.08301429450511932, "learning_rate": 8.996039369243156e-05, "loss": 0.0051, "step": 24500 }, { "epoch": 258.0, "grad_norm": 0.12995539605617523, "learning_rate": 8.995045324925378e-05, "loss": 0.0046, "step": 24510 }, { "epoch": 258.10526315789474, "grad_norm": 0.09372366964817047, "learning_rate": 8.994050843715448e-05, "loss": 0.0053, "step": 24520 }, { "epoch": 258.2105263157895, "grad_norm": 0.11136803776025772, "learning_rate": 8.993055925722121e-05, "loss": 0.0057, "step": 24530 }, { "epoch": 258.3157894736842, "grad_norm": 0.08563277125358582, "learning_rate": 8.992060571054202e-05, "loss": 0.0047, "step": 24540 }, { "epoch": 258.42105263157896, "grad_norm": 0.0716104805469513, "learning_rate": 8.991064779820542e-05, "loss": 0.0046, "step": 24550 }, { "epoch": 258.5263157894737, "grad_norm": 0.08708444982767105, "learning_rate": 8.990068552130036e-05, "loss": 0.0042, "step": 24560 }, { "epoch": 258.63157894736844, "grad_norm": 0.09520867466926575, "learning_rate": 8.989071888091634e-05, "loss": 0.0046, "step": 24570 }, { "epoch": 258.7368421052632, "grad_norm": 0.08152486383914948, "learning_rate": 8.988074787814329e-05, "loss": 0.0047, "step": 24580 }, { "epoch": 258.8421052631579, "grad_norm": 0.09470054507255554, "learning_rate": 8.987077251407158e-05, "loss": 0.0047, "step": 24590 }, { "epoch": 258.94736842105266, "grad_norm": 0.11741986125707626, "learning_rate": 8.986079278979216e-05, "loss": 0.0041, "step": 24600 }, { "epoch": 259.05263157894734, "grad_norm": 0.09281745553016663, "learning_rate": 8.985080870639635e-05, "loss": 0.0042, "step": 24610 }, { "epoch": 259.1578947368421, "grad_norm": 0.06979858875274658, "learning_rate": 8.984082026497603e-05, "loss": 0.0052, "step": 24620 }, { "epoch": 259.2631578947368, "grad_norm": 0.08550279587507248, "learning_rate": 8.98308274666235e-05, "loss": 0.0047, "step": 24630 }, { "epoch": 259.36842105263156, "grad_norm": 0.09997180104255676, "learning_rate": 8.982083031243155e-05, "loss": 0.0055, "step": 24640 }, { "epoch": 259.4736842105263, "grad_norm": 0.0765906572341919, "learning_rate": 8.98108288034935e-05, "loss": 0.0042, "step": 24650 }, { "epoch": 259.57894736842104, "grad_norm": 0.07030662149190903, "learning_rate": 8.980082294090305e-05, "loss": 0.0046, "step": 24660 }, { "epoch": 259.6842105263158, "grad_norm": 0.07387790083885193, "learning_rate": 8.979081272575443e-05, "loss": 0.0043, "step": 24670 }, { "epoch": 259.7894736842105, "grad_norm": 0.09207263588905334, "learning_rate": 8.978079815914236e-05, "loss": 0.0042, "step": 24680 }, { "epoch": 259.89473684210526, "grad_norm": 0.09693892300128937, "learning_rate": 8.977077924216202e-05, "loss": 0.0048, "step": 24690 }, { "epoch": 260.0, "grad_norm": 0.12536367774009705, "learning_rate": 8.976075597590905e-05, "loss": 0.0055, "step": 24700 }, { "epoch": 260.10526315789474, "grad_norm": 0.08259487897157669, "learning_rate": 8.975072836147958e-05, "loss": 0.0044, "step": 24710 }, { "epoch": 260.2105263157895, "grad_norm": 0.1038365587592125, "learning_rate": 8.974069639997025e-05, "loss": 0.0046, "step": 24720 }, { "epoch": 260.3157894736842, "grad_norm": 0.08049435168504715, "learning_rate": 8.973066009247808e-05, "loss": 0.004, "step": 24730 }, { "epoch": 260.42105263157896, "grad_norm": 0.10303998738527298, "learning_rate": 8.972061944010066e-05, "loss": 0.0048, "step": 24740 }, { "epoch": 260.5263157894737, "grad_norm": 0.10094776749610901, "learning_rate": 8.971057444393603e-05, "loss": 0.0037, "step": 24750 }, { "epoch": 260.63157894736844, "grad_norm": 0.10389674454927444, "learning_rate": 8.970052510508268e-05, "loss": 0.0049, "step": 24760 }, { "epoch": 260.7368421052632, "grad_norm": 0.08888756483793259, "learning_rate": 8.969047142463959e-05, "loss": 0.0047, "step": 24770 }, { "epoch": 260.8421052631579, "grad_norm": 0.11569614708423615, "learning_rate": 8.968041340370621e-05, "loss": 0.0047, "step": 24780 }, { "epoch": 260.94736842105266, "grad_norm": 0.11383809894323349, "learning_rate": 8.96703510433825e-05, "loss": 0.0049, "step": 24790 }, { "epoch": 261.05263157894734, "grad_norm": 0.08520704507827759, "learning_rate": 8.966028434476883e-05, "loss": 0.0042, "step": 24800 }, { "epoch": 261.1578947368421, "grad_norm": 0.08884034305810928, "learning_rate": 8.96502133089661e-05, "loss": 0.0046, "step": 24810 }, { "epoch": 261.2631578947368, "grad_norm": 0.1185319647192955, "learning_rate": 8.964013793707564e-05, "loss": 0.0051, "step": 24820 }, { "epoch": 261.36842105263156, "grad_norm": 0.10633552819490433, "learning_rate": 8.963005823019932e-05, "loss": 0.0051, "step": 24830 }, { "epoch": 261.4736842105263, "grad_norm": 0.1153639554977417, "learning_rate": 8.961997418943939e-05, "loss": 0.0052, "step": 24840 }, { "epoch": 261.57894736842104, "grad_norm": 0.07176126539707184, "learning_rate": 8.960988581589865e-05, "loss": 0.0043, "step": 24850 }, { "epoch": 261.6842105263158, "grad_norm": 0.08820091933012009, "learning_rate": 8.959979311068037e-05, "loss": 0.0052, "step": 24860 }, { "epoch": 261.7894736842105, "grad_norm": 0.07814086228609085, "learning_rate": 8.958969607488823e-05, "loss": 0.0048, "step": 24870 }, { "epoch": 261.89473684210526, "grad_norm": 0.06424277275800705, "learning_rate": 8.957959470962647e-05, "loss": 0.0046, "step": 24880 }, { "epoch": 262.0, "grad_norm": 0.0996108129620552, "learning_rate": 8.956948901599971e-05, "loss": 0.0043, "step": 24890 }, { "epoch": 262.10526315789474, "grad_norm": 0.07947339117527008, "learning_rate": 8.955937899511315e-05, "loss": 0.0049, "step": 24900 }, { "epoch": 262.2105263157895, "grad_norm": 0.1045374944806099, "learning_rate": 8.954926464807238e-05, "loss": 0.0049, "step": 24910 }, { "epoch": 262.3157894736842, "grad_norm": 0.10287671536207199, "learning_rate": 8.953914597598347e-05, "loss": 0.0048, "step": 24920 }, { "epoch": 262.42105263157896, "grad_norm": 0.07902420312166214, "learning_rate": 8.952902297995303e-05, "loss": 0.005, "step": 24930 }, { "epoch": 262.5263157894737, "grad_norm": 0.08656543493270874, "learning_rate": 8.951889566108804e-05, "loss": 0.0046, "step": 24940 }, { "epoch": 262.63157894736844, "grad_norm": 0.10315205901861191, "learning_rate": 8.950876402049606e-05, "loss": 0.0043, "step": 24950 }, { "epoch": 262.7368421052632, "grad_norm": 0.08436403423547745, "learning_rate": 8.949862805928504e-05, "loss": 0.0038, "step": 24960 }, { "epoch": 262.8421052631579, "grad_norm": 0.07486289739608765, "learning_rate": 8.948848777856343e-05, "loss": 0.0046, "step": 24970 }, { "epoch": 262.94736842105266, "grad_norm": 0.10446907579898834, "learning_rate": 8.947834317944017e-05, "loss": 0.0047, "step": 24980 }, { "epoch": 263.05263157894734, "grad_norm": 0.0884656310081482, "learning_rate": 8.946819426302466e-05, "loss": 0.0042, "step": 24990 }, { "epoch": 263.1578947368421, "grad_norm": 0.089450903236866, "learning_rate": 8.945804103042676e-05, "loss": 0.0058, "step": 25000 }, { "epoch": 263.2631578947368, "grad_norm": 0.08472003787755966, "learning_rate": 8.944788348275681e-05, "loss": 0.0043, "step": 25010 }, { "epoch": 263.36842105263156, "grad_norm": 0.07965078204870224, "learning_rate": 8.943772162112565e-05, "loss": 0.0048, "step": 25020 }, { "epoch": 263.4736842105263, "grad_norm": 0.09654588997364044, "learning_rate": 8.942755544664454e-05, "loss": 0.0046, "step": 25030 }, { "epoch": 263.57894736842104, "grad_norm": 0.08525487035512924, "learning_rate": 8.941738496042525e-05, "loss": 0.0045, "step": 25040 }, { "epoch": 263.6842105263158, "grad_norm": 0.08361297845840454, "learning_rate": 8.940721016357999e-05, "loss": 0.005, "step": 25050 }, { "epoch": 263.7894736842105, "grad_norm": 0.09318157285451889, "learning_rate": 8.939703105722148e-05, "loss": 0.0049, "step": 25060 }, { "epoch": 263.89473684210526, "grad_norm": 0.10412553697824478, "learning_rate": 8.93868476424629e-05, "loss": 0.0044, "step": 25070 }, { "epoch": 264.0, "grad_norm": 0.13731805980205536, "learning_rate": 8.937665992041786e-05, "loss": 0.0055, "step": 25080 }, { "epoch": 264.10526315789474, "grad_norm": 0.10851281881332397, "learning_rate": 8.93664678922005e-05, "loss": 0.005, "step": 25090 }, { "epoch": 264.2105263157895, "grad_norm": 0.09195223450660706, "learning_rate": 8.93562715589254e-05, "loss": 0.0061, "step": 25100 }, { "epoch": 264.3157894736842, "grad_norm": 0.08452381938695908, "learning_rate": 8.934607092170762e-05, "loss": 0.0046, "step": 25110 }, { "epoch": 264.42105263157896, "grad_norm": 0.07412366569042206, "learning_rate": 8.933586598166266e-05, "loss": 0.0044, "step": 25120 }, { "epoch": 264.5263157894737, "grad_norm": 0.0846882089972496, "learning_rate": 8.932565673990655e-05, "loss": 0.0041, "step": 25130 }, { "epoch": 264.63157894736844, "grad_norm": 0.08040353655815125, "learning_rate": 8.931544319755574e-05, "loss": 0.0052, "step": 25140 }, { "epoch": 264.7368421052632, "grad_norm": 0.10802146047353745, "learning_rate": 8.930522535572718e-05, "loss": 0.0048, "step": 25150 }, { "epoch": 264.8421052631579, "grad_norm": 0.088700070977211, "learning_rate": 8.929500321553826e-05, "loss": 0.0038, "step": 25160 }, { "epoch": 264.94736842105266, "grad_norm": 0.08862722665071487, "learning_rate": 8.928477677810686e-05, "loss": 0.0042, "step": 25170 }, { "epoch": 265.05263157894734, "grad_norm": 0.10761675238609314, "learning_rate": 8.927454604455137e-05, "loss": 0.0043, "step": 25180 }, { "epoch": 265.1578947368421, "grad_norm": 0.13050219416618347, "learning_rate": 8.926431101599053e-05, "loss": 0.0045, "step": 25190 }, { "epoch": 265.2631578947368, "grad_norm": 0.09753599017858505, "learning_rate": 8.925407169354369e-05, "loss": 0.0044, "step": 25200 }, { "epoch": 265.36842105263156, "grad_norm": 0.09452186524868011, "learning_rate": 8.92438280783306e-05, "loss": 0.0048, "step": 25210 }, { "epoch": 265.4736842105263, "grad_norm": 0.09258569777011871, "learning_rate": 8.923358017147146e-05, "loss": 0.0043, "step": 25220 }, { "epoch": 265.57894736842104, "grad_norm": 0.09447649121284485, "learning_rate": 8.922332797408697e-05, "loss": 0.0045, "step": 25230 }, { "epoch": 265.6842105263158, "grad_norm": 0.08117511868476868, "learning_rate": 8.921307148729831e-05, "loss": 0.0042, "step": 25240 }, { "epoch": 265.7894736842105, "grad_norm": 0.11899609863758087, "learning_rate": 8.920281071222712e-05, "loss": 0.0055, "step": 25250 }, { "epoch": 265.89473684210526, "grad_norm": 0.0803011953830719, "learning_rate": 8.919254564999548e-05, "loss": 0.0045, "step": 25260 }, { "epoch": 266.0, "grad_norm": 0.09763781726360321, "learning_rate": 8.918227630172598e-05, "loss": 0.004, "step": 25270 }, { "epoch": 266.10526315789474, "grad_norm": 0.0917486697435379, "learning_rate": 8.917200266854165e-05, "loss": 0.004, "step": 25280 }, { "epoch": 266.2105263157895, "grad_norm": 0.08472423255443573, "learning_rate": 8.9161724751566e-05, "loss": 0.0039, "step": 25290 }, { "epoch": 266.3157894736842, "grad_norm": 0.0654957965016365, "learning_rate": 8.915144255192302e-05, "loss": 0.0049, "step": 25300 }, { "epoch": 266.42105263157896, "grad_norm": 0.08013477176427841, "learning_rate": 8.914115607073714e-05, "loss": 0.0049, "step": 25310 }, { "epoch": 266.5263157894737, "grad_norm": 0.057587336748838425, "learning_rate": 8.913086530913327e-05, "loss": 0.004, "step": 25320 }, { "epoch": 266.63157894736844, "grad_norm": 0.07976482808589935, "learning_rate": 8.912057026823681e-05, "loss": 0.0047, "step": 25330 }, { "epoch": 266.7368421052632, "grad_norm": 0.07261917740106583, "learning_rate": 8.91102709491736e-05, "loss": 0.0054, "step": 25340 }, { "epoch": 266.8421052631579, "grad_norm": 0.11327046900987625, "learning_rate": 8.909996735306996e-05, "loss": 0.0045, "step": 25350 }, { "epoch": 266.94736842105266, "grad_norm": 0.07468683272600174, "learning_rate": 8.908965948105268e-05, "loss": 0.0037, "step": 25360 }, { "epoch": 267.05263157894734, "grad_norm": 0.08391743153333664, "learning_rate": 8.907934733424901e-05, "loss": 0.0044, "step": 25370 }, { "epoch": 267.1578947368421, "grad_norm": 0.07702256739139557, "learning_rate": 8.906903091378666e-05, "loss": 0.0043, "step": 25380 }, { "epoch": 267.2631578947368, "grad_norm": 0.1088130846619606, "learning_rate": 8.905871022079384e-05, "loss": 0.0054, "step": 25390 }, { "epoch": 267.36842105263156, "grad_norm": 0.07949640601873398, "learning_rate": 8.90483852563992e-05, "loss": 0.0042, "step": 25400 }, { "epoch": 267.4736842105263, "grad_norm": 0.09183190017938614, "learning_rate": 8.903805602173185e-05, "loss": 0.004, "step": 25410 }, { "epoch": 267.57894736842104, "grad_norm": 0.10390832275152206, "learning_rate": 8.902772251792137e-05, "loss": 0.0047, "step": 25420 }, { "epoch": 267.6842105263158, "grad_norm": 0.08177808672189713, "learning_rate": 8.901738474609786e-05, "loss": 0.0045, "step": 25430 }, { "epoch": 267.7894736842105, "grad_norm": 0.07041925936937332, "learning_rate": 8.900704270739179e-05, "loss": 0.0054, "step": 25440 }, { "epoch": 267.89473684210526, "grad_norm": 0.07030609250068665, "learning_rate": 8.89966964029342e-05, "loss": 0.0041, "step": 25450 }, { "epoch": 268.0, "grad_norm": 0.12133018672466278, "learning_rate": 8.898634583385652e-05, "loss": 0.0046, "step": 25460 }, { "epoch": 268.10526315789474, "grad_norm": 0.084256611764431, "learning_rate": 8.897599100129065e-05, "loss": 0.0042, "step": 25470 }, { "epoch": 268.2105263157895, "grad_norm": 0.06011296436190605, "learning_rate": 8.896563190636903e-05, "loss": 0.004, "step": 25480 }, { "epoch": 268.3157894736842, "grad_norm": 0.09081108868122101, "learning_rate": 8.895526855022448e-05, "loss": 0.0046, "step": 25490 }, { "epoch": 268.42105263157896, "grad_norm": 0.07770953327417374, "learning_rate": 8.894490093399033e-05, "loss": 0.0045, "step": 25500 }, { "epoch": 268.5263157894737, "grad_norm": 0.08250999450683594, "learning_rate": 8.893452905880035e-05, "loss": 0.0041, "step": 25510 }, { "epoch": 268.63157894736844, "grad_norm": 0.07044582813978195, "learning_rate": 8.892415292578883e-05, "loss": 0.0048, "step": 25520 }, { "epoch": 268.7368421052632, "grad_norm": 0.09372224658727646, "learning_rate": 8.891377253609046e-05, "loss": 0.0045, "step": 25530 }, { "epoch": 268.8421052631579, "grad_norm": 0.07036573439836502, "learning_rate": 8.890338789084043e-05, "loss": 0.0038, "step": 25540 }, { "epoch": 268.94736842105266, "grad_norm": 0.0875639021396637, "learning_rate": 8.88929989911744e-05, "loss": 0.0038, "step": 25550 }, { "epoch": 269.05263157894734, "grad_norm": 0.09813080728054047, "learning_rate": 8.888260583822847e-05, "loss": 0.0047, "step": 25560 }, { "epoch": 269.1578947368421, "grad_norm": 0.08984657377004623, "learning_rate": 8.887220843313921e-05, "loss": 0.0042, "step": 25570 }, { "epoch": 269.2631578947368, "grad_norm": 0.09365209192037582, "learning_rate": 8.88618067770437e-05, "loss": 0.0044, "step": 25580 }, { "epoch": 269.36842105263156, "grad_norm": 0.1171443983912468, "learning_rate": 8.885140087107942e-05, "loss": 0.0046, "step": 25590 }, { "epoch": 269.4736842105263, "grad_norm": 0.10217786580324173, "learning_rate": 8.884099071638436e-05, "loss": 0.0045, "step": 25600 }, { "epoch": 269.57894736842104, "grad_norm": 0.06822846829891205, "learning_rate": 8.883057631409695e-05, "loss": 0.005, "step": 25610 }, { "epoch": 269.6842105263158, "grad_norm": 0.11434365063905716, "learning_rate": 8.882015766535608e-05, "loss": 0.005, "step": 25620 }, { "epoch": 269.7894736842105, "grad_norm": 0.09738904982805252, "learning_rate": 8.880973477130115e-05, "loss": 0.0051, "step": 25630 }, { "epoch": 269.89473684210526, "grad_norm": 0.08646804839372635, "learning_rate": 8.879930763307197e-05, "loss": 0.0054, "step": 25640 }, { "epoch": 270.0, "grad_norm": 0.11940888315439224, "learning_rate": 8.878887625180884e-05, "loss": 0.0051, "step": 25650 }, { "epoch": 270.10526315789474, "grad_norm": 0.08183484524488449, "learning_rate": 8.877844062865253e-05, "loss": 0.0055, "step": 25660 }, { "epoch": 270.2105263157895, "grad_norm": 0.072188600897789, "learning_rate": 8.876800076474424e-05, "loss": 0.0047, "step": 25670 }, { "epoch": 270.3157894736842, "grad_norm": 0.10225999355316162, "learning_rate": 8.875755666122568e-05, "loss": 0.0049, "step": 25680 }, { "epoch": 270.42105263157896, "grad_norm": 0.06360206753015518, "learning_rate": 8.8747108319239e-05, "loss": 0.0049, "step": 25690 }, { "epoch": 270.5263157894737, "grad_norm": 0.09850934147834778, "learning_rate": 8.87366557399268e-05, "loss": 0.0056, "step": 25700 }, { "epoch": 270.63157894736844, "grad_norm": 0.0888322964310646, "learning_rate": 8.872619892443217e-05, "loss": 0.0053, "step": 25710 }, { "epoch": 270.7368421052632, "grad_norm": 0.06876365095376968, "learning_rate": 8.871573787389865e-05, "loss": 0.0049, "step": 25720 }, { "epoch": 270.8421052631579, "grad_norm": 0.07915181666612625, "learning_rate": 8.870527258947024e-05, "loss": 0.0049, "step": 25730 }, { "epoch": 270.94736842105266, "grad_norm": 0.0934557244181633, "learning_rate": 8.869480307229143e-05, "loss": 0.0045, "step": 25740 }, { "epoch": 271.05263157894734, "grad_norm": 0.16328905522823334, "learning_rate": 8.868432932350712e-05, "loss": 0.0095, "step": 25750 }, { "epoch": 271.1578947368421, "grad_norm": 0.11229097843170166, "learning_rate": 8.867385134426272e-05, "loss": 0.0077, "step": 25760 }, { "epoch": 271.2631578947368, "grad_norm": 0.07550697028636932, "learning_rate": 8.866336913570407e-05, "loss": 0.0059, "step": 25770 }, { "epoch": 271.36842105263156, "grad_norm": 0.07357575744390488, "learning_rate": 8.865288269897751e-05, "loss": 0.0057, "step": 25780 }, { "epoch": 271.4736842105263, "grad_norm": 0.08947348594665527, "learning_rate": 8.864239203522981e-05, "loss": 0.0048, "step": 25790 }, { "epoch": 271.57894736842104, "grad_norm": 0.10659067332744598, "learning_rate": 8.863189714560822e-05, "loss": 0.0053, "step": 25800 }, { "epoch": 271.6842105263158, "grad_norm": 0.09006350487470627, "learning_rate": 8.862139803126043e-05, "loss": 0.0051, "step": 25810 }, { "epoch": 271.7894736842105, "grad_norm": 0.08603290468454361, "learning_rate": 8.861089469333463e-05, "loss": 0.0059, "step": 25820 }, { "epoch": 271.89473684210526, "grad_norm": 0.12399038672447205, "learning_rate": 8.860038713297944e-05, "loss": 0.0052, "step": 25830 }, { "epoch": 272.0, "grad_norm": 0.09660585224628448, "learning_rate": 8.858987535134394e-05, "loss": 0.005, "step": 25840 }, { "epoch": 272.10526315789474, "grad_norm": 0.08836778998374939, "learning_rate": 8.857935934957769e-05, "loss": 0.0061, "step": 25850 }, { "epoch": 272.2105263157895, "grad_norm": 0.0962403193116188, "learning_rate": 8.856883912883071e-05, "loss": 0.0054, "step": 25860 }, { "epoch": 272.3157894736842, "grad_norm": 0.0965072438120842, "learning_rate": 8.855831469025346e-05, "loss": 0.0046, "step": 25870 }, { "epoch": 272.42105263157896, "grad_norm": 0.07456550002098083, "learning_rate": 8.854778603499689e-05, "loss": 0.0042, "step": 25880 }, { "epoch": 272.5263157894737, "grad_norm": 0.07623786479234695, "learning_rate": 8.85372531642124e-05, "loss": 0.0051, "step": 25890 }, { "epoch": 272.63157894736844, "grad_norm": 0.09892027080059052, "learning_rate": 8.852671607905185e-05, "loss": 0.0045, "step": 25900 }, { "epoch": 272.7368421052632, "grad_norm": 0.0691121369600296, "learning_rate": 8.851617478066754e-05, "loss": 0.004, "step": 25910 }, { "epoch": 272.8421052631579, "grad_norm": 0.07194157689809799, "learning_rate": 8.850562927021227e-05, "loss": 0.0039, "step": 25920 }, { "epoch": 272.94736842105266, "grad_norm": 0.08033649623394012, "learning_rate": 8.849507954883928e-05, "loss": 0.0045, "step": 25930 }, { "epoch": 273.05263157894734, "grad_norm": 0.09966575354337692, "learning_rate": 8.848452561770226e-05, "loss": 0.0051, "step": 25940 }, { "epoch": 273.1578947368421, "grad_norm": 0.05469442903995514, "learning_rate": 8.847396747795538e-05, "loss": 0.0043, "step": 25950 }, { "epoch": 273.2631578947368, "grad_norm": 0.10359969735145569, "learning_rate": 8.846340513075327e-05, "loss": 0.0055, "step": 25960 }, { "epoch": 273.36842105263156, "grad_norm": 0.0865727961063385, "learning_rate": 8.845283857725099e-05, "loss": 0.0057, "step": 25970 }, { "epoch": 273.4736842105263, "grad_norm": 0.09739156812429428, "learning_rate": 8.844226781860409e-05, "loss": 0.0042, "step": 25980 }, { "epoch": 273.57894736842104, "grad_norm": 0.07304888218641281, "learning_rate": 8.84316928559686e-05, "loss": 0.0047, "step": 25990 }, { "epoch": 273.6842105263158, "grad_norm": 0.08683782070875168, "learning_rate": 8.842111369050094e-05, "loss": 0.0046, "step": 26000 }, { "epoch": 273.7894736842105, "grad_norm": 0.09541759639978409, "learning_rate": 8.841053032335808e-05, "loss": 0.0043, "step": 26010 }, { "epoch": 273.89473684210526, "grad_norm": 0.10679153352975845, "learning_rate": 8.839994275569735e-05, "loss": 0.0052, "step": 26020 }, { "epoch": 274.0, "grad_norm": 0.08639375865459442, "learning_rate": 8.838935098867662e-05, "loss": 0.0045, "step": 26030 }, { "epoch": 274.10526315789474, "grad_norm": 0.10632407665252686, "learning_rate": 8.837875502345418e-05, "loss": 0.0048, "step": 26040 }, { "epoch": 274.2105263157895, "grad_norm": 0.08379576355218887, "learning_rate": 8.83681548611888e-05, "loss": 0.0043, "step": 26050 }, { "epoch": 274.3157894736842, "grad_norm": 0.07894749939441681, "learning_rate": 8.835755050303969e-05, "loss": 0.0049, "step": 26060 }, { "epoch": 274.42105263157896, "grad_norm": 0.1368376612663269, "learning_rate": 8.834694195016653e-05, "loss": 0.0043, "step": 26070 }, { "epoch": 274.5263157894737, "grad_norm": 0.08631683140993118, "learning_rate": 8.833632920372942e-05, "loss": 0.0046, "step": 26080 }, { "epoch": 274.63157894736844, "grad_norm": 0.08674493432044983, "learning_rate": 8.832571226488903e-05, "loss": 0.0048, "step": 26090 }, { "epoch": 274.7368421052632, "grad_norm": 0.10250566899776459, "learning_rate": 8.831509113480634e-05, "loss": 0.0038, "step": 26100 }, { "epoch": 274.8421052631579, "grad_norm": 0.08782687038183212, "learning_rate": 8.83044658146429e-05, "loss": 0.0047, "step": 26110 }, { "epoch": 274.94736842105266, "grad_norm": 0.0806032344698906, "learning_rate": 8.829383630556067e-05, "loss": 0.0046, "step": 26120 }, { "epoch": 275.05263157894734, "grad_norm": 0.08671997487545013, "learning_rate": 8.828320260872207e-05, "loss": 0.0047, "step": 26130 }, { "epoch": 275.1578947368421, "grad_norm": 0.0872565433382988, "learning_rate": 8.827256472529e-05, "loss": 0.0044, "step": 26140 }, { "epoch": 275.2631578947368, "grad_norm": 0.0698537677526474, "learning_rate": 8.826192265642778e-05, "loss": 0.0049, "step": 26150 }, { "epoch": 275.36842105263156, "grad_norm": 0.09929145872592926, "learning_rate": 8.825127640329923e-05, "loss": 0.0048, "step": 26160 }, { "epoch": 275.4736842105263, "grad_norm": 0.0897182896733284, "learning_rate": 8.824062596706861e-05, "loss": 0.005, "step": 26170 }, { "epoch": 275.57894736842104, "grad_norm": 0.08917362987995148, "learning_rate": 8.822997134890062e-05, "loss": 0.004, "step": 26180 }, { "epoch": 275.6842105263158, "grad_norm": 0.10021095722913742, "learning_rate": 8.821931254996044e-05, "loss": 0.0046, "step": 26190 }, { "epoch": 275.7894736842105, "grad_norm": 0.0954337939620018, "learning_rate": 8.82086495714137e-05, "loss": 0.004, "step": 26200 }, { "epoch": 275.89473684210526, "grad_norm": 0.12662161886692047, "learning_rate": 8.81979824144265e-05, "loss": 0.0044, "step": 26210 }, { "epoch": 276.0, "grad_norm": 0.12692326307296753, "learning_rate": 8.818731108016536e-05, "loss": 0.0048, "step": 26220 }, { "epoch": 276.10526315789474, "grad_norm": 0.11836501210927963, "learning_rate": 8.81766355697973e-05, "loss": 0.0042, "step": 26230 }, { "epoch": 276.2105263157895, "grad_norm": 0.08731980621814728, "learning_rate": 8.816595588448977e-05, "loss": 0.0045, "step": 26240 }, { "epoch": 276.3157894736842, "grad_norm": 0.08004482090473175, "learning_rate": 8.81552720254107e-05, "loss": 0.0049, "step": 26250 }, { "epoch": 276.42105263157896, "grad_norm": 0.07911328971385956, "learning_rate": 8.814458399372842e-05, "loss": 0.0041, "step": 26260 }, { "epoch": 276.5263157894737, "grad_norm": 0.07566271722316742, "learning_rate": 8.813389179061181e-05, "loss": 0.0045, "step": 26270 }, { "epoch": 276.63157894736844, "grad_norm": 0.09118282794952393, "learning_rate": 8.812319541723012e-05, "loss": 0.0044, "step": 26280 }, { "epoch": 276.7368421052632, "grad_norm": 0.07796689122915268, "learning_rate": 8.811249487475309e-05, "loss": 0.0042, "step": 26290 }, { "epoch": 276.8421052631579, "grad_norm": 0.08933532238006592, "learning_rate": 8.810179016435092e-05, "loss": 0.0048, "step": 26300 }, { "epoch": 276.94736842105266, "grad_norm": 0.09715192764997482, "learning_rate": 8.809108128719428e-05, "loss": 0.0044, "step": 26310 }, { "epoch": 277.05263157894734, "grad_norm": 0.09559987485408783, "learning_rate": 8.808036824445424e-05, "loss": 0.0039, "step": 26320 }, { "epoch": 277.1578947368421, "grad_norm": 0.07391947507858276, "learning_rate": 8.806965103730238e-05, "loss": 0.0043, "step": 26330 }, { "epoch": 277.2631578947368, "grad_norm": 0.09026072919368744, "learning_rate": 8.805892966691074e-05, "loss": 0.0047, "step": 26340 }, { "epoch": 277.36842105263156, "grad_norm": 0.07462135702371597, "learning_rate": 8.804820413445175e-05, "loss": 0.0039, "step": 26350 }, { "epoch": 277.4736842105263, "grad_norm": 0.07660629600286484, "learning_rate": 8.803747444109837e-05, "loss": 0.0036, "step": 26360 }, { "epoch": 277.57894736842104, "grad_norm": 0.0855351909995079, "learning_rate": 8.802674058802399e-05, "loss": 0.0044, "step": 26370 }, { "epoch": 277.6842105263158, "grad_norm": 0.08825065940618515, "learning_rate": 8.801600257640241e-05, "loss": 0.0055, "step": 26380 }, { "epoch": 277.7894736842105, "grad_norm": 0.07274661213159561, "learning_rate": 8.800526040740795e-05, "loss": 0.0046, "step": 26390 }, { "epoch": 277.89473684210526, "grad_norm": 0.0719640702009201, "learning_rate": 8.799451408221535e-05, "loss": 0.0042, "step": 26400 }, { "epoch": 278.0, "grad_norm": 0.09815608710050583, "learning_rate": 8.798376360199982e-05, "loss": 0.0039, "step": 26410 }, { "epoch": 278.10526315789474, "grad_norm": 0.09522363543510437, "learning_rate": 8.797300896793701e-05, "loss": 0.0051, "step": 26420 }, { "epoch": 278.2105263157895, "grad_norm": 0.07575825601816177, "learning_rate": 8.796225018120302e-05, "loss": 0.0042, "step": 26430 }, { "epoch": 278.3157894736842, "grad_norm": 0.055704668164253235, "learning_rate": 8.795148724297444e-05, "loss": 0.0046, "step": 26440 }, { "epoch": 278.42105263157896, "grad_norm": 0.07340405881404877, "learning_rate": 8.794072015442825e-05, "loss": 0.0057, "step": 26450 }, { "epoch": 278.5263157894737, "grad_norm": 0.07099559903144836, "learning_rate": 8.792994891674198e-05, "loss": 0.0039, "step": 26460 }, { "epoch": 278.63157894736844, "grad_norm": 0.06107150390744209, "learning_rate": 8.79191735310935e-05, "loss": 0.0039, "step": 26470 }, { "epoch": 278.7368421052632, "grad_norm": 0.0732043981552124, "learning_rate": 8.790839399866122e-05, "loss": 0.0045, "step": 26480 }, { "epoch": 278.8421052631579, "grad_norm": 0.07960079610347748, "learning_rate": 8.789761032062397e-05, "loss": 0.0039, "step": 26490 }, { "epoch": 278.94736842105266, "grad_norm": 0.07023688405752182, "learning_rate": 8.788682249816103e-05, "loss": 0.0055, "step": 26500 }, { "epoch": 279.05263157894734, "grad_norm": 0.09830612689256668, "learning_rate": 8.787603053245215e-05, "loss": 0.0043, "step": 26510 }, { "epoch": 279.1578947368421, "grad_norm": 0.12049025297164917, "learning_rate": 8.78652344246775e-05, "loss": 0.0046, "step": 26520 }, { "epoch": 279.2631578947368, "grad_norm": 0.11513478308916092, "learning_rate": 8.785443417601776e-05, "loss": 0.0047, "step": 26530 }, { "epoch": 279.36842105263156, "grad_norm": 0.10652736574411392, "learning_rate": 8.784362978765401e-05, "loss": 0.0048, "step": 26540 }, { "epoch": 279.4736842105263, "grad_norm": 0.1116853579878807, "learning_rate": 8.783282126076779e-05, "loss": 0.0049, "step": 26550 }, { "epoch": 279.57894736842104, "grad_norm": 0.08828974515199661, "learning_rate": 8.782200859654112e-05, "loss": 0.0043, "step": 26560 }, { "epoch": 279.6842105263158, "grad_norm": 0.06569945067167282, "learning_rate": 8.781119179615646e-05, "loss": 0.0039, "step": 26570 }, { "epoch": 279.7894736842105, "grad_norm": 0.08549107611179352, "learning_rate": 8.780037086079674e-05, "loss": 0.0047, "step": 26580 }, { "epoch": 279.89473684210526, "grad_norm": 0.07186080515384674, "learning_rate": 8.778954579164527e-05, "loss": 0.0044, "step": 26590 }, { "epoch": 280.0, "grad_norm": 0.1064726933836937, "learning_rate": 8.777871658988588e-05, "loss": 0.0045, "step": 26600 }, { "epoch": 280.10526315789474, "grad_norm": 0.08620679378509521, "learning_rate": 8.776788325670285e-05, "loss": 0.0046, "step": 26610 }, { "epoch": 280.2105263157895, "grad_norm": 0.08766306191682816, "learning_rate": 8.775704579328089e-05, "loss": 0.0043, "step": 26620 }, { "epoch": 280.3157894736842, "grad_norm": 0.057045456022024155, "learning_rate": 8.774620420080517e-05, "loss": 0.0043, "step": 26630 }, { "epoch": 280.42105263157896, "grad_norm": 0.06701673567295074, "learning_rate": 8.773535848046131e-05, "loss": 0.004, "step": 26640 }, { "epoch": 280.5263157894737, "grad_norm": 0.08726362884044647, "learning_rate": 8.772450863343538e-05, "loss": 0.0044, "step": 26650 }, { "epoch": 280.63157894736844, "grad_norm": 0.07062524557113647, "learning_rate": 8.77136546609139e-05, "loss": 0.0043, "step": 26660 }, { "epoch": 280.7368421052632, "grad_norm": 0.07774404436349869, "learning_rate": 8.770279656408385e-05, "loss": 0.0042, "step": 26670 }, { "epoch": 280.8421052631579, "grad_norm": 0.09628193080425262, "learning_rate": 8.769193434413265e-05, "loss": 0.005, "step": 26680 }, { "epoch": 280.94736842105266, "grad_norm": 0.08007988333702087, "learning_rate": 8.76810680022482e-05, "loss": 0.0036, "step": 26690 }, { "epoch": 281.05263157894734, "grad_norm": 0.08404815196990967, "learning_rate": 8.767019753961878e-05, "loss": 0.0043, "step": 26700 }, { "epoch": 281.1578947368421, "grad_norm": 0.08965503424406052, "learning_rate": 8.765932295743321e-05, "loss": 0.0049, "step": 26710 }, { "epoch": 281.2631578947368, "grad_norm": 0.1022983193397522, "learning_rate": 8.764844425688068e-05, "loss": 0.0052, "step": 26720 }, { "epoch": 281.36842105263156, "grad_norm": 0.08461917191743851, "learning_rate": 8.763756143915092e-05, "loss": 0.0048, "step": 26730 }, { "epoch": 281.4736842105263, "grad_norm": 0.11841841042041779, "learning_rate": 8.7626674505434e-05, "loss": 0.0051, "step": 26740 }, { "epoch": 281.57894736842104, "grad_norm": 0.07353349030017853, "learning_rate": 8.761578345692053e-05, "loss": 0.004, "step": 26750 }, { "epoch": 281.6842105263158, "grad_norm": 0.0725114718079567, "learning_rate": 8.760488829480156e-05, "loss": 0.0043, "step": 26760 }, { "epoch": 281.7894736842105, "grad_norm": 0.09017856419086456, "learning_rate": 8.759398902026854e-05, "loss": 0.0034, "step": 26770 }, { "epoch": 281.89473684210526, "grad_norm": 0.0784936398267746, "learning_rate": 8.758308563451339e-05, "loss": 0.0045, "step": 26780 }, { "epoch": 282.0, "grad_norm": 0.1093662902712822, "learning_rate": 8.75721781387285e-05, "loss": 0.0043, "step": 26790 }, { "epoch": 282.10526315789474, "grad_norm": 0.06947697699069977, "learning_rate": 8.75612665341067e-05, "loss": 0.005, "step": 26800 }, { "epoch": 282.2105263157895, "grad_norm": 0.05782761052250862, "learning_rate": 8.755035082184126e-05, "loss": 0.004, "step": 26810 }, { "epoch": 282.3157894736842, "grad_norm": 0.10339298844337463, "learning_rate": 8.753943100312592e-05, "loss": 0.0052, "step": 26820 }, { "epoch": 282.42105263157896, "grad_norm": 0.09345429390668869, "learning_rate": 8.752850707915484e-05, "loss": 0.005, "step": 26830 }, { "epoch": 282.5263157894737, "grad_norm": 0.07406797260046005, "learning_rate": 8.751757905112264e-05, "loss": 0.0054, "step": 26840 }, { "epoch": 282.63157894736844, "grad_norm": 0.06256195902824402, "learning_rate": 8.75066469202244e-05, "loss": 0.004, "step": 26850 }, { "epoch": 282.7368421052632, "grad_norm": 0.06593518704175949, "learning_rate": 8.749571068765567e-05, "loss": 0.0041, "step": 26860 }, { "epoch": 282.8421052631579, "grad_norm": 0.07157813012599945, "learning_rate": 8.748477035461238e-05, "loss": 0.0054, "step": 26870 }, { "epoch": 282.94736842105266, "grad_norm": 0.08392640948295593, "learning_rate": 8.747382592229095e-05, "loss": 0.0045, "step": 26880 }, { "epoch": 283.05263157894734, "grad_norm": 0.07625623792409897, "learning_rate": 8.746287739188828e-05, "loss": 0.0047, "step": 26890 }, { "epoch": 283.1578947368421, "grad_norm": 0.09244099259376526, "learning_rate": 8.745192476460165e-05, "loss": 0.0044, "step": 26900 }, { "epoch": 283.2631578947368, "grad_norm": 0.10722776502370834, "learning_rate": 8.744096804162882e-05, "loss": 0.0041, "step": 26910 }, { "epoch": 283.36842105263156, "grad_norm": 0.09695279598236084, "learning_rate": 8.743000722416804e-05, "loss": 0.0047, "step": 26920 }, { "epoch": 283.4736842105263, "grad_norm": 0.0954890251159668, "learning_rate": 8.741904231341793e-05, "loss": 0.0056, "step": 26930 }, { "epoch": 283.57894736842104, "grad_norm": 0.11605352908372879, "learning_rate": 8.740807331057762e-05, "loss": 0.0046, "step": 26940 }, { "epoch": 283.6842105263158, "grad_norm": 0.08958913385868073, "learning_rate": 8.739710021684667e-05, "loss": 0.004, "step": 26950 }, { "epoch": 283.7894736842105, "grad_norm": 0.059819020330905914, "learning_rate": 8.738612303342503e-05, "loss": 0.0051, "step": 26960 }, { "epoch": 283.89473684210526, "grad_norm": 0.10945433378219604, "learning_rate": 8.73751417615132e-05, "loss": 0.005, "step": 26970 }, { "epoch": 284.0, "grad_norm": 0.08923549950122833, "learning_rate": 8.736415640231208e-05, "loss": 0.0041, "step": 26980 }, { "epoch": 284.10526315789474, "grad_norm": 0.08741434663534164, "learning_rate": 8.735316695702297e-05, "loss": 0.0042, "step": 26990 }, { "epoch": 284.2105263157895, "grad_norm": 0.07627223432064056, "learning_rate": 8.734217342684769e-05, "loss": 0.004, "step": 27000 }, { "epoch": 284.3157894736842, "grad_norm": 0.09729839116334915, "learning_rate": 8.733117581298847e-05, "loss": 0.0047, "step": 27010 }, { "epoch": 284.42105263157896, "grad_norm": 0.10090164095163345, "learning_rate": 8.732017411664796e-05, "loss": 0.0039, "step": 27020 }, { "epoch": 284.5263157894737, "grad_norm": 0.09336467832326889, "learning_rate": 8.730916833902936e-05, "loss": 0.0046, "step": 27030 }, { "epoch": 284.63157894736844, "grad_norm": 0.09075547754764557, "learning_rate": 8.729815848133618e-05, "loss": 0.0051, "step": 27040 }, { "epoch": 284.7368421052632, "grad_norm": 0.10360216349363327, "learning_rate": 8.728714454477247e-05, "loss": 0.0047, "step": 27050 }, { "epoch": 284.8421052631579, "grad_norm": 0.10794240981340408, "learning_rate": 8.727612653054269e-05, "loss": 0.005, "step": 27060 }, { "epoch": 284.94736842105266, "grad_norm": 0.100019171833992, "learning_rate": 8.726510443985176e-05, "loss": 0.0039, "step": 27070 }, { "epoch": 285.05263157894734, "grad_norm": 0.08680732548236847, "learning_rate": 8.725407827390503e-05, "loss": 0.0046, "step": 27080 }, { "epoch": 285.1578947368421, "grad_norm": 0.07391175627708435, "learning_rate": 8.724304803390833e-05, "loss": 0.0039, "step": 27090 }, { "epoch": 285.2631578947368, "grad_norm": 0.08868984133005142, "learning_rate": 8.723201372106788e-05, "loss": 0.0049, "step": 27100 }, { "epoch": 285.36842105263156, "grad_norm": 0.0903761237859726, "learning_rate": 8.722097533659038e-05, "loss": 0.0052, "step": 27110 }, { "epoch": 285.4736842105263, "grad_norm": 0.07684484124183655, "learning_rate": 8.720993288168299e-05, "loss": 0.0042, "step": 27120 }, { "epoch": 285.57894736842104, "grad_norm": 0.07576607167720795, "learning_rate": 8.719888635755327e-05, "loss": 0.0037, "step": 27130 }, { "epoch": 285.6842105263158, "grad_norm": 0.062362246215343475, "learning_rate": 8.718783576540928e-05, "loss": 0.0047, "step": 27140 }, { "epoch": 285.7894736842105, "grad_norm": 0.07598918676376343, "learning_rate": 8.717678110645948e-05, "loss": 0.0042, "step": 27150 }, { "epoch": 285.89473684210526, "grad_norm": 0.0920020341873169, "learning_rate": 8.716572238191279e-05, "loss": 0.0037, "step": 27160 }, { "epoch": 286.0, "grad_norm": 0.08195938169956207, "learning_rate": 8.715465959297857e-05, "loss": 0.0041, "step": 27170 }, { "epoch": 286.10526315789474, "grad_norm": 0.09584591537714005, "learning_rate": 8.714359274086665e-05, "loss": 0.0048, "step": 27180 }, { "epoch": 286.2105263157895, "grad_norm": 0.06521659344434738, "learning_rate": 8.713252182678726e-05, "loss": 0.0049, "step": 27190 }, { "epoch": 286.3157894736842, "grad_norm": 0.07020696252584457, "learning_rate": 8.712144685195112e-05, "loss": 0.0045, "step": 27200 }, { "epoch": 286.42105263157896, "grad_norm": 0.07095292210578918, "learning_rate": 8.711036781756936e-05, "loss": 0.0048, "step": 27210 }, { "epoch": 286.5263157894737, "grad_norm": 0.07323427498340607, "learning_rate": 8.709928472485357e-05, "loss": 0.0041, "step": 27220 }, { "epoch": 286.63157894736844, "grad_norm": 0.06345857679843903, "learning_rate": 8.708819757501579e-05, "loss": 0.005, "step": 27230 }, { "epoch": 286.7368421052632, "grad_norm": 0.09070886671543121, "learning_rate": 8.707710636926846e-05, "loss": 0.0035, "step": 27240 }, { "epoch": 286.8421052631579, "grad_norm": 0.08069952577352524, "learning_rate": 8.706601110882455e-05, "loss": 0.0045, "step": 27250 }, { "epoch": 286.94736842105266, "grad_norm": 0.0843268409371376, "learning_rate": 8.705491179489738e-05, "loss": 0.0042, "step": 27260 }, { "epoch": 287.05263157894734, "grad_norm": 0.06710461527109146, "learning_rate": 8.704380842870077e-05, "loss": 0.0041, "step": 27270 }, { "epoch": 287.1578947368421, "grad_norm": 0.07940889149904251, "learning_rate": 8.703270101144895e-05, "loss": 0.0042, "step": 27280 }, { "epoch": 287.2631578947368, "grad_norm": 0.07855972647666931, "learning_rate": 8.702158954435664e-05, "loss": 0.0045, "step": 27290 }, { "epoch": 287.36842105263156, "grad_norm": 0.10081668198108673, "learning_rate": 8.701047402863896e-05, "loss": 0.0043, "step": 27300 }, { "epoch": 287.4736842105263, "grad_norm": 0.08883201330900192, "learning_rate": 8.699935446551148e-05, "loss": 0.0046, "step": 27310 }, { "epoch": 287.57894736842104, "grad_norm": 0.07098229229450226, "learning_rate": 8.698823085619022e-05, "loss": 0.0041, "step": 27320 }, { "epoch": 287.6842105263158, "grad_norm": 0.08143884688615799, "learning_rate": 8.697710320189166e-05, "loss": 0.0047, "step": 27330 }, { "epoch": 287.7894736842105, "grad_norm": 0.09120789170265198, "learning_rate": 8.696597150383268e-05, "loss": 0.005, "step": 27340 }, { "epoch": 287.89473684210526, "grad_norm": 0.08324708044528961, "learning_rate": 8.695483576323063e-05, "loss": 0.004, "step": 27350 }, { "epoch": 288.0, "grad_norm": 0.09309989959001541, "learning_rate": 8.69436959813033e-05, "loss": 0.0041, "step": 27360 }, { "epoch": 288.10526315789474, "grad_norm": 0.08922567963600159, "learning_rate": 8.693255215926892e-05, "loss": 0.004, "step": 27370 }, { "epoch": 288.2105263157895, "grad_norm": 0.08090976625680923, "learning_rate": 8.692140429834617e-05, "loss": 0.0038, "step": 27380 }, { "epoch": 288.3157894736842, "grad_norm": 0.0700543001294136, "learning_rate": 8.691025239975415e-05, "loss": 0.0047, "step": 27390 }, { "epoch": 288.42105263157896, "grad_norm": 0.08555468916893005, "learning_rate": 8.689909646471243e-05, "loss": 0.004, "step": 27400 }, { "epoch": 288.5263157894737, "grad_norm": 0.07213220000267029, "learning_rate": 8.688793649444099e-05, "loss": 0.0043, "step": 27410 }, { "epoch": 288.63157894736844, "grad_norm": 0.07009080052375793, "learning_rate": 8.687677249016029e-05, "loss": 0.0045, "step": 27420 }, { "epoch": 288.7368421052632, "grad_norm": 0.0673385038971901, "learning_rate": 8.686560445309118e-05, "loss": 0.0045, "step": 27430 }, { "epoch": 288.8421052631579, "grad_norm": 0.06394622474908829, "learning_rate": 8.685443238445499e-05, "loss": 0.0044, "step": 27440 }, { "epoch": 288.94736842105266, "grad_norm": 0.0697493925690651, "learning_rate": 8.68432562854735e-05, "loss": 0.0042, "step": 27450 }, { "epoch": 289.05263157894734, "grad_norm": 0.07389744371175766, "learning_rate": 8.683207615736887e-05, "loss": 0.0042, "step": 27460 }, { "epoch": 289.1578947368421, "grad_norm": 0.09701427817344666, "learning_rate": 8.682089200136379e-05, "loss": 0.0053, "step": 27470 }, { "epoch": 289.2631578947368, "grad_norm": 0.10197482258081436, "learning_rate": 8.680970381868132e-05, "loss": 0.0043, "step": 27480 }, { "epoch": 289.36842105263156, "grad_norm": 0.0765020027756691, "learning_rate": 8.679851161054498e-05, "loss": 0.004, "step": 27490 }, { "epoch": 289.4736842105263, "grad_norm": 0.08739131689071655, "learning_rate": 8.678731537817873e-05, "loss": 0.0043, "step": 27500 }, { "epoch": 289.57894736842104, "grad_norm": 0.08546791225671768, "learning_rate": 8.677611512280697e-05, "loss": 0.0043, "step": 27510 }, { "epoch": 289.6842105263158, "grad_norm": 0.08385232090950012, "learning_rate": 8.676491084565457e-05, "loss": 0.0043, "step": 27520 }, { "epoch": 289.7894736842105, "grad_norm": 0.07886108011007309, "learning_rate": 8.675370254794678e-05, "loss": 0.0037, "step": 27530 }, { "epoch": 289.89473684210526, "grad_norm": 0.06842994689941406, "learning_rate": 8.674249023090935e-05, "loss": 0.0044, "step": 27540 }, { "epoch": 290.0, "grad_norm": 0.11844057589769363, "learning_rate": 8.673127389576843e-05, "loss": 0.0044, "step": 27550 }, { "epoch": 290.10526315789474, "grad_norm": 0.08349869400262833, "learning_rate": 8.67200535437506e-05, "loss": 0.0048, "step": 27560 }, { "epoch": 290.2105263157895, "grad_norm": 0.07672811299562454, "learning_rate": 8.670882917608296e-05, "loss": 0.0039, "step": 27570 }, { "epoch": 290.3157894736842, "grad_norm": 0.10135731101036072, "learning_rate": 8.669760079399292e-05, "loss": 0.0044, "step": 27580 }, { "epoch": 290.42105263157896, "grad_norm": 0.08707652240991592, "learning_rate": 8.668636839870845e-05, "loss": 0.0043, "step": 27590 }, { "epoch": 290.5263157894737, "grad_norm": 0.07100527733564377, "learning_rate": 8.667513199145789e-05, "loss": 0.0037, "step": 27600 }, { "epoch": 290.63157894736844, "grad_norm": 0.05155849829316139, "learning_rate": 8.666389157347002e-05, "loss": 0.0044, "step": 27610 }, { "epoch": 290.7368421052632, "grad_norm": 0.06777509301900864, "learning_rate": 8.66526471459741e-05, "loss": 0.0044, "step": 27620 }, { "epoch": 290.8421052631579, "grad_norm": 0.0812622532248497, "learning_rate": 8.66413987101998e-05, "loss": 0.0045, "step": 27630 }, { "epoch": 290.94736842105266, "grad_norm": 0.06473856419324875, "learning_rate": 8.663014626737723e-05, "loss": 0.0043, "step": 27640 }, { "epoch": 291.05263157894734, "grad_norm": 0.08117032051086426, "learning_rate": 8.661888981873691e-05, "loss": 0.0042, "step": 27650 }, { "epoch": 291.1578947368421, "grad_norm": 0.09627307951450348, "learning_rate": 8.660762936550988e-05, "loss": 0.0035, "step": 27660 }, { "epoch": 291.2631578947368, "grad_norm": 0.09459449350833893, "learning_rate": 8.659636490892753e-05, "loss": 0.0039, "step": 27670 }, { "epoch": 291.36842105263156, "grad_norm": 0.10355386137962341, "learning_rate": 8.658509645022174e-05, "loss": 0.0046, "step": 27680 }, { "epoch": 291.4736842105263, "grad_norm": 0.08525631576776505, "learning_rate": 8.657382399062481e-05, "loss": 0.0048, "step": 27690 }, { "epoch": 291.57894736842104, "grad_norm": 0.08052193373441696, "learning_rate": 8.656254753136946e-05, "loss": 0.0041, "step": 27700 }, { "epoch": 291.6842105263158, "grad_norm": 0.06088143587112427, "learning_rate": 8.655126707368891e-05, "loss": 0.0037, "step": 27710 }, { "epoch": 291.7894736842105, "grad_norm": 0.08882272243499756, "learning_rate": 8.653998261881672e-05, "loss": 0.0045, "step": 27720 }, { "epoch": 291.89473684210526, "grad_norm": 0.09175068140029907, "learning_rate": 8.652869416798699e-05, "loss": 0.0041, "step": 27730 }, { "epoch": 292.0, "grad_norm": 0.10319516807794571, "learning_rate": 8.651740172243417e-05, "loss": 0.0041, "step": 27740 }, { "epoch": 292.10526315789474, "grad_norm": 0.08209939301013947, "learning_rate": 8.65061052833932e-05, "loss": 0.0048, "step": 27750 }, { "epoch": 292.2105263157895, "grad_norm": 0.1030270978808403, "learning_rate": 8.649480485209945e-05, "loss": 0.0046, "step": 27760 }, { "epoch": 292.3157894736842, "grad_norm": 0.09404334425926208, "learning_rate": 8.64835004297887e-05, "loss": 0.0045, "step": 27770 }, { "epoch": 292.42105263157896, "grad_norm": 0.08179830014705658, "learning_rate": 8.64721920176972e-05, "loss": 0.0041, "step": 27780 }, { "epoch": 292.5263157894737, "grad_norm": 0.06742390990257263, "learning_rate": 8.646087961706164e-05, "loss": 0.0041, "step": 27790 }, { "epoch": 292.63157894736844, "grad_norm": 0.0577036552131176, "learning_rate": 8.644956322911908e-05, "loss": 0.0038, "step": 27800 }, { "epoch": 292.7368421052632, "grad_norm": 0.0827685222029686, "learning_rate": 8.643824285510709e-05, "loss": 0.0043, "step": 27810 }, { "epoch": 292.8421052631579, "grad_norm": 0.08088220655918121, "learning_rate": 8.642691849626364e-05, "loss": 0.0042, "step": 27820 }, { "epoch": 292.94736842105266, "grad_norm": 0.08023693412542343, "learning_rate": 8.641559015382717e-05, "loss": 0.0047, "step": 27830 }, { "epoch": 293.05263157894734, "grad_norm": 0.09407421201467514, "learning_rate": 8.640425782903649e-05, "loss": 0.0044, "step": 27840 }, { "epoch": 293.1578947368421, "grad_norm": 0.0833858773112297, "learning_rate": 8.639292152313091e-05, "loss": 0.0054, "step": 27850 }, { "epoch": 293.2631578947368, "grad_norm": 0.07652278244495392, "learning_rate": 8.638158123735015e-05, "loss": 0.0043, "step": 27860 }, { "epoch": 293.36842105263156, "grad_norm": 0.08520567417144775, "learning_rate": 8.637023697293436e-05, "loss": 0.0041, "step": 27870 }, { "epoch": 293.4736842105263, "grad_norm": 0.0882645696401596, "learning_rate": 8.635888873112414e-05, "loss": 0.004, "step": 27880 }, { "epoch": 293.57894736842104, "grad_norm": 0.06770642846822739, "learning_rate": 8.634753651316052e-05, "loss": 0.0047, "step": 27890 }, { "epoch": 293.6842105263158, "grad_norm": 0.09495127201080322, "learning_rate": 8.633618032028496e-05, "loss": 0.0043, "step": 27900 }, { "epoch": 293.7894736842105, "grad_norm": 0.09083312004804611, "learning_rate": 8.632482015373934e-05, "loss": 0.0049, "step": 27910 }, { "epoch": 293.89473684210526, "grad_norm": 0.11917471140623093, "learning_rate": 8.6313456014766e-05, "loss": 0.0055, "step": 27920 }, { "epoch": 294.0, "grad_norm": 0.07604514807462692, "learning_rate": 8.630208790460771e-05, "loss": 0.0043, "step": 27930 }, { "epoch": 294.10526315789474, "grad_norm": 0.0866646021604538, "learning_rate": 8.629071582450768e-05, "loss": 0.0046, "step": 27940 }, { "epoch": 294.2105263157895, "grad_norm": 0.09679208695888519, "learning_rate": 8.62793397757095e-05, "loss": 0.0041, "step": 27950 }, { "epoch": 294.3157894736842, "grad_norm": 0.08974642306566238, "learning_rate": 8.626795975945729e-05, "loss": 0.0044, "step": 27960 }, { "epoch": 294.42105263157896, "grad_norm": 0.08666490018367767, "learning_rate": 8.625657577699551e-05, "loss": 0.0048, "step": 27970 }, { "epoch": 294.5263157894737, "grad_norm": 0.08710938692092896, "learning_rate": 8.624518782956914e-05, "loss": 0.0048, "step": 27980 }, { "epoch": 294.63157894736844, "grad_norm": 0.08110956102609634, "learning_rate": 8.62337959184235e-05, "loss": 0.0046, "step": 27990 }, { "epoch": 294.7368421052632, "grad_norm": 0.07546861469745636, "learning_rate": 8.622240004480441e-05, "loss": 0.0046, "step": 28000 }, { "epoch": 294.8421052631579, "grad_norm": 0.07231555879116058, "learning_rate": 8.621100020995814e-05, "loss": 0.0041, "step": 28010 }, { "epoch": 294.94736842105266, "grad_norm": 0.0820436179637909, "learning_rate": 8.619959641513132e-05, "loss": 0.0041, "step": 28020 }, { "epoch": 295.05263157894734, "grad_norm": 0.1071946918964386, "learning_rate": 8.618818866157105e-05, "loss": 0.004, "step": 28030 }, { "epoch": 295.1578947368421, "grad_norm": 0.07407122850418091, "learning_rate": 8.617677695052487e-05, "loss": 0.0049, "step": 28040 }, { "epoch": 295.2631578947368, "grad_norm": 0.08254753798246384, "learning_rate": 8.616536128324078e-05, "loss": 0.0043, "step": 28050 }, { "epoch": 295.36842105263156, "grad_norm": 0.08288212865591049, "learning_rate": 8.615394166096712e-05, "loss": 0.0039, "step": 28060 }, { "epoch": 295.4736842105263, "grad_norm": 0.09597957879304886, "learning_rate": 8.614251808495279e-05, "loss": 0.005, "step": 28070 }, { "epoch": 295.57894736842104, "grad_norm": 0.09247954189777374, "learning_rate": 8.6131090556447e-05, "loss": 0.0044, "step": 28080 }, { "epoch": 295.6842105263158, "grad_norm": 0.0988895520567894, "learning_rate": 8.611965907669947e-05, "loss": 0.0039, "step": 28090 }, { "epoch": 295.7894736842105, "grad_norm": 0.10302039235830307, "learning_rate": 8.610822364696034e-05, "loss": 0.0052, "step": 28100 }, { "epoch": 295.89473684210526, "grad_norm": 0.09397164732217789, "learning_rate": 8.609678426848015e-05, "loss": 0.0047, "step": 28110 }, { "epoch": 296.0, "grad_norm": 0.12302346527576447, "learning_rate": 8.60853409425099e-05, "loss": 0.0036, "step": 28120 }, { "epoch": 296.10526315789474, "grad_norm": 0.10123270750045776, "learning_rate": 8.607389367030104e-05, "loss": 0.0039, "step": 28130 }, { "epoch": 296.2105263157895, "grad_norm": 0.08446362614631653, "learning_rate": 8.606244245310538e-05, "loss": 0.004, "step": 28140 }, { "epoch": 296.3157894736842, "grad_norm": 0.09533243626356125, "learning_rate": 8.605098729217525e-05, "loss": 0.0047, "step": 28150 }, { "epoch": 296.42105263157896, "grad_norm": 0.09019669145345688, "learning_rate": 8.603952818876335e-05, "loss": 0.004, "step": 28160 }, { "epoch": 296.5263157894737, "grad_norm": 0.06800145655870438, "learning_rate": 8.602806514412281e-05, "loss": 0.0039, "step": 28170 }, { "epoch": 296.63157894736844, "grad_norm": 0.09416113793849945, "learning_rate": 8.601659815950726e-05, "loss": 0.0041, "step": 28180 }, { "epoch": 296.7368421052632, "grad_norm": 0.08019131422042847, "learning_rate": 8.600512723617067e-05, "loss": 0.0039, "step": 28190 }, { "epoch": 296.8421052631579, "grad_norm": 0.07503518462181091, "learning_rate": 8.59936523753675e-05, "loss": 0.0046, "step": 28200 }, { "epoch": 296.94736842105266, "grad_norm": 0.08357464522123337, "learning_rate": 8.598217357835264e-05, "loss": 0.0041, "step": 28210 }, { "epoch": 297.05263157894734, "grad_norm": 0.07318542897701263, "learning_rate": 8.597069084638135e-05, "loss": 0.004, "step": 28220 }, { "epoch": 297.1578947368421, "grad_norm": 0.09160678833723068, "learning_rate": 8.595920418070939e-05, "loss": 0.0044, "step": 28230 }, { "epoch": 297.2631578947368, "grad_norm": 0.07137010991573334, "learning_rate": 8.594771358259295e-05, "loss": 0.0044, "step": 28240 }, { "epoch": 297.36842105263156, "grad_norm": 0.07888268679380417, "learning_rate": 8.593621905328858e-05, "loss": 0.0043, "step": 28250 }, { "epoch": 297.4736842105263, "grad_norm": 0.07485777139663696, "learning_rate": 8.592472059405333e-05, "loss": 0.0043, "step": 28260 }, { "epoch": 297.57894736842104, "grad_norm": 0.08421693742275238, "learning_rate": 8.591321820614464e-05, "loss": 0.0037, "step": 28270 }, { "epoch": 297.6842105263158, "grad_norm": 0.08750270307064056, "learning_rate": 8.590171189082041e-05, "loss": 0.0041, "step": 28280 }, { "epoch": 297.7894736842105, "grad_norm": 0.07071639597415924, "learning_rate": 8.589020164933894e-05, "loss": 0.0037, "step": 28290 }, { "epoch": 297.89473684210526, "grad_norm": 0.07397415488958359, "learning_rate": 8.587868748295898e-05, "loss": 0.0039, "step": 28300 }, { "epoch": 298.0, "grad_norm": 0.09105014055967331, "learning_rate": 8.586716939293971e-05, "loss": 0.0033, "step": 28310 }, { "epoch": 298.10526315789474, "grad_norm": 0.07183394581079483, "learning_rate": 8.58556473805407e-05, "loss": 0.0039, "step": 28320 }, { "epoch": 298.2105263157895, "grad_norm": 0.07731308788061142, "learning_rate": 8.584412144702202e-05, "loss": 0.004, "step": 28330 }, { "epoch": 298.3157894736842, "grad_norm": 0.06117262318730354, "learning_rate": 8.58325915936441e-05, "loss": 0.0039, "step": 28340 }, { "epoch": 298.42105263157896, "grad_norm": 0.07257434725761414, "learning_rate": 8.582105782166783e-05, "loss": 0.0037, "step": 28350 }, { "epoch": 298.5263157894737, "grad_norm": 0.08132082968950272, "learning_rate": 8.580952013235455e-05, "loss": 0.0045, "step": 28360 }, { "epoch": 298.63157894736844, "grad_norm": 0.08060137927532196, "learning_rate": 8.579797852696596e-05, "loss": 0.0033, "step": 28370 }, { "epoch": 298.7368421052632, "grad_norm": 0.08819885551929474, "learning_rate": 8.578643300676428e-05, "loss": 0.0033, "step": 28380 }, { "epoch": 298.8421052631579, "grad_norm": 0.08014264702796936, "learning_rate": 8.577488357301209e-05, "loss": 0.0042, "step": 28390 }, { "epoch": 298.94736842105266, "grad_norm": 0.09658408910036087, "learning_rate": 8.576333022697242e-05, "loss": 0.0042, "step": 28400 }, { "epoch": 299.05263157894734, "grad_norm": 0.0761386975646019, "learning_rate": 8.575177296990873e-05, "loss": 0.0042, "step": 28410 }, { "epoch": 299.1578947368421, "grad_norm": 0.07199567556381226, "learning_rate": 8.574021180308489e-05, "loss": 0.0039, "step": 28420 }, { "epoch": 299.2631578947368, "grad_norm": 0.07370216399431229, "learning_rate": 8.572864672776523e-05, "loss": 0.0041, "step": 28430 }, { "epoch": 299.36842105263156, "grad_norm": 0.07408387959003448, "learning_rate": 8.571707774521447e-05, "loss": 0.0042, "step": 28440 }, { "epoch": 299.4736842105263, "grad_norm": 0.07966851443052292, "learning_rate": 8.57055048566978e-05, "loss": 0.004, "step": 28450 }, { "epoch": 299.57894736842104, "grad_norm": 0.08526687324047089, "learning_rate": 8.569392806348078e-05, "loss": 0.0042, "step": 28460 }, { "epoch": 299.6842105263158, "grad_norm": 0.09102330356836319, "learning_rate": 8.568234736682947e-05, "loss": 0.0038, "step": 28470 }, { "epoch": 299.7894736842105, "grad_norm": 0.08738188445568085, "learning_rate": 8.567076276801029e-05, "loss": 0.0039, "step": 28480 }, { "epoch": 299.89473684210526, "grad_norm": 0.09365099668502808, "learning_rate": 8.565917426829013e-05, "loss": 0.0038, "step": 28490 }, { "epoch": 300.0, "grad_norm": 0.13528646528720856, "learning_rate": 8.564758186893628e-05, "loss": 0.0053, "step": 28500 }, { "epoch": 300.10526315789474, "grad_norm": 0.1178312674164772, "learning_rate": 8.563598557121649e-05, "loss": 0.0046, "step": 28510 }, { "epoch": 300.2105263157895, "grad_norm": 0.12639722228050232, "learning_rate": 8.562438537639888e-05, "loss": 0.0056, "step": 28520 }, { "epoch": 300.3157894736842, "grad_norm": 0.12224135547876358, "learning_rate": 8.561278128575206e-05, "loss": 0.0041, "step": 28530 }, { "epoch": 300.42105263157896, "grad_norm": 0.07912378013134003, "learning_rate": 8.5601173300545e-05, "loss": 0.0046, "step": 28540 }, { "epoch": 300.5263157894737, "grad_norm": 0.08509700745344162, "learning_rate": 8.558956142204717e-05, "loss": 0.0041, "step": 28550 }, { "epoch": 300.63157894736844, "grad_norm": 0.08413157612085342, "learning_rate": 8.55779456515284e-05, "loss": 0.0043, "step": 28560 }, { "epoch": 300.7368421052632, "grad_norm": 0.07092243432998657, "learning_rate": 8.556632599025898e-05, "loss": 0.0048, "step": 28570 }, { "epoch": 300.8421052631579, "grad_norm": 0.08909817039966583, "learning_rate": 8.555470243950964e-05, "loss": 0.005, "step": 28580 }, { "epoch": 300.94736842105266, "grad_norm": 0.08430489152669907, "learning_rate": 8.554307500055148e-05, "loss": 0.0042, "step": 28590 }, { "epoch": 301.05263157894734, "grad_norm": 0.10042562335729599, "learning_rate": 8.553144367465609e-05, "loss": 0.0048, "step": 28600 }, { "epoch": 301.1578947368421, "grad_norm": 0.09534922242164612, "learning_rate": 8.551980846309544e-05, "loss": 0.0049, "step": 28610 }, { "epoch": 301.2631578947368, "grad_norm": 0.10110700875520706, "learning_rate": 8.550816936714193e-05, "loss": 0.0052, "step": 28620 }, { "epoch": 301.36842105263156, "grad_norm": 0.0782654657959938, "learning_rate": 8.549652638806841e-05, "loss": 0.0046, "step": 28630 }, { "epoch": 301.4736842105263, "grad_norm": 0.09962969273328781, "learning_rate": 8.548487952714812e-05, "loss": 0.0051, "step": 28640 }, { "epoch": 301.57894736842104, "grad_norm": 0.10310707986354828, "learning_rate": 8.547322878565478e-05, "loss": 0.0038, "step": 28650 }, { "epoch": 301.6842105263158, "grad_norm": 0.09107216447591782, "learning_rate": 8.546157416486245e-05, "loss": 0.0044, "step": 28660 }, { "epoch": 301.7894736842105, "grad_norm": 0.07549741119146347, "learning_rate": 8.54499156660457e-05, "loss": 0.0046, "step": 28670 }, { "epoch": 301.89473684210526, "grad_norm": 0.07566586136817932, "learning_rate": 8.543825329047947e-05, "loss": 0.0036, "step": 28680 }, { "epoch": 302.0, "grad_norm": 0.08292429894208908, "learning_rate": 8.542658703943913e-05, "loss": 0.0038, "step": 28690 }, { "epoch": 302.10526315789474, "grad_norm": 0.05470648780465126, "learning_rate": 8.541491691420051e-05, "loss": 0.0042, "step": 28700 }, { "epoch": 302.2105263157895, "grad_norm": 0.059998154640197754, "learning_rate": 8.54032429160398e-05, "loss": 0.0046, "step": 28710 }, { "epoch": 302.3157894736842, "grad_norm": 0.08045626431703568, "learning_rate": 8.539156504623369e-05, "loss": 0.0041, "step": 28720 }, { "epoch": 302.42105263157896, "grad_norm": 0.08212880045175552, "learning_rate": 8.537988330605923e-05, "loss": 0.0043, "step": 28730 }, { "epoch": 302.5263157894737, "grad_norm": 0.07597936689853668, "learning_rate": 8.536819769679393e-05, "loss": 0.0045, "step": 28740 }, { "epoch": 302.63157894736844, "grad_norm": 0.08635888248682022, "learning_rate": 8.53565082197157e-05, "loss": 0.005, "step": 28750 }, { "epoch": 302.7368421052632, "grad_norm": 0.07989040017127991, "learning_rate": 8.534481487610289e-05, "loss": 0.0039, "step": 28760 }, { "epoch": 302.8421052631579, "grad_norm": 0.07664640247821808, "learning_rate": 8.533311766723428e-05, "loss": 0.004, "step": 28770 }, { "epoch": 302.94736842105266, "grad_norm": 0.09248126298189163, "learning_rate": 8.532141659438901e-05, "loss": 0.0046, "step": 28780 }, { "epoch": 303.05263157894734, "grad_norm": 0.10567674040794373, "learning_rate": 8.530971165884675e-05, "loss": 0.0042, "step": 28790 }, { "epoch": 303.1578947368421, "grad_norm": 0.1063741073012352, "learning_rate": 8.529800286188752e-05, "loss": 0.0046, "step": 28800 }, { "epoch": 303.2631578947368, "grad_norm": 0.0954427421092987, "learning_rate": 8.528629020479175e-05, "loss": 0.004, "step": 28810 }, { "epoch": 303.36842105263156, "grad_norm": 0.09198900312185287, "learning_rate": 8.527457368884033e-05, "loss": 0.0049, "step": 28820 }, { "epoch": 303.4736842105263, "grad_norm": 0.07890387624502182, "learning_rate": 8.526285331531458e-05, "loss": 0.0052, "step": 28830 }, { "epoch": 303.57894736842104, "grad_norm": 0.07358583807945251, "learning_rate": 8.525112908549621e-05, "loss": 0.0043, "step": 28840 }, { "epoch": 303.6842105263158, "grad_norm": 0.07226406782865524, "learning_rate": 8.523940100066735e-05, "loss": 0.0049, "step": 28850 }, { "epoch": 303.7894736842105, "grad_norm": 0.10028550028800964, "learning_rate": 8.52276690621106e-05, "loss": 0.0042, "step": 28860 }, { "epoch": 303.89473684210526, "grad_norm": 0.06764130294322968, "learning_rate": 8.521593327110889e-05, "loss": 0.0042, "step": 28870 }, { "epoch": 304.0, "grad_norm": 0.10361812263727188, "learning_rate": 8.520419362894569e-05, "loss": 0.0045, "step": 28880 }, { "epoch": 304.10526315789474, "grad_norm": 0.09012351930141449, "learning_rate": 8.51924501369048e-05, "loss": 0.0046, "step": 28890 }, { "epoch": 304.2105263157895, "grad_norm": 0.06011658534407616, "learning_rate": 8.518070279627047e-05, "loss": 0.0042, "step": 28900 }, { "epoch": 304.3157894736842, "grad_norm": 0.09854382276535034, "learning_rate": 8.516895160832737e-05, "loss": 0.0045, "step": 28910 }, { "epoch": 304.42105263157896, "grad_norm": 0.08008802682161331, "learning_rate": 8.515719657436061e-05, "loss": 0.0049, "step": 28920 }, { "epoch": 304.5263157894737, "grad_norm": 0.08127213269472122, "learning_rate": 8.514543769565568e-05, "loss": 0.004, "step": 28930 }, { "epoch": 304.63157894736844, "grad_norm": 0.07715269178152084, "learning_rate": 8.513367497349853e-05, "loss": 0.0039, "step": 28940 }, { "epoch": 304.7368421052632, "grad_norm": 0.07721704244613647, "learning_rate": 8.51219084091755e-05, "loss": 0.0041, "step": 28950 }, { "epoch": 304.8421052631579, "grad_norm": 0.08471471071243286, "learning_rate": 8.511013800397338e-05, "loss": 0.0042, "step": 28960 }, { "epoch": 304.94736842105266, "grad_norm": 0.08253490179777145, "learning_rate": 8.509836375917937e-05, "loss": 0.0038, "step": 28970 }, { "epoch": 305.05263157894734, "grad_norm": 0.08981560915708542, "learning_rate": 8.508658567608104e-05, "loss": 0.0046, "step": 28980 }, { "epoch": 305.1578947368421, "grad_norm": 0.08504677563905716, "learning_rate": 8.507480375596647e-05, "loss": 0.0058, "step": 28990 }, { "epoch": 305.2631578947368, "grad_norm": 0.08353576809167862, "learning_rate": 8.506301800012408e-05, "loss": 0.005, "step": 29000 }, { "epoch": 305.36842105263156, "grad_norm": 0.07989481836557388, "learning_rate": 8.505122840984278e-05, "loss": 0.0042, "step": 29010 }, { "epoch": 305.4736842105263, "grad_norm": 0.0748315081000328, "learning_rate": 8.503943498641182e-05, "loss": 0.0049, "step": 29020 }, { "epoch": 305.57894736842104, "grad_norm": 0.09580163657665253, "learning_rate": 8.502763773112095e-05, "loss": 0.004, "step": 29030 }, { "epoch": 305.6842105263158, "grad_norm": 0.09042897820472717, "learning_rate": 8.501583664526026e-05, "loss": 0.0045, "step": 29040 }, { "epoch": 305.7894736842105, "grad_norm": 0.07234765589237213, "learning_rate": 8.500403173012032e-05, "loss": 0.0037, "step": 29050 }, { "epoch": 305.89473684210526, "grad_norm": 0.07823028415441513, "learning_rate": 8.499222298699211e-05, "loss": 0.0052, "step": 29060 }, { "epoch": 306.0, "grad_norm": 0.10557185113430023, "learning_rate": 8.498041041716701e-05, "loss": 0.0048, "step": 29070 }, { "epoch": 306.10526315789474, "grad_norm": 0.07874197512865067, "learning_rate": 8.496859402193681e-05, "loss": 0.005, "step": 29080 }, { "epoch": 306.2105263157895, "grad_norm": 0.07647809386253357, "learning_rate": 8.495677380259374e-05, "loss": 0.0043, "step": 29090 }, { "epoch": 306.3157894736842, "grad_norm": 0.07609326392412186, "learning_rate": 8.494494976043045e-05, "loss": 0.0043, "step": 29100 }, { "epoch": 306.42105263157896, "grad_norm": 0.11383099853992462, "learning_rate": 8.493312189673998e-05, "loss": 0.0041, "step": 29110 }, { "epoch": 306.5263157894737, "grad_norm": 0.09498456120491028, "learning_rate": 8.492129021281584e-05, "loss": 0.0049, "step": 29120 }, { "epoch": 306.63157894736844, "grad_norm": 0.0960918739438057, "learning_rate": 8.490945470995188e-05, "loss": 0.0039, "step": 29130 }, { "epoch": 306.7368421052632, "grad_norm": 0.10980014503002167, "learning_rate": 8.489761538944247e-05, "loss": 0.0043, "step": 29140 }, { "epoch": 306.8421052631579, "grad_norm": 0.09853136539459229, "learning_rate": 8.48857722525823e-05, "loss": 0.004, "step": 29150 }, { "epoch": 306.94736842105266, "grad_norm": 0.07835230976343155, "learning_rate": 8.487392530066652e-05, "loss": 0.0047, "step": 29160 }, { "epoch": 307.05263157894734, "grad_norm": 0.06244342774152756, "learning_rate": 8.486207453499069e-05, "loss": 0.0051, "step": 29170 }, { "epoch": 307.1578947368421, "grad_norm": 0.09067866206169128, "learning_rate": 8.485021995685082e-05, "loss": 0.0039, "step": 29180 }, { "epoch": 307.2631578947368, "grad_norm": 0.08668018132448196, "learning_rate": 8.483836156754328e-05, "loss": 0.005, "step": 29190 }, { "epoch": 307.36842105263156, "grad_norm": 0.09297004342079163, "learning_rate": 8.482649936836491e-05, "loss": 0.0049, "step": 29200 }, { "epoch": 307.4736842105263, "grad_norm": 0.0866866186261177, "learning_rate": 8.481463336061293e-05, "loss": 0.0037, "step": 29210 }, { "epoch": 307.57894736842104, "grad_norm": 0.08735837787389755, "learning_rate": 8.480276354558496e-05, "loss": 0.0044, "step": 29220 }, { "epoch": 307.6842105263158, "grad_norm": 0.0698140487074852, "learning_rate": 8.479088992457913e-05, "loss": 0.0045, "step": 29230 }, { "epoch": 307.7894736842105, "grad_norm": 0.0842815488576889, "learning_rate": 8.477901249889387e-05, "loss": 0.0042, "step": 29240 }, { "epoch": 307.89473684210526, "grad_norm": 0.06691791117191315, "learning_rate": 8.47671312698281e-05, "loss": 0.0044, "step": 29250 }, { "epoch": 308.0, "grad_norm": 0.10662936419248581, "learning_rate": 8.475524623868112e-05, "loss": 0.0046, "step": 29260 }, { "epoch": 308.10526315789474, "grad_norm": 0.10438252985477448, "learning_rate": 8.474335740675266e-05, "loss": 0.0046, "step": 29270 }, { "epoch": 308.2105263157895, "grad_norm": 0.10088072717189789, "learning_rate": 8.473146477534289e-05, "loss": 0.0043, "step": 29280 }, { "epoch": 308.3157894736842, "grad_norm": 0.07590822130441666, "learning_rate": 8.471956834575232e-05, "loss": 0.0037, "step": 29290 }, { "epoch": 308.42105263157896, "grad_norm": 0.06876542419195175, "learning_rate": 8.470766811928197e-05, "loss": 0.0033, "step": 29300 }, { "epoch": 308.5263157894737, "grad_norm": 0.06044067069888115, "learning_rate": 8.469576409723323e-05, "loss": 0.004, "step": 29310 }, { "epoch": 308.63157894736844, "grad_norm": 0.06273657828569412, "learning_rate": 8.468385628090788e-05, "loss": 0.0044, "step": 29320 }, { "epoch": 308.7368421052632, "grad_norm": 0.08881422132253647, "learning_rate": 8.467194467160815e-05, "loss": 0.0034, "step": 29330 }, { "epoch": 308.8421052631579, "grad_norm": 0.08894775807857513, "learning_rate": 8.466002927063667e-05, "loss": 0.0045, "step": 29340 }, { "epoch": 308.94736842105266, "grad_norm": 0.08238767832517624, "learning_rate": 8.464811007929651e-05, "loss": 0.0041, "step": 29350 }, { "epoch": 309.05263157894734, "grad_norm": 0.08977605402469635, "learning_rate": 8.463618709889114e-05, "loss": 0.0042, "step": 29360 }, { "epoch": 309.1578947368421, "grad_norm": 0.09103044867515564, "learning_rate": 8.462426033072442e-05, "loss": 0.005, "step": 29370 }, { "epoch": 309.2631578947368, "grad_norm": 0.07721437513828278, "learning_rate": 8.461232977610061e-05, "loss": 0.0038, "step": 29380 }, { "epoch": 309.36842105263156, "grad_norm": 0.0732460618019104, "learning_rate": 8.46003954363245e-05, "loss": 0.0039, "step": 29390 }, { "epoch": 309.4736842105263, "grad_norm": 0.054931554943323135, "learning_rate": 8.458845731270115e-05, "loss": 0.0047, "step": 29400 }, { "epoch": 309.57894736842104, "grad_norm": 0.08753827959299088, "learning_rate": 8.45765154065361e-05, "loss": 0.0042, "step": 29410 }, { "epoch": 309.6842105263158, "grad_norm": 0.07500506937503815, "learning_rate": 8.456456971913532e-05, "loss": 0.0041, "step": 29420 }, { "epoch": 309.7894736842105, "grad_norm": 0.0717223510146141, "learning_rate": 8.455262025180517e-05, "loss": 0.0039, "step": 29430 }, { "epoch": 309.89473684210526, "grad_norm": 0.06192118674516678, "learning_rate": 8.454066700585242e-05, "loss": 0.004, "step": 29440 }, { "epoch": 310.0, "grad_norm": 0.13556401431560516, "learning_rate": 8.452870998258423e-05, "loss": 0.0043, "step": 29450 }, { "epoch": 310.10526315789474, "grad_norm": 0.08052822947502136, "learning_rate": 8.451674918330825e-05, "loss": 0.0044, "step": 29460 }, { "epoch": 310.2105263157895, "grad_norm": 0.07654359191656113, "learning_rate": 8.450478460933246e-05, "loss": 0.004, "step": 29470 }, { "epoch": 310.3157894736842, "grad_norm": 0.09232126921415329, "learning_rate": 8.449281626196532e-05, "loss": 0.0046, "step": 29480 }, { "epoch": 310.42105263157896, "grad_norm": 0.09868331253528595, "learning_rate": 8.448084414251564e-05, "loss": 0.0045, "step": 29490 }, { "epoch": 310.5263157894737, "grad_norm": 0.08419158309698105, "learning_rate": 8.446886825229271e-05, "loss": 0.0038, "step": 29500 }, { "epoch": 310.63157894736844, "grad_norm": 0.08795782178640366, "learning_rate": 8.445688859260615e-05, "loss": 0.0054, "step": 29510 }, { "epoch": 310.7368421052632, "grad_norm": 0.09242936223745346, "learning_rate": 8.444490516476606e-05, "loss": 0.0046, "step": 29520 }, { "epoch": 310.8421052631579, "grad_norm": 0.0997876301407814, "learning_rate": 8.443291797008293e-05, "loss": 0.0043, "step": 29530 }, { "epoch": 310.94736842105266, "grad_norm": 0.07531816512346268, "learning_rate": 8.442092700986765e-05, "loss": 0.0039, "step": 29540 }, { "epoch": 311.05263157894734, "grad_norm": 0.08903396874666214, "learning_rate": 8.440893228543156e-05, "loss": 0.0044, "step": 29550 }, { "epoch": 311.1578947368421, "grad_norm": 0.07642757892608643, "learning_rate": 8.439693379808638e-05, "loss": 0.0047, "step": 29560 }, { "epoch": 311.2631578947368, "grad_norm": 0.07372266799211502, "learning_rate": 8.43849315491442e-05, "loss": 0.0046, "step": 29570 }, { "epoch": 311.36842105263156, "grad_norm": 0.08969183266162872, "learning_rate": 8.437292553991763e-05, "loss": 0.004, "step": 29580 }, { "epoch": 311.4736842105263, "grad_norm": 0.07935890555381775, "learning_rate": 8.436091577171959e-05, "loss": 0.0044, "step": 29590 }, { "epoch": 311.57894736842104, "grad_norm": 0.055786505341529846, "learning_rate": 8.434890224586347e-05, "loss": 0.0041, "step": 29600 }, { "epoch": 311.6842105263158, "grad_norm": 0.06345465779304504, "learning_rate": 8.433688496366303e-05, "loss": 0.0043, "step": 29610 }, { "epoch": 311.7894736842105, "grad_norm": 0.06792891025543213, "learning_rate": 8.432486392643248e-05, "loss": 0.0046, "step": 29620 }, { "epoch": 311.89473684210526, "grad_norm": 0.060793135315179825, "learning_rate": 8.431283913548643e-05, "loss": 0.0043, "step": 29630 }, { "epoch": 312.0, "grad_norm": 0.08633475750684738, "learning_rate": 8.430081059213985e-05, "loss": 0.0047, "step": 29640 }, { "epoch": 312.10526315789474, "grad_norm": 0.0812486782670021, "learning_rate": 8.428877829770823e-05, "loss": 0.0044, "step": 29650 }, { "epoch": 312.2105263157895, "grad_norm": 0.09824888408184052, "learning_rate": 8.427674225350735e-05, "loss": 0.0039, "step": 29660 }, { "epoch": 312.3157894736842, "grad_norm": 0.06389425694942474, "learning_rate": 8.426470246085347e-05, "loss": 0.0044, "step": 29670 }, { "epoch": 312.42105263157896, "grad_norm": 0.07617330551147461, "learning_rate": 8.425265892106324e-05, "loss": 0.0039, "step": 29680 }, { "epoch": 312.5263157894737, "grad_norm": 0.0826912373304367, "learning_rate": 8.424061163545374e-05, "loss": 0.004, "step": 29690 }, { "epoch": 312.63157894736844, "grad_norm": 0.0745914876461029, "learning_rate": 8.422856060534243e-05, "loss": 0.0043, "step": 29700 }, { "epoch": 312.7368421052632, "grad_norm": 0.07333221286535263, "learning_rate": 8.421650583204718e-05, "loss": 0.0045, "step": 29710 }, { "epoch": 312.8421052631579, "grad_norm": 0.08001982420682907, "learning_rate": 8.420444731688633e-05, "loss": 0.0045, "step": 29720 }, { "epoch": 312.94736842105266, "grad_norm": 0.07608243077993393, "learning_rate": 8.419238506117852e-05, "loss": 0.0041, "step": 29730 }, { "epoch": 313.05263157894734, "grad_norm": 0.06642548739910126, "learning_rate": 8.418031906624289e-05, "loss": 0.0043, "step": 29740 }, { "epoch": 313.1578947368421, "grad_norm": 0.23696982860565186, "learning_rate": 8.416824933339898e-05, "loss": 0.0041, "step": 29750 }, { "epoch": 313.2631578947368, "grad_norm": 0.08802656084299088, "learning_rate": 8.415617586396667e-05, "loss": 0.0047, "step": 29760 }, { "epoch": 313.36842105263156, "grad_norm": 0.08058719336986542, "learning_rate": 8.414409865926632e-05, "loss": 0.0045, "step": 29770 }, { "epoch": 313.4736842105263, "grad_norm": 0.11931080371141434, "learning_rate": 8.413201772061867e-05, "loss": 0.0056, "step": 29780 }, { "epoch": 313.57894736842104, "grad_norm": 0.10147644579410553, "learning_rate": 8.411993304934488e-05, "loss": 0.0052, "step": 29790 }, { "epoch": 313.6842105263158, "grad_norm": 0.09704549610614777, "learning_rate": 8.410784464676654e-05, "loss": 0.0039, "step": 29800 }, { "epoch": 313.7894736842105, "grad_norm": 0.081569142639637, "learning_rate": 8.409575251420556e-05, "loss": 0.0038, "step": 29810 }, { "epoch": 313.89473684210526, "grad_norm": 0.08752158284187317, "learning_rate": 8.408365665298435e-05, "loss": 0.0051, "step": 29820 }, { "epoch": 314.0, "grad_norm": 0.08360899984836578, "learning_rate": 8.40715570644257e-05, "loss": 0.0041, "step": 29830 }, { "epoch": 314.10526315789474, "grad_norm": 0.08293022215366364, "learning_rate": 8.40594537498528e-05, "loss": 0.0045, "step": 29840 }, { "epoch": 314.2105263157895, "grad_norm": 0.08975731581449509, "learning_rate": 8.404734671058924e-05, "loss": 0.0048, "step": 29850 }, { "epoch": 314.3157894736842, "grad_norm": 0.07228796184062958, "learning_rate": 8.403523594795902e-05, "loss": 0.0038, "step": 29860 }, { "epoch": 314.42105263157896, "grad_norm": 0.0797145664691925, "learning_rate": 8.402312146328659e-05, "loss": 0.0045, "step": 29870 }, { "epoch": 314.5263157894737, "grad_norm": 0.09061702340841293, "learning_rate": 8.401100325789675e-05, "loss": 0.0044, "step": 29880 }, { "epoch": 314.63157894736844, "grad_norm": 0.0906391367316246, "learning_rate": 8.399888133311472e-05, "loss": 0.0044, "step": 29890 }, { "epoch": 314.7368421052632, "grad_norm": 0.06099357455968857, "learning_rate": 8.398675569026613e-05, "loss": 0.0033, "step": 29900 }, { "epoch": 314.8421052631579, "grad_norm": 0.10482523590326309, "learning_rate": 8.397462633067705e-05, "loss": 0.0049, "step": 29910 }, { "epoch": 314.94736842105266, "grad_norm": 0.09643129259347916, "learning_rate": 8.396249325567392e-05, "loss": 0.004, "step": 29920 }, { "epoch": 315.05263157894734, "grad_norm": 0.0870656669139862, "learning_rate": 8.395035646658357e-05, "loss": 0.004, "step": 29930 }, { "epoch": 315.1578947368421, "grad_norm": 0.07376205176115036, "learning_rate": 8.39382159647333e-05, "loss": 0.0041, "step": 29940 }, { "epoch": 315.2631578947368, "grad_norm": 0.07570836693048477, "learning_rate": 8.392607175145075e-05, "loss": 0.0037, "step": 29950 }, { "epoch": 315.36842105263156, "grad_norm": 0.08791304379701614, "learning_rate": 8.3913923828064e-05, "loss": 0.0042, "step": 29960 }, { "epoch": 315.4736842105263, "grad_norm": 0.0698603168129921, "learning_rate": 8.390177219590152e-05, "loss": 0.0043, "step": 29970 }, { "epoch": 315.57894736842104, "grad_norm": 0.0731324777007103, "learning_rate": 8.388961685629222e-05, "loss": 0.004, "step": 29980 }, { "epoch": 315.6842105263158, "grad_norm": 0.07862588763237, "learning_rate": 8.387745781056536e-05, "loss": 0.0035, "step": 29990 }, { "epoch": 315.7894736842105, "grad_norm": 0.09448936581611633, "learning_rate": 8.386529506005065e-05, "loss": 0.0044, "step": 30000 } ], "logging_steps": 10, "max_steps": 100000, "num_input_tokens_seen": 0, "num_train_epochs": 1053, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 256, "trial_name": null, "trial_params": null }