diff --git "a/checkpoints/GR00T/StackCube/1000/checkpoint-75000/trainer_state.json" "b/checkpoints/GR00T/StackCube/1000/checkpoint-75000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoints/GR00T/StackCube/1000/checkpoint-75000/trainer_state.json" @@ -0,0 +1,52534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.825242718446602, + "eval_steps": 500, + "global_step": 75000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0007766990291262136, + "grad_norm": 11.456265449523926, + "learning_rate": 1.8e-07, + "loss": 0.8619, + "step": 10 + }, + { + "epoch": 0.0015533980582524273, + "grad_norm": 9.962185859680176, + "learning_rate": 3.8e-07, + "loss": 0.8651, + "step": 20 + }, + { + "epoch": 0.002330097087378641, + "grad_norm": 9.519487380981445, + "learning_rate": 5.8e-07, + "loss": 0.8574, + "step": 30 + }, + { + "epoch": 0.0031067961165048546, + "grad_norm": 7.203672885894775, + "learning_rate": 7.8e-07, + "loss": 0.7762, + "step": 40 + }, + { + "epoch": 0.003883495145631068, + "grad_norm": 6.009782791137695, + "learning_rate": 9.8e-07, + "loss": 0.6379, + "step": 50 + }, + { + "epoch": 0.004660194174757282, + "grad_norm": 3.897005081176758, + "learning_rate": 1.18e-06, + "loss": 0.5798, + "step": 60 + }, + { + "epoch": 0.005436893203883495, + "grad_norm": 2.6674468517303467, + "learning_rate": 1.3800000000000001e-06, + "loss": 0.3743, + "step": 70 + }, + { + "epoch": 0.006213592233009709, + "grad_norm": 1.7438561916351318, + "learning_rate": 1.5800000000000003e-06, + "loss": 0.3142, + "step": 80 + }, + { + "epoch": 0.006990291262135922, + "grad_norm": 1.5891557931900024, + "learning_rate": 1.7800000000000001e-06, + "loss": 0.2515, + "step": 90 + }, + { + "epoch": 0.007766990291262136, + "grad_norm": 1.5968161821365356, + "learning_rate": 1.98e-06, + "loss": 0.2126, + "step": 100 + }, + { + "epoch": 0.00854368932038835, + "grad_norm": 1.70491623878479, + "learning_rate": 2.1800000000000003e-06, + "loss": 0.2041, + "step": 110 + }, + { + "epoch": 0.009320388349514564, + "grad_norm": 1.0128200054168701, + "learning_rate": 2.38e-06, + "loss": 0.1866, + "step": 120 + }, + { + "epoch": 0.010097087378640776, + "grad_norm": 1.6038795709609985, + "learning_rate": 2.5800000000000003e-06, + "loss": 0.1742, + "step": 130 + }, + { + "epoch": 0.01087378640776699, + "grad_norm": 1.1186213493347168, + "learning_rate": 2.78e-06, + "loss": 0.1363, + "step": 140 + }, + { + "epoch": 0.011650485436893204, + "grad_norm": 0.9474993944168091, + "learning_rate": 2.9800000000000003e-06, + "loss": 0.1258, + "step": 150 + }, + { + "epoch": 0.012427184466019418, + "grad_norm": 1.2227210998535156, + "learning_rate": 3.1800000000000005e-06, + "loss": 0.1251, + "step": 160 + }, + { + "epoch": 0.01320388349514563, + "grad_norm": 1.0723987817764282, + "learning_rate": 3.38e-06, + "loss": 0.1128, + "step": 170 + }, + { + "epoch": 0.013980582524271845, + "grad_norm": 1.0823431015014648, + "learning_rate": 3.58e-06, + "loss": 0.1023, + "step": 180 + }, + { + "epoch": 0.014757281553398059, + "grad_norm": 0.8663830757141113, + "learning_rate": 3.7800000000000002e-06, + "loss": 0.0853, + "step": 190 + }, + { + "epoch": 0.015533980582524271, + "grad_norm": 1.3565744161605835, + "learning_rate": 3.98e-06, + "loss": 0.0878, + "step": 200 + }, + { + "epoch": 0.016310679611650485, + "grad_norm": 1.8402047157287598, + "learning_rate": 4.18e-06, + "loss": 0.0729, + "step": 210 + }, + { + "epoch": 0.0170873786407767, + "grad_norm": 1.0805026292800903, + "learning_rate": 4.38e-06, + "loss": 0.0764, + "step": 220 + }, + { + "epoch": 0.017864077669902913, + "grad_norm": 1.1092782020568848, + "learning_rate": 4.58e-06, + "loss": 0.0661, + "step": 230 + }, + { + "epoch": 0.018640776699029128, + "grad_norm": 1.0371233224868774, + "learning_rate": 4.780000000000001e-06, + "loss": 0.0612, + "step": 240 + }, + { + "epoch": 0.019417475728155338, + "grad_norm": 1.386696219444275, + "learning_rate": 4.98e-06, + "loss": 0.0671, + "step": 250 + }, + { + "epoch": 0.020194174757281552, + "grad_norm": 1.2284194231033325, + "learning_rate": 5.18e-06, + "loss": 0.0689, + "step": 260 + }, + { + "epoch": 0.020970873786407766, + "grad_norm": 1.7484028339385986, + "learning_rate": 5.38e-06, + "loss": 0.068, + "step": 270 + }, + { + "epoch": 0.02174757281553398, + "grad_norm": 1.1107338666915894, + "learning_rate": 5.580000000000001e-06, + "loss": 0.0623, + "step": 280 + }, + { + "epoch": 0.022524271844660194, + "grad_norm": 0.7357963919639587, + "learning_rate": 5.78e-06, + "loss": 0.0834, + "step": 290 + }, + { + "epoch": 0.02330097087378641, + "grad_norm": 0.932404637336731, + "learning_rate": 5.98e-06, + "loss": 0.0635, + "step": 300 + }, + { + "epoch": 0.024077669902912623, + "grad_norm": 1.1537994146347046, + "learning_rate": 6.18e-06, + "loss": 0.0822, + "step": 310 + }, + { + "epoch": 0.024854368932038837, + "grad_norm": 0.8133753538131714, + "learning_rate": 6.38e-06, + "loss": 0.058, + "step": 320 + }, + { + "epoch": 0.025631067961165047, + "grad_norm": 0.8372452259063721, + "learning_rate": 6.58e-06, + "loss": 0.0543, + "step": 330 + }, + { + "epoch": 0.02640776699029126, + "grad_norm": 0.9861927032470703, + "learning_rate": 6.78e-06, + "loss": 0.0434, + "step": 340 + }, + { + "epoch": 0.027184466019417475, + "grad_norm": 0.7162894010543823, + "learning_rate": 6.98e-06, + "loss": 0.0499, + "step": 350 + }, + { + "epoch": 0.02796116504854369, + "grad_norm": 1.4831666946411133, + "learning_rate": 7.180000000000001e-06, + "loss": 0.0445, + "step": 360 + }, + { + "epoch": 0.028737864077669904, + "grad_norm": 1.3939399719238281, + "learning_rate": 7.3800000000000005e-06, + "loss": 0.0535, + "step": 370 + }, + { + "epoch": 0.029514563106796118, + "grad_norm": 1.2459825277328491, + "learning_rate": 7.580000000000001e-06, + "loss": 0.0526, + "step": 380 + }, + { + "epoch": 0.030291262135922332, + "grad_norm": 0.7645425200462341, + "learning_rate": 7.78e-06, + "loss": 0.0552, + "step": 390 + }, + { + "epoch": 0.031067961165048542, + "grad_norm": 0.6664465069770813, + "learning_rate": 7.98e-06, + "loss": 0.0441, + "step": 400 + }, + { + "epoch": 0.031844660194174756, + "grad_norm": 1.2921053171157837, + "learning_rate": 8.18e-06, + "loss": 0.0503, + "step": 410 + }, + { + "epoch": 0.03262135922330097, + "grad_norm": 0.7930603623390198, + "learning_rate": 8.380000000000001e-06, + "loss": 0.0444, + "step": 420 + }, + { + "epoch": 0.033398058252427185, + "grad_norm": 0.7699264883995056, + "learning_rate": 8.580000000000001e-06, + "loss": 0.0509, + "step": 430 + }, + { + "epoch": 0.0341747572815534, + "grad_norm": 0.8492579460144043, + "learning_rate": 8.78e-06, + "loss": 0.0425, + "step": 440 + }, + { + "epoch": 0.03495145631067961, + "grad_norm": 0.5794140100479126, + "learning_rate": 8.98e-06, + "loss": 0.0449, + "step": 450 + }, + { + "epoch": 0.03572815533980583, + "grad_norm": 0.5802658796310425, + "learning_rate": 9.180000000000002e-06, + "loss": 0.0367, + "step": 460 + }, + { + "epoch": 0.03650485436893204, + "grad_norm": 0.7072301506996155, + "learning_rate": 9.38e-06, + "loss": 0.0441, + "step": 470 + }, + { + "epoch": 0.037281553398058255, + "grad_norm": 0.9142102599143982, + "learning_rate": 9.58e-06, + "loss": 0.0454, + "step": 480 + }, + { + "epoch": 0.03805825242718447, + "grad_norm": 0.7295803427696228, + "learning_rate": 9.78e-06, + "loss": 0.0447, + "step": 490 + }, + { + "epoch": 0.038834951456310676, + "grad_norm": 0.5354565382003784, + "learning_rate": 9.980000000000001e-06, + "loss": 0.0397, + "step": 500 + }, + { + "epoch": 0.03961165048543689, + "grad_norm": 0.5087801814079285, + "learning_rate": 1.018e-05, + "loss": 0.0422, + "step": 510 + }, + { + "epoch": 0.040388349514563104, + "grad_norm": 0.5321855545043945, + "learning_rate": 1.038e-05, + "loss": 0.04, + "step": 520 + }, + { + "epoch": 0.04116504854368932, + "grad_norm": 0.8199273347854614, + "learning_rate": 1.058e-05, + "loss": 0.0419, + "step": 530 + }, + { + "epoch": 0.04194174757281553, + "grad_norm": 0.5017785429954529, + "learning_rate": 1.0780000000000002e-05, + "loss": 0.0325, + "step": 540 + }, + { + "epoch": 0.04271844660194175, + "grad_norm": 0.5398567318916321, + "learning_rate": 1.098e-05, + "loss": 0.037, + "step": 550 + }, + { + "epoch": 0.04349514563106796, + "grad_norm": 0.523453414440155, + "learning_rate": 1.118e-05, + "loss": 0.032, + "step": 560 + }, + { + "epoch": 0.044271844660194175, + "grad_norm": 0.812235951423645, + "learning_rate": 1.1380000000000001e-05, + "loss": 0.04, + "step": 570 + }, + { + "epoch": 0.04504854368932039, + "grad_norm": 0.5188243985176086, + "learning_rate": 1.1580000000000001e-05, + "loss": 0.035, + "step": 580 + }, + { + "epoch": 0.0458252427184466, + "grad_norm": 0.8483536243438721, + "learning_rate": 1.178e-05, + "loss": 0.0376, + "step": 590 + }, + { + "epoch": 0.04660194174757282, + "grad_norm": 0.6280531883239746, + "learning_rate": 1.198e-05, + "loss": 0.0342, + "step": 600 + }, + { + "epoch": 0.04737864077669903, + "grad_norm": 0.7511423230171204, + "learning_rate": 1.2180000000000002e-05, + "loss": 0.0431, + "step": 610 + }, + { + "epoch": 0.048155339805825245, + "grad_norm": 0.7736706733703613, + "learning_rate": 1.238e-05, + "loss": 0.0355, + "step": 620 + }, + { + "epoch": 0.04893203883495146, + "grad_norm": 0.6227216720581055, + "learning_rate": 1.258e-05, + "loss": 0.0366, + "step": 630 + }, + { + "epoch": 0.04970873786407767, + "grad_norm": 0.6571856737136841, + "learning_rate": 1.278e-05, + "loss": 0.0336, + "step": 640 + }, + { + "epoch": 0.05048543689320388, + "grad_norm": 0.8189626336097717, + "learning_rate": 1.2980000000000001e-05, + "loss": 0.0428, + "step": 650 + }, + { + "epoch": 0.051262135922330095, + "grad_norm": 0.7235145568847656, + "learning_rate": 1.3180000000000001e-05, + "loss": 0.0328, + "step": 660 + }, + { + "epoch": 0.05203883495145631, + "grad_norm": 0.6715794801712036, + "learning_rate": 1.338e-05, + "loss": 0.0472, + "step": 670 + }, + { + "epoch": 0.05281553398058252, + "grad_norm": 0.8109439015388489, + "learning_rate": 1.358e-05, + "loss": 0.0289, + "step": 680 + }, + { + "epoch": 0.05359223300970874, + "grad_norm": 0.713049054145813, + "learning_rate": 1.3780000000000002e-05, + "loss": 0.0307, + "step": 690 + }, + { + "epoch": 0.05436893203883495, + "grad_norm": 0.56978839635849, + "learning_rate": 1.3980000000000002e-05, + "loss": 0.0319, + "step": 700 + }, + { + "epoch": 0.055145631067961165, + "grad_norm": 0.38364532589912415, + "learning_rate": 1.4180000000000001e-05, + "loss": 0.038, + "step": 710 + }, + { + "epoch": 0.05592233009708738, + "grad_norm": 1.031825065612793, + "learning_rate": 1.4380000000000001e-05, + "loss": 0.032, + "step": 720 + }, + { + "epoch": 0.05669902912621359, + "grad_norm": 0.6434110403060913, + "learning_rate": 1.4580000000000003e-05, + "loss": 0.031, + "step": 730 + }, + { + "epoch": 0.05747572815533981, + "grad_norm": 0.3871026933193207, + "learning_rate": 1.4779999999999999e-05, + "loss": 0.0305, + "step": 740 + }, + { + "epoch": 0.05825242718446602, + "grad_norm": 0.5035886168479919, + "learning_rate": 1.4979999999999999e-05, + "loss": 0.0318, + "step": 750 + }, + { + "epoch": 0.059029126213592235, + "grad_norm": 0.734967827796936, + "learning_rate": 1.518e-05, + "loss": 0.0266, + "step": 760 + }, + { + "epoch": 0.05980582524271845, + "grad_norm": 0.6754347085952759, + "learning_rate": 1.538e-05, + "loss": 0.0242, + "step": 770 + }, + { + "epoch": 0.060582524271844664, + "grad_norm": 0.4502272605895996, + "learning_rate": 1.558e-05, + "loss": 0.0248, + "step": 780 + }, + { + "epoch": 0.06135922330097087, + "grad_norm": 0.5700598359107971, + "learning_rate": 1.578e-05, + "loss": 0.0256, + "step": 790 + }, + { + "epoch": 0.062135922330097085, + "grad_norm": 0.4572841227054596, + "learning_rate": 1.598e-05, + "loss": 0.0327, + "step": 800 + }, + { + "epoch": 0.0629126213592233, + "grad_norm": 0.3129153251647949, + "learning_rate": 1.618e-05, + "loss": 0.0288, + "step": 810 + }, + { + "epoch": 0.06368932038834951, + "grad_norm": 0.6152353882789612, + "learning_rate": 1.6380000000000002e-05, + "loss": 0.036, + "step": 820 + }, + { + "epoch": 0.06446601941747573, + "grad_norm": 0.39676904678344727, + "learning_rate": 1.658e-05, + "loss": 0.0322, + "step": 830 + }, + { + "epoch": 0.06524271844660194, + "grad_norm": 0.4386296272277832, + "learning_rate": 1.6780000000000002e-05, + "loss": 0.0232, + "step": 840 + }, + { + "epoch": 0.06601941747572816, + "grad_norm": 0.49479958415031433, + "learning_rate": 1.698e-05, + "loss": 0.0313, + "step": 850 + }, + { + "epoch": 0.06679611650485437, + "grad_norm": 0.6992490887641907, + "learning_rate": 1.718e-05, + "loss": 0.0277, + "step": 860 + }, + { + "epoch": 0.06757281553398058, + "grad_norm": 0.6868698000907898, + "learning_rate": 1.7380000000000003e-05, + "loss": 0.0205, + "step": 870 + }, + { + "epoch": 0.0683495145631068, + "grad_norm": 0.5908291935920715, + "learning_rate": 1.758e-05, + "loss": 0.0269, + "step": 880 + }, + { + "epoch": 0.069126213592233, + "grad_norm": 0.525971531867981, + "learning_rate": 1.7780000000000003e-05, + "loss": 0.0225, + "step": 890 + }, + { + "epoch": 0.06990291262135923, + "grad_norm": 0.7407639622688293, + "learning_rate": 1.798e-05, + "loss": 0.0322, + "step": 900 + }, + { + "epoch": 0.07067961165048543, + "grad_norm": 0.7462081909179688, + "learning_rate": 1.818e-05, + "loss": 0.0221, + "step": 910 + }, + { + "epoch": 0.07145631067961165, + "grad_norm": 0.8209527730941772, + "learning_rate": 1.838e-05, + "loss": 0.0294, + "step": 920 + }, + { + "epoch": 0.07223300970873786, + "grad_norm": 0.349356472492218, + "learning_rate": 1.858e-05, + "loss": 0.0245, + "step": 930 + }, + { + "epoch": 0.07300970873786408, + "grad_norm": 0.5736724734306335, + "learning_rate": 1.878e-05, + "loss": 0.026, + "step": 940 + }, + { + "epoch": 0.07378640776699029, + "grad_norm": 0.7484828233718872, + "learning_rate": 1.898e-05, + "loss": 0.0284, + "step": 950 + }, + { + "epoch": 0.07456310679611651, + "grad_norm": 0.8766044974327087, + "learning_rate": 1.918e-05, + "loss": 0.0205, + "step": 960 + }, + { + "epoch": 0.07533980582524272, + "grad_norm": 0.5714200139045715, + "learning_rate": 1.938e-05, + "loss": 0.0234, + "step": 970 + }, + { + "epoch": 0.07611650485436894, + "grad_norm": 0.8250498175621033, + "learning_rate": 1.9580000000000002e-05, + "loss": 0.0234, + "step": 980 + }, + { + "epoch": 0.07689320388349515, + "grad_norm": 0.38166484236717224, + "learning_rate": 1.978e-05, + "loss": 0.0249, + "step": 990 + }, + { + "epoch": 0.07766990291262135, + "grad_norm": 0.644359827041626, + "learning_rate": 1.9980000000000002e-05, + "loss": 0.0275, + "step": 1000 + }, + { + "epoch": 0.07844660194174757, + "grad_norm": 0.47947248816490173, + "learning_rate": 2.0180000000000003e-05, + "loss": 0.0232, + "step": 1010 + }, + { + "epoch": 0.07922330097087378, + "grad_norm": 0.5743876099586487, + "learning_rate": 2.038e-05, + "loss": 0.0215, + "step": 1020 + }, + { + "epoch": 0.08, + "grad_norm": 0.8114569783210754, + "learning_rate": 2.0580000000000003e-05, + "loss": 0.0274, + "step": 1030 + }, + { + "epoch": 0.08077669902912621, + "grad_norm": 0.3896445333957672, + "learning_rate": 2.078e-05, + "loss": 0.0225, + "step": 1040 + }, + { + "epoch": 0.08155339805825243, + "grad_norm": 0.42134711146354675, + "learning_rate": 2.098e-05, + "loss": 0.0231, + "step": 1050 + }, + { + "epoch": 0.08233009708737864, + "grad_norm": 0.8277643322944641, + "learning_rate": 2.118e-05, + "loss": 0.0244, + "step": 1060 + }, + { + "epoch": 0.08310679611650486, + "grad_norm": 0.6429314017295837, + "learning_rate": 2.138e-05, + "loss": 0.0228, + "step": 1070 + }, + { + "epoch": 0.08388349514563107, + "grad_norm": 0.5872289538383484, + "learning_rate": 2.158e-05, + "loss": 0.0261, + "step": 1080 + }, + { + "epoch": 0.08466019417475729, + "grad_norm": 0.37038975954055786, + "learning_rate": 2.178e-05, + "loss": 0.0186, + "step": 1090 + }, + { + "epoch": 0.0854368932038835, + "grad_norm": 0.5239711999893188, + "learning_rate": 2.198e-05, + "loss": 0.0255, + "step": 1100 + }, + { + "epoch": 0.08621359223300971, + "grad_norm": 0.6458615660667419, + "learning_rate": 2.218e-05, + "loss": 0.0228, + "step": 1110 + }, + { + "epoch": 0.08699029126213592, + "grad_norm": 0.6283226609230042, + "learning_rate": 2.2380000000000003e-05, + "loss": 0.0238, + "step": 1120 + }, + { + "epoch": 0.08776699029126214, + "grad_norm": 0.5445438027381897, + "learning_rate": 2.258e-05, + "loss": 0.0274, + "step": 1130 + }, + { + "epoch": 0.08854368932038835, + "grad_norm": 0.6938257217407227, + "learning_rate": 2.2780000000000002e-05, + "loss": 0.0269, + "step": 1140 + }, + { + "epoch": 0.08932038834951456, + "grad_norm": 0.4569568932056427, + "learning_rate": 2.298e-05, + "loss": 0.0233, + "step": 1150 + }, + { + "epoch": 0.09009708737864078, + "grad_norm": 0.3185384273529053, + "learning_rate": 2.318e-05, + "loss": 0.0206, + "step": 1160 + }, + { + "epoch": 0.09087378640776698, + "grad_norm": 0.378008633852005, + "learning_rate": 2.3380000000000003e-05, + "loss": 0.0226, + "step": 1170 + }, + { + "epoch": 0.0916504854368932, + "grad_norm": 0.3828584551811218, + "learning_rate": 2.358e-05, + "loss": 0.0232, + "step": 1180 + }, + { + "epoch": 0.09242718446601941, + "grad_norm": 0.3533577620983124, + "learning_rate": 2.3780000000000003e-05, + "loss": 0.0209, + "step": 1190 + }, + { + "epoch": 0.09320388349514563, + "grad_norm": 0.6514178514480591, + "learning_rate": 2.398e-05, + "loss": 0.0237, + "step": 1200 + }, + { + "epoch": 0.09398058252427184, + "grad_norm": 0.32647907733917236, + "learning_rate": 2.418e-05, + "loss": 0.0192, + "step": 1210 + }, + { + "epoch": 0.09475728155339806, + "grad_norm": 0.6018387675285339, + "learning_rate": 2.438e-05, + "loss": 0.0198, + "step": 1220 + }, + { + "epoch": 0.09553398058252427, + "grad_norm": 0.6085636019706726, + "learning_rate": 2.4580000000000002e-05, + "loss": 0.0253, + "step": 1230 + }, + { + "epoch": 0.09631067961165049, + "grad_norm": 0.6473960876464844, + "learning_rate": 2.478e-05, + "loss": 0.0175, + "step": 1240 + }, + { + "epoch": 0.0970873786407767, + "grad_norm": 0.5022708773612976, + "learning_rate": 2.498e-05, + "loss": 0.0203, + "step": 1250 + }, + { + "epoch": 0.09786407766990292, + "grad_norm": 0.60938560962677, + "learning_rate": 2.5180000000000003e-05, + "loss": 0.0222, + "step": 1260 + }, + { + "epoch": 0.09864077669902913, + "grad_norm": 0.3441193401813507, + "learning_rate": 2.5380000000000004e-05, + "loss": 0.0253, + "step": 1270 + }, + { + "epoch": 0.09941747572815535, + "grad_norm": 0.5875341296195984, + "learning_rate": 2.5580000000000002e-05, + "loss": 0.0224, + "step": 1280 + }, + { + "epoch": 0.10019417475728155, + "grad_norm": 0.8019502758979797, + "learning_rate": 2.5779999999999997e-05, + "loss": 0.0257, + "step": 1290 + }, + { + "epoch": 0.10097087378640776, + "grad_norm": 0.42060619592666626, + "learning_rate": 2.598e-05, + "loss": 0.0228, + "step": 1300 + }, + { + "epoch": 0.10174757281553398, + "grad_norm": 0.271801233291626, + "learning_rate": 2.618e-05, + "loss": 0.0202, + "step": 1310 + }, + { + "epoch": 0.10252427184466019, + "grad_norm": 0.5958484411239624, + "learning_rate": 2.6379999999999998e-05, + "loss": 0.0232, + "step": 1320 + }, + { + "epoch": 0.10330097087378641, + "grad_norm": 0.327237993478775, + "learning_rate": 2.658e-05, + "loss": 0.0188, + "step": 1330 + }, + { + "epoch": 0.10407766990291262, + "grad_norm": 0.5153220295906067, + "learning_rate": 2.678e-05, + "loss": 0.0207, + "step": 1340 + }, + { + "epoch": 0.10485436893203884, + "grad_norm": 0.3281404674053192, + "learning_rate": 2.698e-05, + "loss": 0.0185, + "step": 1350 + }, + { + "epoch": 0.10563106796116505, + "grad_norm": 0.3613488972187042, + "learning_rate": 2.718e-05, + "loss": 0.0229, + "step": 1360 + }, + { + "epoch": 0.10640776699029127, + "grad_norm": 0.4571668207645416, + "learning_rate": 2.738e-05, + "loss": 0.0237, + "step": 1370 + }, + { + "epoch": 0.10718446601941747, + "grad_norm": 0.47137635946273804, + "learning_rate": 2.758e-05, + "loss": 0.0212, + "step": 1380 + }, + { + "epoch": 0.1079611650485437, + "grad_norm": 0.5060802698135376, + "learning_rate": 2.778e-05, + "loss": 0.0191, + "step": 1390 + }, + { + "epoch": 0.1087378640776699, + "grad_norm": 0.5492449402809143, + "learning_rate": 2.798e-05, + "loss": 0.0265, + "step": 1400 + }, + { + "epoch": 0.10951456310679612, + "grad_norm": 0.6164726614952087, + "learning_rate": 2.818e-05, + "loss": 0.0165, + "step": 1410 + }, + { + "epoch": 0.11029126213592233, + "grad_norm": 0.7016597986221313, + "learning_rate": 2.8380000000000003e-05, + "loss": 0.0221, + "step": 1420 + }, + { + "epoch": 0.11106796116504854, + "grad_norm": 0.5741690397262573, + "learning_rate": 2.858e-05, + "loss": 0.0259, + "step": 1430 + }, + { + "epoch": 0.11184466019417476, + "grad_norm": 0.40217480063438416, + "learning_rate": 2.8780000000000002e-05, + "loss": 0.0211, + "step": 1440 + }, + { + "epoch": 0.11262135922330097, + "grad_norm": 0.3694401979446411, + "learning_rate": 2.898e-05, + "loss": 0.0226, + "step": 1450 + }, + { + "epoch": 0.11339805825242719, + "grad_norm": 0.6020891666412354, + "learning_rate": 2.9180000000000002e-05, + "loss": 0.0239, + "step": 1460 + }, + { + "epoch": 0.1141747572815534, + "grad_norm": 0.5407059192657471, + "learning_rate": 2.9380000000000003e-05, + "loss": 0.0193, + "step": 1470 + }, + { + "epoch": 0.11495145631067961, + "grad_norm": 0.4386608600616455, + "learning_rate": 2.958e-05, + "loss": 0.0201, + "step": 1480 + }, + { + "epoch": 0.11572815533980582, + "grad_norm": 0.37629207968711853, + "learning_rate": 2.9780000000000003e-05, + "loss": 0.0191, + "step": 1490 + }, + { + "epoch": 0.11650485436893204, + "grad_norm": 0.6409158706665039, + "learning_rate": 2.998e-05, + "loss": 0.03, + "step": 1500 + }, + { + "epoch": 0.11728155339805825, + "grad_norm": 0.7925148606300354, + "learning_rate": 3.0180000000000002e-05, + "loss": 0.0198, + "step": 1510 + }, + { + "epoch": 0.11805825242718447, + "grad_norm": 0.5788540840148926, + "learning_rate": 3.0380000000000004e-05, + "loss": 0.0233, + "step": 1520 + }, + { + "epoch": 0.11883495145631068, + "grad_norm": 0.6679286360740662, + "learning_rate": 3.058e-05, + "loss": 0.0246, + "step": 1530 + }, + { + "epoch": 0.1196116504854369, + "grad_norm": 0.674258291721344, + "learning_rate": 3.078e-05, + "loss": 0.0284, + "step": 1540 + }, + { + "epoch": 0.1203883495145631, + "grad_norm": 0.44480377435684204, + "learning_rate": 3.0980000000000005e-05, + "loss": 0.0218, + "step": 1550 + }, + { + "epoch": 0.12116504854368933, + "grad_norm": 0.44608497619628906, + "learning_rate": 3.118e-05, + "loss": 0.0206, + "step": 1560 + }, + { + "epoch": 0.12194174757281553, + "grad_norm": 0.3472282886505127, + "learning_rate": 3.138e-05, + "loss": 0.0175, + "step": 1570 + }, + { + "epoch": 0.12271844660194174, + "grad_norm": 0.4155532717704773, + "learning_rate": 3.1580000000000006e-05, + "loss": 0.0228, + "step": 1580 + }, + { + "epoch": 0.12349514563106796, + "grad_norm": 0.5441990494728088, + "learning_rate": 3.1780000000000004e-05, + "loss": 0.0257, + "step": 1590 + }, + { + "epoch": 0.12427184466019417, + "grad_norm": 0.7756466269493103, + "learning_rate": 3.198e-05, + "loss": 0.0207, + "step": 1600 + }, + { + "epoch": 0.1250485436893204, + "grad_norm": 0.6265655159950256, + "learning_rate": 3.218e-05, + "loss": 0.0238, + "step": 1610 + }, + { + "epoch": 0.1258252427184466, + "grad_norm": 0.8335252404212952, + "learning_rate": 3.238e-05, + "loss": 0.0188, + "step": 1620 + }, + { + "epoch": 0.1266019417475728, + "grad_norm": 0.7005942463874817, + "learning_rate": 3.2579999999999996e-05, + "loss": 0.0231, + "step": 1630 + }, + { + "epoch": 0.12737864077669903, + "grad_norm": 0.4774218499660492, + "learning_rate": 3.278e-05, + "loss": 0.0246, + "step": 1640 + }, + { + "epoch": 0.12815533980582525, + "grad_norm": 0.5572545528411865, + "learning_rate": 3.298e-05, + "loss": 0.0266, + "step": 1650 + }, + { + "epoch": 0.12893203883495147, + "grad_norm": 0.6186349987983704, + "learning_rate": 3.318e-05, + "loss": 0.0168, + "step": 1660 + }, + { + "epoch": 0.12970873786407766, + "grad_norm": 0.4300828278064728, + "learning_rate": 3.338e-05, + "loss": 0.0185, + "step": 1670 + }, + { + "epoch": 0.13048543689320388, + "grad_norm": 0.4023299515247345, + "learning_rate": 3.358e-05, + "loss": 0.0274, + "step": 1680 + }, + { + "epoch": 0.1312621359223301, + "grad_norm": 0.3577146530151367, + "learning_rate": 3.378e-05, + "loss": 0.0257, + "step": 1690 + }, + { + "epoch": 0.13203883495145632, + "grad_norm": 0.4407608211040497, + "learning_rate": 3.398e-05, + "loss": 0.0156, + "step": 1700 + }, + { + "epoch": 0.13281553398058252, + "grad_norm": 0.7353548407554626, + "learning_rate": 3.418e-05, + "loss": 0.0203, + "step": 1710 + }, + { + "epoch": 0.13359223300970874, + "grad_norm": 0.3020953834056854, + "learning_rate": 3.438e-05, + "loss": 0.0247, + "step": 1720 + }, + { + "epoch": 0.13436893203883496, + "grad_norm": 0.5842857956886292, + "learning_rate": 3.4580000000000004e-05, + "loss": 0.0149, + "step": 1730 + }, + { + "epoch": 0.13514563106796115, + "grad_norm": 0.3752085566520691, + "learning_rate": 3.478e-05, + "loss": 0.0177, + "step": 1740 + }, + { + "epoch": 0.13592233009708737, + "grad_norm": 0.621293306350708, + "learning_rate": 3.498e-05, + "loss": 0.0219, + "step": 1750 + }, + { + "epoch": 0.1366990291262136, + "grad_norm": 0.40045467019081116, + "learning_rate": 3.518e-05, + "loss": 0.0173, + "step": 1760 + }, + { + "epoch": 0.13747572815533982, + "grad_norm": 0.5520972609519958, + "learning_rate": 3.5380000000000003e-05, + "loss": 0.0223, + "step": 1770 + }, + { + "epoch": 0.138252427184466, + "grad_norm": 0.47574517130851746, + "learning_rate": 3.558e-05, + "loss": 0.0201, + "step": 1780 + }, + { + "epoch": 0.13902912621359223, + "grad_norm": 0.41869792342185974, + "learning_rate": 3.578e-05, + "loss": 0.0203, + "step": 1790 + }, + { + "epoch": 0.13980582524271845, + "grad_norm": 0.3977649211883545, + "learning_rate": 3.5980000000000004e-05, + "loss": 0.0205, + "step": 1800 + }, + { + "epoch": 0.14058252427184467, + "grad_norm": 0.4669925570487976, + "learning_rate": 3.618e-05, + "loss": 0.0176, + "step": 1810 + }, + { + "epoch": 0.14135922330097087, + "grad_norm": 0.37322601675987244, + "learning_rate": 3.638e-05, + "loss": 0.0163, + "step": 1820 + }, + { + "epoch": 0.1421359223300971, + "grad_norm": 0.5274034738540649, + "learning_rate": 3.6580000000000006e-05, + "loss": 0.0162, + "step": 1830 + }, + { + "epoch": 0.1429126213592233, + "grad_norm": 0.3544410169124603, + "learning_rate": 3.6780000000000004e-05, + "loss": 0.0167, + "step": 1840 + }, + { + "epoch": 0.1436893203883495, + "grad_norm": 0.3355095386505127, + "learning_rate": 3.698e-05, + "loss": 0.0223, + "step": 1850 + }, + { + "epoch": 0.14446601941747572, + "grad_norm": 0.44155704975128174, + "learning_rate": 3.7180000000000007e-05, + "loss": 0.0213, + "step": 1860 + }, + { + "epoch": 0.14524271844660194, + "grad_norm": 0.40560266375541687, + "learning_rate": 3.7380000000000005e-05, + "loss": 0.0252, + "step": 1870 + }, + { + "epoch": 0.14601941747572816, + "grad_norm": 0.3546150326728821, + "learning_rate": 3.758e-05, + "loss": 0.0165, + "step": 1880 + }, + { + "epoch": 0.14679611650485436, + "grad_norm": 0.5608948469161987, + "learning_rate": 3.778000000000001e-05, + "loss": 0.0188, + "step": 1890 + }, + { + "epoch": 0.14757281553398058, + "grad_norm": 0.3311212658882141, + "learning_rate": 3.7980000000000006e-05, + "loss": 0.0214, + "step": 1900 + }, + { + "epoch": 0.1483495145631068, + "grad_norm": 0.7319339513778687, + "learning_rate": 3.818e-05, + "loss": 0.0193, + "step": 1910 + }, + { + "epoch": 0.14912621359223302, + "grad_norm": 0.5165913701057434, + "learning_rate": 3.838e-05, + "loss": 0.0179, + "step": 1920 + }, + { + "epoch": 0.1499029126213592, + "grad_norm": 0.4305424392223358, + "learning_rate": 3.858e-05, + "loss": 0.0192, + "step": 1930 + }, + { + "epoch": 0.15067961165048543, + "grad_norm": 0.5532487034797668, + "learning_rate": 3.878e-05, + "loss": 0.0202, + "step": 1940 + }, + { + "epoch": 0.15145631067961166, + "grad_norm": 0.42630642652511597, + "learning_rate": 3.898e-05, + "loss": 0.022, + "step": 1950 + }, + { + "epoch": 0.15223300970873788, + "grad_norm": 0.3187502920627594, + "learning_rate": 3.918e-05, + "loss": 0.0165, + "step": 1960 + }, + { + "epoch": 0.15300970873786407, + "grad_norm": 0.3441694974899292, + "learning_rate": 3.938e-05, + "loss": 0.0179, + "step": 1970 + }, + { + "epoch": 0.1537864077669903, + "grad_norm": 0.4266280233860016, + "learning_rate": 3.958e-05, + "loss": 0.0252, + "step": 1980 + }, + { + "epoch": 0.1545631067961165, + "grad_norm": 0.7132630944252014, + "learning_rate": 3.978e-05, + "loss": 0.0193, + "step": 1990 + }, + { + "epoch": 0.1553398058252427, + "grad_norm": 0.5161796808242798, + "learning_rate": 3.998e-05, + "loss": 0.0194, + "step": 2000 + }, + { + "epoch": 0.15611650485436893, + "grad_norm": 0.7571023106575012, + "learning_rate": 4.018e-05, + "loss": 0.017, + "step": 2010 + }, + { + "epoch": 0.15689320388349515, + "grad_norm": 0.5362123250961304, + "learning_rate": 4.038e-05, + "loss": 0.0247, + "step": 2020 + }, + { + "epoch": 0.15766990291262137, + "grad_norm": 0.4352453351020813, + "learning_rate": 4.058e-05, + "loss": 0.0213, + "step": 2030 + }, + { + "epoch": 0.15844660194174756, + "grad_norm": 0.3270867168903351, + "learning_rate": 4.078e-05, + "loss": 0.0229, + "step": 2040 + }, + { + "epoch": 0.15922330097087378, + "grad_norm": 0.371931791305542, + "learning_rate": 4.0980000000000004e-05, + "loss": 0.0262, + "step": 2050 + }, + { + "epoch": 0.16, + "grad_norm": 0.3360389471054077, + "learning_rate": 4.118e-05, + "loss": 0.0173, + "step": 2060 + }, + { + "epoch": 0.16077669902912622, + "grad_norm": 0.4999483823776245, + "learning_rate": 4.138e-05, + "loss": 0.0153, + "step": 2070 + }, + { + "epoch": 0.16155339805825242, + "grad_norm": 0.25666218996047974, + "learning_rate": 4.1580000000000005e-05, + "loss": 0.0178, + "step": 2080 + }, + { + "epoch": 0.16233009708737864, + "grad_norm": 0.3969675600528717, + "learning_rate": 4.178e-05, + "loss": 0.0227, + "step": 2090 + }, + { + "epoch": 0.16310679611650486, + "grad_norm": 0.3376440107822418, + "learning_rate": 4.198e-05, + "loss": 0.0164, + "step": 2100 + }, + { + "epoch": 0.16388349514563108, + "grad_norm": 0.32502076029777527, + "learning_rate": 4.2180000000000006e-05, + "loss": 0.0186, + "step": 2110 + }, + { + "epoch": 0.16466019417475727, + "grad_norm": 0.5067673921585083, + "learning_rate": 4.2380000000000004e-05, + "loss": 0.0202, + "step": 2120 + }, + { + "epoch": 0.1654368932038835, + "grad_norm": 0.3823685646057129, + "learning_rate": 4.258e-05, + "loss": 0.0154, + "step": 2130 + }, + { + "epoch": 0.16621359223300972, + "grad_norm": 0.44533318281173706, + "learning_rate": 4.278e-05, + "loss": 0.0173, + "step": 2140 + }, + { + "epoch": 0.1669902912621359, + "grad_norm": 0.5523590445518494, + "learning_rate": 4.2980000000000005e-05, + "loss": 0.016, + "step": 2150 + }, + { + "epoch": 0.16776699029126213, + "grad_norm": 0.4371185004711151, + "learning_rate": 4.318e-05, + "loss": 0.0183, + "step": 2160 + }, + { + "epoch": 0.16854368932038835, + "grad_norm": 0.562982439994812, + "learning_rate": 4.338e-05, + "loss": 0.0184, + "step": 2170 + }, + { + "epoch": 0.16932038834951457, + "grad_norm": 0.3983249068260193, + "learning_rate": 4.3580000000000006e-05, + "loss": 0.018, + "step": 2180 + }, + { + "epoch": 0.17009708737864077, + "grad_norm": 0.46984612941741943, + "learning_rate": 4.3780000000000004e-05, + "loss": 0.0191, + "step": 2190 + }, + { + "epoch": 0.170873786407767, + "grad_norm": 0.30866819620132446, + "learning_rate": 4.398e-05, + "loss": 0.0194, + "step": 2200 + }, + { + "epoch": 0.1716504854368932, + "grad_norm": 0.5054110884666443, + "learning_rate": 4.418000000000001e-05, + "loss": 0.0141, + "step": 2210 + }, + { + "epoch": 0.17242718446601943, + "grad_norm": 0.37523916363716125, + "learning_rate": 4.438e-05, + "loss": 0.0224, + "step": 2220 + }, + { + "epoch": 0.17320388349514562, + "grad_norm": 0.5044036507606506, + "learning_rate": 4.458e-05, + "loss": 0.0189, + "step": 2230 + }, + { + "epoch": 0.17398058252427184, + "grad_norm": 0.4628452956676483, + "learning_rate": 4.478e-05, + "loss": 0.0183, + "step": 2240 + }, + { + "epoch": 0.17475728155339806, + "grad_norm": 0.336635023355484, + "learning_rate": 4.498e-05, + "loss": 0.0223, + "step": 2250 + }, + { + "epoch": 0.17553398058252428, + "grad_norm": 0.5211383104324341, + "learning_rate": 4.518e-05, + "loss": 0.0204, + "step": 2260 + }, + { + "epoch": 0.17631067961165048, + "grad_norm": 0.47032609581947327, + "learning_rate": 4.538e-05, + "loss": 0.0148, + "step": 2270 + }, + { + "epoch": 0.1770873786407767, + "grad_norm": 0.5553898811340332, + "learning_rate": 4.558e-05, + "loss": 0.0211, + "step": 2280 + }, + { + "epoch": 0.17786407766990292, + "grad_norm": 0.8387870788574219, + "learning_rate": 4.578e-05, + "loss": 0.022, + "step": 2290 + }, + { + "epoch": 0.1786407766990291, + "grad_norm": 0.45727071166038513, + "learning_rate": 4.5980000000000004e-05, + "loss": 0.0188, + "step": 2300 + }, + { + "epoch": 0.17941747572815533, + "grad_norm": 0.33342084288597107, + "learning_rate": 4.618e-05, + "loss": 0.0225, + "step": 2310 + }, + { + "epoch": 0.18019417475728156, + "grad_norm": 0.3486401438713074, + "learning_rate": 4.638e-05, + "loss": 0.018, + "step": 2320 + }, + { + "epoch": 0.18097087378640778, + "grad_norm": 0.4111383855342865, + "learning_rate": 4.6580000000000005e-05, + "loss": 0.026, + "step": 2330 + }, + { + "epoch": 0.18174757281553397, + "grad_norm": 0.2913767993450165, + "learning_rate": 4.678e-05, + "loss": 0.0155, + "step": 2340 + }, + { + "epoch": 0.1825242718446602, + "grad_norm": 0.4573994576931, + "learning_rate": 4.698e-05, + "loss": 0.0223, + "step": 2350 + }, + { + "epoch": 0.1833009708737864, + "grad_norm": 0.3871082067489624, + "learning_rate": 4.718e-05, + "loss": 0.0194, + "step": 2360 + }, + { + "epoch": 0.18407766990291263, + "grad_norm": 0.43490198254585266, + "learning_rate": 4.7380000000000004e-05, + "loss": 0.0194, + "step": 2370 + }, + { + "epoch": 0.18485436893203883, + "grad_norm": 0.47029393911361694, + "learning_rate": 4.758e-05, + "loss": 0.0168, + "step": 2380 + }, + { + "epoch": 0.18563106796116505, + "grad_norm": 0.3664845824241638, + "learning_rate": 4.778e-05, + "loss": 0.0173, + "step": 2390 + }, + { + "epoch": 0.18640776699029127, + "grad_norm": 0.5873512029647827, + "learning_rate": 4.7980000000000005e-05, + "loss": 0.0212, + "step": 2400 + }, + { + "epoch": 0.1871844660194175, + "grad_norm": 0.6060754656791687, + "learning_rate": 4.818e-05, + "loss": 0.0166, + "step": 2410 + }, + { + "epoch": 0.18796116504854368, + "grad_norm": 0.633400559425354, + "learning_rate": 4.838e-05, + "loss": 0.0212, + "step": 2420 + }, + { + "epoch": 0.1887378640776699, + "grad_norm": 0.3529169261455536, + "learning_rate": 4.8580000000000006e-05, + "loss": 0.0169, + "step": 2430 + }, + { + "epoch": 0.18951456310679612, + "grad_norm": 0.5590711832046509, + "learning_rate": 4.8780000000000004e-05, + "loss": 0.0167, + "step": 2440 + }, + { + "epoch": 0.19029126213592232, + "grad_norm": 0.3475462794303894, + "learning_rate": 4.898e-05, + "loss": 0.0229, + "step": 2450 + }, + { + "epoch": 0.19106796116504854, + "grad_norm": 0.5693970322608948, + "learning_rate": 4.918000000000001e-05, + "loss": 0.0241, + "step": 2460 + }, + { + "epoch": 0.19184466019417476, + "grad_norm": 0.7320810556411743, + "learning_rate": 4.9380000000000005e-05, + "loss": 0.0263, + "step": 2470 + }, + { + "epoch": 0.19262135922330098, + "grad_norm": 0.5333056449890137, + "learning_rate": 4.958e-05, + "loss": 0.0201, + "step": 2480 + }, + { + "epoch": 0.19339805825242717, + "grad_norm": 0.24981577694416046, + "learning_rate": 4.978e-05, + "loss": 0.0144, + "step": 2490 + }, + { + "epoch": 0.1941747572815534, + "grad_norm": 0.28329482674598694, + "learning_rate": 4.9980000000000006e-05, + "loss": 0.0162, + "step": 2500 + }, + { + "epoch": 0.19495145631067962, + "grad_norm": 0.4129181206226349, + "learning_rate": 5.0180000000000004e-05, + "loss": 0.0137, + "step": 2510 + }, + { + "epoch": 0.19572815533980584, + "grad_norm": 0.4096432328224182, + "learning_rate": 5.038e-05, + "loss": 0.0153, + "step": 2520 + }, + { + "epoch": 0.19650485436893203, + "grad_norm": 0.316420316696167, + "learning_rate": 5.058000000000001e-05, + "loss": 0.0162, + "step": 2530 + }, + { + "epoch": 0.19728155339805825, + "grad_norm": 0.6336057186126709, + "learning_rate": 5.0780000000000005e-05, + "loss": 0.0215, + "step": 2540 + }, + { + "epoch": 0.19805825242718447, + "grad_norm": 0.42250218987464905, + "learning_rate": 5.098e-05, + "loss": 0.0165, + "step": 2550 + }, + { + "epoch": 0.1988349514563107, + "grad_norm": 0.4408728778362274, + "learning_rate": 5.118000000000001e-05, + "loss": 0.0177, + "step": 2560 + }, + { + "epoch": 0.1996116504854369, + "grad_norm": 0.397533655166626, + "learning_rate": 5.1380000000000006e-05, + "loss": 0.016, + "step": 2570 + }, + { + "epoch": 0.2003883495145631, + "grad_norm": 0.34488096833229065, + "learning_rate": 5.1580000000000004e-05, + "loss": 0.0187, + "step": 2580 + }, + { + "epoch": 0.20116504854368933, + "grad_norm": 0.41036778688430786, + "learning_rate": 5.178000000000001e-05, + "loss": 0.0217, + "step": 2590 + }, + { + "epoch": 0.20194174757281552, + "grad_norm": 0.6092613339424133, + "learning_rate": 5.198000000000001e-05, + "loss": 0.0167, + "step": 2600 + }, + { + "epoch": 0.20271844660194174, + "grad_norm": 0.5266022086143494, + "learning_rate": 5.2180000000000005e-05, + "loss": 0.0169, + "step": 2610 + }, + { + "epoch": 0.20349514563106796, + "grad_norm": 0.5062000751495361, + "learning_rate": 5.238000000000001e-05, + "loss": 0.0197, + "step": 2620 + }, + { + "epoch": 0.20427184466019419, + "grad_norm": 0.6133871674537659, + "learning_rate": 5.258000000000001e-05, + "loss": 0.0187, + "step": 2630 + }, + { + "epoch": 0.20504854368932038, + "grad_norm": 0.5988008975982666, + "learning_rate": 5.2780000000000006e-05, + "loss": 0.0196, + "step": 2640 + }, + { + "epoch": 0.2058252427184466, + "grad_norm": 0.38887819647789, + "learning_rate": 5.2980000000000004e-05, + "loss": 0.0249, + "step": 2650 + }, + { + "epoch": 0.20660194174757282, + "grad_norm": 0.43990957736968994, + "learning_rate": 5.318000000000001e-05, + "loss": 0.0213, + "step": 2660 + }, + { + "epoch": 0.20737864077669904, + "grad_norm": 0.32881081104278564, + "learning_rate": 5.338000000000001e-05, + "loss": 0.0173, + "step": 2670 + }, + { + "epoch": 0.20815533980582523, + "grad_norm": 0.3500358462333679, + "learning_rate": 5.3580000000000005e-05, + "loss": 0.0299, + "step": 2680 + }, + { + "epoch": 0.20893203883495146, + "grad_norm": 0.35004130005836487, + "learning_rate": 5.378e-05, + "loss": 0.0239, + "step": 2690 + }, + { + "epoch": 0.20970873786407768, + "grad_norm": 0.47113099694252014, + "learning_rate": 5.3979999999999995e-05, + "loss": 0.0208, + "step": 2700 + }, + { + "epoch": 0.21048543689320387, + "grad_norm": 0.334966778755188, + "learning_rate": 5.418e-05, + "loss": 0.023, + "step": 2710 + }, + { + "epoch": 0.2112621359223301, + "grad_norm": 0.3251389265060425, + "learning_rate": 5.438e-05, + "loss": 0.0223, + "step": 2720 + }, + { + "epoch": 0.2120388349514563, + "grad_norm": 0.49215778708457947, + "learning_rate": 5.4579999999999996e-05, + "loss": 0.0179, + "step": 2730 + }, + { + "epoch": 0.21281553398058253, + "grad_norm": 0.31934648752212524, + "learning_rate": 5.478e-05, + "loss": 0.0171, + "step": 2740 + }, + { + "epoch": 0.21359223300970873, + "grad_norm": 0.24403659999370575, + "learning_rate": 5.498e-05, + "loss": 0.0189, + "step": 2750 + }, + { + "epoch": 0.21436893203883495, + "grad_norm": 0.38731658458709717, + "learning_rate": 5.518e-05, + "loss": 0.0193, + "step": 2760 + }, + { + "epoch": 0.21514563106796117, + "grad_norm": 0.5696061253547668, + "learning_rate": 5.538e-05, + "loss": 0.0155, + "step": 2770 + }, + { + "epoch": 0.2159223300970874, + "grad_norm": 0.5511022210121155, + "learning_rate": 5.558e-05, + "loss": 0.0188, + "step": 2780 + }, + { + "epoch": 0.21669902912621358, + "grad_norm": 0.7445398569107056, + "learning_rate": 5.578e-05, + "loss": 0.0215, + "step": 2790 + }, + { + "epoch": 0.2174757281553398, + "grad_norm": 0.4101894199848175, + "learning_rate": 5.5979999999999996e-05, + "loss": 0.0209, + "step": 2800 + }, + { + "epoch": 0.21825242718446602, + "grad_norm": 0.4705612361431122, + "learning_rate": 5.618e-05, + "loss": 0.0183, + "step": 2810 + }, + { + "epoch": 0.21902912621359225, + "grad_norm": 0.5318827033042908, + "learning_rate": 5.638e-05, + "loss": 0.0177, + "step": 2820 + }, + { + "epoch": 0.21980582524271844, + "grad_norm": 0.35572463274002075, + "learning_rate": 5.658e-05, + "loss": 0.0226, + "step": 2830 + }, + { + "epoch": 0.22058252427184466, + "grad_norm": 0.4338955581188202, + "learning_rate": 5.678e-05, + "loss": 0.0234, + "step": 2840 + }, + { + "epoch": 0.22135922330097088, + "grad_norm": 0.43736112117767334, + "learning_rate": 5.698e-05, + "loss": 0.0216, + "step": 2850 + }, + { + "epoch": 0.22213592233009707, + "grad_norm": 0.47033926844596863, + "learning_rate": 5.718e-05, + "loss": 0.0191, + "step": 2860 + }, + { + "epoch": 0.2229126213592233, + "grad_norm": 0.45534801483154297, + "learning_rate": 5.738e-05, + "loss": 0.0221, + "step": 2870 + }, + { + "epoch": 0.22368932038834952, + "grad_norm": 0.4347730278968811, + "learning_rate": 5.758e-05, + "loss": 0.0183, + "step": 2880 + }, + { + "epoch": 0.22446601941747574, + "grad_norm": 0.26509711146354675, + "learning_rate": 5.778e-05, + "loss": 0.0227, + "step": 2890 + }, + { + "epoch": 0.22524271844660193, + "grad_norm": 0.2539368271827698, + "learning_rate": 5.7980000000000004e-05, + "loss": 0.0185, + "step": 2900 + }, + { + "epoch": 0.22601941747572815, + "grad_norm": 0.3634597361087799, + "learning_rate": 5.818e-05, + "loss": 0.0272, + "step": 2910 + }, + { + "epoch": 0.22679611650485437, + "grad_norm": 0.49352386593818665, + "learning_rate": 5.838e-05, + "loss": 0.0219, + "step": 2920 + }, + { + "epoch": 0.2275728155339806, + "grad_norm": 0.3380693197250366, + "learning_rate": 5.858e-05, + "loss": 0.0183, + "step": 2930 + }, + { + "epoch": 0.2283495145631068, + "grad_norm": 0.3729790151119232, + "learning_rate": 5.878e-05, + "loss": 0.0194, + "step": 2940 + }, + { + "epoch": 0.229126213592233, + "grad_norm": 0.40029752254486084, + "learning_rate": 5.898e-05, + "loss": 0.0217, + "step": 2950 + }, + { + "epoch": 0.22990291262135923, + "grad_norm": 0.4819222688674927, + "learning_rate": 5.918e-05, + "loss": 0.021, + "step": 2960 + }, + { + "epoch": 0.23067961165048545, + "grad_norm": 0.4361971914768219, + "learning_rate": 5.9380000000000004e-05, + "loss": 0.0218, + "step": 2970 + }, + { + "epoch": 0.23145631067961164, + "grad_norm": 0.3450440466403961, + "learning_rate": 5.958e-05, + "loss": 0.015, + "step": 2980 + }, + { + "epoch": 0.23223300970873786, + "grad_norm": 0.3759114444255829, + "learning_rate": 5.978e-05, + "loss": 0.0148, + "step": 2990 + }, + { + "epoch": 0.23300970873786409, + "grad_norm": 0.839901328086853, + "learning_rate": 5.9980000000000005e-05, + "loss": 0.0201, + "step": 3000 + }, + { + "epoch": 0.23378640776699028, + "grad_norm": 0.49864163994789124, + "learning_rate": 6.018e-05, + "loss": 0.0219, + "step": 3010 + }, + { + "epoch": 0.2345631067961165, + "grad_norm": 0.43306222558021545, + "learning_rate": 6.038e-05, + "loss": 0.0193, + "step": 3020 + }, + { + "epoch": 0.23533980582524272, + "grad_norm": 0.5157333612442017, + "learning_rate": 6.0580000000000006e-05, + "loss": 0.0156, + "step": 3030 + }, + { + "epoch": 0.23611650485436894, + "grad_norm": 0.3723127245903015, + "learning_rate": 6.0780000000000004e-05, + "loss": 0.0188, + "step": 3040 + }, + { + "epoch": 0.23689320388349513, + "grad_norm": 0.39674243330955505, + "learning_rate": 6.098e-05, + "loss": 0.0211, + "step": 3050 + }, + { + "epoch": 0.23766990291262136, + "grad_norm": 0.35941198468208313, + "learning_rate": 6.118000000000001e-05, + "loss": 0.018, + "step": 3060 + }, + { + "epoch": 0.23844660194174758, + "grad_norm": 0.3171451985836029, + "learning_rate": 6.138e-05, + "loss": 0.0228, + "step": 3070 + }, + { + "epoch": 0.2392233009708738, + "grad_norm": 0.3887479305267334, + "learning_rate": 6.158e-05, + "loss": 0.0191, + "step": 3080 + }, + { + "epoch": 0.24, + "grad_norm": 0.36531150341033936, + "learning_rate": 6.178000000000001e-05, + "loss": 0.0148, + "step": 3090 + }, + { + "epoch": 0.2407766990291262, + "grad_norm": 0.3980225920677185, + "learning_rate": 6.198e-05, + "loss": 0.016, + "step": 3100 + }, + { + "epoch": 0.24155339805825243, + "grad_norm": 0.29956942796707153, + "learning_rate": 6.218e-05, + "loss": 0.0178, + "step": 3110 + }, + { + "epoch": 0.24233009708737865, + "grad_norm": 0.4057750105857849, + "learning_rate": 6.238000000000001e-05, + "loss": 0.0181, + "step": 3120 + }, + { + "epoch": 0.24310679611650485, + "grad_norm": 0.49086105823516846, + "learning_rate": 6.258e-05, + "loss": 0.0147, + "step": 3130 + }, + { + "epoch": 0.24388349514563107, + "grad_norm": 0.4606112241744995, + "learning_rate": 6.278e-05, + "loss": 0.0188, + "step": 3140 + }, + { + "epoch": 0.2446601941747573, + "grad_norm": 0.3884565532207489, + "learning_rate": 6.298000000000001e-05, + "loss": 0.0166, + "step": 3150 + }, + { + "epoch": 0.24543689320388348, + "grad_norm": 0.24379298090934753, + "learning_rate": 6.318e-05, + "loss": 0.0129, + "step": 3160 + }, + { + "epoch": 0.2462135922330097, + "grad_norm": 0.4172167479991913, + "learning_rate": 6.338e-05, + "loss": 0.0156, + "step": 3170 + }, + { + "epoch": 0.24699029126213592, + "grad_norm": 0.5108050107955933, + "learning_rate": 6.358000000000001e-05, + "loss": 0.0154, + "step": 3180 + }, + { + "epoch": 0.24776699029126215, + "grad_norm": 0.35928598046302795, + "learning_rate": 6.378e-05, + "loss": 0.0176, + "step": 3190 + }, + { + "epoch": 0.24854368932038834, + "grad_norm": 0.5025870203971863, + "learning_rate": 6.398000000000001e-05, + "loss": 0.0208, + "step": 3200 + }, + { + "epoch": 0.24932038834951456, + "grad_norm": 0.2754709720611572, + "learning_rate": 6.418000000000001e-05, + "loss": 0.0169, + "step": 3210 + }, + { + "epoch": 0.2500970873786408, + "grad_norm": 0.4218793213367462, + "learning_rate": 6.438e-05, + "loss": 0.0206, + "step": 3220 + }, + { + "epoch": 0.250873786407767, + "grad_norm": 0.3162703514099121, + "learning_rate": 6.458000000000001e-05, + "loss": 0.0167, + "step": 3230 + }, + { + "epoch": 0.2516504854368932, + "grad_norm": 0.41382262110710144, + "learning_rate": 6.478000000000001e-05, + "loss": 0.0139, + "step": 3240 + }, + { + "epoch": 0.2524271844660194, + "grad_norm": 0.32410240173339844, + "learning_rate": 6.498e-05, + "loss": 0.0152, + "step": 3250 + }, + { + "epoch": 0.2532038834951456, + "grad_norm": 0.42656809091567993, + "learning_rate": 6.518000000000001e-05, + "loss": 0.0206, + "step": 3260 + }, + { + "epoch": 0.25398058252427186, + "grad_norm": 0.3926198482513428, + "learning_rate": 6.538000000000001e-05, + "loss": 0.0189, + "step": 3270 + }, + { + "epoch": 0.25475728155339805, + "grad_norm": 0.3821541965007782, + "learning_rate": 6.558e-05, + "loss": 0.0159, + "step": 3280 + }, + { + "epoch": 0.25553398058252424, + "grad_norm": 0.26252004504203796, + "learning_rate": 6.578000000000001e-05, + "loss": 0.0159, + "step": 3290 + }, + { + "epoch": 0.2563106796116505, + "grad_norm": 0.49831733107566833, + "learning_rate": 6.598e-05, + "loss": 0.0205, + "step": 3300 + }, + { + "epoch": 0.2570873786407767, + "grad_norm": 0.43782299757003784, + "learning_rate": 6.618e-05, + "loss": 0.0188, + "step": 3310 + }, + { + "epoch": 0.25786407766990294, + "grad_norm": 0.2951412498950958, + "learning_rate": 6.638e-05, + "loss": 0.0178, + "step": 3320 + }, + { + "epoch": 0.25864077669902913, + "grad_norm": 0.4172593355178833, + "learning_rate": 6.658e-05, + "loss": 0.0196, + "step": 3330 + }, + { + "epoch": 0.2594174757281553, + "grad_norm": 0.4104170501232147, + "learning_rate": 6.678e-05, + "loss": 0.0173, + "step": 3340 + }, + { + "epoch": 0.26019417475728157, + "grad_norm": 0.46474599838256836, + "learning_rate": 6.698e-05, + "loss": 0.0194, + "step": 3350 + }, + { + "epoch": 0.26097087378640776, + "grad_norm": 0.29106003046035767, + "learning_rate": 6.718e-05, + "loss": 0.0178, + "step": 3360 + }, + { + "epoch": 0.26174757281553396, + "grad_norm": 0.3541705906391144, + "learning_rate": 6.738e-05, + "loss": 0.027, + "step": 3370 + }, + { + "epoch": 0.2625242718446602, + "grad_norm": 0.38226065039634705, + "learning_rate": 6.758e-05, + "loss": 0.0249, + "step": 3380 + }, + { + "epoch": 0.2633009708737864, + "grad_norm": 0.43570807576179504, + "learning_rate": 6.778e-05, + "loss": 0.023, + "step": 3390 + }, + { + "epoch": 0.26407766990291265, + "grad_norm": 0.47084301710128784, + "learning_rate": 6.798e-05, + "loss": 0.0196, + "step": 3400 + }, + { + "epoch": 0.26485436893203884, + "grad_norm": 0.453583687543869, + "learning_rate": 6.818e-05, + "loss": 0.0172, + "step": 3410 + }, + { + "epoch": 0.26563106796116503, + "grad_norm": 0.28236234188079834, + "learning_rate": 6.838e-05, + "loss": 0.02, + "step": 3420 + }, + { + "epoch": 0.2664077669902913, + "grad_norm": 0.38836631178855896, + "learning_rate": 6.858e-05, + "loss": 0.016, + "step": 3430 + }, + { + "epoch": 0.2671844660194175, + "grad_norm": 0.42074596881866455, + "learning_rate": 6.878e-05, + "loss": 0.02, + "step": 3440 + }, + { + "epoch": 0.26796116504854367, + "grad_norm": 0.4360883831977844, + "learning_rate": 6.898e-05, + "loss": 0.013, + "step": 3450 + }, + { + "epoch": 0.2687378640776699, + "grad_norm": 0.34713801741600037, + "learning_rate": 6.918e-05, + "loss": 0.024, + "step": 3460 + }, + { + "epoch": 0.2695145631067961, + "grad_norm": 0.49476709961891174, + "learning_rate": 6.938e-05, + "loss": 0.0207, + "step": 3470 + }, + { + "epoch": 0.2702912621359223, + "grad_norm": 0.28371506929397583, + "learning_rate": 6.958e-05, + "loss": 0.0174, + "step": 3480 + }, + { + "epoch": 0.27106796116504855, + "grad_norm": 0.2860935926437378, + "learning_rate": 6.978e-05, + "loss": 0.0214, + "step": 3490 + }, + { + "epoch": 0.27184466019417475, + "grad_norm": 0.41058269143104553, + "learning_rate": 6.998e-05, + "loss": 0.0213, + "step": 3500 + }, + { + "epoch": 0.272621359223301, + "grad_norm": 0.4107618033885956, + "learning_rate": 7.018e-05, + "loss": 0.022, + "step": 3510 + }, + { + "epoch": 0.2733980582524272, + "grad_norm": 0.19366790354251862, + "learning_rate": 7.038e-05, + "loss": 0.015, + "step": 3520 + }, + { + "epoch": 0.2741747572815534, + "grad_norm": 0.6456460356712341, + "learning_rate": 7.058e-05, + "loss": 0.0176, + "step": 3530 + }, + { + "epoch": 0.27495145631067963, + "grad_norm": 0.3181043565273285, + "learning_rate": 7.078e-05, + "loss": 0.0192, + "step": 3540 + }, + { + "epoch": 0.2757281553398058, + "grad_norm": 0.4542637765407562, + "learning_rate": 7.098e-05, + "loss": 0.02, + "step": 3550 + }, + { + "epoch": 0.276504854368932, + "grad_norm": 0.47248828411102295, + "learning_rate": 7.118e-05, + "loss": 0.0217, + "step": 3560 + }, + { + "epoch": 0.27728155339805827, + "grad_norm": 0.36596089601516724, + "learning_rate": 7.138e-05, + "loss": 0.0178, + "step": 3570 + }, + { + "epoch": 0.27805825242718446, + "grad_norm": 0.3998449444770813, + "learning_rate": 7.158e-05, + "loss": 0.0195, + "step": 3580 + }, + { + "epoch": 0.27883495145631065, + "grad_norm": 0.35155826807022095, + "learning_rate": 7.178000000000001e-05, + "loss": 0.0171, + "step": 3590 + }, + { + "epoch": 0.2796116504854369, + "grad_norm": 0.3776254951953888, + "learning_rate": 7.198e-05, + "loss": 0.0221, + "step": 3600 + }, + { + "epoch": 0.2803883495145631, + "grad_norm": 0.32399359345436096, + "learning_rate": 7.218e-05, + "loss": 0.0137, + "step": 3610 + }, + { + "epoch": 0.28116504854368934, + "grad_norm": 0.48552531003952026, + "learning_rate": 7.238000000000001e-05, + "loss": 0.0144, + "step": 3620 + }, + { + "epoch": 0.28194174757281554, + "grad_norm": 0.2151721566915512, + "learning_rate": 7.258e-05, + "loss": 0.0145, + "step": 3630 + }, + { + "epoch": 0.28271844660194173, + "grad_norm": 0.4141179025173187, + "learning_rate": 7.278e-05, + "loss": 0.0154, + "step": 3640 + }, + { + "epoch": 0.283495145631068, + "grad_norm": 0.4806191623210907, + "learning_rate": 7.298000000000001e-05, + "loss": 0.0196, + "step": 3650 + }, + { + "epoch": 0.2842718446601942, + "grad_norm": 0.28997567296028137, + "learning_rate": 7.318e-05, + "loss": 0.0138, + "step": 3660 + }, + { + "epoch": 0.28504854368932037, + "grad_norm": 0.41100966930389404, + "learning_rate": 7.338e-05, + "loss": 0.0219, + "step": 3670 + }, + { + "epoch": 0.2858252427184466, + "grad_norm": 0.37427473068237305, + "learning_rate": 7.358000000000001e-05, + "loss": 0.0187, + "step": 3680 + }, + { + "epoch": 0.2866019417475728, + "grad_norm": 0.539168655872345, + "learning_rate": 7.378e-05, + "loss": 0.0281, + "step": 3690 + }, + { + "epoch": 0.287378640776699, + "grad_norm": 0.5274801254272461, + "learning_rate": 7.398e-05, + "loss": 0.0277, + "step": 3700 + }, + { + "epoch": 0.28815533980582525, + "grad_norm": 0.32059428095817566, + "learning_rate": 7.418000000000001e-05, + "loss": 0.0255, + "step": 3710 + }, + { + "epoch": 0.28893203883495144, + "grad_norm": 0.3620586693286896, + "learning_rate": 7.438e-05, + "loss": 0.0209, + "step": 3720 + }, + { + "epoch": 0.2897087378640777, + "grad_norm": 0.37101057171821594, + "learning_rate": 7.458000000000001e-05, + "loss": 0.016, + "step": 3730 + }, + { + "epoch": 0.2904854368932039, + "grad_norm": 0.3104919195175171, + "learning_rate": 7.478e-05, + "loss": 0.0252, + "step": 3740 + }, + { + "epoch": 0.2912621359223301, + "grad_norm": 0.3826545178890228, + "learning_rate": 7.498e-05, + "loss": 0.0158, + "step": 3750 + }, + { + "epoch": 0.2920388349514563, + "grad_norm": 0.37324270606040955, + "learning_rate": 7.518000000000001e-05, + "loss": 0.0177, + "step": 3760 + }, + { + "epoch": 0.2928155339805825, + "grad_norm": 0.3701549768447876, + "learning_rate": 7.538e-05, + "loss": 0.0179, + "step": 3770 + }, + { + "epoch": 0.2935922330097087, + "grad_norm": 0.3148784041404724, + "learning_rate": 7.558e-05, + "loss": 0.0179, + "step": 3780 + }, + { + "epoch": 0.29436893203883496, + "grad_norm": 0.4308188855648041, + "learning_rate": 7.578000000000001e-05, + "loss": 0.0185, + "step": 3790 + }, + { + "epoch": 0.29514563106796116, + "grad_norm": 0.6232790350914001, + "learning_rate": 7.598e-05, + "loss": 0.0202, + "step": 3800 + }, + { + "epoch": 0.2959223300970874, + "grad_norm": 0.3049946129322052, + "learning_rate": 7.618e-05, + "loss": 0.0184, + "step": 3810 + }, + { + "epoch": 0.2966990291262136, + "grad_norm": 0.3617972135543823, + "learning_rate": 7.638000000000001e-05, + "loss": 0.0265, + "step": 3820 + }, + { + "epoch": 0.2974757281553398, + "grad_norm": 0.4923469126224518, + "learning_rate": 7.658e-05, + "loss": 0.0194, + "step": 3830 + }, + { + "epoch": 0.29825242718446604, + "grad_norm": 0.33236995339393616, + "learning_rate": 7.678000000000001e-05, + "loss": 0.0187, + "step": 3840 + }, + { + "epoch": 0.29902912621359223, + "grad_norm": 0.45059317350387573, + "learning_rate": 7.698000000000001e-05, + "loss": 0.0199, + "step": 3850 + }, + { + "epoch": 0.2998058252427184, + "grad_norm": 0.4633493721485138, + "learning_rate": 7.718e-05, + "loss": 0.0231, + "step": 3860 + }, + { + "epoch": 0.3005825242718447, + "grad_norm": 0.6417273283004761, + "learning_rate": 7.738000000000001e-05, + "loss": 0.0206, + "step": 3870 + }, + { + "epoch": 0.30135922330097087, + "grad_norm": 0.4857272505760193, + "learning_rate": 7.758000000000001e-05, + "loss": 0.0185, + "step": 3880 + }, + { + "epoch": 0.30213592233009706, + "grad_norm": 0.42786455154418945, + "learning_rate": 7.778e-05, + "loss": 0.0221, + "step": 3890 + }, + { + "epoch": 0.3029126213592233, + "grad_norm": 0.6664991974830627, + "learning_rate": 7.798000000000001e-05, + "loss": 0.0192, + "step": 3900 + }, + { + "epoch": 0.3036893203883495, + "grad_norm": 0.708802342414856, + "learning_rate": 7.818000000000001e-05, + "loss": 0.0233, + "step": 3910 + }, + { + "epoch": 0.30446601941747575, + "grad_norm": 0.5586804747581482, + "learning_rate": 7.838e-05, + "loss": 0.0295, + "step": 3920 + }, + { + "epoch": 0.30524271844660195, + "grad_norm": 0.5502609610557556, + "learning_rate": 7.858000000000001e-05, + "loss": 0.0247, + "step": 3930 + }, + { + "epoch": 0.30601941747572814, + "grad_norm": 0.3380940854549408, + "learning_rate": 7.878e-05, + "loss": 0.0231, + "step": 3940 + }, + { + "epoch": 0.3067961165048544, + "grad_norm": 0.45528334379196167, + "learning_rate": 7.897999999999999e-05, + "loss": 0.0231, + "step": 3950 + }, + { + "epoch": 0.3075728155339806, + "grad_norm": 0.43430936336517334, + "learning_rate": 7.918e-05, + "loss": 0.0196, + "step": 3960 + }, + { + "epoch": 0.3083495145631068, + "grad_norm": 0.311321884393692, + "learning_rate": 7.938e-05, + "loss": 0.0273, + "step": 3970 + }, + { + "epoch": 0.309126213592233, + "grad_norm": 0.448284775018692, + "learning_rate": 7.958e-05, + "loss": 0.0266, + "step": 3980 + }, + { + "epoch": 0.3099029126213592, + "grad_norm": 0.3532017171382904, + "learning_rate": 7.978e-05, + "loss": 0.0177, + "step": 3990 + }, + { + "epoch": 0.3106796116504854, + "grad_norm": 0.518733561038971, + "learning_rate": 7.998e-05, + "loss": 0.0183, + "step": 4000 + }, + { + "epoch": 0.31145631067961166, + "grad_norm": 0.4042693078517914, + "learning_rate": 8.018e-05, + "loss": 0.0177, + "step": 4010 + }, + { + "epoch": 0.31223300970873785, + "grad_norm": 0.46018683910369873, + "learning_rate": 8.038e-05, + "loss": 0.0173, + "step": 4020 + }, + { + "epoch": 0.3130097087378641, + "grad_norm": 0.3632228672504425, + "learning_rate": 8.058e-05, + "loss": 0.0187, + "step": 4030 + }, + { + "epoch": 0.3137864077669903, + "grad_norm": 0.2494291514158249, + "learning_rate": 8.078e-05, + "loss": 0.0179, + "step": 4040 + }, + { + "epoch": 0.3145631067961165, + "grad_norm": 0.3945595622062683, + "learning_rate": 8.098e-05, + "loss": 0.016, + "step": 4050 + }, + { + "epoch": 0.31533980582524274, + "grad_norm": 0.2892042398452759, + "learning_rate": 8.118e-05, + "loss": 0.0183, + "step": 4060 + }, + { + "epoch": 0.31611650485436893, + "grad_norm": 0.3707210421562195, + "learning_rate": 8.138e-05, + "loss": 0.0155, + "step": 4070 + }, + { + "epoch": 0.3168932038834951, + "grad_norm": 0.43226101994514465, + "learning_rate": 8.158e-05, + "loss": 0.0168, + "step": 4080 + }, + { + "epoch": 0.31766990291262137, + "grad_norm": 0.3460884094238281, + "learning_rate": 8.178e-05, + "loss": 0.02, + "step": 4090 + }, + { + "epoch": 0.31844660194174756, + "grad_norm": 0.2521960735321045, + "learning_rate": 8.198e-05, + "loss": 0.0202, + "step": 4100 + }, + { + "epoch": 0.3192233009708738, + "grad_norm": 0.3917650878429413, + "learning_rate": 8.218e-05, + "loss": 0.027, + "step": 4110 + }, + { + "epoch": 0.32, + "grad_norm": 0.34652701020240784, + "learning_rate": 8.238000000000001e-05, + "loss": 0.0283, + "step": 4120 + }, + { + "epoch": 0.3207766990291262, + "grad_norm": 0.49304860830307007, + "learning_rate": 8.258e-05, + "loss": 0.0222, + "step": 4130 + }, + { + "epoch": 0.32155339805825245, + "grad_norm": 0.38359859585762024, + "learning_rate": 8.278e-05, + "loss": 0.0205, + "step": 4140 + }, + { + "epoch": 0.32233009708737864, + "grad_norm": 0.4217970073223114, + "learning_rate": 8.298000000000001e-05, + "loss": 0.0208, + "step": 4150 + }, + { + "epoch": 0.32310679611650484, + "grad_norm": 0.34061893820762634, + "learning_rate": 8.318e-05, + "loss": 0.0204, + "step": 4160 + }, + { + "epoch": 0.3238834951456311, + "grad_norm": 0.2875850200653076, + "learning_rate": 8.338e-05, + "loss": 0.0197, + "step": 4170 + }, + { + "epoch": 0.3246601941747573, + "grad_norm": 0.34800806641578674, + "learning_rate": 8.358e-05, + "loss": 0.0174, + "step": 4180 + }, + { + "epoch": 0.32543689320388347, + "grad_norm": 0.3586790859699249, + "learning_rate": 8.378e-05, + "loss": 0.0215, + "step": 4190 + }, + { + "epoch": 0.3262135922330097, + "grad_norm": 0.4744691252708435, + "learning_rate": 8.398e-05, + "loss": 0.0217, + "step": 4200 + }, + { + "epoch": 0.3269902912621359, + "grad_norm": 0.26914793252944946, + "learning_rate": 8.418e-05, + "loss": 0.0146, + "step": 4210 + }, + { + "epoch": 0.32776699029126216, + "grad_norm": 0.39307743310928345, + "learning_rate": 8.438e-05, + "loss": 0.0231, + "step": 4220 + }, + { + "epoch": 0.32854368932038835, + "grad_norm": 0.6224297285079956, + "learning_rate": 8.458e-05, + "loss": 0.0202, + "step": 4230 + }, + { + "epoch": 0.32932038834951455, + "grad_norm": 0.3935736417770386, + "learning_rate": 8.478e-05, + "loss": 0.0225, + "step": 4240 + }, + { + "epoch": 0.3300970873786408, + "grad_norm": 0.3285542130470276, + "learning_rate": 8.498e-05, + "loss": 0.0195, + "step": 4250 + }, + { + "epoch": 0.330873786407767, + "grad_norm": 0.2241489291191101, + "learning_rate": 8.518000000000001e-05, + "loss": 0.0215, + "step": 4260 + }, + { + "epoch": 0.3316504854368932, + "grad_norm": 1.1453784704208374, + "learning_rate": 8.538e-05, + "loss": 0.0214, + "step": 4270 + }, + { + "epoch": 0.33242718446601943, + "grad_norm": 0.2558824121952057, + "learning_rate": 8.558e-05, + "loss": 0.0173, + "step": 4280 + }, + { + "epoch": 0.3332038834951456, + "grad_norm": 0.7094147801399231, + "learning_rate": 8.578000000000001e-05, + "loss": 0.0198, + "step": 4290 + }, + { + "epoch": 0.3339805825242718, + "grad_norm": 0.45847705006599426, + "learning_rate": 8.598e-05, + "loss": 0.0197, + "step": 4300 + }, + { + "epoch": 0.33475728155339807, + "grad_norm": 0.3093036711215973, + "learning_rate": 8.618e-05, + "loss": 0.0196, + "step": 4310 + }, + { + "epoch": 0.33553398058252426, + "grad_norm": 0.5836067795753479, + "learning_rate": 8.638000000000001e-05, + "loss": 0.022, + "step": 4320 + }, + { + "epoch": 0.3363106796116505, + "grad_norm": 0.192386656999588, + "learning_rate": 8.658e-05, + "loss": 0.0157, + "step": 4330 + }, + { + "epoch": 0.3370873786407767, + "grad_norm": 0.31476619839668274, + "learning_rate": 8.678e-05, + "loss": 0.0192, + "step": 4340 + }, + { + "epoch": 0.3378640776699029, + "grad_norm": 0.30343762040138245, + "learning_rate": 8.698000000000001e-05, + "loss": 0.0165, + "step": 4350 + }, + { + "epoch": 0.33864077669902914, + "grad_norm": 0.518138587474823, + "learning_rate": 8.718e-05, + "loss": 0.0167, + "step": 4360 + }, + { + "epoch": 0.33941747572815534, + "grad_norm": 0.42239129543304443, + "learning_rate": 8.738000000000001e-05, + "loss": 0.0189, + "step": 4370 + }, + { + "epoch": 0.34019417475728153, + "grad_norm": 0.3555055260658264, + "learning_rate": 8.758000000000001e-05, + "loss": 0.0176, + "step": 4380 + }, + { + "epoch": 0.3409708737864078, + "grad_norm": 0.581495463848114, + "learning_rate": 8.778e-05, + "loss": 0.0216, + "step": 4390 + }, + { + "epoch": 0.341747572815534, + "grad_norm": 0.6012985706329346, + "learning_rate": 8.798000000000001e-05, + "loss": 0.0196, + "step": 4400 + }, + { + "epoch": 0.3425242718446602, + "grad_norm": 0.3853040039539337, + "learning_rate": 8.818000000000001e-05, + "loss": 0.0176, + "step": 4410 + }, + { + "epoch": 0.3433009708737864, + "grad_norm": 0.4803248941898346, + "learning_rate": 8.838e-05, + "loss": 0.0159, + "step": 4420 + }, + { + "epoch": 0.3440776699029126, + "grad_norm": 0.4943634569644928, + "learning_rate": 8.858000000000001e-05, + "loss": 0.0218, + "step": 4430 + }, + { + "epoch": 0.34485436893203886, + "grad_norm": 0.6410543918609619, + "learning_rate": 8.878000000000001e-05, + "loss": 0.0208, + "step": 4440 + }, + { + "epoch": 0.34563106796116505, + "grad_norm": 0.3285972774028778, + "learning_rate": 8.898e-05, + "loss": 0.0159, + "step": 4450 + }, + { + "epoch": 0.34640776699029124, + "grad_norm": 0.5301697850227356, + "learning_rate": 8.918000000000001e-05, + "loss": 0.0168, + "step": 4460 + }, + { + "epoch": 0.3471844660194175, + "grad_norm": 0.3528461456298828, + "learning_rate": 8.938e-05, + "loss": 0.0184, + "step": 4470 + }, + { + "epoch": 0.3479611650485437, + "grad_norm": 0.38929232954978943, + "learning_rate": 8.958e-05, + "loss": 0.0178, + "step": 4480 + }, + { + "epoch": 0.3487378640776699, + "grad_norm": 0.33004432916641235, + "learning_rate": 8.978000000000001e-05, + "loss": 0.0193, + "step": 4490 + }, + { + "epoch": 0.34951456310679613, + "grad_norm": 0.5427497029304504, + "learning_rate": 8.998e-05, + "loss": 0.0209, + "step": 4500 + }, + { + "epoch": 0.3502912621359223, + "grad_norm": 0.45824670791625977, + "learning_rate": 9.018000000000001e-05, + "loss": 0.0217, + "step": 4510 + }, + { + "epoch": 0.35106796116504857, + "grad_norm": 0.418008416891098, + "learning_rate": 9.038000000000001e-05, + "loss": 0.0184, + "step": 4520 + }, + { + "epoch": 0.35184466019417476, + "grad_norm": 0.49376872181892395, + "learning_rate": 9.058e-05, + "loss": 0.0226, + "step": 4530 + }, + { + "epoch": 0.35262135922330096, + "grad_norm": 0.3304676115512848, + "learning_rate": 9.078000000000001e-05, + "loss": 0.0152, + "step": 4540 + }, + { + "epoch": 0.3533980582524272, + "grad_norm": 0.3805513083934784, + "learning_rate": 9.098000000000001e-05, + "loss": 0.0152, + "step": 4550 + }, + { + "epoch": 0.3541747572815534, + "grad_norm": 0.20017316937446594, + "learning_rate": 9.118e-05, + "loss": 0.0159, + "step": 4560 + }, + { + "epoch": 0.3549514563106796, + "grad_norm": 0.39203280210494995, + "learning_rate": 9.138e-05, + "loss": 0.0205, + "step": 4570 + }, + { + "epoch": 0.35572815533980584, + "grad_norm": 0.4157106876373291, + "learning_rate": 9.158e-05, + "loss": 0.0265, + "step": 4580 + }, + { + "epoch": 0.35650485436893203, + "grad_norm": 0.2803509533405304, + "learning_rate": 9.178e-05, + "loss": 0.022, + "step": 4590 + }, + { + "epoch": 0.3572815533980582, + "grad_norm": 0.40671607851982117, + "learning_rate": 9.198e-05, + "loss": 0.0208, + "step": 4600 + }, + { + "epoch": 0.3580582524271845, + "grad_norm": 0.3946969211101532, + "learning_rate": 9.218e-05, + "loss": 0.0145, + "step": 4610 + }, + { + "epoch": 0.35883495145631067, + "grad_norm": 0.2148197740316391, + "learning_rate": 9.238e-05, + "loss": 0.0217, + "step": 4620 + }, + { + "epoch": 0.3596116504854369, + "grad_norm": 0.35939809679985046, + "learning_rate": 9.258e-05, + "loss": 0.0182, + "step": 4630 + }, + { + "epoch": 0.3603883495145631, + "grad_norm": 0.4486740827560425, + "learning_rate": 9.278e-05, + "loss": 0.0155, + "step": 4640 + }, + { + "epoch": 0.3611650485436893, + "grad_norm": 0.3690606355667114, + "learning_rate": 9.298e-05, + "loss": 0.02, + "step": 4650 + }, + { + "epoch": 0.36194174757281555, + "grad_norm": 0.3252636790275574, + "learning_rate": 9.318e-05, + "loss": 0.0201, + "step": 4660 + }, + { + "epoch": 0.36271844660194175, + "grad_norm": 0.398145467042923, + "learning_rate": 9.338e-05, + "loss": 0.0155, + "step": 4670 + }, + { + "epoch": 0.36349514563106794, + "grad_norm": 0.2620205581188202, + "learning_rate": 9.358e-05, + "loss": 0.0131, + "step": 4680 + }, + { + "epoch": 0.3642718446601942, + "grad_norm": 0.3379886746406555, + "learning_rate": 9.378e-05, + "loss": 0.0209, + "step": 4690 + }, + { + "epoch": 0.3650485436893204, + "grad_norm": 0.37515807151794434, + "learning_rate": 9.398e-05, + "loss": 0.02, + "step": 4700 + }, + { + "epoch": 0.3658252427184466, + "grad_norm": 0.30301737785339355, + "learning_rate": 9.418e-05, + "loss": 0.0175, + "step": 4710 + }, + { + "epoch": 0.3666019417475728, + "grad_norm": 0.32554829120635986, + "learning_rate": 9.438e-05, + "loss": 0.0157, + "step": 4720 + }, + { + "epoch": 0.367378640776699, + "grad_norm": 0.6024938225746155, + "learning_rate": 9.458e-05, + "loss": 0.0168, + "step": 4730 + }, + { + "epoch": 0.36815533980582527, + "grad_norm": 0.2629837095737457, + "learning_rate": 9.478e-05, + "loss": 0.0168, + "step": 4740 + }, + { + "epoch": 0.36893203883495146, + "grad_norm": 0.2924623191356659, + "learning_rate": 9.498e-05, + "loss": 0.0189, + "step": 4750 + }, + { + "epoch": 0.36970873786407765, + "grad_norm": 0.3129458725452423, + "learning_rate": 9.518000000000001e-05, + "loss": 0.0151, + "step": 4760 + }, + { + "epoch": 0.3704854368932039, + "grad_norm": 0.34797415137290955, + "learning_rate": 9.538e-05, + "loss": 0.0186, + "step": 4770 + }, + { + "epoch": 0.3712621359223301, + "grad_norm": 0.32894793152809143, + "learning_rate": 9.558e-05, + "loss": 0.0199, + "step": 4780 + }, + { + "epoch": 0.3720388349514563, + "grad_norm": 0.29041820764541626, + "learning_rate": 9.578000000000001e-05, + "loss": 0.0167, + "step": 4790 + }, + { + "epoch": 0.37281553398058254, + "grad_norm": 0.379032701253891, + "learning_rate": 9.598e-05, + "loss": 0.0187, + "step": 4800 + }, + { + "epoch": 0.37359223300970873, + "grad_norm": 0.38382378220558167, + "learning_rate": 9.618e-05, + "loss": 0.0159, + "step": 4810 + }, + { + "epoch": 0.374368932038835, + "grad_norm": 0.3053785562515259, + "learning_rate": 9.638000000000001e-05, + "loss": 0.016, + "step": 4820 + }, + { + "epoch": 0.37514563106796117, + "grad_norm": 0.48537367582321167, + "learning_rate": 9.658e-05, + "loss": 0.0207, + "step": 4830 + }, + { + "epoch": 0.37592233009708736, + "grad_norm": 0.5468986630439758, + "learning_rate": 9.678e-05, + "loss": 0.0261, + "step": 4840 + }, + { + "epoch": 0.3766990291262136, + "grad_norm": 0.3221736550331116, + "learning_rate": 9.698000000000001e-05, + "loss": 0.0187, + "step": 4850 + }, + { + "epoch": 0.3774757281553398, + "grad_norm": 0.3665720224380493, + "learning_rate": 9.718e-05, + "loss": 0.0232, + "step": 4860 + }, + { + "epoch": 0.378252427184466, + "grad_norm": 0.28872671723365784, + "learning_rate": 9.738e-05, + "loss": 0.0185, + "step": 4870 + }, + { + "epoch": 0.37902912621359225, + "grad_norm": 0.44498804211616516, + "learning_rate": 9.758000000000001e-05, + "loss": 0.0195, + "step": 4880 + }, + { + "epoch": 0.37980582524271844, + "grad_norm": 0.6075417399406433, + "learning_rate": 9.778e-05, + "loss": 0.0235, + "step": 4890 + }, + { + "epoch": 0.38058252427184464, + "grad_norm": 0.3560914695262909, + "learning_rate": 9.798000000000001e-05, + "loss": 0.0194, + "step": 4900 + }, + { + "epoch": 0.3813592233009709, + "grad_norm": 0.32011041045188904, + "learning_rate": 9.818000000000001e-05, + "loss": 0.0172, + "step": 4910 + }, + { + "epoch": 0.3821359223300971, + "grad_norm": 0.4050678312778473, + "learning_rate": 9.838e-05, + "loss": 0.0173, + "step": 4920 + }, + { + "epoch": 0.3829126213592233, + "grad_norm": 0.26400092244148254, + "learning_rate": 9.858000000000001e-05, + "loss": 0.0194, + "step": 4930 + }, + { + "epoch": 0.3836893203883495, + "grad_norm": 0.4033794701099396, + "learning_rate": 9.878e-05, + "loss": 0.0209, + "step": 4940 + }, + { + "epoch": 0.3844660194174757, + "grad_norm": 0.32437577843666077, + "learning_rate": 9.898e-05, + "loss": 0.0217, + "step": 4950 + }, + { + "epoch": 0.38524271844660196, + "grad_norm": 0.4259447753429413, + "learning_rate": 9.918000000000001e-05, + "loss": 0.017, + "step": 4960 + }, + { + "epoch": 0.38601941747572815, + "grad_norm": 0.41726645827293396, + "learning_rate": 9.938e-05, + "loss": 0.0196, + "step": 4970 + }, + { + "epoch": 0.38679611650485435, + "grad_norm": 0.2821904718875885, + "learning_rate": 9.958e-05, + "loss": 0.0178, + "step": 4980 + }, + { + "epoch": 0.3875728155339806, + "grad_norm": 0.3541766107082367, + "learning_rate": 9.978000000000001e-05, + "loss": 0.0209, + "step": 4990 + }, + { + "epoch": 0.3883495145631068, + "grad_norm": 0.34352511167526245, + "learning_rate": 9.998e-05, + "loss": 0.0179, + "step": 5000 + }, + { + "epoch": 0.389126213592233, + "grad_norm": 0.42533746361732483, + "learning_rate": 9.999999778549045e-05, + "loss": 0.0245, + "step": 5010 + }, + { + "epoch": 0.38990291262135923, + "grad_norm": 0.6922399997711182, + "learning_rate": 9.999999013039593e-05, + "loss": 0.0236, + "step": 5020 + }, + { + "epoch": 0.3906796116504854, + "grad_norm": 0.5566357374191284, + "learning_rate": 9.999997700737766e-05, + "loss": 0.0173, + "step": 5030 + }, + { + "epoch": 0.3914563106796117, + "grad_norm": 0.44183579087257385, + "learning_rate": 9.999995841643709e-05, + "loss": 0.0188, + "step": 5040 + }, + { + "epoch": 0.39223300970873787, + "grad_norm": 0.5019325613975525, + "learning_rate": 9.999993435757623e-05, + "loss": 0.0179, + "step": 5050 + }, + { + "epoch": 0.39300970873786406, + "grad_norm": 0.5659012198448181, + "learning_rate": 9.999990483079773e-05, + "loss": 0.0178, + "step": 5060 + }, + { + "epoch": 0.3937864077669903, + "grad_norm": 0.5081613659858704, + "learning_rate": 9.999986983610481e-05, + "loss": 0.0142, + "step": 5070 + }, + { + "epoch": 0.3945631067961165, + "grad_norm": 0.36190396547317505, + "learning_rate": 9.99998293735013e-05, + "loss": 0.0184, + "step": 5080 + }, + { + "epoch": 0.3953398058252427, + "grad_norm": 0.38737571239471436, + "learning_rate": 9.999978344299161e-05, + "loss": 0.0181, + "step": 5090 + }, + { + "epoch": 0.39611650485436894, + "grad_norm": 0.4185486137866974, + "learning_rate": 9.99997320445808e-05, + "loss": 0.0198, + "step": 5100 + }, + { + "epoch": 0.39689320388349514, + "grad_norm": 0.48073673248291016, + "learning_rate": 9.999967517827444e-05, + "loss": 0.0174, + "step": 5110 + }, + { + "epoch": 0.3976699029126214, + "grad_norm": 0.4648996591567993, + "learning_rate": 9.999961284407879e-05, + "loss": 0.0254, + "step": 5120 + }, + { + "epoch": 0.3984466019417476, + "grad_norm": 0.27930697798728943, + "learning_rate": 9.999954504200067e-05, + "loss": 0.0149, + "step": 5130 + }, + { + "epoch": 0.3992233009708738, + "grad_norm": 0.27085909247398376, + "learning_rate": 9.999947177204744e-05, + "loss": 0.0161, + "step": 5140 + }, + { + "epoch": 0.4, + "grad_norm": 0.53117436170578, + "learning_rate": 9.999939303422718e-05, + "loss": 0.0139, + "step": 5150 + }, + { + "epoch": 0.4007766990291262, + "grad_norm": 0.4193735122680664, + "learning_rate": 9.999930882854847e-05, + "loss": 0.0172, + "step": 5160 + }, + { + "epoch": 0.4015533980582524, + "grad_norm": 0.2973770499229431, + "learning_rate": 9.999921915502051e-05, + "loss": 0.0158, + "step": 5170 + }, + { + "epoch": 0.40233009708737866, + "grad_norm": 0.3000485301017761, + "learning_rate": 9.99991240136531e-05, + "loss": 0.02, + "step": 5180 + }, + { + "epoch": 0.40310679611650485, + "grad_norm": 0.47838249802589417, + "learning_rate": 9.999902340445668e-05, + "loss": 0.0183, + "step": 5190 + }, + { + "epoch": 0.40388349514563104, + "grad_norm": 0.7053872346878052, + "learning_rate": 9.999891732744224e-05, + "loss": 0.0199, + "step": 5200 + }, + { + "epoch": 0.4046601941747573, + "grad_norm": 0.5061638355255127, + "learning_rate": 9.999880578262135e-05, + "loss": 0.0166, + "step": 5210 + }, + { + "epoch": 0.4054368932038835, + "grad_norm": 0.42152321338653564, + "learning_rate": 9.999868877000624e-05, + "loss": 0.0216, + "step": 5220 + }, + { + "epoch": 0.40621359223300973, + "grad_norm": 0.5273614525794983, + "learning_rate": 9.99985662896097e-05, + "loss": 0.0191, + "step": 5230 + }, + { + "epoch": 0.40699029126213593, + "grad_norm": 0.41623133420944214, + "learning_rate": 9.999843834144513e-05, + "loss": 0.016, + "step": 5240 + }, + { + "epoch": 0.4077669902912621, + "grad_norm": 0.33954286575317383, + "learning_rate": 9.99983049255265e-05, + "loss": 0.0152, + "step": 5250 + }, + { + "epoch": 0.40854368932038837, + "grad_norm": 0.5013542771339417, + "learning_rate": 9.999816604186843e-05, + "loss": 0.0199, + "step": 5260 + }, + { + "epoch": 0.40932038834951456, + "grad_norm": 0.2988308370113373, + "learning_rate": 9.999802169048609e-05, + "loss": 0.0202, + "step": 5270 + }, + { + "epoch": 0.41009708737864076, + "grad_norm": 0.2821911871433258, + "learning_rate": 9.999787187139527e-05, + "loss": 0.0194, + "step": 5280 + }, + { + "epoch": 0.410873786407767, + "grad_norm": 0.22856755554676056, + "learning_rate": 9.999771658461234e-05, + "loss": 0.0153, + "step": 5290 + }, + { + "epoch": 0.4116504854368932, + "grad_norm": 0.3412010669708252, + "learning_rate": 9.999755583015431e-05, + "loss": 0.0169, + "step": 5300 + }, + { + "epoch": 0.4124271844660194, + "grad_norm": 0.3262406587600708, + "learning_rate": 9.999738960803874e-05, + "loss": 0.0154, + "step": 5310 + }, + { + "epoch": 0.41320388349514564, + "grad_norm": 0.372711718082428, + "learning_rate": 9.99972179182838e-05, + "loss": 0.0202, + "step": 5320 + }, + { + "epoch": 0.41398058252427183, + "grad_norm": 0.27083754539489746, + "learning_rate": 9.99970407609083e-05, + "loss": 0.0143, + "step": 5330 + }, + { + "epoch": 0.4147572815533981, + "grad_norm": 0.3865097761154175, + "learning_rate": 9.999685813593159e-05, + "loss": 0.0173, + "step": 5340 + }, + { + "epoch": 0.4155339805825243, + "grad_norm": 0.32765260338783264, + "learning_rate": 9.999667004337362e-05, + "loss": 0.0212, + "step": 5350 + }, + { + "epoch": 0.41631067961165047, + "grad_norm": 0.3723406195640564, + "learning_rate": 9.9996476483255e-05, + "loss": 0.0217, + "step": 5360 + }, + { + "epoch": 0.4170873786407767, + "grad_norm": 0.514020562171936, + "learning_rate": 9.999627745559688e-05, + "loss": 0.0187, + "step": 5370 + }, + { + "epoch": 0.4178640776699029, + "grad_norm": 0.3512232303619385, + "learning_rate": 9.999607296042101e-05, + "loss": 0.0195, + "step": 5380 + }, + { + "epoch": 0.4186407766990291, + "grad_norm": 0.40483468770980835, + "learning_rate": 9.99958629977498e-05, + "loss": 0.0124, + "step": 5390 + }, + { + "epoch": 0.41941747572815535, + "grad_norm": 0.39295729994773865, + "learning_rate": 9.999564756760615e-05, + "loss": 0.0143, + "step": 5400 + }, + { + "epoch": 0.42019417475728155, + "grad_norm": 0.23651814460754395, + "learning_rate": 9.999542667001366e-05, + "loss": 0.017, + "step": 5410 + }, + { + "epoch": 0.42097087378640774, + "grad_norm": 0.24674992263317108, + "learning_rate": 9.999520030499647e-05, + "loss": 0.0177, + "step": 5420 + }, + { + "epoch": 0.421747572815534, + "grad_norm": 0.334553986787796, + "learning_rate": 9.999496847257936e-05, + "loss": 0.0153, + "step": 5430 + }, + { + "epoch": 0.4225242718446602, + "grad_norm": 0.3107491135597229, + "learning_rate": 9.999473117278764e-05, + "loss": 0.0155, + "step": 5440 + }, + { + "epoch": 0.42330097087378643, + "grad_norm": 0.40088707208633423, + "learning_rate": 9.999448840564731e-05, + "loss": 0.0206, + "step": 5450 + }, + { + "epoch": 0.4240776699029126, + "grad_norm": 0.3833012878894806, + "learning_rate": 9.999424017118488e-05, + "loss": 0.0174, + "step": 5460 + }, + { + "epoch": 0.4248543689320388, + "grad_norm": 0.25979143381118774, + "learning_rate": 9.999398646942751e-05, + "loss": 0.0211, + "step": 5470 + }, + { + "epoch": 0.42563106796116507, + "grad_norm": 0.35884901881217957, + "learning_rate": 9.999372730040296e-05, + "loss": 0.016, + "step": 5480 + }, + { + "epoch": 0.42640776699029126, + "grad_norm": 0.25630512833595276, + "learning_rate": 9.999346266413953e-05, + "loss": 0.0146, + "step": 5490 + }, + { + "epoch": 0.42718446601941745, + "grad_norm": 0.27231550216674805, + "learning_rate": 9.99931925606662e-05, + "loss": 0.0138, + "step": 5500 + }, + { + "epoch": 0.4279611650485437, + "grad_norm": 0.3622545301914215, + "learning_rate": 9.99929169900125e-05, + "loss": 0.017, + "step": 5510 + }, + { + "epoch": 0.4287378640776699, + "grad_norm": 0.2990235388278961, + "learning_rate": 9.999263595220855e-05, + "loss": 0.015, + "step": 5520 + }, + { + "epoch": 0.42951456310679614, + "grad_norm": 0.476974219083786, + "learning_rate": 9.99923494472851e-05, + "loss": 0.0254, + "step": 5530 + }, + { + "epoch": 0.43029126213592234, + "grad_norm": 0.4809616506099701, + "learning_rate": 9.999205747527348e-05, + "loss": 0.02, + "step": 5540 + }, + { + "epoch": 0.43106796116504853, + "grad_norm": 0.3310031294822693, + "learning_rate": 9.999176003620561e-05, + "loss": 0.0162, + "step": 5550 + }, + { + "epoch": 0.4318446601941748, + "grad_norm": 0.4237656891345978, + "learning_rate": 9.999145713011405e-05, + "loss": 0.0204, + "step": 5560 + }, + { + "epoch": 0.43262135922330097, + "grad_norm": 0.3468461334705353, + "learning_rate": 9.999114875703186e-05, + "loss": 0.0219, + "step": 5570 + }, + { + "epoch": 0.43339805825242717, + "grad_norm": 0.2844405770301819, + "learning_rate": 9.999083491699281e-05, + "loss": 0.0199, + "step": 5580 + }, + { + "epoch": 0.4341747572815534, + "grad_norm": 0.2318791002035141, + "learning_rate": 9.999051561003123e-05, + "loss": 0.0169, + "step": 5590 + }, + { + "epoch": 0.4349514563106796, + "grad_norm": 0.3718665540218353, + "learning_rate": 9.999019083618202e-05, + "loss": 0.0159, + "step": 5600 + }, + { + "epoch": 0.4357281553398058, + "grad_norm": 0.36030954122543335, + "learning_rate": 9.99898605954807e-05, + "loss": 0.0159, + "step": 5610 + }, + { + "epoch": 0.43650485436893205, + "grad_norm": 0.3723878562450409, + "learning_rate": 9.998952488796338e-05, + "loss": 0.0194, + "step": 5620 + }, + { + "epoch": 0.43728155339805824, + "grad_norm": 0.3795473575592041, + "learning_rate": 9.998918371366676e-05, + "loss": 0.0157, + "step": 5630 + }, + { + "epoch": 0.4380582524271845, + "grad_norm": 0.37124502658843994, + "learning_rate": 9.99888370726282e-05, + "loss": 0.0201, + "step": 5640 + }, + { + "epoch": 0.4388349514563107, + "grad_norm": 0.3385646641254425, + "learning_rate": 9.998848496488556e-05, + "loss": 0.0138, + "step": 5650 + }, + { + "epoch": 0.4396116504854369, + "grad_norm": 0.4100220203399658, + "learning_rate": 9.998812739047736e-05, + "loss": 0.0183, + "step": 5660 + }, + { + "epoch": 0.4403883495145631, + "grad_norm": 0.39605239033699036, + "learning_rate": 9.99877643494427e-05, + "loss": 0.0173, + "step": 5670 + }, + { + "epoch": 0.4411650485436893, + "grad_norm": 0.35381075739860535, + "learning_rate": 9.998739584182128e-05, + "loss": 0.022, + "step": 5680 + }, + { + "epoch": 0.4419417475728155, + "grad_norm": 0.2913196384906769, + "learning_rate": 9.998702186765342e-05, + "loss": 0.0214, + "step": 5690 + }, + { + "epoch": 0.44271844660194176, + "grad_norm": 0.492634117603302, + "learning_rate": 9.998664242698e-05, + "loss": 0.017, + "step": 5700 + }, + { + "epoch": 0.44349514563106796, + "grad_norm": 0.6042423844337463, + "learning_rate": 9.998625751984251e-05, + "loss": 0.0183, + "step": 5710 + }, + { + "epoch": 0.44427184466019415, + "grad_norm": 0.6611431837081909, + "learning_rate": 9.998586714628307e-05, + "loss": 0.0218, + "step": 5720 + }, + { + "epoch": 0.4450485436893204, + "grad_norm": 0.45785191655158997, + "learning_rate": 9.998547130634432e-05, + "loss": 0.0193, + "step": 5730 + }, + { + "epoch": 0.4458252427184466, + "grad_norm": 0.4323107600212097, + "learning_rate": 9.99850700000696e-05, + "loss": 0.0196, + "step": 5740 + }, + { + "epoch": 0.44660194174757284, + "grad_norm": 0.3789357542991638, + "learning_rate": 9.998466322750278e-05, + "loss": 0.0179, + "step": 5750 + }, + { + "epoch": 0.44737864077669903, + "grad_norm": 0.2994386851787567, + "learning_rate": 9.998425098868834e-05, + "loss": 0.0196, + "step": 5760 + }, + { + "epoch": 0.4481553398058252, + "grad_norm": 0.31660890579223633, + "learning_rate": 9.998383328367136e-05, + "loss": 0.0206, + "step": 5770 + }, + { + "epoch": 0.4489320388349515, + "grad_norm": 0.39807331562042236, + "learning_rate": 9.99834101124975e-05, + "loss": 0.0156, + "step": 5780 + }, + { + "epoch": 0.44970873786407767, + "grad_norm": 0.34803056716918945, + "learning_rate": 9.998298147521309e-05, + "loss": 0.0165, + "step": 5790 + }, + { + "epoch": 0.45048543689320386, + "grad_norm": 0.31745588779449463, + "learning_rate": 9.998254737186496e-05, + "loss": 0.0263, + "step": 5800 + }, + { + "epoch": 0.4512621359223301, + "grad_norm": 0.42990484833717346, + "learning_rate": 9.99821078025006e-05, + "loss": 0.0196, + "step": 5810 + }, + { + "epoch": 0.4520388349514563, + "grad_norm": 0.35135582089424133, + "learning_rate": 9.998166276716807e-05, + "loss": 0.0194, + "step": 5820 + }, + { + "epoch": 0.45281553398058255, + "grad_norm": 0.6046357154846191, + "learning_rate": 9.998121226591606e-05, + "loss": 0.0209, + "step": 5830 + }, + { + "epoch": 0.45359223300970875, + "grad_norm": 0.33618128299713135, + "learning_rate": 9.998075629879382e-05, + "loss": 0.0157, + "step": 5840 + }, + { + "epoch": 0.45436893203883494, + "grad_norm": 0.38687023520469666, + "learning_rate": 9.99802948658512e-05, + "loss": 0.0246, + "step": 5850 + }, + { + "epoch": 0.4551456310679612, + "grad_norm": 0.43360620737075806, + "learning_rate": 9.99798279671387e-05, + "loss": 0.0184, + "step": 5860 + }, + { + "epoch": 0.4559223300970874, + "grad_norm": 0.43554285168647766, + "learning_rate": 9.997935560270734e-05, + "loss": 0.0186, + "step": 5870 + }, + { + "epoch": 0.4566990291262136, + "grad_norm": 0.312788724899292, + "learning_rate": 9.997887777260879e-05, + "loss": 0.0233, + "step": 5880 + }, + { + "epoch": 0.4574757281553398, + "grad_norm": 0.22001473605632782, + "learning_rate": 9.997839447689532e-05, + "loss": 0.0153, + "step": 5890 + }, + { + "epoch": 0.458252427184466, + "grad_norm": 0.2678442895412445, + "learning_rate": 9.997790571561978e-05, + "loss": 0.0202, + "step": 5900 + }, + { + "epoch": 0.4590291262135922, + "grad_norm": 0.28205493092536926, + "learning_rate": 9.99774114888356e-05, + "loss": 0.0152, + "step": 5910 + }, + { + "epoch": 0.45980582524271846, + "grad_norm": 0.28551286458969116, + "learning_rate": 9.997691179659684e-05, + "loss": 0.015, + "step": 5920 + }, + { + "epoch": 0.46058252427184465, + "grad_norm": 0.4330602288246155, + "learning_rate": 9.997640663895815e-05, + "loss": 0.0161, + "step": 5930 + }, + { + "epoch": 0.4613592233009709, + "grad_norm": 0.2519950866699219, + "learning_rate": 9.997589601597477e-05, + "loss": 0.0154, + "step": 5940 + }, + { + "epoch": 0.4621359223300971, + "grad_norm": 0.4286029040813446, + "learning_rate": 9.997537992770252e-05, + "loss": 0.0147, + "step": 5950 + }, + { + "epoch": 0.4629126213592233, + "grad_norm": 0.4654032588005066, + "learning_rate": 9.997485837419788e-05, + "loss": 0.0136, + "step": 5960 + }, + { + "epoch": 0.46368932038834954, + "grad_norm": 0.3887253701686859, + "learning_rate": 9.997433135551786e-05, + "loss": 0.015, + "step": 5970 + }, + { + "epoch": 0.46446601941747573, + "grad_norm": 0.284051775932312, + "learning_rate": 9.997379887172009e-05, + "loss": 0.0201, + "step": 5980 + }, + { + "epoch": 0.4652427184466019, + "grad_norm": 0.3888261616230011, + "learning_rate": 9.997326092286281e-05, + "loss": 0.0167, + "step": 5990 + }, + { + "epoch": 0.46601941747572817, + "grad_norm": 0.394752562046051, + "learning_rate": 9.997271750900486e-05, + "loss": 0.014, + "step": 6000 + }, + { + "epoch": 0.46679611650485436, + "grad_norm": 0.45245540142059326, + "learning_rate": 9.997216863020565e-05, + "loss": 0.0175, + "step": 6010 + }, + { + "epoch": 0.46757281553398056, + "grad_norm": 0.24775007367134094, + "learning_rate": 9.99716142865252e-05, + "loss": 0.0175, + "step": 6020 + }, + { + "epoch": 0.4683495145631068, + "grad_norm": 0.3006967008113861, + "learning_rate": 9.997105447802415e-05, + "loss": 0.0153, + "step": 6030 + }, + { + "epoch": 0.469126213592233, + "grad_norm": 0.39332863688468933, + "learning_rate": 9.997048920476373e-05, + "loss": 0.018, + "step": 6040 + }, + { + "epoch": 0.46990291262135925, + "grad_norm": 0.5469543933868408, + "learning_rate": 9.996991846680572e-05, + "loss": 0.0154, + "step": 6050 + }, + { + "epoch": 0.47067961165048544, + "grad_norm": 0.9452661275863647, + "learning_rate": 9.996934226421257e-05, + "loss": 0.0183, + "step": 6060 + }, + { + "epoch": 0.47145631067961163, + "grad_norm": 0.339528888463974, + "learning_rate": 9.996876059704726e-05, + "loss": 0.0194, + "step": 6070 + }, + { + "epoch": 0.4722330097087379, + "grad_norm": 0.33733731508255005, + "learning_rate": 9.996817346537343e-05, + "loss": 0.0213, + "step": 6080 + }, + { + "epoch": 0.4730097087378641, + "grad_norm": 0.3759877681732178, + "learning_rate": 9.996758086925526e-05, + "loss": 0.0189, + "step": 6090 + }, + { + "epoch": 0.47378640776699027, + "grad_norm": 0.7006130218505859, + "learning_rate": 9.996698280875759e-05, + "loss": 0.0142, + "step": 6100 + }, + { + "epoch": 0.4745631067961165, + "grad_norm": 0.36795172095298767, + "learning_rate": 9.99663792839458e-05, + "loss": 0.0223, + "step": 6110 + }, + { + "epoch": 0.4753398058252427, + "grad_norm": 0.4065096080303192, + "learning_rate": 9.99657702948859e-05, + "loss": 0.025, + "step": 6120 + }, + { + "epoch": 0.4761165048543689, + "grad_norm": 0.3302776515483856, + "learning_rate": 9.996515584164448e-05, + "loss": 0.0204, + "step": 6130 + }, + { + "epoch": 0.47689320388349515, + "grad_norm": 0.407590389251709, + "learning_rate": 9.996453592428873e-05, + "loss": 0.0194, + "step": 6140 + }, + { + "epoch": 0.47766990291262135, + "grad_norm": 0.3784017562866211, + "learning_rate": 9.996391054288646e-05, + "loss": 0.0186, + "step": 6150 + }, + { + "epoch": 0.4784466019417476, + "grad_norm": 0.33953142166137695, + "learning_rate": 9.996327969750605e-05, + "loss": 0.0199, + "step": 6160 + }, + { + "epoch": 0.4792233009708738, + "grad_norm": 0.3095296323299408, + "learning_rate": 9.996264338821649e-05, + "loss": 0.0183, + "step": 6170 + }, + { + "epoch": 0.48, + "grad_norm": 0.29083287715911865, + "learning_rate": 9.996200161508735e-05, + "loss": 0.0172, + "step": 6180 + }, + { + "epoch": 0.48077669902912623, + "grad_norm": 0.5361764430999756, + "learning_rate": 9.996135437818885e-05, + "loss": 0.0212, + "step": 6190 + }, + { + "epoch": 0.4815533980582524, + "grad_norm": 0.3308776319026947, + "learning_rate": 9.996070167759175e-05, + "loss": 0.018, + "step": 6200 + }, + { + "epoch": 0.4823300970873786, + "grad_norm": 0.36937201023101807, + "learning_rate": 9.996004351336743e-05, + "loss": 0.0162, + "step": 6210 + }, + { + "epoch": 0.48310679611650487, + "grad_norm": 0.33533692359924316, + "learning_rate": 9.995937988558785e-05, + "loss": 0.0218, + "step": 6220 + }, + { + "epoch": 0.48388349514563106, + "grad_norm": 0.26746514439582825, + "learning_rate": 9.995871079432561e-05, + "loss": 0.0163, + "step": 6230 + }, + { + "epoch": 0.4846601941747573, + "grad_norm": 0.3843660354614258, + "learning_rate": 9.995803623965389e-05, + "loss": 0.0142, + "step": 6240 + }, + { + "epoch": 0.4854368932038835, + "grad_norm": 0.3685079514980316, + "learning_rate": 9.995735622164641e-05, + "loss": 0.0209, + "step": 6250 + }, + { + "epoch": 0.4862135922330097, + "grad_norm": 0.5080148577690125, + "learning_rate": 9.995667074037758e-05, + "loss": 0.0208, + "step": 6260 + }, + { + "epoch": 0.48699029126213594, + "grad_norm": 0.2911519706249237, + "learning_rate": 9.995597979592232e-05, + "loss": 0.0159, + "step": 6270 + }, + { + "epoch": 0.48776699029126214, + "grad_norm": 0.524261474609375, + "learning_rate": 9.995528338835625e-05, + "loss": 0.0152, + "step": 6280 + }, + { + "epoch": 0.48854368932038833, + "grad_norm": 0.35969310998916626, + "learning_rate": 9.995458151775547e-05, + "loss": 0.018, + "step": 6290 + }, + { + "epoch": 0.4893203883495146, + "grad_norm": 0.306270569562912, + "learning_rate": 9.995387418419677e-05, + "loss": 0.0172, + "step": 6300 + }, + { + "epoch": 0.49009708737864077, + "grad_norm": 0.5358296036720276, + "learning_rate": 9.99531613877575e-05, + "loss": 0.0178, + "step": 6310 + }, + { + "epoch": 0.49087378640776697, + "grad_norm": 0.4129957854747772, + "learning_rate": 9.995244312851559e-05, + "loss": 0.0157, + "step": 6320 + }, + { + "epoch": 0.4916504854368932, + "grad_norm": 0.40087011456489563, + "learning_rate": 9.995171940654961e-05, + "loss": 0.0172, + "step": 6330 + }, + { + "epoch": 0.4924271844660194, + "grad_norm": 0.3285282552242279, + "learning_rate": 9.995099022193871e-05, + "loss": 0.019, + "step": 6340 + }, + { + "epoch": 0.49320388349514566, + "grad_norm": 0.331952303647995, + "learning_rate": 9.995025557476261e-05, + "loss": 0.0158, + "step": 6350 + }, + { + "epoch": 0.49398058252427185, + "grad_norm": 0.3217718303203583, + "learning_rate": 9.994951546510165e-05, + "loss": 0.0178, + "step": 6360 + }, + { + "epoch": 0.49475728155339804, + "grad_norm": 0.40921276807785034, + "learning_rate": 9.994876989303679e-05, + "loss": 0.0147, + "step": 6370 + }, + { + "epoch": 0.4955339805825243, + "grad_norm": 0.46952885389328003, + "learning_rate": 9.994801885864955e-05, + "loss": 0.0169, + "step": 6380 + }, + { + "epoch": 0.4963106796116505, + "grad_norm": 0.3232288360595703, + "learning_rate": 9.994726236202205e-05, + "loss": 0.0165, + "step": 6390 + }, + { + "epoch": 0.4970873786407767, + "grad_norm": 0.45588919520378113, + "learning_rate": 9.994650040323704e-05, + "loss": 0.0188, + "step": 6400 + }, + { + "epoch": 0.4978640776699029, + "grad_norm": 0.4443589150905609, + "learning_rate": 9.994573298237784e-05, + "loss": 0.0172, + "step": 6410 + }, + { + "epoch": 0.4986407766990291, + "grad_norm": 0.4163509011268616, + "learning_rate": 9.994496009952837e-05, + "loss": 0.0179, + "step": 6420 + }, + { + "epoch": 0.4994174757281553, + "grad_norm": 0.2959410846233368, + "learning_rate": 9.994418175477316e-05, + "loss": 0.0181, + "step": 6430 + }, + { + "epoch": 0.5001941747572816, + "grad_norm": 0.29841989278793335, + "learning_rate": 9.994339794819733e-05, + "loss": 0.0147, + "step": 6440 + }, + { + "epoch": 0.5009708737864078, + "grad_norm": 0.5465672612190247, + "learning_rate": 9.994260867988658e-05, + "loss": 0.0134, + "step": 6450 + }, + { + "epoch": 0.501747572815534, + "grad_norm": 0.38591986894607544, + "learning_rate": 9.994181394992723e-05, + "loss": 0.0121, + "step": 6460 + }, + { + "epoch": 0.5025242718446602, + "grad_norm": 0.3596358895301819, + "learning_rate": 9.994101375840618e-05, + "loss": 0.0163, + "step": 6470 + }, + { + "epoch": 0.5033009708737864, + "grad_norm": 0.28583818674087524, + "learning_rate": 9.994020810541098e-05, + "loss": 0.0142, + "step": 6480 + }, + { + "epoch": 0.5040776699029126, + "grad_norm": 0.2487356960773468, + "learning_rate": 9.99393969910297e-05, + "loss": 0.0141, + "step": 6490 + }, + { + "epoch": 0.5048543689320388, + "grad_norm": 0.33444178104400635, + "learning_rate": 9.993858041535104e-05, + "loss": 0.0133, + "step": 6500 + }, + { + "epoch": 0.5056310679611651, + "grad_norm": 0.16363854706287384, + "learning_rate": 9.99377583784643e-05, + "loss": 0.0204, + "step": 6510 + }, + { + "epoch": 0.5064077669902912, + "grad_norm": 0.34240055084228516, + "learning_rate": 9.993693088045939e-05, + "loss": 0.0212, + "step": 6520 + }, + { + "epoch": 0.5071844660194175, + "grad_norm": 0.3379019498825073, + "learning_rate": 9.99360979214268e-05, + "loss": 0.0187, + "step": 6530 + }, + { + "epoch": 0.5079611650485437, + "grad_norm": 0.2640601098537445, + "learning_rate": 9.99352595014576e-05, + "loss": 0.0172, + "step": 6540 + }, + { + "epoch": 0.5087378640776699, + "grad_norm": 0.43914639949798584, + "learning_rate": 9.993441562064354e-05, + "loss": 0.0196, + "step": 6550 + }, + { + "epoch": 0.5095145631067961, + "grad_norm": 0.30464524030685425, + "learning_rate": 9.993356627907685e-05, + "loss": 0.018, + "step": 6560 + }, + { + "epoch": 0.5102912621359224, + "grad_norm": 0.3284054100513458, + "learning_rate": 9.99327114768504e-05, + "loss": 0.0174, + "step": 6570 + }, + { + "epoch": 0.5110679611650485, + "grad_norm": 0.31302401423454285, + "learning_rate": 9.99318512140577e-05, + "loss": 0.0202, + "step": 6580 + }, + { + "epoch": 0.5118446601941747, + "grad_norm": 0.5139011144638062, + "learning_rate": 9.993098549079284e-05, + "loss": 0.0163, + "step": 6590 + }, + { + "epoch": 0.512621359223301, + "grad_norm": 0.37007373571395874, + "learning_rate": 9.993011430715047e-05, + "loss": 0.0179, + "step": 6600 + }, + { + "epoch": 0.5133980582524272, + "grad_norm": 0.3782326877117157, + "learning_rate": 9.992923766322586e-05, + "loss": 0.0202, + "step": 6610 + }, + { + "epoch": 0.5141747572815534, + "grad_norm": 0.36150774359703064, + "learning_rate": 9.99283555591149e-05, + "loss": 0.0164, + "step": 6620 + }, + { + "epoch": 0.5149514563106796, + "grad_norm": 0.5067179799079895, + "learning_rate": 9.992746799491404e-05, + "loss": 0.0179, + "step": 6630 + }, + { + "epoch": 0.5157281553398059, + "grad_norm": 0.347018837928772, + "learning_rate": 9.992657497072033e-05, + "loss": 0.0155, + "step": 6640 + }, + { + "epoch": 0.516504854368932, + "grad_norm": 0.37512338161468506, + "learning_rate": 9.992567648663147e-05, + "loss": 0.0143, + "step": 6650 + }, + { + "epoch": 0.5172815533980583, + "grad_norm": 0.41735517978668213, + "learning_rate": 9.992477254274568e-05, + "loss": 0.0182, + "step": 6660 + }, + { + "epoch": 0.5180582524271845, + "grad_norm": 0.3262370228767395, + "learning_rate": 9.992386313916183e-05, + "loss": 0.0167, + "step": 6670 + }, + { + "epoch": 0.5188349514563106, + "grad_norm": 0.46614131331443787, + "learning_rate": 9.992294827597934e-05, + "loss": 0.0167, + "step": 6680 + }, + { + "epoch": 0.5196116504854369, + "grad_norm": 0.29545122385025024, + "learning_rate": 9.992202795329831e-05, + "loss": 0.0175, + "step": 6690 + }, + { + "epoch": 0.5203883495145631, + "grad_norm": 0.3142915666103363, + "learning_rate": 9.992110217121936e-05, + "loss": 0.0154, + "step": 6700 + }, + { + "epoch": 0.5211650485436893, + "grad_norm": 0.21099340915679932, + "learning_rate": 9.992017092984372e-05, + "loss": 0.0142, + "step": 6710 + }, + { + "epoch": 0.5219417475728155, + "grad_norm": 0.5174623727798462, + "learning_rate": 9.991923422927326e-05, + "loss": 0.0148, + "step": 6720 + }, + { + "epoch": 0.5227184466019418, + "grad_norm": 0.46377032995224, + "learning_rate": 9.991829206961037e-05, + "loss": 0.0194, + "step": 6730 + }, + { + "epoch": 0.5234951456310679, + "grad_norm": 0.4720034599304199, + "learning_rate": 9.991734445095813e-05, + "loss": 0.0205, + "step": 6740 + }, + { + "epoch": 0.5242718446601942, + "grad_norm": 0.4895906448364258, + "learning_rate": 9.991639137342015e-05, + "loss": 0.0195, + "step": 6750 + }, + { + "epoch": 0.5250485436893204, + "grad_norm": 0.31778690218925476, + "learning_rate": 9.991543283710064e-05, + "loss": 0.0177, + "step": 6760 + }, + { + "epoch": 0.5258252427184466, + "grad_norm": 0.34155014157295227, + "learning_rate": 9.991446884210445e-05, + "loss": 0.0179, + "step": 6770 + }, + { + "epoch": 0.5266019417475728, + "grad_norm": 0.3463625907897949, + "learning_rate": 9.9913499388537e-05, + "loss": 0.0164, + "step": 6780 + }, + { + "epoch": 0.527378640776699, + "grad_norm": 0.34625837206840515, + "learning_rate": 9.99125244765043e-05, + "loss": 0.0141, + "step": 6790 + }, + { + "epoch": 0.5281553398058253, + "grad_norm": 0.4008655846118927, + "learning_rate": 9.991154410611296e-05, + "loss": 0.0177, + "step": 6800 + }, + { + "epoch": 0.5289320388349514, + "grad_norm": 0.28023552894592285, + "learning_rate": 9.99105582774702e-05, + "loss": 0.0144, + "step": 6810 + }, + { + "epoch": 0.5297087378640777, + "grad_norm": 0.2500804662704468, + "learning_rate": 9.990956699068384e-05, + "loss": 0.0139, + "step": 6820 + }, + { + "epoch": 0.5304854368932039, + "grad_norm": 0.3512135446071625, + "learning_rate": 9.990857024586224e-05, + "loss": 0.0145, + "step": 6830 + }, + { + "epoch": 0.5312621359223301, + "grad_norm": 0.35832706093788147, + "learning_rate": 9.990756804311446e-05, + "loss": 0.0123, + "step": 6840 + }, + { + "epoch": 0.5320388349514563, + "grad_norm": 0.2420641928911209, + "learning_rate": 9.990656038255006e-05, + "loss": 0.0142, + "step": 6850 + }, + { + "epoch": 0.5328155339805826, + "grad_norm": 0.47498124837875366, + "learning_rate": 9.990554726427926e-05, + "loss": 0.0175, + "step": 6860 + }, + { + "epoch": 0.5335922330097087, + "grad_norm": 0.5251485109329224, + "learning_rate": 9.990452868841284e-05, + "loss": 0.0194, + "step": 6870 + }, + { + "epoch": 0.534368932038835, + "grad_norm": 0.3562118709087372, + "learning_rate": 9.99035046550622e-05, + "loss": 0.0184, + "step": 6880 + }, + { + "epoch": 0.5351456310679612, + "grad_norm": 0.2649781107902527, + "learning_rate": 9.99024751643393e-05, + "loss": 0.0171, + "step": 6890 + }, + { + "epoch": 0.5359223300970873, + "grad_norm": 0.3406604528427124, + "learning_rate": 9.990144021635677e-05, + "loss": 0.0144, + "step": 6900 + }, + { + "epoch": 0.5366990291262136, + "grad_norm": 0.2764207422733307, + "learning_rate": 9.990039981122775e-05, + "loss": 0.016, + "step": 6910 + }, + { + "epoch": 0.5374757281553398, + "grad_norm": 0.26267901062965393, + "learning_rate": 9.989935394906602e-05, + "loss": 0.02, + "step": 6920 + }, + { + "epoch": 0.538252427184466, + "grad_norm": 0.31150928139686584, + "learning_rate": 9.989830262998598e-05, + "loss": 0.018, + "step": 6930 + }, + { + "epoch": 0.5390291262135922, + "grad_norm": 0.43066272139549255, + "learning_rate": 9.989724585410259e-05, + "loss": 0.0159, + "step": 6940 + }, + { + "epoch": 0.5398058252427185, + "grad_norm": 0.4236268997192383, + "learning_rate": 9.989618362153139e-05, + "loss": 0.0185, + "step": 6950 + }, + { + "epoch": 0.5405825242718446, + "grad_norm": 0.42831772565841675, + "learning_rate": 9.989511593238859e-05, + "loss": 0.0179, + "step": 6960 + }, + { + "epoch": 0.5413592233009709, + "grad_norm": 0.5109051465988159, + "learning_rate": 9.98940427867909e-05, + "loss": 0.0177, + "step": 6970 + }, + { + "epoch": 0.5421359223300971, + "grad_norm": 0.27260082960128784, + "learning_rate": 9.989296418485573e-05, + "loss": 0.0172, + "step": 6980 + }, + { + "epoch": 0.5429126213592232, + "grad_norm": 0.30345478653907776, + "learning_rate": 9.989188012670101e-05, + "loss": 0.0157, + "step": 6990 + }, + { + "epoch": 0.5436893203883495, + "grad_norm": 0.40549635887145996, + "learning_rate": 9.989079061244528e-05, + "loss": 0.0158, + "step": 7000 + }, + { + "epoch": 0.5444660194174757, + "grad_norm": 0.27612951397895813, + "learning_rate": 9.988969564220769e-05, + "loss": 0.0137, + "step": 7010 + }, + { + "epoch": 0.545242718446602, + "grad_norm": 0.26261597871780396, + "learning_rate": 9.988859521610801e-05, + "loss": 0.0126, + "step": 7020 + }, + { + "epoch": 0.5460194174757281, + "grad_norm": 0.3542844355106354, + "learning_rate": 9.988748933426656e-05, + "loss": 0.0178, + "step": 7030 + }, + { + "epoch": 0.5467961165048544, + "grad_norm": 0.17473441362380981, + "learning_rate": 9.988637799680428e-05, + "loss": 0.0177, + "step": 7040 + }, + { + "epoch": 0.5475728155339806, + "grad_norm": 0.3954033851623535, + "learning_rate": 9.98852612038427e-05, + "loss": 0.0219, + "step": 7050 + }, + { + "epoch": 0.5483495145631068, + "grad_norm": 0.3033815622329712, + "learning_rate": 9.988413895550397e-05, + "loss": 0.0131, + "step": 7060 + }, + { + "epoch": 0.549126213592233, + "grad_norm": 0.36701810359954834, + "learning_rate": 9.98830112519108e-05, + "loss": 0.0189, + "step": 7070 + }, + { + "epoch": 0.5499029126213593, + "grad_norm": 0.28242605924606323, + "learning_rate": 9.98818780931865e-05, + "loss": 0.0157, + "step": 7080 + }, + { + "epoch": 0.5506796116504854, + "grad_norm": 0.42808806896209717, + "learning_rate": 9.988073947945502e-05, + "loss": 0.0153, + "step": 7090 + }, + { + "epoch": 0.5514563106796116, + "grad_norm": 0.22296378016471863, + "learning_rate": 9.987959541084087e-05, + "loss": 0.0159, + "step": 7100 + }, + { + "epoch": 0.5522330097087379, + "grad_norm": 0.28445670008659363, + "learning_rate": 9.987844588746915e-05, + "loss": 0.0183, + "step": 7110 + }, + { + "epoch": 0.553009708737864, + "grad_norm": 0.3410952687263489, + "learning_rate": 9.987729090946558e-05, + "loss": 0.0145, + "step": 7120 + }, + { + "epoch": 0.5537864077669903, + "grad_norm": 0.4294724464416504, + "learning_rate": 9.987613047695647e-05, + "loss": 0.0149, + "step": 7130 + }, + { + "epoch": 0.5545631067961165, + "grad_norm": 0.23027269542217255, + "learning_rate": 9.987496459006871e-05, + "loss": 0.0169, + "step": 7140 + }, + { + "epoch": 0.5553398058252427, + "grad_norm": 0.2998911142349243, + "learning_rate": 9.987379324892982e-05, + "loss": 0.0168, + "step": 7150 + }, + { + "epoch": 0.5561165048543689, + "grad_norm": 0.14063705503940582, + "learning_rate": 9.987261645366788e-05, + "loss": 0.0123, + "step": 7160 + }, + { + "epoch": 0.5568932038834952, + "grad_norm": 0.41211482882499695, + "learning_rate": 9.987143420441158e-05, + "loss": 0.016, + "step": 7170 + }, + { + "epoch": 0.5576699029126213, + "grad_norm": 0.35137468576431274, + "learning_rate": 9.987024650129022e-05, + "loss": 0.014, + "step": 7180 + }, + { + "epoch": 0.5584466019417476, + "grad_norm": 0.39560216665267944, + "learning_rate": 9.986905334443368e-05, + "loss": 0.0198, + "step": 7190 + }, + { + "epoch": 0.5592233009708738, + "grad_norm": 0.3400043845176697, + "learning_rate": 9.986785473397245e-05, + "loss": 0.0177, + "step": 7200 + }, + { + "epoch": 0.56, + "grad_norm": 0.29368317127227783, + "learning_rate": 9.98666506700376e-05, + "loss": 0.0153, + "step": 7210 + }, + { + "epoch": 0.5607766990291262, + "grad_norm": 0.2613251805305481, + "learning_rate": 9.986544115276081e-05, + "loss": 0.0161, + "step": 7220 + }, + { + "epoch": 0.5615533980582524, + "grad_norm": 0.28277018666267395, + "learning_rate": 9.986422618227433e-05, + "loss": 0.0195, + "step": 7230 + }, + { + "epoch": 0.5623300970873787, + "grad_norm": 0.3667984902858734, + "learning_rate": 9.986300575871106e-05, + "loss": 0.0145, + "step": 7240 + }, + { + "epoch": 0.5631067961165048, + "grad_norm": 0.3810027241706848, + "learning_rate": 9.986177988220444e-05, + "loss": 0.0151, + "step": 7250 + }, + { + "epoch": 0.5638834951456311, + "grad_norm": 0.3792873024940491, + "learning_rate": 9.986054855288856e-05, + "loss": 0.0164, + "step": 7260 + }, + { + "epoch": 0.5646601941747573, + "grad_norm": 0.6326590180397034, + "learning_rate": 9.985931177089802e-05, + "loss": 0.015, + "step": 7270 + }, + { + "epoch": 0.5654368932038835, + "grad_norm": 0.24789796769618988, + "learning_rate": 9.985806953636814e-05, + "loss": 0.0124, + "step": 7280 + }, + { + "epoch": 0.5662135922330097, + "grad_norm": 0.41628801822662354, + "learning_rate": 9.985682184943471e-05, + "loss": 0.0158, + "step": 7290 + }, + { + "epoch": 0.566990291262136, + "grad_norm": 0.33503293991088867, + "learning_rate": 9.98555687102342e-05, + "loss": 0.0166, + "step": 7300 + }, + { + "epoch": 0.5677669902912621, + "grad_norm": 0.3798504173755646, + "learning_rate": 9.985431011890367e-05, + "loss": 0.0168, + "step": 7310 + }, + { + "epoch": 0.5685436893203883, + "grad_norm": 0.2787573039531708, + "learning_rate": 9.985304607558075e-05, + "loss": 0.0146, + "step": 7320 + }, + { + "epoch": 0.5693203883495146, + "grad_norm": 0.2719564437866211, + "learning_rate": 9.985177658040364e-05, + "loss": 0.0187, + "step": 7330 + }, + { + "epoch": 0.5700970873786407, + "grad_norm": 0.33143502473831177, + "learning_rate": 9.985050163351119e-05, + "loss": 0.0169, + "step": 7340 + }, + { + "epoch": 0.570873786407767, + "grad_norm": 0.29812484979629517, + "learning_rate": 9.984922123504286e-05, + "loss": 0.0123, + "step": 7350 + }, + { + "epoch": 0.5716504854368932, + "grad_norm": 0.30803051590919495, + "learning_rate": 9.984793538513862e-05, + "loss": 0.0176, + "step": 7360 + }, + { + "epoch": 0.5724271844660194, + "grad_norm": 0.332546204328537, + "learning_rate": 9.984664408393912e-05, + "loss": 0.0143, + "step": 7370 + }, + { + "epoch": 0.5732038834951456, + "grad_norm": 0.6037254333496094, + "learning_rate": 9.984534733158556e-05, + "loss": 0.0143, + "step": 7380 + }, + { + "epoch": 0.5739805825242719, + "grad_norm": 0.24710653722286224, + "learning_rate": 9.984404512821977e-05, + "loss": 0.0167, + "step": 7390 + }, + { + "epoch": 0.574757281553398, + "grad_norm": 0.2487608939409256, + "learning_rate": 9.984273747398411e-05, + "loss": 0.0173, + "step": 7400 + }, + { + "epoch": 0.5755339805825243, + "grad_norm": 0.2980274260044098, + "learning_rate": 9.984142436902165e-05, + "loss": 0.0118, + "step": 7410 + }, + { + "epoch": 0.5763106796116505, + "grad_norm": 0.3089301884174347, + "learning_rate": 9.984010581347596e-05, + "loss": 0.0214, + "step": 7420 + }, + { + "epoch": 0.5770873786407767, + "grad_norm": 0.42305371165275574, + "learning_rate": 9.983878180749121e-05, + "loss": 0.0178, + "step": 7430 + }, + { + "epoch": 0.5778640776699029, + "grad_norm": 0.25175926089286804, + "learning_rate": 9.983745235121222e-05, + "loss": 0.0168, + "step": 7440 + }, + { + "epoch": 0.5786407766990291, + "grad_norm": 0.3087265193462372, + "learning_rate": 9.983611744478438e-05, + "loss": 0.0135, + "step": 7450 + }, + { + "epoch": 0.5794174757281554, + "grad_norm": 0.26874831318855286, + "learning_rate": 9.983477708835365e-05, + "loss": 0.0166, + "step": 7460 + }, + { + "epoch": 0.5801941747572815, + "grad_norm": 0.5688969492912292, + "learning_rate": 9.983343128206664e-05, + "loss": 0.0169, + "step": 7470 + }, + { + "epoch": 0.5809708737864078, + "grad_norm": 0.5264084935188293, + "learning_rate": 9.983208002607049e-05, + "loss": 0.0162, + "step": 7480 + }, + { + "epoch": 0.581747572815534, + "grad_norm": 0.3427838385105133, + "learning_rate": 9.9830723320513e-05, + "loss": 0.0178, + "step": 7490 + }, + { + "epoch": 0.5825242718446602, + "grad_norm": 0.3336305618286133, + "learning_rate": 9.982936116554254e-05, + "loss": 0.0145, + "step": 7500 + }, + { + "epoch": 0.5833009708737864, + "grad_norm": 0.3156367540359497, + "learning_rate": 9.982799356130803e-05, + "loss": 0.0133, + "step": 7510 + }, + { + "epoch": 0.5840776699029127, + "grad_norm": 0.17890219390392303, + "learning_rate": 9.982662050795908e-05, + "loss": 0.0166, + "step": 7520 + }, + { + "epoch": 0.5848543689320388, + "grad_norm": 0.4394797086715698, + "learning_rate": 9.982524200564583e-05, + "loss": 0.015, + "step": 7530 + }, + { + "epoch": 0.585631067961165, + "grad_norm": 0.316152423620224, + "learning_rate": 9.982385805451901e-05, + "loss": 0.0143, + "step": 7540 + }, + { + "epoch": 0.5864077669902913, + "grad_norm": 0.19562052190303802, + "learning_rate": 9.982246865472998e-05, + "loss": 0.0159, + "step": 7550 + }, + { + "epoch": 0.5871844660194174, + "grad_norm": 0.2780504524707794, + "learning_rate": 9.982107380643069e-05, + "loss": 0.0172, + "step": 7560 + }, + { + "epoch": 0.5879611650485437, + "grad_norm": 0.2988039553165436, + "learning_rate": 9.981967350977368e-05, + "loss": 0.017, + "step": 7570 + }, + { + "epoch": 0.5887378640776699, + "grad_norm": 0.33791035413742065, + "learning_rate": 9.981826776491208e-05, + "loss": 0.0202, + "step": 7580 + }, + { + "epoch": 0.5895145631067961, + "grad_norm": 0.2815684974193573, + "learning_rate": 9.98168565719996e-05, + "loss": 0.0149, + "step": 7590 + }, + { + "epoch": 0.5902912621359223, + "grad_norm": 0.3649086356163025, + "learning_rate": 9.98154399311906e-05, + "loss": 0.0168, + "step": 7600 + }, + { + "epoch": 0.5910679611650486, + "grad_norm": 0.26662760972976685, + "learning_rate": 9.981401784263997e-05, + "loss": 0.0148, + "step": 7610 + }, + { + "epoch": 0.5918446601941748, + "grad_norm": 0.26444947719573975, + "learning_rate": 9.981259030650326e-05, + "loss": 0.0168, + "step": 7620 + }, + { + "epoch": 0.592621359223301, + "grad_norm": 0.4364876449108124, + "learning_rate": 9.981115732293655e-05, + "loss": 0.0137, + "step": 7630 + }, + { + "epoch": 0.5933980582524272, + "grad_norm": 0.35444700717926025, + "learning_rate": 9.980971889209659e-05, + "loss": 0.0176, + "step": 7640 + }, + { + "epoch": 0.5941747572815534, + "grad_norm": 0.25106340646743774, + "learning_rate": 9.980827501414064e-05, + "loss": 0.0165, + "step": 7650 + }, + { + "epoch": 0.5949514563106796, + "grad_norm": 0.4254549443721771, + "learning_rate": 9.980682568922663e-05, + "loss": 0.0158, + "step": 7660 + }, + { + "epoch": 0.5957281553398058, + "grad_norm": 0.3347803056240082, + "learning_rate": 9.980537091751304e-05, + "loss": 0.0195, + "step": 7670 + }, + { + "epoch": 0.5965048543689321, + "grad_norm": 0.3162393569946289, + "learning_rate": 9.980391069915897e-05, + "loss": 0.015, + "step": 7680 + }, + { + "epoch": 0.5972815533980582, + "grad_norm": 0.3111127018928528, + "learning_rate": 9.98024450343241e-05, + "loss": 0.0156, + "step": 7690 + }, + { + "epoch": 0.5980582524271845, + "grad_norm": 0.3425641357898712, + "learning_rate": 9.980097392316872e-05, + "loss": 0.0158, + "step": 7700 + }, + { + "epoch": 0.5988349514563107, + "grad_norm": 0.303239107131958, + "learning_rate": 9.97994973658537e-05, + "loss": 0.0164, + "step": 7710 + }, + { + "epoch": 0.5996116504854369, + "grad_norm": 0.22363735735416412, + "learning_rate": 9.979801536254054e-05, + "loss": 0.0135, + "step": 7720 + }, + { + "epoch": 0.6003883495145631, + "grad_norm": 0.29646068811416626, + "learning_rate": 9.979652791339127e-05, + "loss": 0.0143, + "step": 7730 + }, + { + "epoch": 0.6011650485436894, + "grad_norm": 0.31545495986938477, + "learning_rate": 9.97950350185686e-05, + "loss": 0.015, + "step": 7740 + }, + { + "epoch": 0.6019417475728155, + "grad_norm": 0.2570723593235016, + "learning_rate": 9.979353667823574e-05, + "loss": 0.0145, + "step": 7750 + }, + { + "epoch": 0.6027184466019417, + "grad_norm": 0.322750985622406, + "learning_rate": 9.979203289255658e-05, + "loss": 0.0151, + "step": 7760 + }, + { + "epoch": 0.603495145631068, + "grad_norm": 0.22678884863853455, + "learning_rate": 9.979052366169557e-05, + "loss": 0.0116, + "step": 7770 + }, + { + "epoch": 0.6042718446601941, + "grad_norm": 0.3446555435657501, + "learning_rate": 9.978900898581775e-05, + "loss": 0.0151, + "step": 7780 + }, + { + "epoch": 0.6050485436893204, + "grad_norm": 0.21116575598716736, + "learning_rate": 9.978748886508875e-05, + "loss": 0.0151, + "step": 7790 + }, + { + "epoch": 0.6058252427184466, + "grad_norm": 0.24494978785514832, + "learning_rate": 9.978596329967484e-05, + "loss": 0.0165, + "step": 7800 + }, + { + "epoch": 0.6066019417475729, + "grad_norm": 0.35108110308647156, + "learning_rate": 9.978443228974284e-05, + "loss": 0.0164, + "step": 7810 + }, + { + "epoch": 0.607378640776699, + "grad_norm": 0.2550705671310425, + "learning_rate": 9.978289583546015e-05, + "loss": 0.0203, + "step": 7820 + }, + { + "epoch": 0.6081553398058253, + "grad_norm": 0.2022348791360855, + "learning_rate": 9.978135393699484e-05, + "loss": 0.0184, + "step": 7830 + }, + { + "epoch": 0.6089320388349515, + "grad_norm": 0.25020432472229004, + "learning_rate": 9.977980659451548e-05, + "loss": 0.0171, + "step": 7840 + }, + { + "epoch": 0.6097087378640776, + "grad_norm": 0.3362933397293091, + "learning_rate": 9.977825380819135e-05, + "loss": 0.0187, + "step": 7850 + }, + { + "epoch": 0.6104854368932039, + "grad_norm": 0.32346847653388977, + "learning_rate": 9.97766955781922e-05, + "loss": 0.0157, + "step": 7860 + }, + { + "epoch": 0.6112621359223301, + "grad_norm": 0.26583409309387207, + "learning_rate": 9.977513190468848e-05, + "loss": 0.0173, + "step": 7870 + }, + { + "epoch": 0.6120388349514563, + "grad_norm": 0.3060167729854584, + "learning_rate": 9.977356278785116e-05, + "loss": 0.0173, + "step": 7880 + }, + { + "epoch": 0.6128155339805825, + "grad_norm": 0.3035072684288025, + "learning_rate": 9.977198822785184e-05, + "loss": 0.0137, + "step": 7890 + }, + { + "epoch": 0.6135922330097088, + "grad_norm": 0.2905961573123932, + "learning_rate": 9.977040822486273e-05, + "loss": 0.013, + "step": 7900 + }, + { + "epoch": 0.6143689320388349, + "grad_norm": 0.28336474299430847, + "learning_rate": 9.97688227790566e-05, + "loss": 0.018, + "step": 7910 + }, + { + "epoch": 0.6151456310679612, + "grad_norm": 0.35250160098075867, + "learning_rate": 9.976723189060684e-05, + "loss": 0.0141, + "step": 7920 + }, + { + "epoch": 0.6159223300970874, + "grad_norm": 0.3245476484298706, + "learning_rate": 9.976563555968742e-05, + "loss": 0.0129, + "step": 7930 + }, + { + "epoch": 0.6166990291262135, + "grad_norm": 0.29178136587142944, + "learning_rate": 9.976403378647292e-05, + "loss": 0.0168, + "step": 7940 + }, + { + "epoch": 0.6174757281553398, + "grad_norm": 0.3510233759880066, + "learning_rate": 9.97624265711385e-05, + "loss": 0.0203, + "step": 7950 + }, + { + "epoch": 0.618252427184466, + "grad_norm": 0.3490883409976959, + "learning_rate": 9.976081391385993e-05, + "loss": 0.0176, + "step": 7960 + }, + { + "epoch": 0.6190291262135922, + "grad_norm": 0.42360273003578186, + "learning_rate": 9.975919581481356e-05, + "loss": 0.0151, + "step": 7970 + }, + { + "epoch": 0.6198058252427184, + "grad_norm": 0.5113619565963745, + "learning_rate": 9.975757227417634e-05, + "loss": 0.0232, + "step": 7980 + }, + { + "epoch": 0.6205825242718447, + "grad_norm": 0.8022323846817017, + "learning_rate": 9.975594329212586e-05, + "loss": 0.022, + "step": 7990 + }, + { + "epoch": 0.6213592233009708, + "grad_norm": 0.3410646319389343, + "learning_rate": 9.97543088688402e-05, + "loss": 0.0199, + "step": 8000 + }, + { + "epoch": 0.6221359223300971, + "grad_norm": 0.4583568572998047, + "learning_rate": 9.975266900449814e-05, + "loss": 0.0177, + "step": 8010 + }, + { + "epoch": 0.6229126213592233, + "grad_norm": 0.28258320689201355, + "learning_rate": 9.975102369927898e-05, + "loss": 0.0197, + "step": 8020 + }, + { + "epoch": 0.6236893203883496, + "grad_norm": 0.3217106759548187, + "learning_rate": 9.974937295336269e-05, + "loss": 0.0166, + "step": 8030 + }, + { + "epoch": 0.6244660194174757, + "grad_norm": 0.3444070816040039, + "learning_rate": 9.974771676692975e-05, + "loss": 0.0151, + "step": 8040 + }, + { + "epoch": 0.625242718446602, + "grad_norm": 0.2805786728858948, + "learning_rate": 9.974605514016131e-05, + "loss": 0.018, + "step": 8050 + }, + { + "epoch": 0.6260194174757282, + "grad_norm": 0.33412832021713257, + "learning_rate": 9.974438807323907e-05, + "loss": 0.019, + "step": 8060 + }, + { + "epoch": 0.6267961165048543, + "grad_norm": 0.2893219590187073, + "learning_rate": 9.974271556634535e-05, + "loss": 0.014, + "step": 8070 + }, + { + "epoch": 0.6275728155339806, + "grad_norm": 0.24753254652023315, + "learning_rate": 9.974103761966302e-05, + "loss": 0.0139, + "step": 8080 + }, + { + "epoch": 0.6283495145631068, + "grad_norm": 0.3173404633998871, + "learning_rate": 9.973935423337563e-05, + "loss": 0.0134, + "step": 8090 + }, + { + "epoch": 0.629126213592233, + "grad_norm": 0.38256287574768066, + "learning_rate": 9.973766540766722e-05, + "loss": 0.0176, + "step": 8100 + }, + { + "epoch": 0.6299029126213592, + "grad_norm": 0.531424880027771, + "learning_rate": 9.97359711427225e-05, + "loss": 0.0148, + "step": 8110 + }, + { + "epoch": 0.6306796116504855, + "grad_norm": 0.31608572602272034, + "learning_rate": 9.973427143872677e-05, + "loss": 0.0158, + "step": 8120 + }, + { + "epoch": 0.6314563106796116, + "grad_norm": 0.22516775131225586, + "learning_rate": 9.973256629586589e-05, + "loss": 0.0131, + "step": 8130 + }, + { + "epoch": 0.6322330097087379, + "grad_norm": 0.36061549186706543, + "learning_rate": 9.973085571432632e-05, + "loss": 0.016, + "step": 8140 + }, + { + "epoch": 0.6330097087378641, + "grad_norm": 0.3289807438850403, + "learning_rate": 9.972913969429513e-05, + "loss": 0.0163, + "step": 8150 + }, + { + "epoch": 0.6337864077669902, + "grad_norm": 0.19661888480186462, + "learning_rate": 9.972741823596e-05, + "loss": 0.0135, + "step": 8160 + }, + { + "epoch": 0.6345631067961165, + "grad_norm": 0.3306589722633362, + "learning_rate": 9.972569133950917e-05, + "loss": 0.0118, + "step": 8170 + }, + { + "epoch": 0.6353398058252427, + "grad_norm": 0.2810363471508026, + "learning_rate": 9.972395900513151e-05, + "loss": 0.0154, + "step": 8180 + }, + { + "epoch": 0.6361165048543689, + "grad_norm": 0.4981288015842438, + "learning_rate": 9.972222123301645e-05, + "loss": 0.018, + "step": 8190 + }, + { + "epoch": 0.6368932038834951, + "grad_norm": 0.5190226435661316, + "learning_rate": 9.972047802335403e-05, + "loss": 0.017, + "step": 8200 + }, + { + "epoch": 0.6376699029126214, + "grad_norm": 0.31964296102523804, + "learning_rate": 9.971872937633488e-05, + "loss": 0.0156, + "step": 8210 + }, + { + "epoch": 0.6384466019417476, + "grad_norm": 0.49651870131492615, + "learning_rate": 9.971697529215024e-05, + "loss": 0.0139, + "step": 8220 + }, + { + "epoch": 0.6392233009708738, + "grad_norm": 0.3881778120994568, + "learning_rate": 9.971521577099192e-05, + "loss": 0.0147, + "step": 8230 + }, + { + "epoch": 0.64, + "grad_norm": 0.45811912417411804, + "learning_rate": 9.971345081305236e-05, + "loss": 0.0138, + "step": 8240 + }, + { + "epoch": 0.6407766990291263, + "grad_norm": 0.29081302881240845, + "learning_rate": 9.971168041852456e-05, + "loss": 0.0159, + "step": 8250 + }, + { + "epoch": 0.6415533980582524, + "grad_norm": 0.5996077656745911, + "learning_rate": 9.970990458760215e-05, + "loss": 0.0148, + "step": 8260 + }, + { + "epoch": 0.6423300970873786, + "grad_norm": 0.4130687415599823, + "learning_rate": 9.970812332047929e-05, + "loss": 0.0204, + "step": 8270 + }, + { + "epoch": 0.6431067961165049, + "grad_norm": 0.5191580653190613, + "learning_rate": 9.97063366173508e-05, + "loss": 0.0173, + "step": 8280 + }, + { + "epoch": 0.643883495145631, + "grad_norm": 0.38587403297424316, + "learning_rate": 9.970454447841207e-05, + "loss": 0.0151, + "step": 8290 + }, + { + "epoch": 0.6446601941747573, + "grad_norm": 0.3755660057067871, + "learning_rate": 9.970274690385909e-05, + "loss": 0.0151, + "step": 8300 + }, + { + "epoch": 0.6454368932038835, + "grad_norm": 0.3450181484222412, + "learning_rate": 9.970094389388844e-05, + "loss": 0.0176, + "step": 8310 + }, + { + "epoch": 0.6462135922330097, + "grad_norm": 0.3748031556606293, + "learning_rate": 9.969913544869728e-05, + "loss": 0.0138, + "step": 8320 + }, + { + "epoch": 0.6469902912621359, + "grad_norm": 0.33277133107185364, + "learning_rate": 9.96973215684834e-05, + "loss": 0.0135, + "step": 8330 + }, + { + "epoch": 0.6477669902912622, + "grad_norm": 0.1890735626220703, + "learning_rate": 9.969550225344513e-05, + "loss": 0.0158, + "step": 8340 + }, + { + "epoch": 0.6485436893203883, + "grad_norm": 0.26355743408203125, + "learning_rate": 9.969367750378147e-05, + "loss": 0.0141, + "step": 8350 + }, + { + "epoch": 0.6493203883495146, + "grad_norm": 0.37233951687812805, + "learning_rate": 9.969184731969194e-05, + "loss": 0.0127, + "step": 8360 + }, + { + "epoch": 0.6500970873786408, + "grad_norm": 0.4224410653114319, + "learning_rate": 9.96900117013767e-05, + "loss": 0.0156, + "step": 8370 + }, + { + "epoch": 0.6508737864077669, + "grad_norm": 0.3273191750049591, + "learning_rate": 9.96881706490365e-05, + "loss": 0.0158, + "step": 8380 + }, + { + "epoch": 0.6516504854368932, + "grad_norm": 0.33521127700805664, + "learning_rate": 9.968632416287265e-05, + "loss": 0.0146, + "step": 8390 + }, + { + "epoch": 0.6524271844660194, + "grad_norm": 0.2230314463376999, + "learning_rate": 9.96844722430871e-05, + "loss": 0.0131, + "step": 8400 + }, + { + "epoch": 0.6532038834951456, + "grad_norm": 0.3418956398963928, + "learning_rate": 9.968261488988235e-05, + "loss": 0.0149, + "step": 8410 + }, + { + "epoch": 0.6539805825242718, + "grad_norm": 0.25055137276649475, + "learning_rate": 9.968075210346155e-05, + "loss": 0.0155, + "step": 8420 + }, + { + "epoch": 0.6547572815533981, + "grad_norm": 0.32306933403015137, + "learning_rate": 9.967888388402839e-05, + "loss": 0.0146, + "step": 8430 + }, + { + "epoch": 0.6555339805825243, + "grad_norm": 0.5882017612457275, + "learning_rate": 9.967701023178717e-05, + "loss": 0.0126, + "step": 8440 + }, + { + "epoch": 0.6563106796116505, + "grad_norm": 0.271259605884552, + "learning_rate": 9.967513114694282e-05, + "loss": 0.0149, + "step": 8450 + }, + { + "epoch": 0.6570873786407767, + "grad_norm": 0.3759942054748535, + "learning_rate": 9.967324662970079e-05, + "loss": 0.0148, + "step": 8460 + }, + { + "epoch": 0.657864077669903, + "grad_norm": 0.4494543969631195, + "learning_rate": 9.96713566802672e-05, + "loss": 0.0125, + "step": 8470 + }, + { + "epoch": 0.6586407766990291, + "grad_norm": 0.3638138771057129, + "learning_rate": 9.966946129884873e-05, + "loss": 0.0179, + "step": 8480 + }, + { + "epoch": 0.6594174757281553, + "grad_norm": 0.2409653663635254, + "learning_rate": 9.966756048565265e-05, + "loss": 0.0167, + "step": 8490 + }, + { + "epoch": 0.6601941747572816, + "grad_norm": 0.42240190505981445, + "learning_rate": 9.966565424088681e-05, + "loss": 0.0126, + "step": 8500 + }, + { + "epoch": 0.6609708737864077, + "grad_norm": 0.3392994701862335, + "learning_rate": 9.96637425647597e-05, + "loss": 0.0134, + "step": 8510 + }, + { + "epoch": 0.661747572815534, + "grad_norm": 0.37188202142715454, + "learning_rate": 9.966182545748038e-05, + "loss": 0.0141, + "step": 8520 + }, + { + "epoch": 0.6625242718446602, + "grad_norm": 0.3791256844997406, + "learning_rate": 9.96599029192585e-05, + "loss": 0.0146, + "step": 8530 + }, + { + "epoch": 0.6633009708737864, + "grad_norm": 0.3231237232685089, + "learning_rate": 9.965797495030428e-05, + "loss": 0.0136, + "step": 8540 + }, + { + "epoch": 0.6640776699029126, + "grad_norm": 0.20833691954612732, + "learning_rate": 9.96560415508286e-05, + "loss": 0.0127, + "step": 8550 + }, + { + "epoch": 0.6648543689320389, + "grad_norm": 0.22374531626701355, + "learning_rate": 9.965410272104286e-05, + "loss": 0.0125, + "step": 8560 + }, + { + "epoch": 0.665631067961165, + "grad_norm": 0.33289870619773865, + "learning_rate": 9.96521584611591e-05, + "loss": 0.0156, + "step": 8570 + }, + { + "epoch": 0.6664077669902913, + "grad_norm": 0.27489137649536133, + "learning_rate": 9.965020877138994e-05, + "loss": 0.0155, + "step": 8580 + }, + { + "epoch": 0.6671844660194175, + "grad_norm": 0.36780381202697754, + "learning_rate": 9.964825365194861e-05, + "loss": 0.0119, + "step": 8590 + }, + { + "epoch": 0.6679611650485436, + "grad_norm": 0.3872430622577667, + "learning_rate": 9.96462931030489e-05, + "loss": 0.016, + "step": 8600 + }, + { + "epoch": 0.6687378640776699, + "grad_norm": 0.287748247385025, + "learning_rate": 9.96443271249052e-05, + "loss": 0.0136, + "step": 8610 + }, + { + "epoch": 0.6695145631067961, + "grad_norm": 0.2723401188850403, + "learning_rate": 9.964235571773255e-05, + "loss": 0.0137, + "step": 8620 + }, + { + "epoch": 0.6702912621359224, + "grad_norm": 0.22571538388729095, + "learning_rate": 9.96403788817465e-05, + "loss": 0.011, + "step": 8630 + }, + { + "epoch": 0.6710679611650485, + "grad_norm": 0.2954121232032776, + "learning_rate": 9.963839661716325e-05, + "loss": 0.0131, + "step": 8640 + }, + { + "epoch": 0.6718446601941748, + "grad_norm": 0.2501232922077179, + "learning_rate": 9.963640892419958e-05, + "loss": 0.0154, + "step": 8650 + }, + { + "epoch": 0.672621359223301, + "grad_norm": 0.34844639897346497, + "learning_rate": 9.963441580307286e-05, + "loss": 0.0143, + "step": 8660 + }, + { + "epoch": 0.6733980582524272, + "grad_norm": 0.380234956741333, + "learning_rate": 9.963241725400104e-05, + "loss": 0.0138, + "step": 8670 + }, + { + "epoch": 0.6741747572815534, + "grad_norm": 0.6316119432449341, + "learning_rate": 9.963041327720271e-05, + "loss": 0.0209, + "step": 8680 + }, + { + "epoch": 0.6749514563106797, + "grad_norm": 0.24040722846984863, + "learning_rate": 9.962840387289697e-05, + "loss": 0.0125, + "step": 8690 + }, + { + "epoch": 0.6757281553398058, + "grad_norm": 0.35760223865509033, + "learning_rate": 9.962638904130363e-05, + "loss": 0.0207, + "step": 8700 + }, + { + "epoch": 0.676504854368932, + "grad_norm": 0.3477059304714203, + "learning_rate": 9.962436878264298e-05, + "loss": 0.0165, + "step": 8710 + }, + { + "epoch": 0.6772815533980583, + "grad_norm": 0.33241838216781616, + "learning_rate": 9.962234309713598e-05, + "loss": 0.0175, + "step": 8720 + }, + { + "epoch": 0.6780582524271844, + "grad_norm": 0.3006706237792969, + "learning_rate": 9.962031198500414e-05, + "loss": 0.0196, + "step": 8730 + }, + { + "epoch": 0.6788349514563107, + "grad_norm": 0.3834460377693176, + "learning_rate": 9.961827544646958e-05, + "loss": 0.0167, + "step": 8740 + }, + { + "epoch": 0.6796116504854369, + "grad_norm": 0.3594569265842438, + "learning_rate": 9.961623348175501e-05, + "loss": 0.0169, + "step": 8750 + }, + { + "epoch": 0.6803883495145631, + "grad_norm": 0.5181947946548462, + "learning_rate": 9.961418609108377e-05, + "loss": 0.0141, + "step": 8760 + }, + { + "epoch": 0.6811650485436893, + "grad_norm": 0.37276768684387207, + "learning_rate": 9.961213327467971e-05, + "loss": 0.0149, + "step": 8770 + }, + { + "epoch": 0.6819417475728156, + "grad_norm": 0.5892850160598755, + "learning_rate": 9.961007503276736e-05, + "loss": 0.0162, + "step": 8780 + }, + { + "epoch": 0.6827184466019417, + "grad_norm": 0.34722983837127686, + "learning_rate": 9.960801136557179e-05, + "loss": 0.0173, + "step": 8790 + }, + { + "epoch": 0.683495145631068, + "grad_norm": 0.5586703419685364, + "learning_rate": 9.960594227331866e-05, + "loss": 0.0146, + "step": 8800 + }, + { + "epoch": 0.6842718446601942, + "grad_norm": 0.4622187316417694, + "learning_rate": 9.960386775623429e-05, + "loss": 0.0181, + "step": 8810 + }, + { + "epoch": 0.6850485436893204, + "grad_norm": 0.3767361044883728, + "learning_rate": 9.96017878145455e-05, + "loss": 0.0155, + "step": 8820 + }, + { + "epoch": 0.6858252427184466, + "grad_norm": 0.22971542179584503, + "learning_rate": 9.959970244847977e-05, + "loss": 0.0166, + "step": 8830 + }, + { + "epoch": 0.6866019417475728, + "grad_norm": 0.30863428115844727, + "learning_rate": 9.959761165826518e-05, + "loss": 0.0123, + "step": 8840 + }, + { + "epoch": 0.6873786407766991, + "grad_norm": 0.23264813423156738, + "learning_rate": 9.959551544413033e-05, + "loss": 0.0158, + "step": 8850 + }, + { + "epoch": 0.6881553398058252, + "grad_norm": 0.22902080416679382, + "learning_rate": 9.959341380630448e-05, + "loss": 0.0165, + "step": 8860 + }, + { + "epoch": 0.6889320388349515, + "grad_norm": 0.20343707501888275, + "learning_rate": 9.959130674501746e-05, + "loss": 0.0141, + "step": 8870 + }, + { + "epoch": 0.6897087378640777, + "grad_norm": 0.22864899039268494, + "learning_rate": 9.958919426049968e-05, + "loss": 0.0167, + "step": 8880 + }, + { + "epoch": 0.6904854368932039, + "grad_norm": 0.36849528551101685, + "learning_rate": 9.958707635298219e-05, + "loss": 0.0146, + "step": 8890 + }, + { + "epoch": 0.6912621359223301, + "grad_norm": 0.24914893507957458, + "learning_rate": 9.958495302269657e-05, + "loss": 0.0169, + "step": 8900 + }, + { + "epoch": 0.6920388349514563, + "grad_norm": 0.28930625319480896, + "learning_rate": 9.958282426987503e-05, + "loss": 0.0161, + "step": 8910 + }, + { + "epoch": 0.6928155339805825, + "grad_norm": 0.40943753719329834, + "learning_rate": 9.95806900947504e-05, + "loss": 0.022, + "step": 8920 + }, + { + "epoch": 0.6935922330097087, + "grad_norm": 0.22602874040603638, + "learning_rate": 9.957855049755604e-05, + "loss": 0.0142, + "step": 8930 + }, + { + "epoch": 0.694368932038835, + "grad_norm": 0.2603902518749237, + "learning_rate": 9.957640547852593e-05, + "loss": 0.015, + "step": 8940 + }, + { + "epoch": 0.6951456310679611, + "grad_norm": 0.22436439990997314, + "learning_rate": 9.957425503789466e-05, + "loss": 0.019, + "step": 8950 + }, + { + "epoch": 0.6959223300970874, + "grad_norm": 0.3534005582332611, + "learning_rate": 9.957209917589738e-05, + "loss": 0.0133, + "step": 8960 + }, + { + "epoch": 0.6966990291262136, + "grad_norm": 0.27318495512008667, + "learning_rate": 9.956993789276987e-05, + "loss": 0.0109, + "step": 8970 + }, + { + "epoch": 0.6974757281553398, + "grad_norm": 0.3159703314304352, + "learning_rate": 9.956777118874847e-05, + "loss": 0.0212, + "step": 8980 + }, + { + "epoch": 0.698252427184466, + "grad_norm": 0.5181014537811279, + "learning_rate": 9.956559906407016e-05, + "loss": 0.0132, + "step": 8990 + }, + { + "epoch": 0.6990291262135923, + "grad_norm": 0.249860942363739, + "learning_rate": 9.956342151897245e-05, + "loss": 0.0131, + "step": 9000 + }, + { + "epoch": 0.6998058252427184, + "grad_norm": 0.3023490309715271, + "learning_rate": 9.956123855369346e-05, + "loss": 0.0122, + "step": 9010 + }, + { + "epoch": 0.7005825242718446, + "grad_norm": 0.27376091480255127, + "learning_rate": 9.955905016847196e-05, + "loss": 0.0159, + "step": 9020 + }, + { + "epoch": 0.7013592233009709, + "grad_norm": 0.3739902377128601, + "learning_rate": 9.955685636354723e-05, + "loss": 0.0135, + "step": 9030 + }, + { + "epoch": 0.7021359223300971, + "grad_norm": 0.32351812720298767, + "learning_rate": 9.95546571391592e-05, + "loss": 0.0132, + "step": 9040 + }, + { + "epoch": 0.7029126213592233, + "grad_norm": 0.3556567430496216, + "learning_rate": 9.955245249554837e-05, + "loss": 0.0149, + "step": 9050 + }, + { + "epoch": 0.7036893203883495, + "grad_norm": 0.3606586754322052, + "learning_rate": 9.955024243295582e-05, + "loss": 0.0188, + "step": 9060 + }, + { + "epoch": 0.7044660194174758, + "grad_norm": 0.3143426477909088, + "learning_rate": 9.954802695162328e-05, + "loss": 0.0108, + "step": 9070 + }, + { + "epoch": 0.7052427184466019, + "grad_norm": 0.42628347873687744, + "learning_rate": 9.954580605179302e-05, + "loss": 0.0205, + "step": 9080 + }, + { + "epoch": 0.7060194174757282, + "grad_norm": 0.21248868107795715, + "learning_rate": 9.954357973370788e-05, + "loss": 0.0162, + "step": 9090 + }, + { + "epoch": 0.7067961165048544, + "grad_norm": 0.21635748445987701, + "learning_rate": 9.954134799761135e-05, + "loss": 0.0157, + "step": 9100 + }, + { + "epoch": 0.7075728155339805, + "grad_norm": 0.2251570224761963, + "learning_rate": 9.953911084374748e-05, + "loss": 0.0136, + "step": 9110 + }, + { + "epoch": 0.7083495145631068, + "grad_norm": 0.29277512431144714, + "learning_rate": 9.953686827236093e-05, + "loss": 0.0147, + "step": 9120 + }, + { + "epoch": 0.709126213592233, + "grad_norm": 0.7427107095718384, + "learning_rate": 9.953462028369695e-05, + "loss": 0.0151, + "step": 9130 + }, + { + "epoch": 0.7099029126213592, + "grad_norm": 0.178310364484787, + "learning_rate": 9.953236687800136e-05, + "loss": 0.0164, + "step": 9140 + }, + { + "epoch": 0.7106796116504854, + "grad_norm": 0.33375856280326843, + "learning_rate": 9.95301080555206e-05, + "loss": 0.019, + "step": 9150 + }, + { + "epoch": 0.7114563106796117, + "grad_norm": 0.32048726081848145, + "learning_rate": 9.952784381650171e-05, + "loss": 0.0163, + "step": 9160 + }, + { + "epoch": 0.7122330097087378, + "grad_norm": 0.6426693797111511, + "learning_rate": 9.952557416119226e-05, + "loss": 0.0177, + "step": 9170 + }, + { + "epoch": 0.7130097087378641, + "grad_norm": 0.432279109954834, + "learning_rate": 9.95232990898405e-05, + "loss": 0.0193, + "step": 9180 + }, + { + "epoch": 0.7137864077669903, + "grad_norm": 0.32453039288520813, + "learning_rate": 9.95210186026952e-05, + "loss": 0.0117, + "step": 9190 + }, + { + "epoch": 0.7145631067961165, + "grad_norm": 0.4127701222896576, + "learning_rate": 9.951873270000576e-05, + "loss": 0.0189, + "step": 9200 + }, + { + "epoch": 0.7153398058252427, + "grad_norm": 0.2629903256893158, + "learning_rate": 9.951644138202216e-05, + "loss": 0.0155, + "step": 9210 + }, + { + "epoch": 0.716116504854369, + "grad_norm": 0.36856982111930847, + "learning_rate": 9.951414464899498e-05, + "loss": 0.0179, + "step": 9220 + }, + { + "epoch": 0.7168932038834952, + "grad_norm": 0.40755775570869446, + "learning_rate": 9.951184250117538e-05, + "loss": 0.0156, + "step": 9230 + }, + { + "epoch": 0.7176699029126213, + "grad_norm": 0.2967313826084137, + "learning_rate": 9.950953493881513e-05, + "loss": 0.0133, + "step": 9240 + }, + { + "epoch": 0.7184466019417476, + "grad_norm": 0.4234502911567688, + "learning_rate": 9.950722196216658e-05, + "loss": 0.0163, + "step": 9250 + }, + { + "epoch": 0.7192233009708738, + "grad_norm": 0.36015620827674866, + "learning_rate": 9.950490357148265e-05, + "loss": 0.0165, + "step": 9260 + }, + { + "epoch": 0.72, + "grad_norm": 0.4420570135116577, + "learning_rate": 9.950257976701692e-05, + "loss": 0.0181, + "step": 9270 + }, + { + "epoch": 0.7207766990291262, + "grad_norm": 0.30256137251853943, + "learning_rate": 9.950025054902348e-05, + "loss": 0.0226, + "step": 9280 + }, + { + "epoch": 0.7215533980582525, + "grad_norm": 0.24786187708377838, + "learning_rate": 9.949791591775706e-05, + "loss": 0.0146, + "step": 9290 + }, + { + "epoch": 0.7223300970873786, + "grad_norm": 0.2011193037033081, + "learning_rate": 9.949557587347298e-05, + "loss": 0.0186, + "step": 9300 + }, + { + "epoch": 0.7231067961165049, + "grad_norm": 0.3371869623661041, + "learning_rate": 9.949323041642713e-05, + "loss": 0.0142, + "step": 9310 + }, + { + "epoch": 0.7238834951456311, + "grad_norm": 0.1607266217470169, + "learning_rate": 9.949087954687602e-05, + "loss": 0.0152, + "step": 9320 + }, + { + "epoch": 0.7246601941747572, + "grad_norm": 0.4565463662147522, + "learning_rate": 9.948852326507672e-05, + "loss": 0.014, + "step": 9330 + }, + { + "epoch": 0.7254368932038835, + "grad_norm": 0.25878462195396423, + "learning_rate": 9.948616157128694e-05, + "loss": 0.0173, + "step": 9340 + }, + { + "epoch": 0.7262135922330097, + "grad_norm": 0.40987950563430786, + "learning_rate": 9.948379446576493e-05, + "loss": 0.0146, + "step": 9350 + }, + { + "epoch": 0.7269902912621359, + "grad_norm": 0.30094361305236816, + "learning_rate": 9.948142194876952e-05, + "loss": 0.0155, + "step": 9360 + }, + { + "epoch": 0.7277669902912621, + "grad_norm": 0.5316991209983826, + "learning_rate": 9.947904402056024e-05, + "loss": 0.0175, + "step": 9370 + }, + { + "epoch": 0.7285436893203884, + "grad_norm": 0.15502102673053741, + "learning_rate": 9.947666068139708e-05, + "loss": 0.0129, + "step": 9380 + }, + { + "epoch": 0.7293203883495145, + "grad_norm": 0.3450351655483246, + "learning_rate": 9.947427193154071e-05, + "loss": 0.016, + "step": 9390 + }, + { + "epoch": 0.7300970873786408, + "grad_norm": 0.5987892150878906, + "learning_rate": 9.947187777125233e-05, + "loss": 0.0139, + "step": 9400 + }, + { + "epoch": 0.730873786407767, + "grad_norm": 0.5194984674453735, + "learning_rate": 9.946947820079377e-05, + "loss": 0.0127, + "step": 9410 + }, + { + "epoch": 0.7316504854368931, + "grad_norm": 0.2032226175069809, + "learning_rate": 9.946707322042747e-05, + "loss": 0.0154, + "step": 9420 + }, + { + "epoch": 0.7324271844660194, + "grad_norm": 0.27418071031570435, + "learning_rate": 9.94646628304164e-05, + "loss": 0.0139, + "step": 9430 + }, + { + "epoch": 0.7332038834951456, + "grad_norm": 0.3581126928329468, + "learning_rate": 9.946224703102418e-05, + "loss": 0.0155, + "step": 9440 + }, + { + "epoch": 0.7339805825242719, + "grad_norm": 0.2519194185733795, + "learning_rate": 9.945982582251498e-05, + "loss": 0.017, + "step": 9450 + }, + { + "epoch": 0.734757281553398, + "grad_norm": 0.5239279866218567, + "learning_rate": 9.94573992051536e-05, + "loss": 0.0179, + "step": 9460 + }, + { + "epoch": 0.7355339805825243, + "grad_norm": 0.40536609292030334, + "learning_rate": 9.94549671792054e-05, + "loss": 0.017, + "step": 9470 + }, + { + "epoch": 0.7363106796116505, + "grad_norm": 0.2922605276107788, + "learning_rate": 9.945252974493635e-05, + "loss": 0.0162, + "step": 9480 + }, + { + "epoch": 0.7370873786407767, + "grad_norm": 0.39521291851997375, + "learning_rate": 9.9450086902613e-05, + "loss": 0.018, + "step": 9490 + }, + { + "epoch": 0.7378640776699029, + "grad_norm": 0.32192081212997437, + "learning_rate": 9.944763865250248e-05, + "loss": 0.0175, + "step": 9500 + }, + { + "epoch": 0.7386407766990292, + "grad_norm": 0.4617466628551483, + "learning_rate": 9.944518499487254e-05, + "loss": 0.0168, + "step": 9510 + }, + { + "epoch": 0.7394174757281553, + "grad_norm": 0.35807493329048157, + "learning_rate": 9.944272592999151e-05, + "loss": 0.0186, + "step": 9520 + }, + { + "epoch": 0.7401941747572816, + "grad_norm": 0.319609135389328, + "learning_rate": 9.94402614581283e-05, + "loss": 0.0149, + "step": 9530 + }, + { + "epoch": 0.7409708737864078, + "grad_norm": 0.4375326335430145, + "learning_rate": 9.943779157955244e-05, + "loss": 0.0177, + "step": 9540 + }, + { + "epoch": 0.7417475728155339, + "grad_norm": 0.29009494185447693, + "learning_rate": 9.943531629453403e-05, + "loss": 0.0197, + "step": 9550 + }, + { + "epoch": 0.7425242718446602, + "grad_norm": 0.34495213627815247, + "learning_rate": 9.943283560334375e-05, + "loss": 0.0146, + "step": 9560 + }, + { + "epoch": 0.7433009708737864, + "grad_norm": 0.3505113124847412, + "learning_rate": 9.943034950625288e-05, + "loss": 0.0141, + "step": 9570 + }, + { + "epoch": 0.7440776699029126, + "grad_norm": 0.35716211795806885, + "learning_rate": 9.942785800353332e-05, + "loss": 0.0137, + "step": 9580 + }, + { + "epoch": 0.7448543689320388, + "grad_norm": 0.35291656851768494, + "learning_rate": 9.942536109545751e-05, + "loss": 0.0148, + "step": 9590 + }, + { + "epoch": 0.7456310679611651, + "grad_norm": 0.37154579162597656, + "learning_rate": 9.942285878229853e-05, + "loss": 0.0136, + "step": 9600 + }, + { + "epoch": 0.7464077669902912, + "grad_norm": 0.24978919327259064, + "learning_rate": 9.942035106433001e-05, + "loss": 0.0156, + "step": 9610 + }, + { + "epoch": 0.7471844660194175, + "grad_norm": 0.26566630601882935, + "learning_rate": 9.94178379418262e-05, + "loss": 0.0139, + "step": 9620 + }, + { + "epoch": 0.7479611650485437, + "grad_norm": 0.25021517276763916, + "learning_rate": 9.941531941506194e-05, + "loss": 0.0145, + "step": 9630 + }, + { + "epoch": 0.74873786407767, + "grad_norm": 0.24953296780586243, + "learning_rate": 9.941279548431263e-05, + "loss": 0.0126, + "step": 9640 + }, + { + "epoch": 0.7495145631067961, + "grad_norm": 0.4345526099205017, + "learning_rate": 9.941026614985431e-05, + "loss": 0.0138, + "step": 9650 + }, + { + "epoch": 0.7502912621359223, + "grad_norm": 0.470493346452713, + "learning_rate": 9.940773141196357e-05, + "loss": 0.0156, + "step": 9660 + }, + { + "epoch": 0.7510679611650486, + "grad_norm": 0.4378665089607239, + "learning_rate": 9.94051912709176e-05, + "loss": 0.0179, + "step": 9670 + }, + { + "epoch": 0.7518446601941747, + "grad_norm": 0.5883545279502869, + "learning_rate": 9.940264572699421e-05, + "loss": 0.0171, + "step": 9680 + }, + { + "epoch": 0.752621359223301, + "grad_norm": 0.4145841896533966, + "learning_rate": 9.940009478047174e-05, + "loss": 0.0154, + "step": 9690 + }, + { + "epoch": 0.7533980582524272, + "grad_norm": 0.3733755648136139, + "learning_rate": 9.939753843162918e-05, + "loss": 0.0162, + "step": 9700 + }, + { + "epoch": 0.7541747572815534, + "grad_norm": 0.292828232049942, + "learning_rate": 9.939497668074609e-05, + "loss": 0.0153, + "step": 9710 + }, + { + "epoch": 0.7549514563106796, + "grad_norm": 0.3043445348739624, + "learning_rate": 9.93924095281026e-05, + "loss": 0.0155, + "step": 9720 + }, + { + "epoch": 0.7557281553398059, + "grad_norm": 0.27115151286125183, + "learning_rate": 9.938983697397948e-05, + "loss": 0.0135, + "step": 9730 + }, + { + "epoch": 0.756504854368932, + "grad_norm": 0.27715420722961426, + "learning_rate": 9.938725901865805e-05, + "loss": 0.0111, + "step": 9740 + }, + { + "epoch": 0.7572815533980582, + "grad_norm": 0.3343384861946106, + "learning_rate": 9.93846756624202e-05, + "loss": 0.0131, + "step": 9750 + }, + { + "epoch": 0.7580582524271845, + "grad_norm": 0.2952840328216553, + "learning_rate": 9.938208690554849e-05, + "loss": 0.0136, + "step": 9760 + }, + { + "epoch": 0.7588349514563106, + "grad_norm": 0.33777013421058655, + "learning_rate": 9.9379492748326e-05, + "loss": 0.0141, + "step": 9770 + }, + { + "epoch": 0.7596116504854369, + "grad_norm": 0.6053186058998108, + "learning_rate": 9.937689319103641e-05, + "loss": 0.0156, + "step": 9780 + }, + { + "epoch": 0.7603883495145631, + "grad_norm": 0.300719678401947, + "learning_rate": 9.937428823396404e-05, + "loss": 0.0142, + "step": 9790 + }, + { + "epoch": 0.7611650485436893, + "grad_norm": 0.372396320104599, + "learning_rate": 9.937167787739372e-05, + "loss": 0.0139, + "step": 9800 + }, + { + "epoch": 0.7619417475728155, + "grad_norm": 0.30810418725013733, + "learning_rate": 9.936906212161095e-05, + "loss": 0.0162, + "step": 9810 + }, + { + "epoch": 0.7627184466019418, + "grad_norm": 0.43991342186927795, + "learning_rate": 9.936644096690176e-05, + "loss": 0.0138, + "step": 9820 + }, + { + "epoch": 0.7634951456310679, + "grad_norm": 0.23604442179203033, + "learning_rate": 9.936381441355282e-05, + "loss": 0.017, + "step": 9830 + }, + { + "epoch": 0.7642718446601942, + "grad_norm": 0.28154048323631287, + "learning_rate": 9.936118246185136e-05, + "loss": 0.015, + "step": 9840 + }, + { + "epoch": 0.7650485436893204, + "grad_norm": 0.32028430700302124, + "learning_rate": 9.935854511208518e-05, + "loss": 0.0204, + "step": 9850 + }, + { + "epoch": 0.7658252427184467, + "grad_norm": 0.31074783205986023, + "learning_rate": 9.935590236454272e-05, + "loss": 0.0176, + "step": 9860 + }, + { + "epoch": 0.7666019417475728, + "grad_norm": 0.3637944459915161, + "learning_rate": 9.935325421951298e-05, + "loss": 0.0144, + "step": 9870 + }, + { + "epoch": 0.767378640776699, + "grad_norm": 0.32912561297416687, + "learning_rate": 9.935060067728557e-05, + "loss": 0.0176, + "step": 9880 + }, + { + "epoch": 0.7681553398058253, + "grad_norm": 0.3322298526763916, + "learning_rate": 9.934794173815067e-05, + "loss": 0.0156, + "step": 9890 + }, + { + "epoch": 0.7689320388349514, + "grad_norm": 0.26900628209114075, + "learning_rate": 9.934527740239906e-05, + "loss": 0.0164, + "step": 9900 + }, + { + "epoch": 0.7697087378640777, + "grad_norm": 0.2097969800233841, + "learning_rate": 9.934260767032209e-05, + "loss": 0.0169, + "step": 9910 + }, + { + "epoch": 0.7704854368932039, + "grad_norm": 0.22065645456314087, + "learning_rate": 9.933993254221172e-05, + "loss": 0.0185, + "step": 9920 + }, + { + "epoch": 0.7712621359223301, + "grad_norm": 0.4505429267883301, + "learning_rate": 9.933725201836053e-05, + "loss": 0.0146, + "step": 9930 + }, + { + "epoch": 0.7720388349514563, + "grad_norm": 0.31839337944984436, + "learning_rate": 9.933456609906162e-05, + "loss": 0.015, + "step": 9940 + }, + { + "epoch": 0.7728155339805826, + "grad_norm": 0.2762009799480438, + "learning_rate": 9.933187478460875e-05, + "loss": 0.0119, + "step": 9950 + }, + { + "epoch": 0.7735922330097087, + "grad_norm": 0.3012867271900177, + "learning_rate": 9.93291780752962e-05, + "loss": 0.0136, + "step": 9960 + }, + { + "epoch": 0.774368932038835, + "grad_norm": 0.49426189064979553, + "learning_rate": 9.932647597141893e-05, + "loss": 0.0184, + "step": 9970 + }, + { + "epoch": 0.7751456310679612, + "grad_norm": 0.2249089628458023, + "learning_rate": 9.932376847327239e-05, + "loss": 0.0144, + "step": 9980 + }, + { + "epoch": 0.7759223300970873, + "grad_norm": 0.42843958735466003, + "learning_rate": 9.932105558115268e-05, + "loss": 0.0188, + "step": 9990 + }, + { + "epoch": 0.7766990291262136, + "grad_norm": 0.260844349861145, + "learning_rate": 9.931833729535651e-05, + "loss": 0.0147, + "step": 10000 + }, + { + "epoch": 0.7774757281553398, + "grad_norm": 0.23403282463550568, + "learning_rate": 9.931561361618111e-05, + "loss": 0.0186, + "step": 10010 + }, + { + "epoch": 0.778252427184466, + "grad_norm": 0.3231751322746277, + "learning_rate": 9.931288454392435e-05, + "loss": 0.0117, + "step": 10020 + }, + { + "epoch": 0.7790291262135922, + "grad_norm": 0.767547607421875, + "learning_rate": 9.931015007888467e-05, + "loss": 0.0172, + "step": 10030 + }, + { + "epoch": 0.7798058252427185, + "grad_norm": 0.32892364263534546, + "learning_rate": 9.930741022136112e-05, + "loss": 0.018, + "step": 10040 + }, + { + "epoch": 0.7805825242718447, + "grad_norm": 0.25563347339630127, + "learning_rate": 9.930466497165333e-05, + "loss": 0.0159, + "step": 10050 + }, + { + "epoch": 0.7813592233009709, + "grad_norm": 0.27139902114868164, + "learning_rate": 9.93019143300615e-05, + "loss": 0.0141, + "step": 10060 + }, + { + "epoch": 0.7821359223300971, + "grad_norm": 0.25345975160598755, + "learning_rate": 9.929915829688644e-05, + "loss": 0.015, + "step": 10070 + }, + { + "epoch": 0.7829126213592233, + "grad_norm": 0.1816626936197281, + "learning_rate": 9.929639687242955e-05, + "loss": 0.0147, + "step": 10080 + }, + { + "epoch": 0.7836893203883495, + "grad_norm": 0.36826291680336, + "learning_rate": 9.929363005699281e-05, + "loss": 0.0156, + "step": 10090 + }, + { + "epoch": 0.7844660194174757, + "grad_norm": 0.366060346364975, + "learning_rate": 9.92908578508788e-05, + "loss": 0.0111, + "step": 10100 + }, + { + "epoch": 0.785242718446602, + "grad_norm": 0.36719992756843567, + "learning_rate": 9.928808025439069e-05, + "loss": 0.0148, + "step": 10110 + }, + { + "epoch": 0.7860194174757281, + "grad_norm": 0.343594491481781, + "learning_rate": 9.928529726783223e-05, + "loss": 0.0157, + "step": 10120 + }, + { + "epoch": 0.7867961165048544, + "grad_norm": 0.3014799654483795, + "learning_rate": 9.928250889150774e-05, + "loss": 0.0136, + "step": 10130 + }, + { + "epoch": 0.7875728155339806, + "grad_norm": 0.4949723482131958, + "learning_rate": 9.92797151257222e-05, + "loss": 0.0153, + "step": 10140 + }, + { + "epoch": 0.7883495145631068, + "grad_norm": 0.20539478957653046, + "learning_rate": 9.927691597078108e-05, + "loss": 0.0155, + "step": 10150 + }, + { + "epoch": 0.789126213592233, + "grad_norm": 0.27197813987731934, + "learning_rate": 9.927411142699053e-05, + "loss": 0.0158, + "step": 10160 + }, + { + "epoch": 0.7899029126213593, + "grad_norm": 0.2953568398952484, + "learning_rate": 9.927130149465725e-05, + "loss": 0.0149, + "step": 10170 + }, + { + "epoch": 0.7906796116504854, + "grad_norm": 0.19169186055660248, + "learning_rate": 9.92684861740885e-05, + "loss": 0.0118, + "step": 10180 + }, + { + "epoch": 0.7914563106796116, + "grad_norm": 0.3277527987957001, + "learning_rate": 9.926566546559217e-05, + "loss": 0.0143, + "step": 10190 + }, + { + "epoch": 0.7922330097087379, + "grad_norm": 0.3082920014858246, + "learning_rate": 9.926283936947673e-05, + "loss": 0.0169, + "step": 10200 + }, + { + "epoch": 0.793009708737864, + "grad_norm": 0.3385584056377411, + "learning_rate": 9.926000788605126e-05, + "loss": 0.0156, + "step": 10210 + }, + { + "epoch": 0.7937864077669903, + "grad_norm": 0.2681563198566437, + "learning_rate": 9.92571710156254e-05, + "loss": 0.012, + "step": 10220 + }, + { + "epoch": 0.7945631067961165, + "grad_norm": 0.22182415425777435, + "learning_rate": 9.925432875850936e-05, + "loss": 0.0165, + "step": 10230 + }, + { + "epoch": 0.7953398058252428, + "grad_norm": 0.3558785319328308, + "learning_rate": 9.925148111501396e-05, + "loss": 0.0112, + "step": 10240 + }, + { + "epoch": 0.7961165048543689, + "grad_norm": 0.478115051984787, + "learning_rate": 9.924862808545066e-05, + "loss": 0.0126, + "step": 10250 + }, + { + "epoch": 0.7968932038834952, + "grad_norm": 0.35173407196998596, + "learning_rate": 9.924576967013141e-05, + "loss": 0.015, + "step": 10260 + }, + { + "epoch": 0.7976699029126214, + "grad_norm": 0.20749825239181519, + "learning_rate": 9.924290586936887e-05, + "loss": 0.0119, + "step": 10270 + }, + { + "epoch": 0.7984466019417475, + "grad_norm": 0.2967933416366577, + "learning_rate": 9.924003668347614e-05, + "loss": 0.0167, + "step": 10280 + }, + { + "epoch": 0.7992233009708738, + "grad_norm": 0.30588361620903015, + "learning_rate": 9.923716211276704e-05, + "loss": 0.0149, + "step": 10290 + }, + { + "epoch": 0.8, + "grad_norm": 0.37905153632164, + "learning_rate": 9.923428215755594e-05, + "loss": 0.0161, + "step": 10300 + }, + { + "epoch": 0.8007766990291262, + "grad_norm": 0.1990775614976883, + "learning_rate": 9.923139681815775e-05, + "loss": 0.0142, + "step": 10310 + }, + { + "epoch": 0.8015533980582524, + "grad_norm": 0.2616576552391052, + "learning_rate": 9.922850609488801e-05, + "loss": 0.0149, + "step": 10320 + }, + { + "epoch": 0.8023300970873787, + "grad_norm": 0.17714186012744904, + "learning_rate": 9.922560998806287e-05, + "loss": 0.0098, + "step": 10330 + }, + { + "epoch": 0.8031067961165048, + "grad_norm": 0.4670056402683258, + "learning_rate": 9.922270849799905e-05, + "loss": 0.0145, + "step": 10340 + }, + { + "epoch": 0.8038834951456311, + "grad_norm": 0.3072328567504883, + "learning_rate": 9.92198016250138e-05, + "loss": 0.0158, + "step": 10350 + }, + { + "epoch": 0.8046601941747573, + "grad_norm": 0.20670084655284882, + "learning_rate": 9.921688936942506e-05, + "loss": 0.0102, + "step": 10360 + }, + { + "epoch": 0.8054368932038835, + "grad_norm": 0.23630300164222717, + "learning_rate": 9.921397173155129e-05, + "loss": 0.0128, + "step": 10370 + }, + { + "epoch": 0.8062135922330097, + "grad_norm": 0.2912234365940094, + "learning_rate": 9.921104871171157e-05, + "loss": 0.0157, + "step": 10380 + }, + { + "epoch": 0.806990291262136, + "grad_norm": 0.3213827311992645, + "learning_rate": 9.920812031022554e-05, + "loss": 0.0148, + "step": 10390 + }, + { + "epoch": 0.8077669902912621, + "grad_norm": 0.3181355595588684, + "learning_rate": 9.920518652741348e-05, + "loss": 0.0157, + "step": 10400 + }, + { + "epoch": 0.8085436893203883, + "grad_norm": 0.6385346055030823, + "learning_rate": 9.920224736359618e-05, + "loss": 0.0124, + "step": 10410 + }, + { + "epoch": 0.8093203883495146, + "grad_norm": 0.3357437252998352, + "learning_rate": 9.91993028190951e-05, + "loss": 0.0107, + "step": 10420 + }, + { + "epoch": 0.8100970873786407, + "grad_norm": 0.3164859414100647, + "learning_rate": 9.919635289423222e-05, + "loss": 0.0153, + "step": 10430 + }, + { + "epoch": 0.810873786407767, + "grad_norm": 0.4933522343635559, + "learning_rate": 9.919339758933015e-05, + "loss": 0.0164, + "step": 10440 + }, + { + "epoch": 0.8116504854368932, + "grad_norm": 0.461402028799057, + "learning_rate": 9.919043690471209e-05, + "loss": 0.0158, + "step": 10450 + }, + { + "epoch": 0.8124271844660195, + "grad_norm": 0.2141295075416565, + "learning_rate": 9.91874708407018e-05, + "loss": 0.0162, + "step": 10460 + }, + { + "epoch": 0.8132038834951456, + "grad_norm": 0.2633765935897827, + "learning_rate": 9.918449939762367e-05, + "loss": 0.0147, + "step": 10470 + }, + { + "epoch": 0.8139805825242719, + "grad_norm": 0.2641275227069855, + "learning_rate": 9.91815225758026e-05, + "loss": 0.0152, + "step": 10480 + }, + { + "epoch": 0.8147572815533981, + "grad_norm": 0.4268720746040344, + "learning_rate": 9.917854037556419e-05, + "loss": 0.0134, + "step": 10490 + }, + { + "epoch": 0.8155339805825242, + "grad_norm": 0.23238234221935272, + "learning_rate": 9.917555279723454e-05, + "loss": 0.0106, + "step": 10500 + }, + { + "epoch": 0.8163106796116505, + "grad_norm": 0.2820134162902832, + "learning_rate": 9.917255984114036e-05, + "loss": 0.0128, + "step": 10510 + }, + { + "epoch": 0.8170873786407767, + "grad_norm": 0.23804427683353424, + "learning_rate": 9.916956150760896e-05, + "loss": 0.0134, + "step": 10520 + }, + { + "epoch": 0.8178640776699029, + "grad_norm": 0.3753458261489868, + "learning_rate": 9.916655779696826e-05, + "loss": 0.0179, + "step": 10530 + }, + { + "epoch": 0.8186407766990291, + "grad_norm": 0.3264981508255005, + "learning_rate": 9.916354870954671e-05, + "loss": 0.0143, + "step": 10540 + }, + { + "epoch": 0.8194174757281554, + "grad_norm": 0.330022394657135, + "learning_rate": 9.91605342456734e-05, + "loss": 0.0116, + "step": 10550 + }, + { + "epoch": 0.8201941747572815, + "grad_norm": 0.3607065677642822, + "learning_rate": 9.915751440567795e-05, + "loss": 0.0144, + "step": 10560 + }, + { + "epoch": 0.8209708737864078, + "grad_norm": 0.5386129021644592, + "learning_rate": 9.915448918989066e-05, + "loss": 0.0136, + "step": 10570 + }, + { + "epoch": 0.821747572815534, + "grad_norm": 0.24062387645244598, + "learning_rate": 9.915145859864232e-05, + "loss": 0.0111, + "step": 10580 + }, + { + "epoch": 0.8225242718446601, + "grad_norm": 0.18035408854484558, + "learning_rate": 9.914842263226437e-05, + "loss": 0.0144, + "step": 10590 + }, + { + "epoch": 0.8233009708737864, + "grad_norm": 0.2360474169254303, + "learning_rate": 9.914538129108882e-05, + "loss": 0.0143, + "step": 10600 + }, + { + "epoch": 0.8240776699029126, + "grad_norm": 0.2412513643503189, + "learning_rate": 9.914233457544825e-05, + "loss": 0.015, + "step": 10610 + }, + { + "epoch": 0.8248543689320388, + "grad_norm": 0.19771525263786316, + "learning_rate": 9.913928248567586e-05, + "loss": 0.0117, + "step": 10620 + }, + { + "epoch": 0.825631067961165, + "grad_norm": 0.37196147441864014, + "learning_rate": 9.913622502210542e-05, + "loss": 0.0118, + "step": 10630 + }, + { + "epoch": 0.8264077669902913, + "grad_norm": 0.34724223613739014, + "learning_rate": 9.913316218507128e-05, + "loss": 0.0142, + "step": 10640 + }, + { + "epoch": 0.8271844660194175, + "grad_norm": 0.35854339599609375, + "learning_rate": 9.91300939749084e-05, + "loss": 0.0107, + "step": 10650 + }, + { + "epoch": 0.8279611650485437, + "grad_norm": 0.3117939531803131, + "learning_rate": 9.91270203919523e-05, + "loss": 0.0158, + "step": 10660 + }, + { + "epoch": 0.8287378640776699, + "grad_norm": 0.34022200107574463, + "learning_rate": 9.912394143653912e-05, + "loss": 0.0118, + "step": 10670 + }, + { + "epoch": 0.8295145631067962, + "grad_norm": 0.4235425293445587, + "learning_rate": 9.912085710900555e-05, + "loss": 0.0139, + "step": 10680 + }, + { + "epoch": 0.8302912621359223, + "grad_norm": 0.33308014273643494, + "learning_rate": 9.911776740968892e-05, + "loss": 0.0101, + "step": 10690 + }, + { + "epoch": 0.8310679611650486, + "grad_norm": 0.26800602674484253, + "learning_rate": 9.911467233892709e-05, + "loss": 0.0102, + "step": 10700 + }, + { + "epoch": 0.8318446601941748, + "grad_norm": 0.31846463680267334, + "learning_rate": 9.911157189705853e-05, + "loss": 0.0141, + "step": 10710 + }, + { + "epoch": 0.8326213592233009, + "grad_norm": 0.28702178597450256, + "learning_rate": 9.910846608442229e-05, + "loss": 0.0121, + "step": 10720 + }, + { + "epoch": 0.8333980582524272, + "grad_norm": 0.3075157403945923, + "learning_rate": 9.910535490135805e-05, + "loss": 0.0137, + "step": 10730 + }, + { + "epoch": 0.8341747572815534, + "grad_norm": 0.19426432251930237, + "learning_rate": 9.910223834820603e-05, + "loss": 0.0168, + "step": 10740 + }, + { + "epoch": 0.8349514563106796, + "grad_norm": 0.21569643914699554, + "learning_rate": 9.909911642530703e-05, + "loss": 0.0117, + "step": 10750 + }, + { + "epoch": 0.8357281553398058, + "grad_norm": 0.5081427097320557, + "learning_rate": 9.909598913300249e-05, + "loss": 0.0127, + "step": 10760 + }, + { + "epoch": 0.8365048543689321, + "grad_norm": 0.25627976655960083, + "learning_rate": 9.909285647163438e-05, + "loss": 0.0119, + "step": 10770 + }, + { + "epoch": 0.8372815533980582, + "grad_norm": 0.3052299916744232, + "learning_rate": 9.908971844154531e-05, + "loss": 0.0142, + "step": 10780 + }, + { + "epoch": 0.8380582524271845, + "grad_norm": 0.24613609910011292, + "learning_rate": 9.908657504307843e-05, + "loss": 0.0148, + "step": 10790 + }, + { + "epoch": 0.8388349514563107, + "grad_norm": 0.3783813714981079, + "learning_rate": 9.908342627657751e-05, + "loss": 0.0152, + "step": 10800 + }, + { + "epoch": 0.8396116504854368, + "grad_norm": 0.301085889339447, + "learning_rate": 9.908027214238689e-05, + "loss": 0.0145, + "step": 10810 + }, + { + "epoch": 0.8403883495145631, + "grad_norm": 0.5620572566986084, + "learning_rate": 9.90771126408515e-05, + "loss": 0.0152, + "step": 10820 + }, + { + "epoch": 0.8411650485436893, + "grad_norm": 0.4049241840839386, + "learning_rate": 9.907394777231685e-05, + "loss": 0.0164, + "step": 10830 + }, + { + "epoch": 0.8419417475728155, + "grad_norm": 0.19663415849208832, + "learning_rate": 9.907077753712905e-05, + "loss": 0.0107, + "step": 10840 + }, + { + "epoch": 0.8427184466019417, + "grad_norm": 0.20675309002399445, + "learning_rate": 9.906760193563482e-05, + "loss": 0.0132, + "step": 10850 + }, + { + "epoch": 0.843495145631068, + "grad_norm": 0.45466965436935425, + "learning_rate": 9.906442096818139e-05, + "loss": 0.0142, + "step": 10860 + }, + { + "epoch": 0.8442718446601942, + "grad_norm": 0.37439751625061035, + "learning_rate": 9.906123463511665e-05, + "loss": 0.0119, + "step": 10870 + }, + { + "epoch": 0.8450485436893204, + "grad_norm": 0.30580148100852966, + "learning_rate": 9.905804293678907e-05, + "loss": 0.0134, + "step": 10880 + }, + { + "epoch": 0.8458252427184466, + "grad_norm": 0.5365626215934753, + "learning_rate": 9.905484587354766e-05, + "loss": 0.013, + "step": 10890 + }, + { + "epoch": 0.8466019417475729, + "grad_norm": 0.2281510829925537, + "learning_rate": 9.905164344574205e-05, + "loss": 0.0122, + "step": 10900 + }, + { + "epoch": 0.847378640776699, + "grad_norm": 0.5750395655632019, + "learning_rate": 9.904843565372248e-05, + "loss": 0.014, + "step": 10910 + }, + { + "epoch": 0.8481553398058252, + "grad_norm": 0.29391926527023315, + "learning_rate": 9.904522249783972e-05, + "loss": 0.0143, + "step": 10920 + }, + { + "epoch": 0.8489320388349515, + "grad_norm": 0.22353020310401917, + "learning_rate": 9.904200397844517e-05, + "loss": 0.0106, + "step": 10930 + }, + { + "epoch": 0.8497087378640776, + "grad_norm": 0.2586445212364197, + "learning_rate": 9.903878009589078e-05, + "loss": 0.0099, + "step": 10940 + }, + { + "epoch": 0.8504854368932039, + "grad_norm": 0.20462393760681152, + "learning_rate": 9.903555085052915e-05, + "loss": 0.018, + "step": 10950 + }, + { + "epoch": 0.8512621359223301, + "grad_norm": 0.308570921421051, + "learning_rate": 9.903231624271338e-05, + "loss": 0.0122, + "step": 10960 + }, + { + "epoch": 0.8520388349514563, + "grad_norm": 0.3183949291706085, + "learning_rate": 9.902907627279724e-05, + "loss": 0.0145, + "step": 10970 + }, + { + "epoch": 0.8528155339805825, + "grad_norm": 0.3802608251571655, + "learning_rate": 9.902583094113504e-05, + "loss": 0.011, + "step": 10980 + }, + { + "epoch": 0.8535922330097088, + "grad_norm": 0.19623805582523346, + "learning_rate": 9.902258024808168e-05, + "loss": 0.0142, + "step": 10990 + }, + { + "epoch": 0.8543689320388349, + "grad_norm": 0.3581583797931671, + "learning_rate": 9.901932419399264e-05, + "loss": 0.0114, + "step": 11000 + }, + { + "epoch": 0.8551456310679612, + "grad_norm": 0.22340510785579681, + "learning_rate": 9.9016062779224e-05, + "loss": 0.0109, + "step": 11010 + }, + { + "epoch": 0.8559223300970874, + "grad_norm": 0.22217176854610443, + "learning_rate": 9.901279600413242e-05, + "loss": 0.0122, + "step": 11020 + }, + { + "epoch": 0.8566990291262135, + "grad_norm": 0.3226583003997803, + "learning_rate": 9.900952386907518e-05, + "loss": 0.017, + "step": 11030 + }, + { + "epoch": 0.8574757281553398, + "grad_norm": 0.21669039130210876, + "learning_rate": 9.90062463744101e-05, + "loss": 0.0144, + "step": 11040 + }, + { + "epoch": 0.858252427184466, + "grad_norm": 0.256137877702713, + "learning_rate": 9.900296352049558e-05, + "loss": 0.0136, + "step": 11050 + }, + { + "epoch": 0.8590291262135923, + "grad_norm": 0.24693305790424347, + "learning_rate": 9.899967530769065e-05, + "loss": 0.0108, + "step": 11060 + }, + { + "epoch": 0.8598058252427184, + "grad_norm": 0.25959935784339905, + "learning_rate": 9.899638173635489e-05, + "loss": 0.0143, + "step": 11070 + }, + { + "epoch": 0.8605825242718447, + "grad_norm": 0.5370497107505798, + "learning_rate": 9.899308280684849e-05, + "loss": 0.011, + "step": 11080 + }, + { + "epoch": 0.8613592233009709, + "grad_norm": 0.20366613566875458, + "learning_rate": 9.898977851953222e-05, + "loss": 0.0131, + "step": 11090 + }, + { + "epoch": 0.8621359223300971, + "grad_norm": 0.36568495631217957, + "learning_rate": 9.898646887476741e-05, + "loss": 0.0119, + "step": 11100 + }, + { + "epoch": 0.8629126213592233, + "grad_norm": 0.3685816526412964, + "learning_rate": 9.898315387291603e-05, + "loss": 0.0122, + "step": 11110 + }, + { + "epoch": 0.8636893203883496, + "grad_norm": 0.3492642343044281, + "learning_rate": 9.89798335143406e-05, + "loss": 0.0127, + "step": 11120 + }, + { + "epoch": 0.8644660194174757, + "grad_norm": 0.23852616548538208, + "learning_rate": 9.897650779940419e-05, + "loss": 0.0122, + "step": 11130 + }, + { + "epoch": 0.8652427184466019, + "grad_norm": 0.5466972589492798, + "learning_rate": 9.897317672847054e-05, + "loss": 0.014, + "step": 11140 + }, + { + "epoch": 0.8660194174757282, + "grad_norm": 0.385654479265213, + "learning_rate": 9.89698403019039e-05, + "loss": 0.0147, + "step": 11150 + }, + { + "epoch": 0.8667961165048543, + "grad_norm": 0.40754348039627075, + "learning_rate": 9.896649852006917e-05, + "loss": 0.0176, + "step": 11160 + }, + { + "epoch": 0.8675728155339806, + "grad_norm": 0.5352248549461365, + "learning_rate": 9.896315138333177e-05, + "loss": 0.0181, + "step": 11170 + }, + { + "epoch": 0.8683495145631068, + "grad_norm": 0.2482679784297943, + "learning_rate": 9.895979889205774e-05, + "loss": 0.0125, + "step": 11180 + }, + { + "epoch": 0.869126213592233, + "grad_norm": 0.3348122239112854, + "learning_rate": 9.895644104661372e-05, + "loss": 0.0213, + "step": 11190 + }, + { + "epoch": 0.8699029126213592, + "grad_norm": 0.3059256970882416, + "learning_rate": 9.895307784736691e-05, + "loss": 0.0203, + "step": 11200 + }, + { + "epoch": 0.8706796116504855, + "grad_norm": 0.27117738127708435, + "learning_rate": 9.894970929468512e-05, + "loss": 0.0151, + "step": 11210 + }, + { + "epoch": 0.8714563106796116, + "grad_norm": 0.2851846516132355, + "learning_rate": 9.89463353889367e-05, + "loss": 0.0154, + "step": 11220 + }, + { + "epoch": 0.8722330097087378, + "grad_norm": 0.40486276149749756, + "learning_rate": 9.894295613049065e-05, + "loss": 0.015, + "step": 11230 + }, + { + "epoch": 0.8730097087378641, + "grad_norm": 0.3732682466506958, + "learning_rate": 9.893957151971649e-05, + "loss": 0.0128, + "step": 11240 + }, + { + "epoch": 0.8737864077669902, + "grad_norm": 0.38575050234794617, + "learning_rate": 9.893618155698436e-05, + "loss": 0.0176, + "step": 11250 + }, + { + "epoch": 0.8745631067961165, + "grad_norm": 0.3485153913497925, + "learning_rate": 9.8932786242665e-05, + "loss": 0.0137, + "step": 11260 + }, + { + "epoch": 0.8753398058252427, + "grad_norm": 0.2525748014450073, + "learning_rate": 9.89293855771297e-05, + "loss": 0.0106, + "step": 11270 + }, + { + "epoch": 0.876116504854369, + "grad_norm": 0.2201000154018402, + "learning_rate": 9.892597956075036e-05, + "loss": 0.0128, + "step": 11280 + }, + { + "epoch": 0.8768932038834951, + "grad_norm": 0.1892174929380417, + "learning_rate": 9.892256819389947e-05, + "loss": 0.0126, + "step": 11290 + }, + { + "epoch": 0.8776699029126214, + "grad_norm": 0.258849173784256, + "learning_rate": 9.891915147695006e-05, + "loss": 0.0139, + "step": 11300 + }, + { + "epoch": 0.8784466019417476, + "grad_norm": 0.24122735857963562, + "learning_rate": 9.891572941027577e-05, + "loss": 0.0153, + "step": 11310 + }, + { + "epoch": 0.8792233009708738, + "grad_norm": 0.341532826423645, + "learning_rate": 9.89123019942509e-05, + "loss": 0.0121, + "step": 11320 + }, + { + "epoch": 0.88, + "grad_norm": 0.5244576930999756, + "learning_rate": 9.89088692292502e-05, + "loss": 0.0168, + "step": 11330 + }, + { + "epoch": 0.8807766990291263, + "grad_norm": 0.3139859735965729, + "learning_rate": 9.89054311156491e-05, + "loss": 0.0129, + "step": 11340 + }, + { + "epoch": 0.8815533980582524, + "grad_norm": 0.2849518656730652, + "learning_rate": 9.890198765382357e-05, + "loss": 0.0164, + "step": 11350 + }, + { + "epoch": 0.8823300970873786, + "grad_norm": 0.3598756194114685, + "learning_rate": 9.889853884415021e-05, + "loss": 0.0179, + "step": 11360 + }, + { + "epoch": 0.8831067961165049, + "grad_norm": 0.45807570219039917, + "learning_rate": 9.889508468700614e-05, + "loss": 0.0174, + "step": 11370 + }, + { + "epoch": 0.883883495145631, + "grad_norm": 0.2753118872642517, + "learning_rate": 9.889162518276915e-05, + "loss": 0.0144, + "step": 11380 + }, + { + "epoch": 0.8846601941747573, + "grad_norm": 0.4248497486114502, + "learning_rate": 9.888816033181752e-05, + "loss": 0.0135, + "step": 11390 + }, + { + "epoch": 0.8854368932038835, + "grad_norm": 0.5464569926261902, + "learning_rate": 9.888469013453018e-05, + "loss": 0.0165, + "step": 11400 + }, + { + "epoch": 0.8862135922330097, + "grad_norm": 0.34448933601379395, + "learning_rate": 9.888121459128663e-05, + "loss": 0.0155, + "step": 11410 + }, + { + "epoch": 0.8869902912621359, + "grad_norm": 0.23077011108398438, + "learning_rate": 9.887773370246693e-05, + "loss": 0.0158, + "step": 11420 + }, + { + "epoch": 0.8877669902912622, + "grad_norm": 0.19758445024490356, + "learning_rate": 9.887424746845177e-05, + "loss": 0.0128, + "step": 11430 + }, + { + "epoch": 0.8885436893203883, + "grad_norm": 0.411414235830307, + "learning_rate": 9.887075588962239e-05, + "loss": 0.0116, + "step": 11440 + }, + { + "epoch": 0.8893203883495145, + "grad_norm": 0.320865273475647, + "learning_rate": 9.88672589663606e-05, + "loss": 0.0127, + "step": 11450 + }, + { + "epoch": 0.8900970873786408, + "grad_norm": 0.28062400221824646, + "learning_rate": 9.886375669904886e-05, + "loss": 0.0201, + "step": 11460 + }, + { + "epoch": 0.890873786407767, + "grad_norm": 0.18737582862377167, + "learning_rate": 9.886024908807014e-05, + "loss": 0.0157, + "step": 11470 + }, + { + "epoch": 0.8916504854368932, + "grad_norm": 0.2971031963825226, + "learning_rate": 9.885673613380806e-05, + "loss": 0.0148, + "step": 11480 + }, + { + "epoch": 0.8924271844660194, + "grad_norm": 0.2786371409893036, + "learning_rate": 9.885321783664676e-05, + "loss": 0.0166, + "step": 11490 + }, + { + "epoch": 0.8932038834951457, + "grad_norm": 0.3124575614929199, + "learning_rate": 9.884969419697101e-05, + "loss": 0.0146, + "step": 11500 + }, + { + "epoch": 0.8939805825242718, + "grad_norm": 0.2327699512243271, + "learning_rate": 9.884616521516614e-05, + "loss": 0.0103, + "step": 11510 + }, + { + "epoch": 0.8947572815533981, + "grad_norm": 0.26086440682411194, + "learning_rate": 9.88426308916181e-05, + "loss": 0.0168, + "step": 11520 + }, + { + "epoch": 0.8955339805825243, + "grad_norm": 0.27378225326538086, + "learning_rate": 9.883909122671335e-05, + "loss": 0.015, + "step": 11530 + }, + { + "epoch": 0.8963106796116505, + "grad_norm": 0.22880546748638153, + "learning_rate": 9.883554622083904e-05, + "loss": 0.0147, + "step": 11540 + }, + { + "epoch": 0.8970873786407767, + "grad_norm": 0.22396279871463776, + "learning_rate": 9.88319958743828e-05, + "loss": 0.0116, + "step": 11550 + }, + { + "epoch": 0.897864077669903, + "grad_norm": 0.4106120765209198, + "learning_rate": 9.882844018773291e-05, + "loss": 0.0116, + "step": 11560 + }, + { + "epoch": 0.8986407766990291, + "grad_norm": 0.1534367948770523, + "learning_rate": 9.882487916127823e-05, + "loss": 0.0127, + "step": 11570 + }, + { + "epoch": 0.8994174757281553, + "grad_norm": 0.35480937361717224, + "learning_rate": 9.882131279540815e-05, + "loss": 0.0135, + "step": 11580 + }, + { + "epoch": 0.9001941747572816, + "grad_norm": 0.32063889503479004, + "learning_rate": 9.881774109051271e-05, + "loss": 0.0149, + "step": 11590 + }, + { + "epoch": 0.9009708737864077, + "grad_norm": 0.30327165126800537, + "learning_rate": 9.881416404698252e-05, + "loss": 0.0117, + "step": 11600 + }, + { + "epoch": 0.901747572815534, + "grad_norm": 0.28779134154319763, + "learning_rate": 9.881058166520873e-05, + "loss": 0.0136, + "step": 11610 + }, + { + "epoch": 0.9025242718446602, + "grad_norm": 0.3043694496154785, + "learning_rate": 9.880699394558311e-05, + "loss": 0.0144, + "step": 11620 + }, + { + "epoch": 0.9033009708737864, + "grad_norm": 0.17267554998397827, + "learning_rate": 9.880340088849801e-05, + "loss": 0.0114, + "step": 11630 + }, + { + "epoch": 0.9040776699029126, + "grad_norm": 0.3436731696128845, + "learning_rate": 9.879980249434637e-05, + "loss": 0.0153, + "step": 11640 + }, + { + "epoch": 0.9048543689320389, + "grad_norm": 0.3565048277378082, + "learning_rate": 9.879619876352168e-05, + "loss": 0.0128, + "step": 11650 + }, + { + "epoch": 0.9056310679611651, + "grad_norm": 0.29795193672180176, + "learning_rate": 9.879258969641809e-05, + "loss": 0.0115, + "step": 11660 + }, + { + "epoch": 0.9064077669902912, + "grad_norm": 0.3642078638076782, + "learning_rate": 9.878897529343023e-05, + "loss": 0.0154, + "step": 11670 + }, + { + "epoch": 0.9071844660194175, + "grad_norm": 0.20030789077281952, + "learning_rate": 9.878535555495338e-05, + "loss": 0.0114, + "step": 11680 + }, + { + "epoch": 0.9079611650485437, + "grad_norm": 0.26841971278190613, + "learning_rate": 9.87817304813834e-05, + "loss": 0.0132, + "step": 11690 + }, + { + "epoch": 0.9087378640776699, + "grad_norm": 0.3456032872200012, + "learning_rate": 9.877810007311671e-05, + "loss": 0.0136, + "step": 11700 + }, + { + "epoch": 0.9095145631067961, + "grad_norm": 0.36344531178474426, + "learning_rate": 9.877446433055035e-05, + "loss": 0.0155, + "step": 11710 + }, + { + "epoch": 0.9102912621359224, + "grad_norm": 0.21942411363124847, + "learning_rate": 9.877082325408191e-05, + "loss": 0.0154, + "step": 11720 + }, + { + "epoch": 0.9110679611650485, + "grad_norm": 0.3640318214893341, + "learning_rate": 9.876717684410954e-05, + "loss": 0.0157, + "step": 11730 + }, + { + "epoch": 0.9118446601941748, + "grad_norm": 0.19311584532260895, + "learning_rate": 9.876352510103204e-05, + "loss": 0.0129, + "step": 11740 + }, + { + "epoch": 0.912621359223301, + "grad_norm": 0.2516866624355316, + "learning_rate": 9.875986802524875e-05, + "loss": 0.0152, + "step": 11750 + }, + { + "epoch": 0.9133980582524271, + "grad_norm": 0.31161561608314514, + "learning_rate": 9.87562056171596e-05, + "loss": 0.0129, + "step": 11760 + }, + { + "epoch": 0.9141747572815534, + "grad_norm": 0.2560758590698242, + "learning_rate": 9.875253787716511e-05, + "loss": 0.0126, + "step": 11770 + }, + { + "epoch": 0.9149514563106796, + "grad_norm": 0.3189917504787445, + "learning_rate": 9.874886480566637e-05, + "loss": 0.011, + "step": 11780 + }, + { + "epoch": 0.9157281553398058, + "grad_norm": 0.39199692010879517, + "learning_rate": 9.874518640306507e-05, + "loss": 0.015, + "step": 11790 + }, + { + "epoch": 0.916504854368932, + "grad_norm": 0.36287933588027954, + "learning_rate": 9.874150266976347e-05, + "loss": 0.0118, + "step": 11800 + }, + { + "epoch": 0.9172815533980583, + "grad_norm": 0.3316769003868103, + "learning_rate": 9.873781360616443e-05, + "loss": 0.013, + "step": 11810 + }, + { + "epoch": 0.9180582524271844, + "grad_norm": 0.49396803975105286, + "learning_rate": 9.873411921267137e-05, + "loss": 0.013, + "step": 11820 + }, + { + "epoch": 0.9188349514563107, + "grad_norm": 0.5461536049842834, + "learning_rate": 9.873041948968829e-05, + "loss": 0.0112, + "step": 11830 + }, + { + "epoch": 0.9196116504854369, + "grad_norm": 0.2987590432167053, + "learning_rate": 9.872671443761981e-05, + "loss": 0.013, + "step": 11840 + }, + { + "epoch": 0.920388349514563, + "grad_norm": 0.21426555514335632, + "learning_rate": 9.872300405687109e-05, + "loss": 0.0124, + "step": 11850 + }, + { + "epoch": 0.9211650485436893, + "grad_norm": 0.3048076033592224, + "learning_rate": 9.871928834784792e-05, + "loss": 0.0161, + "step": 11860 + }, + { + "epoch": 0.9219417475728156, + "grad_norm": 0.3899257481098175, + "learning_rate": 9.871556731095661e-05, + "loss": 0.0169, + "step": 11870 + }, + { + "epoch": 0.9227184466019418, + "grad_norm": 0.41799449920654297, + "learning_rate": 9.871184094660411e-05, + "loss": 0.0147, + "step": 11880 + }, + { + "epoch": 0.9234951456310679, + "grad_norm": 0.28261369466781616, + "learning_rate": 9.870810925519791e-05, + "loss": 0.0139, + "step": 11890 + }, + { + "epoch": 0.9242718446601942, + "grad_norm": 0.2974977493286133, + "learning_rate": 9.870437223714612e-05, + "loss": 0.0166, + "step": 11900 + }, + { + "epoch": 0.9250485436893204, + "grad_norm": 0.17371727526187897, + "learning_rate": 9.87006298928574e-05, + "loss": 0.0099, + "step": 11910 + }, + { + "epoch": 0.9258252427184466, + "grad_norm": 0.3076588213443756, + "learning_rate": 9.869688222274103e-05, + "loss": 0.0112, + "step": 11920 + }, + { + "epoch": 0.9266019417475728, + "grad_norm": 0.17382100224494934, + "learning_rate": 9.869312922720681e-05, + "loss": 0.0102, + "step": 11930 + }, + { + "epoch": 0.9273786407766991, + "grad_norm": 0.35511621832847595, + "learning_rate": 9.868937090666521e-05, + "loss": 0.0146, + "step": 11940 + }, + { + "epoch": 0.9281553398058252, + "grad_norm": 0.2593424618244171, + "learning_rate": 9.86856072615272e-05, + "loss": 0.0133, + "step": 11950 + }, + { + "epoch": 0.9289320388349515, + "grad_norm": 0.27473247051239014, + "learning_rate": 9.868183829220438e-05, + "loss": 0.0157, + "step": 11960 + }, + { + "epoch": 0.9297087378640777, + "grad_norm": 0.5791239738464355, + "learning_rate": 9.867806399910893e-05, + "loss": 0.0177, + "step": 11970 + }, + { + "epoch": 0.9304854368932038, + "grad_norm": 0.37051841616630554, + "learning_rate": 9.867428438265356e-05, + "loss": 0.013, + "step": 11980 + }, + { + "epoch": 0.9312621359223301, + "grad_norm": 0.7062299847602844, + "learning_rate": 9.867049944325165e-05, + "loss": 0.0141, + "step": 11990 + }, + { + "epoch": 0.9320388349514563, + "grad_norm": 0.3789532482624054, + "learning_rate": 9.86667091813171e-05, + "loss": 0.0134, + "step": 12000 + }, + { + "epoch": 0.9328155339805825, + "grad_norm": 0.37176746129989624, + "learning_rate": 9.866291359726438e-05, + "loss": 0.0118, + "step": 12010 + }, + { + "epoch": 0.9335922330097087, + "grad_norm": 0.2631697952747345, + "learning_rate": 9.865911269150861e-05, + "loss": 0.0168, + "step": 12020 + }, + { + "epoch": 0.934368932038835, + "grad_norm": 0.37093016505241394, + "learning_rate": 9.865530646446544e-05, + "loss": 0.0147, + "step": 12030 + }, + { + "epoch": 0.9351456310679611, + "grad_norm": 0.3111679255962372, + "learning_rate": 9.86514949165511e-05, + "loss": 0.0154, + "step": 12040 + }, + { + "epoch": 0.9359223300970874, + "grad_norm": 0.36392295360565186, + "learning_rate": 9.864767804818243e-05, + "loss": 0.0197, + "step": 12050 + }, + { + "epoch": 0.9366990291262136, + "grad_norm": 0.42294323444366455, + "learning_rate": 9.86438558597768e-05, + "loss": 0.0183, + "step": 12060 + }, + { + "epoch": 0.9374757281553399, + "grad_norm": 0.16893315315246582, + "learning_rate": 9.864002835175225e-05, + "loss": 0.0126, + "step": 12070 + }, + { + "epoch": 0.938252427184466, + "grad_norm": 0.28236258029937744, + "learning_rate": 9.863619552452734e-05, + "loss": 0.0116, + "step": 12080 + }, + { + "epoch": 0.9390291262135922, + "grad_norm": 0.27081435918807983, + "learning_rate": 9.863235737852119e-05, + "loss": 0.014, + "step": 12090 + }, + { + "epoch": 0.9398058252427185, + "grad_norm": 0.23070690035820007, + "learning_rate": 9.862851391415356e-05, + "loss": 0.0157, + "step": 12100 + }, + { + "epoch": 0.9405825242718446, + "grad_norm": 0.18139676749706268, + "learning_rate": 9.862466513184477e-05, + "loss": 0.0121, + "step": 12110 + }, + { + "epoch": 0.9413592233009709, + "grad_norm": 0.21257083117961884, + "learning_rate": 9.86208110320157e-05, + "loss": 0.0125, + "step": 12120 + }, + { + "epoch": 0.9421359223300971, + "grad_norm": 0.31379950046539307, + "learning_rate": 9.861695161508784e-05, + "loss": 0.0214, + "step": 12130 + }, + { + "epoch": 0.9429126213592233, + "grad_norm": 0.3777964413166046, + "learning_rate": 9.861308688148324e-05, + "loss": 0.0141, + "step": 12140 + }, + { + "epoch": 0.9436893203883495, + "grad_norm": 0.2109776735305786, + "learning_rate": 9.860921683162455e-05, + "loss": 0.0182, + "step": 12150 + }, + { + "epoch": 0.9444660194174758, + "grad_norm": 0.2848191261291504, + "learning_rate": 9.860534146593499e-05, + "loss": 0.0187, + "step": 12160 + }, + { + "epoch": 0.9452427184466019, + "grad_norm": 0.2776665985584259, + "learning_rate": 9.860146078483836e-05, + "loss": 0.0123, + "step": 12170 + }, + { + "epoch": 0.9460194174757282, + "grad_norm": 0.307785302400589, + "learning_rate": 9.859757478875905e-05, + "loss": 0.016, + "step": 12180 + }, + { + "epoch": 0.9467961165048544, + "grad_norm": 0.37513259053230286, + "learning_rate": 9.859368347812204e-05, + "loss": 0.0179, + "step": 12190 + }, + { + "epoch": 0.9475728155339805, + "grad_norm": 0.2747090458869934, + "learning_rate": 9.858978685335285e-05, + "loss": 0.0106, + "step": 12200 + }, + { + "epoch": 0.9483495145631068, + "grad_norm": 0.3390193283557892, + "learning_rate": 9.858588491487763e-05, + "loss": 0.0181, + "step": 12210 + }, + { + "epoch": 0.949126213592233, + "grad_norm": 0.3340936005115509, + "learning_rate": 9.858197766312308e-05, + "loss": 0.0154, + "step": 12220 + }, + { + "epoch": 0.9499029126213592, + "grad_norm": 0.31813669204711914, + "learning_rate": 9.857806509851649e-05, + "loss": 0.021, + "step": 12230 + }, + { + "epoch": 0.9506796116504854, + "grad_norm": 0.3126957416534424, + "learning_rate": 9.857414722148574e-05, + "loss": 0.0126, + "step": 12240 + }, + { + "epoch": 0.9514563106796117, + "grad_norm": 0.3843240737915039, + "learning_rate": 9.857022403245928e-05, + "loss": 0.0137, + "step": 12250 + }, + { + "epoch": 0.9522330097087378, + "grad_norm": 0.28428688645362854, + "learning_rate": 9.856629553186615e-05, + "loss": 0.0154, + "step": 12260 + }, + { + "epoch": 0.9530097087378641, + "grad_norm": 0.37226444482803345, + "learning_rate": 9.856236172013595e-05, + "loss": 0.0141, + "step": 12270 + }, + { + "epoch": 0.9537864077669903, + "grad_norm": 0.4140941798686981, + "learning_rate": 9.85584225976989e-05, + "loss": 0.0137, + "step": 12280 + }, + { + "epoch": 0.9545631067961166, + "grad_norm": 0.4513002634048462, + "learning_rate": 9.855447816498575e-05, + "loss": 0.0176, + "step": 12290 + }, + { + "epoch": 0.9553398058252427, + "grad_norm": 0.4452679455280304, + "learning_rate": 9.855052842242787e-05, + "loss": 0.0131, + "step": 12300 + }, + { + "epoch": 0.9561165048543689, + "grad_norm": 0.2968914210796356, + "learning_rate": 9.85465733704572e-05, + "loss": 0.0108, + "step": 12310 + }, + { + "epoch": 0.9568932038834952, + "grad_norm": 0.33187055587768555, + "learning_rate": 9.854261300950624e-05, + "loss": 0.015, + "step": 12320 + }, + { + "epoch": 0.9576699029126213, + "grad_norm": 0.27465060353279114, + "learning_rate": 9.853864734000813e-05, + "loss": 0.0098, + "step": 12330 + }, + { + "epoch": 0.9584466019417476, + "grad_norm": 0.1896515041589737, + "learning_rate": 9.85346763623965e-05, + "loss": 0.0092, + "step": 12340 + }, + { + "epoch": 0.9592233009708738, + "grad_norm": 0.2727273404598236, + "learning_rate": 9.853070007710564e-05, + "loss": 0.0107, + "step": 12350 + }, + { + "epoch": 0.96, + "grad_norm": 0.17373360693454742, + "learning_rate": 9.85267184845704e-05, + "loss": 0.0123, + "step": 12360 + }, + { + "epoch": 0.9607766990291262, + "grad_norm": 0.2739451229572296, + "learning_rate": 9.852273158522616e-05, + "loss": 0.0124, + "step": 12370 + }, + { + "epoch": 0.9615533980582525, + "grad_norm": 0.3181130290031433, + "learning_rate": 9.851873937950896e-05, + "loss": 0.0176, + "step": 12380 + }, + { + "epoch": 0.9623300970873786, + "grad_norm": 0.36771872639656067, + "learning_rate": 9.851474186785537e-05, + "loss": 0.015, + "step": 12390 + }, + { + "epoch": 0.9631067961165048, + "grad_norm": 0.2930164337158203, + "learning_rate": 9.851073905070254e-05, + "loss": 0.0157, + "step": 12400 + }, + { + "epoch": 0.9638834951456311, + "grad_norm": 0.36099135875701904, + "learning_rate": 9.850673092848824e-05, + "loss": 0.0111, + "step": 12410 + }, + { + "epoch": 0.9646601941747572, + "grad_norm": 0.6623901128768921, + "learning_rate": 9.850271750165077e-05, + "loss": 0.015, + "step": 12420 + }, + { + "epoch": 0.9654368932038835, + "grad_norm": 0.3428417146205902, + "learning_rate": 9.849869877062902e-05, + "loss": 0.0149, + "step": 12430 + }, + { + "epoch": 0.9662135922330097, + "grad_norm": 0.34169304370880127, + "learning_rate": 9.849467473586252e-05, + "loss": 0.0161, + "step": 12440 + }, + { + "epoch": 0.9669902912621359, + "grad_norm": 0.26332560181617737, + "learning_rate": 9.849064539779127e-05, + "loss": 0.0124, + "step": 12450 + }, + { + "epoch": 0.9677669902912621, + "grad_norm": 0.29839226603507996, + "learning_rate": 9.848661075685594e-05, + "loss": 0.0129, + "step": 12460 + }, + { + "epoch": 0.9685436893203884, + "grad_norm": 0.3067491054534912, + "learning_rate": 9.848257081349778e-05, + "loss": 0.0139, + "step": 12470 + }, + { + "epoch": 0.9693203883495146, + "grad_norm": 0.21039627492427826, + "learning_rate": 9.847852556815856e-05, + "loss": 0.0138, + "step": 12480 + }, + { + "epoch": 0.9700970873786408, + "grad_norm": 0.2333678901195526, + "learning_rate": 9.847447502128067e-05, + "loss": 0.0131, + "step": 12490 + }, + { + "epoch": 0.970873786407767, + "grad_norm": 0.24382920563220978, + "learning_rate": 9.847041917330708e-05, + "loss": 0.0123, + "step": 12500 + }, + { + "epoch": 0.9716504854368933, + "grad_norm": 0.36398929357528687, + "learning_rate": 9.846635802468132e-05, + "loss": 0.0139, + "step": 12510 + }, + { + "epoch": 0.9724271844660194, + "grad_norm": 0.35657474398612976, + "learning_rate": 9.84622915758475e-05, + "loss": 0.015, + "step": 12520 + }, + { + "epoch": 0.9732038834951456, + "grad_norm": 0.15099184215068817, + "learning_rate": 9.845821982725034e-05, + "loss": 0.0097, + "step": 12530 + }, + { + "epoch": 0.9739805825242719, + "grad_norm": 0.1797579526901245, + "learning_rate": 9.845414277933514e-05, + "loss": 0.0114, + "step": 12540 + }, + { + "epoch": 0.974757281553398, + "grad_norm": 0.19040608406066895, + "learning_rate": 9.845006043254771e-05, + "loss": 0.0135, + "step": 12550 + }, + { + "epoch": 0.9755339805825243, + "grad_norm": 0.5010889768600464, + "learning_rate": 9.844597278733451e-05, + "loss": 0.0109, + "step": 12560 + }, + { + "epoch": 0.9763106796116505, + "grad_norm": 0.25450047850608826, + "learning_rate": 9.844187984414259e-05, + "loss": 0.0179, + "step": 12570 + }, + { + "epoch": 0.9770873786407767, + "grad_norm": 0.3446997404098511, + "learning_rate": 9.84377816034195e-05, + "loss": 0.0153, + "step": 12580 + }, + { + "epoch": 0.9778640776699029, + "grad_norm": 0.3531416058540344, + "learning_rate": 9.843367806561345e-05, + "loss": 0.0148, + "step": 12590 + }, + { + "epoch": 0.9786407766990292, + "grad_norm": 0.31728413701057434, + "learning_rate": 9.842956923117317e-05, + "loss": 0.0158, + "step": 12600 + }, + { + "epoch": 0.9794174757281553, + "grad_norm": 0.18654288351535797, + "learning_rate": 9.842545510054802e-05, + "loss": 0.011, + "step": 12610 + }, + { + "epoch": 0.9801941747572815, + "grad_norm": 0.28303366899490356, + "learning_rate": 9.842133567418792e-05, + "loss": 0.0136, + "step": 12620 + }, + { + "epoch": 0.9809708737864078, + "grad_norm": 0.350962370634079, + "learning_rate": 9.841721095254333e-05, + "loss": 0.0136, + "step": 12630 + }, + { + "epoch": 0.9817475728155339, + "grad_norm": 0.23658119142055511, + "learning_rate": 9.841308093606537e-05, + "loss": 0.0098, + "step": 12640 + }, + { + "epoch": 0.9825242718446602, + "grad_norm": 0.36501067876815796, + "learning_rate": 9.840894562520565e-05, + "loss": 0.0117, + "step": 12650 + }, + { + "epoch": 0.9833009708737864, + "grad_norm": 0.5076348781585693, + "learning_rate": 9.840480502041642e-05, + "loss": 0.0154, + "step": 12660 + }, + { + "epoch": 0.9840776699029126, + "grad_norm": 0.4043689966201782, + "learning_rate": 9.840065912215049e-05, + "loss": 0.0144, + "step": 12670 + }, + { + "epoch": 0.9848543689320388, + "grad_norm": 0.28447622060775757, + "learning_rate": 9.839650793086124e-05, + "loss": 0.0149, + "step": 12680 + }, + { + "epoch": 0.9856310679611651, + "grad_norm": 0.29509517550468445, + "learning_rate": 9.839235144700265e-05, + "loss": 0.0176, + "step": 12690 + }, + { + "epoch": 0.9864077669902913, + "grad_norm": 0.3909014165401459, + "learning_rate": 9.838818967102926e-05, + "loss": 0.0138, + "step": 12700 + }, + { + "epoch": 0.9871844660194175, + "grad_norm": 0.2343320995569229, + "learning_rate": 9.83840226033962e-05, + "loss": 0.0133, + "step": 12710 + }, + { + "epoch": 0.9879611650485437, + "grad_norm": 0.3745079040527344, + "learning_rate": 9.837985024455918e-05, + "loss": 0.0142, + "step": 12720 + }, + { + "epoch": 0.98873786407767, + "grad_norm": 0.38069912791252136, + "learning_rate": 9.837567259497447e-05, + "loss": 0.0134, + "step": 12730 + }, + { + "epoch": 0.9895145631067961, + "grad_norm": 0.5193870663642883, + "learning_rate": 9.837148965509894e-05, + "loss": 0.0155, + "step": 12740 + }, + { + "epoch": 0.9902912621359223, + "grad_norm": 0.33459290862083435, + "learning_rate": 9.836730142539001e-05, + "loss": 0.0139, + "step": 12750 + }, + { + "epoch": 0.9910679611650486, + "grad_norm": 0.2543620467185974, + "learning_rate": 9.836310790630574e-05, + "loss": 0.0155, + "step": 12760 + }, + { + "epoch": 0.9918446601941747, + "grad_norm": 0.2753019332885742, + "learning_rate": 9.83589090983047e-05, + "loss": 0.013, + "step": 12770 + }, + { + "epoch": 0.992621359223301, + "grad_norm": 0.313123494386673, + "learning_rate": 9.835470500184605e-05, + "loss": 0.013, + "step": 12780 + }, + { + "epoch": 0.9933980582524272, + "grad_norm": 0.41914454102516174, + "learning_rate": 9.835049561738957e-05, + "loss": 0.0185, + "step": 12790 + }, + { + "epoch": 0.9941747572815534, + "grad_norm": 0.23748953640460968, + "learning_rate": 9.834628094539558e-05, + "loss": 0.0182, + "step": 12800 + }, + { + "epoch": 0.9949514563106796, + "grad_norm": 0.2486427277326584, + "learning_rate": 9.834206098632499e-05, + "loss": 0.0145, + "step": 12810 + }, + { + "epoch": 0.9957281553398059, + "grad_norm": 0.4608488380908966, + "learning_rate": 9.833783574063931e-05, + "loss": 0.0133, + "step": 12820 + }, + { + "epoch": 0.996504854368932, + "grad_norm": 0.2609032988548279, + "learning_rate": 9.833360520880058e-05, + "loss": 0.0165, + "step": 12830 + }, + { + "epoch": 0.9972815533980582, + "grad_norm": 0.40296670794487, + "learning_rate": 9.832936939127144e-05, + "loss": 0.0149, + "step": 12840 + }, + { + "epoch": 0.9980582524271845, + "grad_norm": 0.26429224014282227, + "learning_rate": 9.832512828851515e-05, + "loss": 0.0122, + "step": 12850 + }, + { + "epoch": 0.9988349514563106, + "grad_norm": 0.2713635563850403, + "learning_rate": 9.832088190099546e-05, + "loss": 0.0141, + "step": 12860 + }, + { + "epoch": 0.9996116504854369, + "grad_norm": 0.298799991607666, + "learning_rate": 9.831663022917679e-05, + "loss": 0.0152, + "step": 12870 + }, + { + "epoch": 1.0003883495145631, + "grad_norm": 0.1940838098526001, + "learning_rate": 9.831237327352407e-05, + "loss": 0.011, + "step": 12880 + }, + { + "epoch": 1.0011650485436894, + "grad_norm": 0.22308559715747833, + "learning_rate": 9.830811103450286e-05, + "loss": 0.0119, + "step": 12890 + }, + { + "epoch": 1.0019417475728156, + "grad_norm": 0.19629502296447754, + "learning_rate": 9.830384351257924e-05, + "loss": 0.0128, + "step": 12900 + }, + { + "epoch": 1.0027184466019416, + "grad_norm": 0.2724462151527405, + "learning_rate": 9.829957070821993e-05, + "loss": 0.0128, + "step": 12910 + }, + { + "epoch": 1.003495145631068, + "grad_norm": 0.3369176685810089, + "learning_rate": 9.829529262189218e-05, + "loss": 0.0131, + "step": 12920 + }, + { + "epoch": 1.0042718446601941, + "grad_norm": 0.40180936455726624, + "learning_rate": 9.829100925406385e-05, + "loss": 0.0125, + "step": 12930 + }, + { + "epoch": 1.0050485436893204, + "grad_norm": 0.23559445142745972, + "learning_rate": 9.828672060520333e-05, + "loss": 0.0105, + "step": 12940 + }, + { + "epoch": 1.0058252427184466, + "grad_norm": 0.34896549582481384, + "learning_rate": 9.828242667577966e-05, + "loss": 0.0134, + "step": 12950 + }, + { + "epoch": 1.006601941747573, + "grad_norm": 0.21377454698085785, + "learning_rate": 9.82781274662624e-05, + "loss": 0.0104, + "step": 12960 + }, + { + "epoch": 1.007378640776699, + "grad_norm": 0.1851074993610382, + "learning_rate": 9.82738229771217e-05, + "loss": 0.0121, + "step": 12970 + }, + { + "epoch": 1.0081553398058252, + "grad_norm": 0.3873532712459564, + "learning_rate": 9.826951320882829e-05, + "loss": 0.0141, + "step": 12980 + }, + { + "epoch": 1.0089320388349514, + "grad_norm": 0.3055707514286041, + "learning_rate": 9.826519816185351e-05, + "loss": 0.0146, + "step": 12990 + }, + { + "epoch": 1.0097087378640777, + "grad_norm": 0.3959999978542328, + "learning_rate": 9.826087783666921e-05, + "loss": 0.0117, + "step": 13000 + }, + { + "epoch": 1.010485436893204, + "grad_norm": 0.3025612533092499, + "learning_rate": 9.825655223374787e-05, + "loss": 0.0125, + "step": 13010 + }, + { + "epoch": 1.0112621359223302, + "grad_norm": 0.2892164885997772, + "learning_rate": 9.825222135356253e-05, + "loss": 0.0115, + "step": 13020 + }, + { + "epoch": 1.0120388349514564, + "grad_norm": 0.15238802134990692, + "learning_rate": 9.82478851965868e-05, + "loss": 0.0154, + "step": 13030 + }, + { + "epoch": 1.0128155339805824, + "grad_norm": 0.4370860159397125, + "learning_rate": 9.82435437632949e-05, + "loss": 0.0149, + "step": 13040 + }, + { + "epoch": 1.0135922330097087, + "grad_norm": 0.39402711391448975, + "learning_rate": 9.823919705416158e-05, + "loss": 0.0139, + "step": 13050 + }, + { + "epoch": 1.014368932038835, + "grad_norm": 0.2405788153409958, + "learning_rate": 9.82348450696622e-05, + "loss": 0.0135, + "step": 13060 + }, + { + "epoch": 1.0151456310679612, + "grad_norm": 0.2306615710258484, + "learning_rate": 9.823048781027268e-05, + "loss": 0.0119, + "step": 13070 + }, + { + "epoch": 1.0159223300970874, + "grad_norm": 0.34758099913597107, + "learning_rate": 9.822612527646953e-05, + "loss": 0.0125, + "step": 13080 + }, + { + "epoch": 1.0166990291262137, + "grad_norm": 0.3933582901954651, + "learning_rate": 9.822175746872984e-05, + "loss": 0.0116, + "step": 13090 + }, + { + "epoch": 1.0174757281553397, + "grad_norm": 0.31624022126197815, + "learning_rate": 9.821738438753123e-05, + "loss": 0.0109, + "step": 13100 + }, + { + "epoch": 1.018252427184466, + "grad_norm": 0.21370826661586761, + "learning_rate": 9.821300603335196e-05, + "loss": 0.0122, + "step": 13110 + }, + { + "epoch": 1.0190291262135922, + "grad_norm": 0.23461541533470154, + "learning_rate": 9.820862240667085e-05, + "loss": 0.014, + "step": 13120 + }, + { + "epoch": 1.0198058252427185, + "grad_norm": 0.4385398328304291, + "learning_rate": 9.820423350796726e-05, + "loss": 0.0149, + "step": 13130 + }, + { + "epoch": 1.0205825242718447, + "grad_norm": 0.2412104308605194, + "learning_rate": 9.819983933772118e-05, + "loss": 0.0146, + "step": 13140 + }, + { + "epoch": 1.021359223300971, + "grad_norm": 0.24417069554328918, + "learning_rate": 9.819543989641314e-05, + "loss": 0.0125, + "step": 13150 + }, + { + "epoch": 1.022135922330097, + "grad_norm": 0.498424232006073, + "learning_rate": 9.819103518452423e-05, + "loss": 0.016, + "step": 13160 + }, + { + "epoch": 1.0229126213592232, + "grad_norm": 0.2888980209827423, + "learning_rate": 9.818662520253618e-05, + "loss": 0.0146, + "step": 13170 + }, + { + "epoch": 1.0236893203883495, + "grad_norm": 0.1694532185792923, + "learning_rate": 9.818220995093126e-05, + "loss": 0.0118, + "step": 13180 + }, + { + "epoch": 1.0244660194174757, + "grad_norm": 0.26678767800331116, + "learning_rate": 9.817778943019228e-05, + "loss": 0.0099, + "step": 13190 + }, + { + "epoch": 1.025242718446602, + "grad_norm": 0.1916746348142624, + "learning_rate": 9.81733636408027e-05, + "loss": 0.0202, + "step": 13200 + }, + { + "epoch": 1.0260194174757282, + "grad_norm": 0.2292085736989975, + "learning_rate": 9.816893258324649e-05, + "loss": 0.0145, + "step": 13210 + }, + { + "epoch": 1.0267961165048545, + "grad_norm": 0.37615418434143066, + "learning_rate": 9.816449625800823e-05, + "loss": 0.0146, + "step": 13220 + }, + { + "epoch": 1.0275728155339805, + "grad_norm": 0.6329358220100403, + "learning_rate": 9.816005466557308e-05, + "loss": 0.0174, + "step": 13230 + }, + { + "epoch": 1.0283495145631067, + "grad_norm": 0.2408079355955124, + "learning_rate": 9.815560780642674e-05, + "loss": 0.0138, + "step": 13240 + }, + { + "epoch": 1.029126213592233, + "grad_norm": 0.23456622660160065, + "learning_rate": 9.815115568105555e-05, + "loss": 0.0151, + "step": 13250 + }, + { + "epoch": 1.0299029126213592, + "grad_norm": 0.2967127561569214, + "learning_rate": 9.814669828994638e-05, + "loss": 0.0161, + "step": 13260 + }, + { + "epoch": 1.0306796116504855, + "grad_norm": 0.3182600140571594, + "learning_rate": 9.814223563358665e-05, + "loss": 0.0165, + "step": 13270 + }, + { + "epoch": 1.0314563106796117, + "grad_norm": 0.2769196629524231, + "learning_rate": 9.813776771246443e-05, + "loss": 0.0153, + "step": 13280 + }, + { + "epoch": 1.0322330097087378, + "grad_norm": 0.40487638115882874, + "learning_rate": 9.813329452706829e-05, + "loss": 0.0157, + "step": 13290 + }, + { + "epoch": 1.033009708737864, + "grad_norm": 0.3758496344089508, + "learning_rate": 9.812881607788744e-05, + "loss": 0.0164, + "step": 13300 + }, + { + "epoch": 1.0337864077669903, + "grad_norm": 0.20832735300064087, + "learning_rate": 9.812433236541163e-05, + "loss": 0.0135, + "step": 13310 + }, + { + "epoch": 1.0345631067961165, + "grad_norm": 0.3657478094100952, + "learning_rate": 9.811984339013116e-05, + "loss": 0.0147, + "step": 13320 + }, + { + "epoch": 1.0353398058252428, + "grad_norm": 0.49358269572257996, + "learning_rate": 9.811534915253698e-05, + "loss": 0.0156, + "step": 13330 + }, + { + "epoch": 1.036116504854369, + "grad_norm": 0.21290284395217896, + "learning_rate": 9.811084965312056e-05, + "loss": 0.012, + "step": 13340 + }, + { + "epoch": 1.036893203883495, + "grad_norm": 0.24200648069381714, + "learning_rate": 9.810634489237396e-05, + "loss": 0.0149, + "step": 13350 + }, + { + "epoch": 1.0376699029126213, + "grad_norm": 0.2725851237773895, + "learning_rate": 9.81018348707898e-05, + "loss": 0.0128, + "step": 13360 + }, + { + "epoch": 1.0384466019417475, + "grad_norm": 0.2737564742565155, + "learning_rate": 9.809731958886131e-05, + "loss": 0.0134, + "step": 13370 + }, + { + "epoch": 1.0392233009708738, + "grad_norm": 0.5198431611061096, + "learning_rate": 9.809279904708224e-05, + "loss": 0.0104, + "step": 13380 + }, + { + "epoch": 1.04, + "grad_norm": 0.13211673498153687, + "learning_rate": 9.808827324594699e-05, + "loss": 0.0159, + "step": 13390 + }, + { + "epoch": 1.0407766990291263, + "grad_norm": 0.33782079815864563, + "learning_rate": 9.808374218595046e-05, + "loss": 0.012, + "step": 13400 + }, + { + "epoch": 1.0415533980582525, + "grad_norm": 0.2653517723083496, + "learning_rate": 9.80792058675882e-05, + "loss": 0.013, + "step": 13410 + }, + { + "epoch": 1.0423300970873786, + "grad_norm": 0.34212005138397217, + "learning_rate": 9.807466429135627e-05, + "loss": 0.0133, + "step": 13420 + }, + { + "epoch": 1.0431067961165048, + "grad_norm": 0.1892663836479187, + "learning_rate": 9.807011745775132e-05, + "loss": 0.0143, + "step": 13430 + }, + { + "epoch": 1.043883495145631, + "grad_norm": 0.2318224459886551, + "learning_rate": 9.806556536727061e-05, + "loss": 0.015, + "step": 13440 + }, + { + "epoch": 1.0446601941747573, + "grad_norm": 0.36741676926612854, + "learning_rate": 9.806100802041193e-05, + "loss": 0.0133, + "step": 13450 + }, + { + "epoch": 1.0454368932038836, + "grad_norm": 0.19106441736221313, + "learning_rate": 9.805644541767368e-05, + "loss": 0.0152, + "step": 13460 + }, + { + "epoch": 1.0462135922330098, + "grad_norm": 0.304009348154068, + "learning_rate": 9.805187755955478e-05, + "loss": 0.0137, + "step": 13470 + }, + { + "epoch": 1.0469902912621358, + "grad_norm": 0.2027643322944641, + "learning_rate": 9.804730444655483e-05, + "loss": 0.0109, + "step": 13480 + }, + { + "epoch": 1.047766990291262, + "grad_norm": 0.3156839907169342, + "learning_rate": 9.804272607917388e-05, + "loss": 0.0189, + "step": 13490 + }, + { + "epoch": 1.0485436893203883, + "grad_norm": 0.2935045063495636, + "learning_rate": 9.803814245791265e-05, + "loss": 0.0154, + "step": 13500 + }, + { + "epoch": 1.0493203883495146, + "grad_norm": 0.2196035534143448, + "learning_rate": 9.803355358327239e-05, + "loss": 0.015, + "step": 13510 + }, + { + "epoch": 1.0500970873786408, + "grad_norm": 0.18505771458148956, + "learning_rate": 9.802895945575492e-05, + "loss": 0.0116, + "step": 13520 + }, + { + "epoch": 1.050873786407767, + "grad_norm": 0.22431236505508423, + "learning_rate": 9.802436007586266e-05, + "loss": 0.013, + "step": 13530 + }, + { + "epoch": 1.051650485436893, + "grad_norm": 0.7175360918045044, + "learning_rate": 9.801975544409858e-05, + "loss": 0.016, + "step": 13540 + }, + { + "epoch": 1.0524271844660193, + "grad_norm": 0.2868236303329468, + "learning_rate": 9.801514556096625e-05, + "loss": 0.0095, + "step": 13550 + }, + { + "epoch": 1.0532038834951456, + "grad_norm": 0.28871285915374756, + "learning_rate": 9.801053042696977e-05, + "loss": 0.0143, + "step": 13560 + }, + { + "epoch": 1.0539805825242718, + "grad_norm": 0.37210050225257874, + "learning_rate": 9.800591004261388e-05, + "loss": 0.0108, + "step": 13570 + }, + { + "epoch": 1.054757281553398, + "grad_norm": 0.22997038066387177, + "learning_rate": 9.800128440840385e-05, + "loss": 0.0144, + "step": 13580 + }, + { + "epoch": 1.0555339805825243, + "grad_norm": 0.5639575123786926, + "learning_rate": 9.799665352484552e-05, + "loss": 0.0123, + "step": 13590 + }, + { + "epoch": 1.0563106796116504, + "grad_norm": 0.3577196002006531, + "learning_rate": 9.799201739244532e-05, + "loss": 0.013, + "step": 13600 + }, + { + "epoch": 1.0570873786407766, + "grad_norm": 0.14390605688095093, + "learning_rate": 9.798737601171025e-05, + "loss": 0.011, + "step": 13610 + }, + { + "epoch": 1.0578640776699029, + "grad_norm": 0.18834128975868225, + "learning_rate": 9.79827293831479e-05, + "loss": 0.0094, + "step": 13620 + }, + { + "epoch": 1.0586407766990291, + "grad_norm": 0.2294120192527771, + "learning_rate": 9.797807750726638e-05, + "loss": 0.0142, + "step": 13630 + }, + { + "epoch": 1.0594174757281554, + "grad_norm": 0.5338706970214844, + "learning_rate": 9.797342038457446e-05, + "loss": 0.0144, + "step": 13640 + }, + { + "epoch": 1.0601941747572816, + "grad_norm": 0.30950871109962463, + "learning_rate": 9.796875801558141e-05, + "loss": 0.0136, + "step": 13650 + }, + { + "epoch": 1.0609708737864079, + "grad_norm": 0.27177396416664124, + "learning_rate": 9.79640904007971e-05, + "loss": 0.0158, + "step": 13660 + }, + { + "epoch": 1.061747572815534, + "grad_norm": 0.2920258045196533, + "learning_rate": 9.795941754073199e-05, + "loss": 0.0105, + "step": 13670 + }, + { + "epoch": 1.0625242718446601, + "grad_norm": 0.32728996872901917, + "learning_rate": 9.795473943589705e-05, + "loss": 0.0126, + "step": 13680 + }, + { + "epoch": 1.0633009708737864, + "grad_norm": 0.2933209538459778, + "learning_rate": 9.795005608680394e-05, + "loss": 0.0141, + "step": 13690 + }, + { + "epoch": 1.0640776699029126, + "grad_norm": 0.39565911889076233, + "learning_rate": 9.794536749396477e-05, + "loss": 0.0117, + "step": 13700 + }, + { + "epoch": 1.0648543689320389, + "grad_norm": 0.25173869729042053, + "learning_rate": 9.79406736578923e-05, + "loss": 0.012, + "step": 13710 + }, + { + "epoch": 1.0656310679611651, + "grad_norm": 0.4372747242450714, + "learning_rate": 9.793597457909984e-05, + "loss": 0.0084, + "step": 13720 + }, + { + "epoch": 1.0664077669902912, + "grad_norm": 0.28705307841300964, + "learning_rate": 9.793127025810127e-05, + "loss": 0.0137, + "step": 13730 + }, + { + "epoch": 1.0671844660194174, + "grad_norm": 0.3110854923725128, + "learning_rate": 9.792656069541104e-05, + "loss": 0.0109, + "step": 13740 + }, + { + "epoch": 1.0679611650485437, + "grad_norm": 0.3353462219238281, + "learning_rate": 9.79218458915442e-05, + "loss": 0.0156, + "step": 13750 + }, + { + "epoch": 1.06873786407767, + "grad_norm": 0.340254008769989, + "learning_rate": 9.791712584701634e-05, + "loss": 0.0111, + "step": 13760 + }, + { + "epoch": 1.0695145631067962, + "grad_norm": 0.34509894251823425, + "learning_rate": 9.791240056234364e-05, + "loss": 0.0114, + "step": 13770 + }, + { + "epoch": 1.0702912621359224, + "grad_norm": 0.2385881394147873, + "learning_rate": 9.790767003804283e-05, + "loss": 0.0111, + "step": 13780 + }, + { + "epoch": 1.0710679611650487, + "grad_norm": 0.3101560175418854, + "learning_rate": 9.790293427463126e-05, + "loss": 0.0114, + "step": 13790 + }, + { + "epoch": 1.0718446601941747, + "grad_norm": 0.32974645495414734, + "learning_rate": 9.789819327262684e-05, + "loss": 0.0141, + "step": 13800 + }, + { + "epoch": 1.072621359223301, + "grad_norm": 0.2714821994304657, + "learning_rate": 9.7893447032548e-05, + "loss": 0.0117, + "step": 13810 + }, + { + "epoch": 1.0733980582524272, + "grad_norm": 0.20237642526626587, + "learning_rate": 9.78886955549138e-05, + "loss": 0.0137, + "step": 13820 + }, + { + "epoch": 1.0741747572815534, + "grad_norm": 0.3076236844062805, + "learning_rate": 9.788393884024387e-05, + "loss": 0.0133, + "step": 13830 + }, + { + "epoch": 1.0749514563106797, + "grad_norm": 0.24819795787334442, + "learning_rate": 9.787917688905836e-05, + "loss": 0.0104, + "step": 13840 + }, + { + "epoch": 1.075728155339806, + "grad_norm": 0.30523908138275146, + "learning_rate": 9.787440970187807e-05, + "loss": 0.0096, + "step": 13850 + }, + { + "epoch": 1.076504854368932, + "grad_norm": 0.3660496771335602, + "learning_rate": 9.786963727922429e-05, + "loss": 0.016, + "step": 13860 + }, + { + "epoch": 1.0772815533980582, + "grad_norm": 0.2119644433259964, + "learning_rate": 9.786485962161897e-05, + "loss": 0.0131, + "step": 13870 + }, + { + "epoch": 1.0780582524271844, + "grad_norm": 0.22889189422130585, + "learning_rate": 9.786007672958455e-05, + "loss": 0.0118, + "step": 13880 + }, + { + "epoch": 1.0788349514563107, + "grad_norm": 0.20869310200214386, + "learning_rate": 9.78552886036441e-05, + "loss": 0.0127, + "step": 13890 + }, + { + "epoch": 1.079611650485437, + "grad_norm": 0.17839401960372925, + "learning_rate": 9.785049524432124e-05, + "loss": 0.0127, + "step": 13900 + }, + { + "epoch": 1.0803883495145632, + "grad_norm": 0.21342037618160248, + "learning_rate": 9.784569665214016e-05, + "loss": 0.0126, + "step": 13910 + }, + { + "epoch": 1.0811650485436892, + "grad_norm": 0.25348854064941406, + "learning_rate": 9.784089282762563e-05, + "loss": 0.012, + "step": 13920 + }, + { + "epoch": 1.0819417475728155, + "grad_norm": 0.14735397696495056, + "learning_rate": 9.7836083771303e-05, + "loss": 0.0091, + "step": 13930 + }, + { + "epoch": 1.0827184466019417, + "grad_norm": 0.87553870677948, + "learning_rate": 9.783126948369817e-05, + "loss": 0.014, + "step": 13940 + }, + { + "epoch": 1.083495145631068, + "grad_norm": 0.3856087028980255, + "learning_rate": 9.78264499653376e-05, + "loss": 0.0141, + "step": 13950 + }, + { + "epoch": 1.0842718446601942, + "grad_norm": 0.2865378260612488, + "learning_rate": 9.782162521674838e-05, + "loss": 0.0148, + "step": 13960 + }, + { + "epoch": 1.0850485436893205, + "grad_norm": 0.28800538182258606, + "learning_rate": 9.781679523845812e-05, + "loss": 0.0132, + "step": 13970 + }, + { + "epoch": 1.0858252427184465, + "grad_norm": 0.267716646194458, + "learning_rate": 9.781196003099502e-05, + "loss": 0.0135, + "step": 13980 + }, + { + "epoch": 1.0866019417475727, + "grad_norm": 0.2969583570957184, + "learning_rate": 9.780711959488786e-05, + "loss": 0.0152, + "step": 13990 + }, + { + "epoch": 1.087378640776699, + "grad_norm": 0.3411621153354645, + "learning_rate": 9.780227393066599e-05, + "loss": 0.0123, + "step": 14000 + }, + { + "epoch": 1.0881553398058252, + "grad_norm": 0.3014488220214844, + "learning_rate": 9.77974230388593e-05, + "loss": 0.0138, + "step": 14010 + }, + { + "epoch": 1.0889320388349515, + "grad_norm": 0.45955854654312134, + "learning_rate": 9.779256691999829e-05, + "loss": 0.013, + "step": 14020 + }, + { + "epoch": 1.0897087378640777, + "grad_norm": 0.24564996361732483, + "learning_rate": 9.778770557461403e-05, + "loss": 0.0138, + "step": 14030 + }, + { + "epoch": 1.090485436893204, + "grad_norm": 0.3588430881500244, + "learning_rate": 9.778283900323812e-05, + "loss": 0.0141, + "step": 14040 + }, + { + "epoch": 1.09126213592233, + "grad_norm": 0.2838679254055023, + "learning_rate": 9.777796720640277e-05, + "loss": 0.0112, + "step": 14050 + }, + { + "epoch": 1.0920388349514563, + "grad_norm": 0.2305087149143219, + "learning_rate": 9.777309018464078e-05, + "loss": 0.0153, + "step": 14060 + }, + { + "epoch": 1.0928155339805825, + "grad_norm": 0.2850010395050049, + "learning_rate": 9.776820793848547e-05, + "loss": 0.0168, + "step": 14070 + }, + { + "epoch": 1.0935922330097088, + "grad_norm": 0.2782677710056305, + "learning_rate": 9.776332046847075e-05, + "loss": 0.0128, + "step": 14080 + }, + { + "epoch": 1.094368932038835, + "grad_norm": 0.2787250876426697, + "learning_rate": 9.775842777513111e-05, + "loss": 0.0105, + "step": 14090 + }, + { + "epoch": 1.0951456310679613, + "grad_norm": 0.32620903849601746, + "learning_rate": 9.775352985900163e-05, + "loss": 0.0123, + "step": 14100 + }, + { + "epoch": 1.0959223300970873, + "grad_norm": 0.2610512375831604, + "learning_rate": 9.774862672061791e-05, + "loss": 0.0097, + "step": 14110 + }, + { + "epoch": 1.0966990291262135, + "grad_norm": 0.19593876600265503, + "learning_rate": 9.774371836051616e-05, + "loss": 0.0148, + "step": 14120 + }, + { + "epoch": 1.0974757281553398, + "grad_norm": 0.3427547216415405, + "learning_rate": 9.773880477923315e-05, + "loss": 0.0136, + "step": 14130 + }, + { + "epoch": 1.098252427184466, + "grad_norm": 0.4353368580341339, + "learning_rate": 9.773388597730623e-05, + "loss": 0.0172, + "step": 14140 + }, + { + "epoch": 1.0990291262135923, + "grad_norm": 0.39866405725479126, + "learning_rate": 9.77289619552733e-05, + "loss": 0.011, + "step": 14150 + }, + { + "epoch": 1.0998058252427185, + "grad_norm": 0.2834815979003906, + "learning_rate": 9.772403271367285e-05, + "loss": 0.0145, + "step": 14160 + }, + { + "epoch": 1.1005825242718448, + "grad_norm": 0.2973640561103821, + "learning_rate": 9.771909825304396e-05, + "loss": 0.014, + "step": 14170 + }, + { + "epoch": 1.1013592233009708, + "grad_norm": 0.19235819578170776, + "learning_rate": 9.771415857392619e-05, + "loss": 0.0117, + "step": 14180 + }, + { + "epoch": 1.102135922330097, + "grad_norm": 0.3522842824459076, + "learning_rate": 9.770921367685978e-05, + "loss": 0.0117, + "step": 14190 + }, + { + "epoch": 1.1029126213592233, + "grad_norm": 0.34091654419898987, + "learning_rate": 9.770426356238551e-05, + "loss": 0.017, + "step": 14200 + }, + { + "epoch": 1.1036893203883495, + "grad_norm": 0.29848456382751465, + "learning_rate": 9.769930823104469e-05, + "loss": 0.0118, + "step": 14210 + }, + { + "epoch": 1.1044660194174758, + "grad_norm": 0.27188339829444885, + "learning_rate": 9.769434768337926e-05, + "loss": 0.0132, + "step": 14220 + }, + { + "epoch": 1.1052427184466018, + "grad_norm": 0.2880343496799469, + "learning_rate": 9.768938191993164e-05, + "loss": 0.0138, + "step": 14230 + }, + { + "epoch": 1.106019417475728, + "grad_norm": 0.21763227880001068, + "learning_rate": 9.768441094124494e-05, + "loss": 0.0095, + "step": 14240 + }, + { + "epoch": 1.1067961165048543, + "grad_norm": 0.2700042128562927, + "learning_rate": 9.767943474786275e-05, + "loss": 0.0129, + "step": 14250 + }, + { + "epoch": 1.1075728155339806, + "grad_norm": 0.8090839982032776, + "learning_rate": 9.767445334032923e-05, + "loss": 0.0119, + "step": 14260 + }, + { + "epoch": 1.1083495145631068, + "grad_norm": 0.25494056940078735, + "learning_rate": 9.766946671918919e-05, + "loss": 0.0099, + "step": 14270 + }, + { + "epoch": 1.109126213592233, + "grad_norm": 0.3371877670288086, + "learning_rate": 9.766447488498796e-05, + "loss": 0.0111, + "step": 14280 + }, + { + "epoch": 1.1099029126213593, + "grad_norm": 0.18960358202457428, + "learning_rate": 9.765947783827139e-05, + "loss": 0.009, + "step": 14290 + }, + { + "epoch": 1.1106796116504853, + "grad_norm": 0.19980040192604065, + "learning_rate": 9.765447557958599e-05, + "loss": 0.0121, + "step": 14300 + }, + { + "epoch": 1.1114563106796116, + "grad_norm": 0.2755138576030731, + "learning_rate": 9.764946810947879e-05, + "loss": 0.0095, + "step": 14310 + }, + { + "epoch": 1.1122330097087378, + "grad_norm": 0.24444729089736938, + "learning_rate": 9.764445542849738e-05, + "loss": 0.015, + "step": 14320 + }, + { + "epoch": 1.113009708737864, + "grad_norm": 0.6591444611549377, + "learning_rate": 9.763943753718998e-05, + "loss": 0.0185, + "step": 14330 + }, + { + "epoch": 1.1137864077669903, + "grad_norm": 0.19946126639842987, + "learning_rate": 9.76344144361053e-05, + "loss": 0.0114, + "step": 14340 + }, + { + "epoch": 1.1145631067961166, + "grad_norm": 0.2112717181444168, + "learning_rate": 9.762938612579269e-05, + "loss": 0.0117, + "step": 14350 + }, + { + "epoch": 1.1153398058252426, + "grad_norm": 0.2715565860271454, + "learning_rate": 9.762435260680202e-05, + "loss": 0.0137, + "step": 14360 + }, + { + "epoch": 1.1161165048543689, + "grad_norm": 0.24723194539546967, + "learning_rate": 9.761931387968373e-05, + "loss": 0.0116, + "step": 14370 + }, + { + "epoch": 1.1168932038834951, + "grad_norm": 0.35962656140327454, + "learning_rate": 9.76142699449889e-05, + "loss": 0.0162, + "step": 14380 + }, + { + "epoch": 1.1176699029126214, + "grad_norm": 0.2811196744441986, + "learning_rate": 9.760922080326908e-05, + "loss": 0.0128, + "step": 14390 + }, + { + "epoch": 1.1184466019417476, + "grad_norm": 0.20855800807476044, + "learning_rate": 9.760416645507644e-05, + "loss": 0.0145, + "step": 14400 + }, + { + "epoch": 1.1192233009708739, + "grad_norm": 0.2809492349624634, + "learning_rate": 9.759910690096375e-05, + "loss": 0.0125, + "step": 14410 + }, + { + "epoch": 1.12, + "grad_norm": 0.3249956965446472, + "learning_rate": 9.759404214148429e-05, + "loss": 0.0118, + "step": 14420 + }, + { + "epoch": 1.1207766990291261, + "grad_norm": 0.23805196583271027, + "learning_rate": 9.758897217719191e-05, + "loss": 0.0101, + "step": 14430 + }, + { + "epoch": 1.1215533980582524, + "grad_norm": 0.3346310257911682, + "learning_rate": 9.758389700864113e-05, + "loss": 0.0137, + "step": 14440 + }, + { + "epoch": 1.1223300970873786, + "grad_norm": 0.35273927450180054, + "learning_rate": 9.757881663638688e-05, + "loss": 0.0163, + "step": 14450 + }, + { + "epoch": 1.1231067961165049, + "grad_norm": 0.4687587022781372, + "learning_rate": 9.757373106098478e-05, + "loss": 0.0137, + "step": 14460 + }, + { + "epoch": 1.1238834951456311, + "grad_norm": 0.34834107756614685, + "learning_rate": 9.756864028299097e-05, + "loss": 0.0161, + "step": 14470 + }, + { + "epoch": 1.1246601941747574, + "grad_norm": 0.3825992941856384, + "learning_rate": 9.75635443029622e-05, + "loss": 0.0191, + "step": 14480 + }, + { + "epoch": 1.1254368932038834, + "grad_norm": 0.3243505358695984, + "learning_rate": 9.755844312145572e-05, + "loss": 0.0131, + "step": 14490 + }, + { + "epoch": 1.1262135922330097, + "grad_norm": 0.3059026598930359, + "learning_rate": 9.755333673902941e-05, + "loss": 0.0135, + "step": 14500 + }, + { + "epoch": 1.126990291262136, + "grad_norm": 0.4078558683395386, + "learning_rate": 9.75482251562417e-05, + "loss": 0.0117, + "step": 14510 + }, + { + "epoch": 1.1277669902912622, + "grad_norm": 0.43311116099357605, + "learning_rate": 9.754310837365155e-05, + "loss": 0.0126, + "step": 14520 + }, + { + "epoch": 1.1285436893203884, + "grad_norm": 0.3226035237312317, + "learning_rate": 9.753798639181856e-05, + "loss": 0.0128, + "step": 14530 + }, + { + "epoch": 1.1293203883495146, + "grad_norm": 0.30252572894096375, + "learning_rate": 9.753285921130286e-05, + "loss": 0.0151, + "step": 14540 + }, + { + "epoch": 1.130097087378641, + "grad_norm": 0.319465309381485, + "learning_rate": 9.752772683266512e-05, + "loss": 0.0133, + "step": 14550 + }, + { + "epoch": 1.130873786407767, + "grad_norm": 0.32288858294487, + "learning_rate": 9.752258925646665e-05, + "loss": 0.014, + "step": 14560 + }, + { + "epoch": 1.1316504854368932, + "grad_norm": 0.2840300500392914, + "learning_rate": 9.751744648326926e-05, + "loss": 0.0166, + "step": 14570 + }, + { + "epoch": 1.1324271844660194, + "grad_norm": 0.1791188269853592, + "learning_rate": 9.751229851363536e-05, + "loss": 0.013, + "step": 14580 + }, + { + "epoch": 1.1332038834951457, + "grad_norm": 0.21488714218139648, + "learning_rate": 9.750714534812793e-05, + "loss": 0.0122, + "step": 14590 + }, + { + "epoch": 1.133980582524272, + "grad_norm": 0.4266347885131836, + "learning_rate": 9.750198698731053e-05, + "loss": 0.0116, + "step": 14600 + }, + { + "epoch": 1.134757281553398, + "grad_norm": 0.1726216822862625, + "learning_rate": 9.749682343174722e-05, + "loss": 0.0127, + "step": 14610 + }, + { + "epoch": 1.1355339805825242, + "grad_norm": 0.2138567566871643, + "learning_rate": 9.749165468200272e-05, + "loss": 0.0113, + "step": 14620 + }, + { + "epoch": 1.1363106796116504, + "grad_norm": 0.22902967035770416, + "learning_rate": 9.748648073864229e-05, + "loss": 0.0169, + "step": 14630 + }, + { + "epoch": 1.1370873786407767, + "grad_norm": 0.4522528350353241, + "learning_rate": 9.748130160223168e-05, + "loss": 0.0139, + "step": 14640 + }, + { + "epoch": 1.137864077669903, + "grad_norm": 0.28949064016342163, + "learning_rate": 9.747611727333734e-05, + "loss": 0.0121, + "step": 14650 + }, + { + "epoch": 1.1386407766990292, + "grad_norm": 0.31074193120002747, + "learning_rate": 9.74709277525262e-05, + "loss": 0.0135, + "step": 14660 + }, + { + "epoch": 1.1394174757281554, + "grad_norm": 0.30175575613975525, + "learning_rate": 9.746573304036576e-05, + "loss": 0.0134, + "step": 14670 + }, + { + "epoch": 1.1401941747572815, + "grad_norm": 0.32140910625457764, + "learning_rate": 9.746053313742412e-05, + "loss": 0.0117, + "step": 14680 + }, + { + "epoch": 1.1409708737864077, + "grad_norm": 0.2489815503358841, + "learning_rate": 9.745532804426994e-05, + "loss": 0.0129, + "step": 14690 + }, + { + "epoch": 1.141747572815534, + "grad_norm": 0.5761396884918213, + "learning_rate": 9.745011776147242e-05, + "loss": 0.0111, + "step": 14700 + }, + { + "epoch": 1.1425242718446602, + "grad_norm": 0.5909214615821838, + "learning_rate": 9.744490228960138e-05, + "loss": 0.0154, + "step": 14710 + }, + { + "epoch": 1.1433009708737865, + "grad_norm": 0.23644274473190308, + "learning_rate": 9.743968162922713e-05, + "loss": 0.0141, + "step": 14720 + }, + { + "epoch": 1.1440776699029127, + "grad_norm": 0.852524995803833, + "learning_rate": 9.743445578092064e-05, + "loss": 0.0133, + "step": 14730 + }, + { + "epoch": 1.1448543689320387, + "grad_norm": 0.32966381311416626, + "learning_rate": 9.742922474525338e-05, + "loss": 0.0176, + "step": 14740 + }, + { + "epoch": 1.145631067961165, + "grad_norm": 0.37139806151390076, + "learning_rate": 9.742398852279741e-05, + "loss": 0.0131, + "step": 14750 + }, + { + "epoch": 1.1464077669902912, + "grad_norm": 0.17771175503730774, + "learning_rate": 9.741874711412535e-05, + "loss": 0.0149, + "step": 14760 + }, + { + "epoch": 1.1471844660194175, + "grad_norm": 0.37630319595336914, + "learning_rate": 9.741350051981042e-05, + "loss": 0.0109, + "step": 14770 + }, + { + "epoch": 1.1479611650485437, + "grad_norm": 0.25444626808166504, + "learning_rate": 9.740824874042633e-05, + "loss": 0.0114, + "step": 14780 + }, + { + "epoch": 1.14873786407767, + "grad_norm": 0.21465523540973663, + "learning_rate": 9.740299177654746e-05, + "loss": 0.012, + "step": 14790 + }, + { + "epoch": 1.1495145631067962, + "grad_norm": 0.25583919882774353, + "learning_rate": 9.739772962874867e-05, + "loss": 0.0082, + "step": 14800 + }, + { + "epoch": 1.1502912621359223, + "grad_norm": 0.3342750668525696, + "learning_rate": 9.739246229760541e-05, + "loss": 0.0095, + "step": 14810 + }, + { + "epoch": 1.1510679611650485, + "grad_norm": 0.2250155359506607, + "learning_rate": 9.738718978369376e-05, + "loss": 0.0128, + "step": 14820 + }, + { + "epoch": 1.1518446601941748, + "grad_norm": 0.3227555453777313, + "learning_rate": 9.738191208759025e-05, + "loss": 0.0116, + "step": 14830 + }, + { + "epoch": 1.152621359223301, + "grad_norm": 0.4262251853942871, + "learning_rate": 9.73766292098721e-05, + "loss": 0.0138, + "step": 14840 + }, + { + "epoch": 1.1533980582524272, + "grad_norm": 0.2935245633125305, + "learning_rate": 9.737134115111699e-05, + "loss": 0.0125, + "step": 14850 + }, + { + "epoch": 1.1541747572815533, + "grad_norm": 0.2215176224708557, + "learning_rate": 9.736604791190323e-05, + "loss": 0.0113, + "step": 14860 + }, + { + "epoch": 1.1549514563106795, + "grad_norm": 0.38106581568717957, + "learning_rate": 9.73607494928097e-05, + "loss": 0.0118, + "step": 14870 + }, + { + "epoch": 1.1557281553398058, + "grad_norm": 0.1755438596010208, + "learning_rate": 9.735544589441581e-05, + "loss": 0.0121, + "step": 14880 + }, + { + "epoch": 1.156504854368932, + "grad_norm": 0.30726751685142517, + "learning_rate": 9.735013711730154e-05, + "loss": 0.0143, + "step": 14890 + }, + { + "epoch": 1.1572815533980583, + "grad_norm": 0.1662568747997284, + "learning_rate": 9.734482316204747e-05, + "loss": 0.012, + "step": 14900 + }, + { + "epoch": 1.1580582524271845, + "grad_norm": 0.35843661427497864, + "learning_rate": 9.733950402923473e-05, + "loss": 0.0143, + "step": 14910 + }, + { + "epoch": 1.1588349514563108, + "grad_norm": 0.8666385412216187, + "learning_rate": 9.7334179719445e-05, + "loss": 0.0153, + "step": 14920 + }, + { + "epoch": 1.159611650485437, + "grad_norm": 0.19507744908332825, + "learning_rate": 9.732885023326053e-05, + "loss": 0.0129, + "step": 14930 + }, + { + "epoch": 1.160388349514563, + "grad_norm": 0.3301110565662384, + "learning_rate": 9.732351557126418e-05, + "loss": 0.0136, + "step": 14940 + }, + { + "epoch": 1.1611650485436893, + "grad_norm": 0.30799201130867004, + "learning_rate": 9.731817573403929e-05, + "loss": 0.01, + "step": 14950 + }, + { + "epoch": 1.1619417475728155, + "grad_norm": 0.3573544919490814, + "learning_rate": 9.731283072216985e-05, + "loss": 0.011, + "step": 14960 + }, + { + "epoch": 1.1627184466019418, + "grad_norm": 0.3567051887512207, + "learning_rate": 9.730748053624039e-05, + "loss": 0.0131, + "step": 14970 + }, + { + "epoch": 1.163495145631068, + "grad_norm": 0.22592371702194214, + "learning_rate": 9.730212517683598e-05, + "loss": 0.011, + "step": 14980 + }, + { + "epoch": 1.164271844660194, + "grad_norm": 0.24830658733844757, + "learning_rate": 9.729676464454228e-05, + "loss": 0.0104, + "step": 14990 + }, + { + "epoch": 1.1650485436893203, + "grad_norm": 0.2884840965270996, + "learning_rate": 9.72913989399455e-05, + "loss": 0.0093, + "step": 15000 + }, + { + "epoch": 1.1658252427184466, + "grad_norm": 0.24981389939785004, + "learning_rate": 9.728602806363242e-05, + "loss": 0.0112, + "step": 15010 + }, + { + "epoch": 1.1666019417475728, + "grad_norm": 0.14644424617290497, + "learning_rate": 9.728065201619043e-05, + "loss": 0.0109, + "step": 15020 + }, + { + "epoch": 1.167378640776699, + "grad_norm": 0.2134678214788437, + "learning_rate": 9.727527079820742e-05, + "loss": 0.0132, + "step": 15030 + }, + { + "epoch": 1.1681553398058253, + "grad_norm": 0.37416505813598633, + "learning_rate": 9.726988441027186e-05, + "loss": 0.0121, + "step": 15040 + }, + { + "epoch": 1.1689320388349516, + "grad_norm": 0.3351801931858063, + "learning_rate": 9.726449285297281e-05, + "loss": 0.0141, + "step": 15050 + }, + { + "epoch": 1.1697087378640776, + "grad_norm": 0.29035213589668274, + "learning_rate": 9.72590961268999e-05, + "loss": 0.0134, + "step": 15060 + }, + { + "epoch": 1.1704854368932038, + "grad_norm": 0.2779758870601654, + "learning_rate": 9.725369423264328e-05, + "loss": 0.0163, + "step": 15070 + }, + { + "epoch": 1.17126213592233, + "grad_norm": 0.36697179079055786, + "learning_rate": 9.72482871707937e-05, + "loss": 0.0112, + "step": 15080 + }, + { + "epoch": 1.1720388349514563, + "grad_norm": 0.41297581791877747, + "learning_rate": 9.724287494194247e-05, + "loss": 0.0142, + "step": 15090 + }, + { + "epoch": 1.1728155339805826, + "grad_norm": 0.197710782289505, + "learning_rate": 9.723745754668147e-05, + "loss": 0.013, + "step": 15100 + }, + { + "epoch": 1.1735922330097088, + "grad_norm": 0.2116517871618271, + "learning_rate": 9.723203498560313e-05, + "loss": 0.0138, + "step": 15110 + }, + { + "epoch": 1.1743689320388349, + "grad_norm": 0.30957862734794617, + "learning_rate": 9.722660725930046e-05, + "loss": 0.0142, + "step": 15120 + }, + { + "epoch": 1.175145631067961, + "grad_norm": 0.2671459913253784, + "learning_rate": 9.722117436836702e-05, + "loss": 0.0108, + "step": 15130 + }, + { + "epoch": 1.1759223300970874, + "grad_norm": 0.20557422935962677, + "learning_rate": 9.721573631339696e-05, + "loss": 0.0101, + "step": 15140 + }, + { + "epoch": 1.1766990291262136, + "grad_norm": 0.15171678364276886, + "learning_rate": 9.721029309498494e-05, + "loss": 0.0112, + "step": 15150 + }, + { + "epoch": 1.1774757281553399, + "grad_norm": 0.25932198762893677, + "learning_rate": 9.720484471372627e-05, + "loss": 0.0103, + "step": 15160 + }, + { + "epoch": 1.178252427184466, + "grad_norm": 0.300493985414505, + "learning_rate": 9.719939117021673e-05, + "loss": 0.012, + "step": 15170 + }, + { + "epoch": 1.1790291262135923, + "grad_norm": 0.33013713359832764, + "learning_rate": 9.719393246505275e-05, + "loss": 0.0114, + "step": 15180 + }, + { + "epoch": 1.1798058252427184, + "grad_norm": 0.21153895556926727, + "learning_rate": 9.718846859883128e-05, + "loss": 0.0119, + "step": 15190 + }, + { + "epoch": 1.1805825242718446, + "grad_norm": 0.18809372186660767, + "learning_rate": 9.718299957214982e-05, + "loss": 0.0113, + "step": 15200 + }, + { + "epoch": 1.1813592233009709, + "grad_norm": 0.19634835422039032, + "learning_rate": 9.717752538560646e-05, + "loss": 0.0116, + "step": 15210 + }, + { + "epoch": 1.1821359223300971, + "grad_norm": 0.4340144693851471, + "learning_rate": 9.717204603979986e-05, + "loss": 0.0107, + "step": 15220 + }, + { + "epoch": 1.1829126213592234, + "grad_norm": 0.4026137590408325, + "learning_rate": 9.716656153532922e-05, + "loss": 0.0094, + "step": 15230 + }, + { + "epoch": 1.1836893203883494, + "grad_norm": 0.5406955480575562, + "learning_rate": 9.716107187279434e-05, + "loss": 0.0132, + "step": 15240 + }, + { + "epoch": 1.1844660194174756, + "grad_norm": 0.3540995419025421, + "learning_rate": 9.715557705279555e-05, + "loss": 0.0142, + "step": 15250 + }, + { + "epoch": 1.185242718446602, + "grad_norm": 0.27674081921577454, + "learning_rate": 9.715007707593372e-05, + "loss": 0.0128, + "step": 15260 + }, + { + "epoch": 1.1860194174757281, + "grad_norm": 0.33658367395401, + "learning_rate": 9.714457194281036e-05, + "loss": 0.0142, + "step": 15270 + }, + { + "epoch": 1.1867961165048544, + "grad_norm": 0.36035260558128357, + "learning_rate": 9.713906165402751e-05, + "loss": 0.0137, + "step": 15280 + }, + { + "epoch": 1.1875728155339806, + "grad_norm": 0.27272501587867737, + "learning_rate": 9.713354621018774e-05, + "loss": 0.0141, + "step": 15290 + }, + { + "epoch": 1.188349514563107, + "grad_norm": 0.20592625439167023, + "learning_rate": 9.712802561189422e-05, + "loss": 0.0108, + "step": 15300 + }, + { + "epoch": 1.189126213592233, + "grad_norm": 0.2957897186279297, + "learning_rate": 9.712249985975069e-05, + "loss": 0.013, + "step": 15310 + }, + { + "epoch": 1.1899029126213592, + "grad_norm": 0.2695428729057312, + "learning_rate": 9.71169689543614e-05, + "loss": 0.0154, + "step": 15320 + }, + { + "epoch": 1.1906796116504854, + "grad_norm": 0.3210349678993225, + "learning_rate": 9.711143289633123e-05, + "loss": 0.013, + "step": 15330 + }, + { + "epoch": 1.1914563106796117, + "grad_norm": 0.197113037109375, + "learning_rate": 9.710589168626561e-05, + "loss": 0.0114, + "step": 15340 + }, + { + "epoch": 1.192233009708738, + "grad_norm": 0.2738254964351654, + "learning_rate": 9.710034532477048e-05, + "loss": 0.012, + "step": 15350 + }, + { + "epoch": 1.1930097087378642, + "grad_norm": 0.2382541447877884, + "learning_rate": 9.709479381245239e-05, + "loss": 0.012, + "step": 15360 + }, + { + "epoch": 1.1937864077669902, + "grad_norm": 0.2578085660934448, + "learning_rate": 9.708923714991847e-05, + "loss": 0.013, + "step": 15370 + }, + { + "epoch": 1.1945631067961164, + "grad_norm": 0.27399885654449463, + "learning_rate": 9.708367533777638e-05, + "loss": 0.0146, + "step": 15380 + }, + { + "epoch": 1.1953398058252427, + "grad_norm": 0.34505462646484375, + "learning_rate": 9.707810837663431e-05, + "loss": 0.0155, + "step": 15390 + }, + { + "epoch": 1.196116504854369, + "grad_norm": 0.2390601485967636, + "learning_rate": 9.707253626710113e-05, + "loss": 0.0162, + "step": 15400 + }, + { + "epoch": 1.1968932038834952, + "grad_norm": 0.2882860004901886, + "learning_rate": 9.706695900978613e-05, + "loss": 0.0125, + "step": 15410 + }, + { + "epoch": 1.1976699029126214, + "grad_norm": 0.3296816051006317, + "learning_rate": 9.706137660529926e-05, + "loss": 0.0169, + "step": 15420 + }, + { + "epoch": 1.1984466019417477, + "grad_norm": 0.36013880372047424, + "learning_rate": 9.705578905425101e-05, + "loss": 0.0137, + "step": 15430 + }, + { + "epoch": 1.1992233009708737, + "grad_norm": 0.24412186443805695, + "learning_rate": 9.705019635725241e-05, + "loss": 0.0105, + "step": 15440 + }, + { + "epoch": 1.2, + "grad_norm": 0.1904483139514923, + "learning_rate": 9.704459851491508e-05, + "loss": 0.0115, + "step": 15450 + }, + { + "epoch": 1.2007766990291262, + "grad_norm": 0.1862114667892456, + "learning_rate": 9.703899552785118e-05, + "loss": 0.0115, + "step": 15460 + }, + { + "epoch": 1.2015533980582525, + "grad_norm": 0.7166153788566589, + "learning_rate": 9.703338739667346e-05, + "loss": 0.018, + "step": 15470 + }, + { + "epoch": 1.2023300970873787, + "grad_norm": 0.3291536569595337, + "learning_rate": 9.70277741219952e-05, + "loss": 0.0105, + "step": 15480 + }, + { + "epoch": 1.203106796116505, + "grad_norm": 0.24112707376480103, + "learning_rate": 9.702215570443027e-05, + "loss": 0.015, + "step": 15490 + }, + { + "epoch": 1.203883495145631, + "grad_norm": 0.26109153032302856, + "learning_rate": 9.701653214459309e-05, + "loss": 0.0136, + "step": 15500 + }, + { + "epoch": 1.2046601941747572, + "grad_norm": 0.3237490653991699, + "learning_rate": 9.701090344309865e-05, + "loss": 0.0127, + "step": 15510 + }, + { + "epoch": 1.2054368932038835, + "grad_norm": 0.27851977944374084, + "learning_rate": 9.700526960056247e-05, + "loss": 0.0126, + "step": 15520 + }, + { + "epoch": 1.2062135922330097, + "grad_norm": 0.3831906020641327, + "learning_rate": 9.699963061760068e-05, + "loss": 0.0175, + "step": 15530 + }, + { + "epoch": 1.206990291262136, + "grad_norm": 0.35299190878868103, + "learning_rate": 9.699398649482997e-05, + "loss": 0.0127, + "step": 15540 + }, + { + "epoch": 1.2077669902912622, + "grad_norm": 0.35513707995414734, + "learning_rate": 9.698833723286753e-05, + "loss": 0.0152, + "step": 15550 + }, + { + "epoch": 1.2085436893203885, + "grad_norm": 0.18663601577281952, + "learning_rate": 9.698268283233118e-05, + "loss": 0.0119, + "step": 15560 + }, + { + "epoch": 1.2093203883495145, + "grad_norm": 0.2400321662425995, + "learning_rate": 9.697702329383929e-05, + "loss": 0.0111, + "step": 15570 + }, + { + "epoch": 1.2100970873786407, + "grad_norm": 0.2594195604324341, + "learning_rate": 9.697135861801074e-05, + "loss": 0.0104, + "step": 15580 + }, + { + "epoch": 1.210873786407767, + "grad_norm": 0.2730589509010315, + "learning_rate": 9.696568880546505e-05, + "loss": 0.0112, + "step": 15590 + }, + { + "epoch": 1.2116504854368932, + "grad_norm": 0.20140697062015533, + "learning_rate": 9.696001385682223e-05, + "loss": 0.0124, + "step": 15600 + }, + { + "epoch": 1.2124271844660195, + "grad_norm": 0.25468575954437256, + "learning_rate": 9.695433377270291e-05, + "loss": 0.0148, + "step": 15610 + }, + { + "epoch": 1.2132038834951455, + "grad_norm": 0.1912050098180771, + "learning_rate": 9.694864855372824e-05, + "loss": 0.0112, + "step": 15620 + }, + { + "epoch": 1.2139805825242718, + "grad_norm": 0.23445528745651245, + "learning_rate": 9.694295820051995e-05, + "loss": 0.0128, + "step": 15630 + }, + { + "epoch": 1.214757281553398, + "grad_norm": 0.22640806436538696, + "learning_rate": 9.693726271370032e-05, + "loss": 0.0101, + "step": 15640 + }, + { + "epoch": 1.2155339805825243, + "grad_norm": 0.3095328211784363, + "learning_rate": 9.693156209389221e-05, + "loss": 0.0126, + "step": 15650 + }, + { + "epoch": 1.2163106796116505, + "grad_norm": 0.17094187438488007, + "learning_rate": 9.692585634171905e-05, + "loss": 0.0119, + "step": 15660 + }, + { + "epoch": 1.2170873786407768, + "grad_norm": 0.3337049186229706, + "learning_rate": 9.692014545780476e-05, + "loss": 0.0102, + "step": 15670 + }, + { + "epoch": 1.217864077669903, + "grad_norm": 0.3177102208137512, + "learning_rate": 9.691442944277393e-05, + "loss": 0.0083, + "step": 15680 + }, + { + "epoch": 1.218640776699029, + "grad_norm": 0.40698036551475525, + "learning_rate": 9.690870829725162e-05, + "loss": 0.0108, + "step": 15690 + }, + { + "epoch": 1.2194174757281553, + "grad_norm": 0.2961624562740326, + "learning_rate": 9.69029820218635e-05, + "loss": 0.0105, + "step": 15700 + }, + { + "epoch": 1.2201941747572815, + "grad_norm": 0.23617912828922272, + "learning_rate": 9.689725061723579e-05, + "loss": 0.0104, + "step": 15710 + }, + { + "epoch": 1.2209708737864078, + "grad_norm": 0.15540093183517456, + "learning_rate": 9.689151408399527e-05, + "loss": 0.0132, + "step": 15720 + }, + { + "epoch": 1.221747572815534, + "grad_norm": 0.258646696805954, + "learning_rate": 9.688577242276924e-05, + "loss": 0.0109, + "step": 15730 + }, + { + "epoch": 1.2225242718446603, + "grad_norm": 0.2637520134449005, + "learning_rate": 9.688002563418566e-05, + "loss": 0.0097, + "step": 15740 + }, + { + "epoch": 1.2233009708737863, + "grad_norm": 0.6032989621162415, + "learning_rate": 9.687427371887293e-05, + "loss": 0.0114, + "step": 15750 + }, + { + "epoch": 1.2240776699029126, + "grad_norm": 0.35634416341781616, + "learning_rate": 9.686851667746012e-05, + "loss": 0.0149, + "step": 15760 + }, + { + "epoch": 1.2248543689320388, + "grad_norm": 0.6873892545700073, + "learning_rate": 9.686275451057677e-05, + "loss": 0.0098, + "step": 15770 + }, + { + "epoch": 1.225631067961165, + "grad_norm": 0.2452804148197174, + "learning_rate": 9.685698721885308e-05, + "loss": 0.0108, + "step": 15780 + }, + { + "epoch": 1.2264077669902913, + "grad_norm": 0.29456502199172974, + "learning_rate": 9.68512148029197e-05, + "loss": 0.012, + "step": 15790 + }, + { + "epoch": 1.2271844660194176, + "grad_norm": 0.3199583888053894, + "learning_rate": 9.684543726340791e-05, + "loss": 0.0133, + "step": 15800 + }, + { + "epoch": 1.2279611650485438, + "grad_norm": 0.214458167552948, + "learning_rate": 9.683965460094952e-05, + "loss": 0.0101, + "step": 15810 + }, + { + "epoch": 1.2287378640776698, + "grad_norm": 0.2139192670583725, + "learning_rate": 9.683386681617694e-05, + "loss": 0.0153, + "step": 15820 + }, + { + "epoch": 1.229514563106796, + "grad_norm": 0.23742835223674774, + "learning_rate": 9.68280739097231e-05, + "loss": 0.0099, + "step": 15830 + }, + { + "epoch": 1.2302912621359223, + "grad_norm": 0.23050525784492493, + "learning_rate": 9.682227588222148e-05, + "loss": 0.0105, + "step": 15840 + }, + { + "epoch": 1.2310679611650486, + "grad_norm": 0.3961739242076874, + "learning_rate": 9.681647273430618e-05, + "loss": 0.0139, + "step": 15850 + }, + { + "epoch": 1.2318446601941748, + "grad_norm": 0.7297412753105164, + "learning_rate": 9.681066446661182e-05, + "loss": 0.0098, + "step": 15860 + }, + { + "epoch": 1.2326213592233008, + "grad_norm": 0.4011157751083374, + "learning_rate": 9.680485107977357e-05, + "loss": 0.01, + "step": 15870 + }, + { + "epoch": 1.233398058252427, + "grad_norm": 0.20991988480091095, + "learning_rate": 9.679903257442716e-05, + "loss": 0.0132, + "step": 15880 + }, + { + "epoch": 1.2341747572815533, + "grad_norm": 0.34722402691841125, + "learning_rate": 9.679320895120891e-05, + "loss": 0.0134, + "step": 15890 + }, + { + "epoch": 1.2349514563106796, + "grad_norm": 0.524823784828186, + "learning_rate": 9.67873802107557e-05, + "loss": 0.016, + "step": 15900 + }, + { + "epoch": 1.2357281553398058, + "grad_norm": 0.8536986112594604, + "learning_rate": 9.67815463537049e-05, + "loss": 0.0155, + "step": 15910 + }, + { + "epoch": 1.236504854368932, + "grad_norm": 0.28504180908203125, + "learning_rate": 9.677570738069457e-05, + "loss": 0.01, + "step": 15920 + }, + { + "epoch": 1.2372815533980583, + "grad_norm": 0.3387596607208252, + "learning_rate": 9.676986329236318e-05, + "loss": 0.0171, + "step": 15930 + }, + { + "epoch": 1.2380582524271844, + "grad_norm": 0.33118167519569397, + "learning_rate": 9.676401408934987e-05, + "loss": 0.0111, + "step": 15940 + }, + { + "epoch": 1.2388349514563106, + "grad_norm": 0.26824697852134705, + "learning_rate": 9.675815977229428e-05, + "loss": 0.013, + "step": 15950 + }, + { + "epoch": 1.2396116504854369, + "grad_norm": 0.2998082637786865, + "learning_rate": 9.675230034183664e-05, + "loss": 0.0167, + "step": 15960 + }, + { + "epoch": 1.2403883495145631, + "grad_norm": 0.252854585647583, + "learning_rate": 9.674643579861773e-05, + "loss": 0.0129, + "step": 15970 + }, + { + "epoch": 1.2411650485436894, + "grad_norm": 0.19224630296230316, + "learning_rate": 9.674056614327886e-05, + "loss": 0.0139, + "step": 15980 + }, + { + "epoch": 1.2419417475728156, + "grad_norm": 0.31879106163978577, + "learning_rate": 9.673469137646198e-05, + "loss": 0.0136, + "step": 15990 + }, + { + "epoch": 1.2427184466019416, + "grad_norm": 0.3033474385738373, + "learning_rate": 9.67288114988095e-05, + "loss": 0.0137, + "step": 16000 + }, + { + "epoch": 1.2434951456310679, + "grad_norm": 0.3685727119445801, + "learning_rate": 9.672292651096447e-05, + "loss": 0.0134, + "step": 16010 + }, + { + "epoch": 1.2442718446601941, + "grad_norm": 0.22624656558036804, + "learning_rate": 9.671703641357042e-05, + "loss": 0.0117, + "step": 16020 + }, + { + "epoch": 1.2450485436893204, + "grad_norm": 0.25759512186050415, + "learning_rate": 9.67111412072715e-05, + "loss": 0.0148, + "step": 16030 + }, + { + "epoch": 1.2458252427184466, + "grad_norm": 0.2118954211473465, + "learning_rate": 9.670524089271242e-05, + "loss": 0.0122, + "step": 16040 + }, + { + "epoch": 1.2466019417475729, + "grad_norm": 0.25817790627479553, + "learning_rate": 9.669933547053842e-05, + "loss": 0.0122, + "step": 16050 + }, + { + "epoch": 1.2473786407766991, + "grad_norm": 0.2787051498889923, + "learning_rate": 9.669342494139531e-05, + "loss": 0.0117, + "step": 16060 + }, + { + "epoch": 1.2481553398058252, + "grad_norm": 0.35278651118278503, + "learning_rate": 9.668750930592943e-05, + "loss": 0.0088, + "step": 16070 + }, + { + "epoch": 1.2489320388349514, + "grad_norm": 0.2938457131385803, + "learning_rate": 9.668158856478775e-05, + "loss": 0.0152, + "step": 16080 + }, + { + "epoch": 1.2497087378640777, + "grad_norm": 0.38019952178001404, + "learning_rate": 9.66756627186177e-05, + "loss": 0.0118, + "step": 16090 + }, + { + "epoch": 1.250485436893204, + "grad_norm": 0.2544981837272644, + "learning_rate": 9.666973176806737e-05, + "loss": 0.0126, + "step": 16100 + }, + { + "epoch": 1.2512621359223302, + "grad_norm": 0.35351553559303284, + "learning_rate": 9.666379571378534e-05, + "loss": 0.014, + "step": 16110 + }, + { + "epoch": 1.2520388349514562, + "grad_norm": 0.35684216022491455, + "learning_rate": 9.665785455642076e-05, + "loss": 0.0144, + "step": 16120 + }, + { + "epoch": 1.2528155339805824, + "grad_norm": 0.5292831063270569, + "learning_rate": 9.665190829662337e-05, + "loss": 0.0151, + "step": 16130 + }, + { + "epoch": 1.2535922330097087, + "grad_norm": 0.26493269205093384, + "learning_rate": 9.664595693504342e-05, + "loss": 0.0138, + "step": 16140 + }, + { + "epoch": 1.254368932038835, + "grad_norm": 0.5126725435256958, + "learning_rate": 9.664000047233175e-05, + "loss": 0.0132, + "step": 16150 + }, + { + "epoch": 1.2551456310679612, + "grad_norm": 0.30687662959098816, + "learning_rate": 9.663403890913976e-05, + "loss": 0.0152, + "step": 16160 + }, + { + "epoch": 1.2559223300970874, + "grad_norm": 0.5758340358734131, + "learning_rate": 9.662807224611938e-05, + "loss": 0.0118, + "step": 16170 + }, + { + "epoch": 1.2566990291262137, + "grad_norm": 0.21270795166492462, + "learning_rate": 9.662210048392311e-05, + "loss": 0.0126, + "step": 16180 + }, + { + "epoch": 1.25747572815534, + "grad_norm": 0.2931859493255615, + "learning_rate": 9.661612362320405e-05, + "loss": 0.011, + "step": 16190 + }, + { + "epoch": 1.258252427184466, + "grad_norm": 0.498800128698349, + "learning_rate": 9.661014166461579e-05, + "loss": 0.0117, + "step": 16200 + }, + { + "epoch": 1.2590291262135922, + "grad_norm": 0.25776755809783936, + "learning_rate": 9.66041546088125e-05, + "loss": 0.0106, + "step": 16210 + }, + { + "epoch": 1.2598058252427184, + "grad_norm": 0.5057174563407898, + "learning_rate": 9.659816245644895e-05, + "loss": 0.0096, + "step": 16220 + }, + { + "epoch": 1.2605825242718447, + "grad_norm": 0.4290523827075958, + "learning_rate": 9.65921652081804e-05, + "loss": 0.0163, + "step": 16230 + }, + { + "epoch": 1.261359223300971, + "grad_norm": 0.44085320830345154, + "learning_rate": 9.658616286466271e-05, + "loss": 0.014, + "step": 16240 + }, + { + "epoch": 1.262135922330097, + "grad_norm": 0.40593260526657104, + "learning_rate": 9.65801554265523e-05, + "loss": 0.014, + "step": 16250 + }, + { + "epoch": 1.2629126213592232, + "grad_norm": 0.26893752813339233, + "learning_rate": 9.657414289450612e-05, + "loss": 0.0114, + "step": 16260 + }, + { + "epoch": 1.2636893203883495, + "grad_norm": 0.28931090235710144, + "learning_rate": 9.656812526918171e-05, + "loss": 0.0119, + "step": 16270 + }, + { + "epoch": 1.2644660194174757, + "grad_norm": 0.3469919264316559, + "learning_rate": 9.656210255123712e-05, + "loss": 0.0143, + "step": 16280 + }, + { + "epoch": 1.265242718446602, + "grad_norm": 0.19444473087787628, + "learning_rate": 9.6556074741331e-05, + "loss": 0.0126, + "step": 16290 + }, + { + "epoch": 1.2660194174757282, + "grad_norm": 0.26679372787475586, + "learning_rate": 9.655004184012256e-05, + "loss": 0.0131, + "step": 16300 + }, + { + "epoch": 1.2667961165048545, + "grad_norm": 0.2583162188529968, + "learning_rate": 9.654400384827152e-05, + "loss": 0.0119, + "step": 16310 + }, + { + "epoch": 1.2675728155339807, + "grad_norm": 0.34779319167137146, + "learning_rate": 9.653796076643818e-05, + "loss": 0.011, + "step": 16320 + }, + { + "epoch": 1.2683495145631067, + "grad_norm": 0.40716928243637085, + "learning_rate": 9.653191259528344e-05, + "loss": 0.0183, + "step": 16330 + }, + { + "epoch": 1.269126213592233, + "grad_norm": 0.23570798337459564, + "learning_rate": 9.65258593354687e-05, + "loss": 0.0108, + "step": 16340 + }, + { + "epoch": 1.2699029126213592, + "grad_norm": 0.4132896661758423, + "learning_rate": 9.651980098765591e-05, + "loss": 0.0103, + "step": 16350 + }, + { + "epoch": 1.2706796116504855, + "grad_norm": 0.30625373125076294, + "learning_rate": 9.651373755250765e-05, + "loss": 0.0129, + "step": 16360 + }, + { + "epoch": 1.2714563106796115, + "grad_norm": 0.243314728140831, + "learning_rate": 9.650766903068697e-05, + "loss": 0.0118, + "step": 16370 + }, + { + "epoch": 1.2722330097087378, + "grad_norm": 0.2681092917919159, + "learning_rate": 9.650159542285753e-05, + "loss": 0.0158, + "step": 16380 + }, + { + "epoch": 1.273009708737864, + "grad_norm": 0.6243059635162354, + "learning_rate": 9.649551672968353e-05, + "loss": 0.0147, + "step": 16390 + }, + { + "epoch": 1.2737864077669903, + "grad_norm": 0.3584565818309784, + "learning_rate": 9.648943295182973e-05, + "loss": 0.0134, + "step": 16400 + }, + { + "epoch": 1.2745631067961165, + "grad_norm": 0.23915264010429382, + "learning_rate": 9.648334408996144e-05, + "loss": 0.0163, + "step": 16410 + }, + { + "epoch": 1.2753398058252428, + "grad_norm": 0.3979370594024658, + "learning_rate": 9.647725014474452e-05, + "loss": 0.0119, + "step": 16420 + }, + { + "epoch": 1.276116504854369, + "grad_norm": 0.2899194061756134, + "learning_rate": 9.64711511168454e-05, + "loss": 0.0156, + "step": 16430 + }, + { + "epoch": 1.2768932038834953, + "grad_norm": 0.29548490047454834, + "learning_rate": 9.646504700693108e-05, + "loss": 0.0117, + "step": 16440 + }, + { + "epoch": 1.2776699029126213, + "grad_norm": 0.24523238837718964, + "learning_rate": 9.645893781566907e-05, + "loss": 0.011, + "step": 16450 + }, + { + "epoch": 1.2784466019417475, + "grad_norm": 0.2742880582809448, + "learning_rate": 9.645282354372744e-05, + "loss": 0.012, + "step": 16460 + }, + { + "epoch": 1.2792233009708738, + "grad_norm": 0.25801795721054077, + "learning_rate": 9.644670419177491e-05, + "loss": 0.0127, + "step": 16470 + }, + { + "epoch": 1.28, + "grad_norm": 0.22577832639217377, + "learning_rate": 9.644057976048062e-05, + "loss": 0.0131, + "step": 16480 + }, + { + "epoch": 1.2807766990291263, + "grad_norm": 0.2628299295902252, + "learning_rate": 9.643445025051435e-05, + "loss": 0.0133, + "step": 16490 + }, + { + "epoch": 1.2815533980582523, + "grad_norm": 0.3690747618675232, + "learning_rate": 9.642831566254641e-05, + "loss": 0.0116, + "step": 16500 + }, + { + "epoch": 1.2823300970873786, + "grad_norm": 0.3870394229888916, + "learning_rate": 9.642217599724769e-05, + "loss": 0.0132, + "step": 16510 + }, + { + "epoch": 1.2831067961165048, + "grad_norm": 0.4252617359161377, + "learning_rate": 9.64160312552896e-05, + "loss": 0.0125, + "step": 16520 + }, + { + "epoch": 1.283883495145631, + "grad_norm": 0.300498366355896, + "learning_rate": 9.64098814373441e-05, + "loss": 0.0139, + "step": 16530 + }, + { + "epoch": 1.2846601941747573, + "grad_norm": 0.23946385085582733, + "learning_rate": 9.640372654408374e-05, + "loss": 0.0136, + "step": 16540 + }, + { + "epoch": 1.2854368932038835, + "grad_norm": 0.3570847809314728, + "learning_rate": 9.639756657618162e-05, + "loss": 0.0115, + "step": 16550 + }, + { + "epoch": 1.2862135922330098, + "grad_norm": 0.32873108983039856, + "learning_rate": 9.639140153431138e-05, + "loss": 0.0132, + "step": 16560 + }, + { + "epoch": 1.286990291262136, + "grad_norm": 0.27998560667037964, + "learning_rate": 9.638523141914721e-05, + "loss": 0.0111, + "step": 16570 + }, + { + "epoch": 1.287766990291262, + "grad_norm": 0.3845936357975006, + "learning_rate": 9.637905623136388e-05, + "loss": 0.0127, + "step": 16580 + }, + { + "epoch": 1.2885436893203883, + "grad_norm": 0.27228227257728577, + "learning_rate": 9.637287597163669e-05, + "loss": 0.0116, + "step": 16590 + }, + { + "epoch": 1.2893203883495146, + "grad_norm": 0.2395179718732834, + "learning_rate": 9.63666906406415e-05, + "loss": 0.0117, + "step": 16600 + }, + { + "epoch": 1.2900970873786408, + "grad_norm": 0.3149220645427704, + "learning_rate": 9.636050023905473e-05, + "loss": 0.0129, + "step": 16610 + }, + { + "epoch": 1.290873786407767, + "grad_norm": 0.3167288601398468, + "learning_rate": 9.635430476755336e-05, + "loss": 0.0128, + "step": 16620 + }, + { + "epoch": 1.291650485436893, + "grad_norm": 0.4750005900859833, + "learning_rate": 9.63481042268149e-05, + "loss": 0.0121, + "step": 16630 + }, + { + "epoch": 1.2924271844660193, + "grad_norm": 0.2763836085796356, + "learning_rate": 9.634189861751745e-05, + "loss": 0.0159, + "step": 16640 + }, + { + "epoch": 1.2932038834951456, + "grad_norm": 0.23258553445339203, + "learning_rate": 9.633568794033967e-05, + "loss": 0.0123, + "step": 16650 + }, + { + "epoch": 1.2939805825242718, + "grad_norm": 0.21467064321041107, + "learning_rate": 9.63294721959607e-05, + "loss": 0.0106, + "step": 16660 + }, + { + "epoch": 1.294757281553398, + "grad_norm": 0.3415873646736145, + "learning_rate": 9.63232513850603e-05, + "loss": 0.0128, + "step": 16670 + }, + { + "epoch": 1.2955339805825243, + "grad_norm": 0.207296684384346, + "learning_rate": 9.631702550831878e-05, + "loss": 0.0121, + "step": 16680 + }, + { + "epoch": 1.2963106796116506, + "grad_norm": 0.4387165307998657, + "learning_rate": 9.631079456641698e-05, + "loss": 0.0123, + "step": 16690 + }, + { + "epoch": 1.2970873786407768, + "grad_norm": 0.5301819443702698, + "learning_rate": 9.630455856003632e-05, + "loss": 0.0114, + "step": 16700 + }, + { + "epoch": 1.2978640776699029, + "grad_norm": 0.3848535418510437, + "learning_rate": 9.629831748985876e-05, + "loss": 0.0137, + "step": 16710 + }, + { + "epoch": 1.298640776699029, + "grad_norm": 0.22012563049793243, + "learning_rate": 9.629207135656679e-05, + "loss": 0.013, + "step": 16720 + }, + { + "epoch": 1.2994174757281554, + "grad_norm": 0.21784836053848267, + "learning_rate": 9.628582016084353e-05, + "loss": 0.0098, + "step": 16730 + }, + { + "epoch": 1.3001941747572816, + "grad_norm": 0.261178195476532, + "learning_rate": 9.627956390337254e-05, + "loss": 0.0131, + "step": 16740 + }, + { + "epoch": 1.3009708737864076, + "grad_norm": 0.20291242003440857, + "learning_rate": 9.627330258483802e-05, + "loss": 0.0115, + "step": 16750 + }, + { + "epoch": 1.3017475728155339, + "grad_norm": 0.22305797040462494, + "learning_rate": 9.62670362059247e-05, + "loss": 0.0115, + "step": 16760 + }, + { + "epoch": 1.3025242718446601, + "grad_norm": 0.3010100722312927, + "learning_rate": 9.626076476731786e-05, + "loss": 0.0134, + "step": 16770 + }, + { + "epoch": 1.3033009708737864, + "grad_norm": 0.26566341519355774, + "learning_rate": 9.625448826970336e-05, + "loss": 0.013, + "step": 16780 + }, + { + "epoch": 1.3040776699029126, + "grad_norm": 0.3389512002468109, + "learning_rate": 9.624820671376755e-05, + "loss": 0.0134, + "step": 16790 + }, + { + "epoch": 1.3048543689320389, + "grad_norm": 0.18131853640079498, + "learning_rate": 9.62419201001974e-05, + "loss": 0.0128, + "step": 16800 + }, + { + "epoch": 1.3056310679611651, + "grad_norm": 0.2486528754234314, + "learning_rate": 9.623562842968037e-05, + "loss": 0.014, + "step": 16810 + }, + { + "epoch": 1.3064077669902914, + "grad_norm": 0.35210949182510376, + "learning_rate": 9.622933170290454e-05, + "loss": 0.0121, + "step": 16820 + }, + { + "epoch": 1.3071844660194174, + "grad_norm": 0.25938862562179565, + "learning_rate": 9.622302992055849e-05, + "loss": 0.0134, + "step": 16830 + }, + { + "epoch": 1.3079611650485437, + "grad_norm": 0.2811667025089264, + "learning_rate": 9.62167230833314e-05, + "loss": 0.0132, + "step": 16840 + }, + { + "epoch": 1.30873786407767, + "grad_norm": 0.29263269901275635, + "learning_rate": 9.621041119191295e-05, + "loss": 0.0105, + "step": 16850 + }, + { + "epoch": 1.3095145631067961, + "grad_norm": 0.45471084117889404, + "learning_rate": 9.620409424699342e-05, + "loss": 0.0136, + "step": 16860 + }, + { + "epoch": 1.3102912621359224, + "grad_norm": 0.29286515712738037, + "learning_rate": 9.619777224926359e-05, + "loss": 0.0095, + "step": 16870 + }, + { + "epoch": 1.3110679611650484, + "grad_norm": 0.2917861342430115, + "learning_rate": 9.619144519941485e-05, + "loss": 0.0113, + "step": 16880 + }, + { + "epoch": 1.3118446601941747, + "grad_norm": 0.28571778535842896, + "learning_rate": 9.618511309813912e-05, + "loss": 0.0123, + "step": 16890 + }, + { + "epoch": 1.312621359223301, + "grad_norm": 0.25790974497795105, + "learning_rate": 9.617877594612886e-05, + "loss": 0.0129, + "step": 16900 + }, + { + "epoch": 1.3133980582524272, + "grad_norm": 0.314420610666275, + "learning_rate": 9.617243374407707e-05, + "loss": 0.0134, + "step": 16910 + }, + { + "epoch": 1.3141747572815534, + "grad_norm": 0.3860366642475128, + "learning_rate": 9.616608649267736e-05, + "loss": 0.0128, + "step": 16920 + }, + { + "epoch": 1.3149514563106797, + "grad_norm": 0.18914934992790222, + "learning_rate": 9.615973419262385e-05, + "loss": 0.009, + "step": 16930 + }, + { + "epoch": 1.315728155339806, + "grad_norm": 0.3335777521133423, + "learning_rate": 9.615337684461119e-05, + "loss": 0.0117, + "step": 16940 + }, + { + "epoch": 1.3165048543689322, + "grad_norm": 0.21269653737545013, + "learning_rate": 9.614701444933465e-05, + "loss": 0.0122, + "step": 16950 + }, + { + "epoch": 1.3172815533980582, + "grad_norm": 0.21181142330169678, + "learning_rate": 9.614064700748997e-05, + "loss": 0.0136, + "step": 16960 + }, + { + "epoch": 1.3180582524271844, + "grad_norm": 0.2777940332889557, + "learning_rate": 9.613427451977352e-05, + "loss": 0.0141, + "step": 16970 + }, + { + "epoch": 1.3188349514563107, + "grad_norm": 0.27617326378822327, + "learning_rate": 9.612789698688216e-05, + "loss": 0.0114, + "step": 16980 + }, + { + "epoch": 1.319611650485437, + "grad_norm": 0.2889269292354584, + "learning_rate": 9.612151440951334e-05, + "loss": 0.0175, + "step": 16990 + }, + { + "epoch": 1.3203883495145632, + "grad_norm": 0.3562390208244324, + "learning_rate": 9.611512678836506e-05, + "loss": 0.0138, + "step": 17000 + }, + { + "epoch": 1.3211650485436892, + "grad_norm": 0.21606408059597015, + "learning_rate": 9.610873412413584e-05, + "loss": 0.0135, + "step": 17010 + }, + { + "epoch": 1.3219417475728155, + "grad_norm": 0.36523348093032837, + "learning_rate": 9.610233641752476e-05, + "loss": 0.0138, + "step": 17020 + }, + { + "epoch": 1.3227184466019417, + "grad_norm": 0.3324788808822632, + "learning_rate": 9.609593366923151e-05, + "loss": 0.0125, + "step": 17030 + }, + { + "epoch": 1.323495145631068, + "grad_norm": 0.3549651503562927, + "learning_rate": 9.608952587995625e-05, + "loss": 0.0147, + "step": 17040 + }, + { + "epoch": 1.3242718446601942, + "grad_norm": 0.18249602615833282, + "learning_rate": 9.608311305039972e-05, + "loss": 0.0094, + "step": 17050 + }, + { + "epoch": 1.3250485436893205, + "grad_norm": 0.25750184059143066, + "learning_rate": 9.607669518126326e-05, + "loss": 0.0106, + "step": 17060 + }, + { + "epoch": 1.3258252427184467, + "grad_norm": 0.321380078792572, + "learning_rate": 9.607027227324866e-05, + "loss": 0.0111, + "step": 17070 + }, + { + "epoch": 1.3266019417475727, + "grad_norm": 0.29174795746803284, + "learning_rate": 9.606384432705837e-05, + "loss": 0.0113, + "step": 17080 + }, + { + "epoch": 1.327378640776699, + "grad_norm": 0.22092364728450775, + "learning_rate": 9.60574113433953e-05, + "loss": 0.0105, + "step": 17090 + }, + { + "epoch": 1.3281553398058252, + "grad_norm": 0.30832424759864807, + "learning_rate": 9.6050973322963e-05, + "loss": 0.0134, + "step": 17100 + }, + { + "epoch": 1.3289320388349515, + "grad_norm": 0.3014914393424988, + "learning_rate": 9.604453026646547e-05, + "loss": 0.0101, + "step": 17110 + }, + { + "epoch": 1.3297087378640777, + "grad_norm": 0.18783946335315704, + "learning_rate": 9.603808217460735e-05, + "loss": 0.01, + "step": 17120 + }, + { + "epoch": 1.3304854368932038, + "grad_norm": 0.3383469879627228, + "learning_rate": 9.603162904809377e-05, + "loss": 0.012, + "step": 17130 + }, + { + "epoch": 1.33126213592233, + "grad_norm": 0.22989879548549652, + "learning_rate": 9.602517088763045e-05, + "loss": 0.014, + "step": 17140 + }, + { + "epoch": 1.3320388349514563, + "grad_norm": 0.31693893671035767, + "learning_rate": 9.601870769392365e-05, + "loss": 0.0107, + "step": 17150 + }, + { + "epoch": 1.3328155339805825, + "grad_norm": 0.2531783878803253, + "learning_rate": 9.601223946768017e-05, + "loss": 0.0112, + "step": 17160 + }, + { + "epoch": 1.3335922330097087, + "grad_norm": 0.21306024491786957, + "learning_rate": 9.600576620960734e-05, + "loss": 0.0091, + "step": 17170 + }, + { + "epoch": 1.334368932038835, + "grad_norm": 0.20792968571186066, + "learning_rate": 9.599928792041308e-05, + "loss": 0.0112, + "step": 17180 + }, + { + "epoch": 1.3351456310679612, + "grad_norm": 0.2862526476383209, + "learning_rate": 9.599280460080587e-05, + "loss": 0.0094, + "step": 17190 + }, + { + "epoch": 1.3359223300970875, + "grad_norm": 0.5957197546958923, + "learning_rate": 9.59863162514947e-05, + "loss": 0.0118, + "step": 17200 + }, + { + "epoch": 1.3366990291262135, + "grad_norm": 0.33054134249687195, + "learning_rate": 9.597982287318911e-05, + "loss": 0.0139, + "step": 17210 + }, + { + "epoch": 1.3374757281553398, + "grad_norm": 0.24079830944538116, + "learning_rate": 9.597332446659923e-05, + "loss": 0.0137, + "step": 17220 + }, + { + "epoch": 1.338252427184466, + "grad_norm": 0.1410139501094818, + "learning_rate": 9.59668210324357e-05, + "loss": 0.0116, + "step": 17230 + }, + { + "epoch": 1.3390291262135923, + "grad_norm": 0.19044077396392822, + "learning_rate": 9.596031257140974e-05, + "loss": 0.0114, + "step": 17240 + }, + { + "epoch": 1.3398058252427185, + "grad_norm": 0.28314971923828125, + "learning_rate": 9.59537990842331e-05, + "loss": 0.0166, + "step": 17250 + }, + { + "epoch": 1.3405825242718445, + "grad_norm": 0.4838104844093323, + "learning_rate": 9.594728057161806e-05, + "loss": 0.0117, + "step": 17260 + }, + { + "epoch": 1.3413592233009708, + "grad_norm": 0.32112181186676025, + "learning_rate": 9.594075703427752e-05, + "loss": 0.014, + "step": 17270 + }, + { + "epoch": 1.342135922330097, + "grad_norm": 0.31610187888145447, + "learning_rate": 9.593422847292486e-05, + "loss": 0.013, + "step": 17280 + }, + { + "epoch": 1.3429126213592233, + "grad_norm": 0.3468317687511444, + "learning_rate": 9.592769488827402e-05, + "loss": 0.0116, + "step": 17290 + }, + { + "epoch": 1.3436893203883495, + "grad_norm": 0.4074651598930359, + "learning_rate": 9.592115628103952e-05, + "loss": 0.015, + "step": 17300 + }, + { + "epoch": 1.3444660194174758, + "grad_norm": 0.4004952907562256, + "learning_rate": 9.591461265193643e-05, + "loss": 0.0158, + "step": 17310 + }, + { + "epoch": 1.345242718446602, + "grad_norm": 0.2872414290904999, + "learning_rate": 9.590806400168032e-05, + "loss": 0.0092, + "step": 17320 + }, + { + "epoch": 1.3460194174757283, + "grad_norm": 0.19893237948417664, + "learning_rate": 9.590151033098735e-05, + "loss": 0.0136, + "step": 17330 + }, + { + "epoch": 1.3467961165048543, + "grad_norm": 0.46067142486572266, + "learning_rate": 9.589495164057423e-05, + "loss": 0.0118, + "step": 17340 + }, + { + "epoch": 1.3475728155339806, + "grad_norm": 0.2047572135925293, + "learning_rate": 9.58883879311582e-05, + "loss": 0.0098, + "step": 17350 + }, + { + "epoch": 1.3483495145631068, + "grad_norm": 0.41908085346221924, + "learning_rate": 9.588181920345705e-05, + "loss": 0.0149, + "step": 17360 + }, + { + "epoch": 1.349126213592233, + "grad_norm": 0.22518102824687958, + "learning_rate": 9.587524545818913e-05, + "loss": 0.0097, + "step": 17370 + }, + { + "epoch": 1.349902912621359, + "grad_norm": 0.22959692776203156, + "learning_rate": 9.586866669607335e-05, + "loss": 0.0116, + "step": 17380 + }, + { + "epoch": 1.3506796116504853, + "grad_norm": 0.2473696917295456, + "learning_rate": 9.586208291782915e-05, + "loss": 0.0131, + "step": 17390 + }, + { + "epoch": 1.3514563106796116, + "grad_norm": 0.2210330367088318, + "learning_rate": 9.58554941241765e-05, + "loss": 0.0103, + "step": 17400 + }, + { + "epoch": 1.3522330097087378, + "grad_norm": 0.3554847836494446, + "learning_rate": 9.584890031583596e-05, + "loss": 0.01, + "step": 17410 + }, + { + "epoch": 1.353009708737864, + "grad_norm": 0.46824175119400024, + "learning_rate": 9.584230149352861e-05, + "loss": 0.0129, + "step": 17420 + }, + { + "epoch": 1.3537864077669903, + "grad_norm": 0.24184659123420715, + "learning_rate": 9.58356976579761e-05, + "loss": 0.0151, + "step": 17430 + }, + { + "epoch": 1.3545631067961166, + "grad_norm": 0.26725590229034424, + "learning_rate": 9.58290888099006e-05, + "loss": 0.0138, + "step": 17440 + }, + { + "epoch": 1.3553398058252428, + "grad_norm": 0.2847893238067627, + "learning_rate": 9.582247495002486e-05, + "loss": 0.0106, + "step": 17450 + }, + { + "epoch": 1.3561165048543689, + "grad_norm": 0.18888720870018005, + "learning_rate": 9.581585607907214e-05, + "loss": 0.0107, + "step": 17460 + }, + { + "epoch": 1.356893203883495, + "grad_norm": 0.271089643239975, + "learning_rate": 9.580923219776628e-05, + "loss": 0.0159, + "step": 17470 + }, + { + "epoch": 1.3576699029126214, + "grad_norm": 0.2693699896335602, + "learning_rate": 9.580260330683167e-05, + "loss": 0.0129, + "step": 17480 + }, + { + "epoch": 1.3584466019417476, + "grad_norm": 0.19819103181362152, + "learning_rate": 9.579596940699322e-05, + "loss": 0.0096, + "step": 17490 + }, + { + "epoch": 1.3592233009708738, + "grad_norm": 0.2982953190803528, + "learning_rate": 9.578933049897643e-05, + "loss": 0.0108, + "step": 17500 + }, + { + "epoch": 1.3599999999999999, + "grad_norm": 0.25770947337150574, + "learning_rate": 9.578268658350728e-05, + "loss": 0.0111, + "step": 17510 + }, + { + "epoch": 1.3607766990291261, + "grad_norm": 0.3338336944580078, + "learning_rate": 9.577603766131235e-05, + "loss": 0.011, + "step": 17520 + }, + { + "epoch": 1.3615533980582524, + "grad_norm": 0.4688740670681, + "learning_rate": 9.576938373311878e-05, + "loss": 0.0142, + "step": 17530 + }, + { + "epoch": 1.3623300970873786, + "grad_norm": 0.39039453864097595, + "learning_rate": 9.576272479965421e-05, + "loss": 0.0134, + "step": 17540 + }, + { + "epoch": 1.3631067961165049, + "grad_norm": 0.4150950610637665, + "learning_rate": 9.575606086164687e-05, + "loss": 0.0125, + "step": 17550 + }, + { + "epoch": 1.3638834951456311, + "grad_norm": 0.26262128353118896, + "learning_rate": 9.57493919198255e-05, + "loss": 0.0126, + "step": 17560 + }, + { + "epoch": 1.3646601941747574, + "grad_norm": 0.2828134000301361, + "learning_rate": 9.57427179749194e-05, + "loss": 0.0119, + "step": 17570 + }, + { + "epoch": 1.3654368932038836, + "grad_norm": 0.2619640529155731, + "learning_rate": 9.573603902765846e-05, + "loss": 0.0138, + "step": 17580 + }, + { + "epoch": 1.3662135922330096, + "grad_norm": 0.25754356384277344, + "learning_rate": 9.572935507877304e-05, + "loss": 0.0114, + "step": 17590 + }, + { + "epoch": 1.366990291262136, + "grad_norm": 0.21066495776176453, + "learning_rate": 9.57226661289941e-05, + "loss": 0.0137, + "step": 17600 + }, + { + "epoch": 1.3677669902912621, + "grad_norm": 0.24101558327674866, + "learning_rate": 9.571597217905315e-05, + "loss": 0.0105, + "step": 17610 + }, + { + "epoch": 1.3685436893203884, + "grad_norm": 0.2622159719467163, + "learning_rate": 9.57092732296822e-05, + "loss": 0.0105, + "step": 17620 + }, + { + "epoch": 1.3693203883495146, + "grad_norm": 0.1695616990327835, + "learning_rate": 9.570256928161385e-05, + "loss": 0.0088, + "step": 17630 + }, + { + "epoch": 1.3700970873786407, + "grad_norm": 0.45360279083251953, + "learning_rate": 9.569586033558126e-05, + "loss": 0.0099, + "step": 17640 + }, + { + "epoch": 1.370873786407767, + "grad_norm": 0.2114020138978958, + "learning_rate": 9.568914639231807e-05, + "loss": 0.0089, + "step": 17650 + }, + { + "epoch": 1.3716504854368932, + "grad_norm": 0.264797568321228, + "learning_rate": 9.568242745255852e-05, + "loss": 0.0136, + "step": 17660 + }, + { + "epoch": 1.3724271844660194, + "grad_norm": 0.2783481776714325, + "learning_rate": 9.567570351703739e-05, + "loss": 0.0144, + "step": 17670 + }, + { + "epoch": 1.3732038834951457, + "grad_norm": 0.19512680172920227, + "learning_rate": 9.566897458649001e-05, + "loss": 0.011, + "step": 17680 + }, + { + "epoch": 1.373980582524272, + "grad_norm": 0.2426459938287735, + "learning_rate": 9.566224066165221e-05, + "loss": 0.0106, + "step": 17690 + }, + { + "epoch": 1.3747572815533982, + "grad_norm": 0.40676578879356384, + "learning_rate": 9.565550174326043e-05, + "loss": 0.0167, + "step": 17700 + }, + { + "epoch": 1.3755339805825244, + "grad_norm": 0.20101287961006165, + "learning_rate": 9.564875783205162e-05, + "loss": 0.0128, + "step": 17710 + }, + { + "epoch": 1.3763106796116504, + "grad_norm": 0.32368695735931396, + "learning_rate": 9.564200892876328e-05, + "loss": 0.0118, + "step": 17720 + }, + { + "epoch": 1.3770873786407767, + "grad_norm": 0.2951362133026123, + "learning_rate": 9.563525503413348e-05, + "loss": 0.0099, + "step": 17730 + }, + { + "epoch": 1.377864077669903, + "grad_norm": 0.3856777548789978, + "learning_rate": 9.562849614890079e-05, + "loss": 0.0137, + "step": 17740 + }, + { + "epoch": 1.3786407766990292, + "grad_norm": 0.3347018361091614, + "learning_rate": 9.562173227380436e-05, + "loss": 0.0101, + "step": 17750 + }, + { + "epoch": 1.3794174757281552, + "grad_norm": 0.3675691783428192, + "learning_rate": 9.561496340958389e-05, + "loss": 0.0083, + "step": 17760 + }, + { + "epoch": 1.3801941747572815, + "grad_norm": 0.45515140891075134, + "learning_rate": 9.560818955697959e-05, + "loss": 0.0138, + "step": 17770 + }, + { + "epoch": 1.3809708737864077, + "grad_norm": 0.3182057738304138, + "learning_rate": 9.560141071673228e-05, + "loss": 0.0112, + "step": 17780 + }, + { + "epoch": 1.381747572815534, + "grad_norm": 0.45654138922691345, + "learning_rate": 9.559462688958323e-05, + "loss": 0.0112, + "step": 17790 + }, + { + "epoch": 1.3825242718446602, + "grad_norm": 0.2588914930820465, + "learning_rate": 9.558783807627434e-05, + "loss": 0.0127, + "step": 17800 + }, + { + "epoch": 1.3833009708737865, + "grad_norm": 0.19072650372982025, + "learning_rate": 9.558104427754801e-05, + "loss": 0.0099, + "step": 17810 + }, + { + "epoch": 1.3840776699029127, + "grad_norm": 0.2844918668270111, + "learning_rate": 9.557424549414722e-05, + "loss": 0.011, + "step": 17820 + }, + { + "epoch": 1.384854368932039, + "grad_norm": 0.20842193067073822, + "learning_rate": 9.556744172681546e-05, + "loss": 0.0124, + "step": 17830 + }, + { + "epoch": 1.385631067961165, + "grad_norm": 0.26616305112838745, + "learning_rate": 9.556063297629677e-05, + "loss": 0.0146, + "step": 17840 + }, + { + "epoch": 1.3864077669902912, + "grad_norm": 0.30787667632102966, + "learning_rate": 9.555381924333578e-05, + "loss": 0.0187, + "step": 17850 + }, + { + "epoch": 1.3871844660194175, + "grad_norm": 0.2798207104206085, + "learning_rate": 9.554700052867758e-05, + "loss": 0.0106, + "step": 17860 + }, + { + "epoch": 1.3879611650485437, + "grad_norm": 0.3782559633255005, + "learning_rate": 9.554017683306789e-05, + "loss": 0.0114, + "step": 17870 + }, + { + "epoch": 1.38873786407767, + "grad_norm": 0.7127171158790588, + "learning_rate": 9.553334815725294e-05, + "loss": 0.0139, + "step": 17880 + }, + { + "epoch": 1.389514563106796, + "grad_norm": 0.2401258945465088, + "learning_rate": 9.552651450197949e-05, + "loss": 0.0117, + "step": 17890 + }, + { + "epoch": 1.3902912621359222, + "grad_norm": 0.3123614192008972, + "learning_rate": 9.551967586799486e-05, + "loss": 0.0134, + "step": 17900 + }, + { + "epoch": 1.3910679611650485, + "grad_norm": 0.16814544796943665, + "learning_rate": 9.551283225604692e-05, + "loss": 0.0092, + "step": 17910 + }, + { + "epoch": 1.3918446601941747, + "grad_norm": 0.22611701488494873, + "learning_rate": 9.550598366688406e-05, + "loss": 0.013, + "step": 17920 + }, + { + "epoch": 1.392621359223301, + "grad_norm": 0.3384135365486145, + "learning_rate": 9.549913010125526e-05, + "loss": 0.0146, + "step": 17930 + }, + { + "epoch": 1.3933980582524272, + "grad_norm": 0.29262775182724, + "learning_rate": 9.549227155990999e-05, + "loss": 0.012, + "step": 17940 + }, + { + "epoch": 1.3941747572815535, + "grad_norm": 0.39673563838005066, + "learning_rate": 9.548540804359828e-05, + "loss": 0.0119, + "step": 17950 + }, + { + "epoch": 1.3949514563106797, + "grad_norm": 0.3810230493545532, + "learning_rate": 9.547853955307077e-05, + "loss": 0.0113, + "step": 17960 + }, + { + "epoch": 1.3957281553398058, + "grad_norm": 0.2897545099258423, + "learning_rate": 9.547166608907853e-05, + "loss": 0.0182, + "step": 17970 + }, + { + "epoch": 1.396504854368932, + "grad_norm": 0.34435319900512695, + "learning_rate": 9.546478765237326e-05, + "loss": 0.0108, + "step": 17980 + }, + { + "epoch": 1.3972815533980583, + "grad_norm": 0.3460085988044739, + "learning_rate": 9.545790424370715e-05, + "loss": 0.0141, + "step": 17990 + }, + { + "epoch": 1.3980582524271845, + "grad_norm": 0.2393694967031479, + "learning_rate": 9.5451015863833e-05, + "loss": 0.0149, + "step": 18000 + }, + { + "epoch": 1.3988349514563108, + "grad_norm": 0.3756355047225952, + "learning_rate": 9.544412251350408e-05, + "loss": 0.0149, + "step": 18010 + }, + { + "epoch": 1.3996116504854368, + "grad_norm": 0.39321544766426086, + "learning_rate": 9.543722419347422e-05, + "loss": 0.0192, + "step": 18020 + }, + { + "epoch": 1.400388349514563, + "grad_norm": 0.2343810647726059, + "learning_rate": 9.543032090449788e-05, + "loss": 0.0159, + "step": 18030 + }, + { + "epoch": 1.4011650485436893, + "grad_norm": 0.22380079329013824, + "learning_rate": 9.542341264732992e-05, + "loss": 0.0136, + "step": 18040 + }, + { + "epoch": 1.4019417475728155, + "grad_norm": 0.236505925655365, + "learning_rate": 9.541649942272585e-05, + "loss": 0.0138, + "step": 18050 + }, + { + "epoch": 1.4027184466019418, + "grad_norm": 0.23313623666763306, + "learning_rate": 9.54095812314417e-05, + "loss": 0.0142, + "step": 18060 + }, + { + "epoch": 1.403495145631068, + "grad_norm": 0.28119489550590515, + "learning_rate": 9.540265807423401e-05, + "loss": 0.0124, + "step": 18070 + }, + { + "epoch": 1.4042718446601943, + "grad_norm": 0.2559293508529663, + "learning_rate": 9.53957299518599e-05, + "loss": 0.0136, + "step": 18080 + }, + { + "epoch": 1.4050485436893203, + "grad_norm": 0.24615149199962616, + "learning_rate": 9.5388796865077e-05, + "loss": 0.012, + "step": 18090 + }, + { + "epoch": 1.4058252427184466, + "grad_norm": 0.5290189981460571, + "learning_rate": 9.538185881464353e-05, + "loss": 0.0141, + "step": 18100 + }, + { + "epoch": 1.4066019417475728, + "grad_norm": 0.35321924090385437, + "learning_rate": 9.537491580131821e-05, + "loss": 0.0143, + "step": 18110 + }, + { + "epoch": 1.407378640776699, + "grad_norm": 0.24332265555858612, + "learning_rate": 9.53679678258603e-05, + "loss": 0.0121, + "step": 18120 + }, + { + "epoch": 1.4081553398058253, + "grad_norm": 0.5436052083969116, + "learning_rate": 9.536101488902966e-05, + "loss": 0.0133, + "step": 18130 + }, + { + "epoch": 1.4089320388349513, + "grad_norm": 0.3638934791088104, + "learning_rate": 9.535405699158663e-05, + "loss": 0.0167, + "step": 18140 + }, + { + "epoch": 1.4097087378640776, + "grad_norm": 0.2575422525405884, + "learning_rate": 9.53470941342921e-05, + "loss": 0.0144, + "step": 18150 + }, + { + "epoch": 1.4104854368932038, + "grad_norm": 0.21422536671161652, + "learning_rate": 9.534012631790756e-05, + "loss": 0.0122, + "step": 18160 + }, + { + "epoch": 1.41126213592233, + "grad_norm": 0.21065615117549896, + "learning_rate": 9.533315354319494e-05, + "loss": 0.0124, + "step": 18170 + }, + { + "epoch": 1.4120388349514563, + "grad_norm": 0.2345971018075943, + "learning_rate": 9.532617581091682e-05, + "loss": 0.0135, + "step": 18180 + }, + { + "epoch": 1.4128155339805826, + "grad_norm": 0.2308947890996933, + "learning_rate": 9.531919312183629e-05, + "loss": 0.0117, + "step": 18190 + }, + { + "epoch": 1.4135922330097088, + "grad_norm": 0.9184838533401489, + "learning_rate": 9.531220547671688e-05, + "loss": 0.0121, + "step": 18200 + }, + { + "epoch": 1.414368932038835, + "grad_norm": 0.17526127398014069, + "learning_rate": 9.530521287632285e-05, + "loss": 0.0107, + "step": 18210 + }, + { + "epoch": 1.415145631067961, + "grad_norm": 0.3254309594631195, + "learning_rate": 9.529821532141884e-05, + "loss": 0.0133, + "step": 18220 + }, + { + "epoch": 1.4159223300970873, + "grad_norm": 0.49275389313697815, + "learning_rate": 9.52912128127701e-05, + "loss": 0.0146, + "step": 18230 + }, + { + "epoch": 1.4166990291262136, + "grad_norm": 0.37840765714645386, + "learning_rate": 9.528420535114244e-05, + "loss": 0.0162, + "step": 18240 + }, + { + "epoch": 1.4174757281553398, + "grad_norm": 0.29764673113822937, + "learning_rate": 9.527719293730215e-05, + "loss": 0.0169, + "step": 18250 + }, + { + "epoch": 1.418252427184466, + "grad_norm": 0.18950864672660828, + "learning_rate": 9.527017557201611e-05, + "loss": 0.0116, + "step": 18260 + }, + { + "epoch": 1.4190291262135921, + "grad_norm": 0.30923089385032654, + "learning_rate": 9.526315325605176e-05, + "loss": 0.0144, + "step": 18270 + }, + { + "epoch": 1.4198058252427184, + "grad_norm": 0.2264733463525772, + "learning_rate": 9.525612599017699e-05, + "loss": 0.0115, + "step": 18280 + }, + { + "epoch": 1.4205825242718446, + "grad_norm": 0.3518330156803131, + "learning_rate": 9.524909377516033e-05, + "loss": 0.015, + "step": 18290 + }, + { + "epoch": 1.4213592233009709, + "grad_norm": 0.3086264431476593, + "learning_rate": 9.524205661177081e-05, + "loss": 0.0124, + "step": 18300 + }, + { + "epoch": 1.4221359223300971, + "grad_norm": 0.3513212203979492, + "learning_rate": 9.523501450077801e-05, + "loss": 0.0145, + "step": 18310 + }, + { + "epoch": 1.4229126213592234, + "grad_norm": 0.1868438720703125, + "learning_rate": 9.522796744295202e-05, + "loss": 0.0119, + "step": 18320 + }, + { + "epoch": 1.4236893203883496, + "grad_norm": 0.34576186537742615, + "learning_rate": 9.522091543906352e-05, + "loss": 0.0125, + "step": 18330 + }, + { + "epoch": 1.4244660194174759, + "grad_norm": 0.35369086265563965, + "learning_rate": 9.521385848988369e-05, + "loss": 0.0117, + "step": 18340 + }, + { + "epoch": 1.4252427184466019, + "grad_norm": 0.552251935005188, + "learning_rate": 9.520679659618428e-05, + "loss": 0.0115, + "step": 18350 + }, + { + "epoch": 1.4260194174757281, + "grad_norm": 0.21475984156131744, + "learning_rate": 9.519972975873754e-05, + "loss": 0.0123, + "step": 18360 + }, + { + "epoch": 1.4267961165048544, + "grad_norm": 0.3105992078781128, + "learning_rate": 9.519265797831633e-05, + "loss": 0.0109, + "step": 18370 + }, + { + "epoch": 1.4275728155339806, + "grad_norm": 0.23757322132587433, + "learning_rate": 9.518558125569399e-05, + "loss": 0.0101, + "step": 18380 + }, + { + "epoch": 1.4283495145631067, + "grad_norm": 0.28191396594047546, + "learning_rate": 9.517849959164442e-05, + "loss": 0.0126, + "step": 18390 + }, + { + "epoch": 1.429126213592233, + "grad_norm": 0.399008184671402, + "learning_rate": 9.517141298694205e-05, + "loss": 0.013, + "step": 18400 + }, + { + "epoch": 1.4299029126213592, + "grad_norm": 0.17706167697906494, + "learning_rate": 9.516432144236188e-05, + "loss": 0.0136, + "step": 18410 + }, + { + "epoch": 1.4306796116504854, + "grad_norm": 0.1972285658121109, + "learning_rate": 9.515722495867941e-05, + "loss": 0.0111, + "step": 18420 + }, + { + "epoch": 1.4314563106796117, + "grad_norm": 0.17690810561180115, + "learning_rate": 9.515012353667072e-05, + "loss": 0.0151, + "step": 18430 + }, + { + "epoch": 1.432233009708738, + "grad_norm": 0.15552842617034912, + "learning_rate": 9.51430171771124e-05, + "loss": 0.0111, + "step": 18440 + }, + { + "epoch": 1.4330097087378642, + "grad_norm": 0.2949543297290802, + "learning_rate": 9.513590588078159e-05, + "loss": 0.0122, + "step": 18450 + }, + { + "epoch": 1.4337864077669904, + "grad_norm": 0.3432469069957733, + "learning_rate": 9.512878964845597e-05, + "loss": 0.0145, + "step": 18460 + }, + { + "epoch": 1.4345631067961164, + "grad_norm": 0.2799486517906189, + "learning_rate": 9.512166848091377e-05, + "loss": 0.0102, + "step": 18470 + }, + { + "epoch": 1.4353398058252427, + "grad_norm": 0.3065233528614044, + "learning_rate": 9.511454237893376e-05, + "loss": 0.0133, + "step": 18480 + }, + { + "epoch": 1.436116504854369, + "grad_norm": 0.22135722637176514, + "learning_rate": 9.51074113432952e-05, + "loss": 0.0119, + "step": 18490 + }, + { + "epoch": 1.4368932038834952, + "grad_norm": 0.21748751401901245, + "learning_rate": 9.510027537477797e-05, + "loss": 0.0118, + "step": 18500 + }, + { + "epoch": 1.4376699029126214, + "grad_norm": 0.2891216576099396, + "learning_rate": 9.509313447416242e-05, + "loss": 0.0141, + "step": 18510 + }, + { + "epoch": 1.4384466019417474, + "grad_norm": 0.21960972249507904, + "learning_rate": 9.508598864222949e-05, + "loss": 0.0126, + "step": 18520 + }, + { + "epoch": 1.4392233009708737, + "grad_norm": 0.18387819826602936, + "learning_rate": 9.507883787976062e-05, + "loss": 0.0111, + "step": 18530 + }, + { + "epoch": 1.44, + "grad_norm": 0.5515927076339722, + "learning_rate": 9.507168218753781e-05, + "loss": 0.0149, + "step": 18540 + }, + { + "epoch": 1.4407766990291262, + "grad_norm": 0.4787263572216034, + "learning_rate": 9.506452156634362e-05, + "loss": 0.0113, + "step": 18550 + }, + { + "epoch": 1.4415533980582524, + "grad_norm": 0.30556395649909973, + "learning_rate": 9.505735601696109e-05, + "loss": 0.0129, + "step": 18560 + }, + { + "epoch": 1.4423300970873787, + "grad_norm": 0.341556578874588, + "learning_rate": 9.505018554017385e-05, + "loss": 0.0117, + "step": 18570 + }, + { + "epoch": 1.443106796116505, + "grad_norm": 0.22390629351139069, + "learning_rate": 9.504301013676604e-05, + "loss": 0.0094, + "step": 18580 + }, + { + "epoch": 1.4438834951456312, + "grad_norm": 0.5636851787567139, + "learning_rate": 9.503582980752238e-05, + "loss": 0.0111, + "step": 18590 + }, + { + "epoch": 1.4446601941747572, + "grad_norm": 0.3945988416671753, + "learning_rate": 9.502864455322809e-05, + "loss": 0.0129, + "step": 18600 + }, + { + "epoch": 1.4454368932038835, + "grad_norm": 0.3951624929904938, + "learning_rate": 9.502145437466891e-05, + "loss": 0.0119, + "step": 18610 + }, + { + "epoch": 1.4462135922330097, + "grad_norm": 0.2562118172645569, + "learning_rate": 9.501425927263116e-05, + "loss": 0.0107, + "step": 18620 + }, + { + "epoch": 1.446990291262136, + "grad_norm": 0.2809574604034424, + "learning_rate": 9.500705924790172e-05, + "loss": 0.0128, + "step": 18630 + }, + { + "epoch": 1.4477669902912622, + "grad_norm": 0.26918259263038635, + "learning_rate": 9.499985430126794e-05, + "loss": 0.0117, + "step": 18640 + }, + { + "epoch": 1.4485436893203882, + "grad_norm": 0.1682724803686142, + "learning_rate": 9.499264443351775e-05, + "loss": 0.0114, + "step": 18650 + }, + { + "epoch": 1.4493203883495145, + "grad_norm": 0.1608191430568695, + "learning_rate": 9.498542964543961e-05, + "loss": 0.01, + "step": 18660 + }, + { + "epoch": 1.4500970873786407, + "grad_norm": 0.267206609249115, + "learning_rate": 9.497820993782252e-05, + "loss": 0.0156, + "step": 18670 + }, + { + "epoch": 1.450873786407767, + "grad_norm": 0.23219725489616394, + "learning_rate": 9.497098531145601e-05, + "loss": 0.0112, + "step": 18680 + }, + { + "epoch": 1.4516504854368932, + "grad_norm": 0.4516846537590027, + "learning_rate": 9.496375576713017e-05, + "loss": 0.0125, + "step": 18690 + }, + { + "epoch": 1.4524271844660195, + "grad_norm": 0.33930641412734985, + "learning_rate": 9.49565213056356e-05, + "loss": 0.0142, + "step": 18700 + }, + { + "epoch": 1.4532038834951457, + "grad_norm": 0.38303425908088684, + "learning_rate": 9.494928192776342e-05, + "loss": 0.0161, + "step": 18710 + }, + { + "epoch": 1.453980582524272, + "grad_norm": 0.7267704010009766, + "learning_rate": 9.494203763430538e-05, + "loss": 0.0121, + "step": 18720 + }, + { + "epoch": 1.454757281553398, + "grad_norm": 0.32614341378211975, + "learning_rate": 9.493478842605366e-05, + "loss": 0.0125, + "step": 18730 + }, + { + "epoch": 1.4555339805825243, + "grad_norm": 0.23987767100334167, + "learning_rate": 9.492753430380105e-05, + "loss": 0.0128, + "step": 18740 + }, + { + "epoch": 1.4563106796116505, + "grad_norm": 0.31185346841812134, + "learning_rate": 9.492027526834083e-05, + "loss": 0.01, + "step": 18750 + }, + { + "epoch": 1.4570873786407768, + "grad_norm": 0.24933095276355743, + "learning_rate": 9.491301132046684e-05, + "loss": 0.0126, + "step": 18760 + }, + { + "epoch": 1.4578640776699028, + "grad_norm": 0.25532710552215576, + "learning_rate": 9.490574246097345e-05, + "loss": 0.0164, + "step": 18770 + }, + { + "epoch": 1.458640776699029, + "grad_norm": 0.23153892159461975, + "learning_rate": 9.48984686906556e-05, + "loss": 0.0095, + "step": 18780 + }, + { + "epoch": 1.4594174757281553, + "grad_norm": 0.2966865301132202, + "learning_rate": 9.489119001030871e-05, + "loss": 0.0122, + "step": 18790 + }, + { + "epoch": 1.4601941747572815, + "grad_norm": 0.2751288414001465, + "learning_rate": 9.488390642072878e-05, + "loss": 0.0088, + "step": 18800 + }, + { + "epoch": 1.4609708737864078, + "grad_norm": 0.2758534848690033, + "learning_rate": 9.48766179227123e-05, + "loss": 0.01, + "step": 18810 + }, + { + "epoch": 1.461747572815534, + "grad_norm": 0.23475630581378937, + "learning_rate": 9.486932451705636e-05, + "loss": 0.0158, + "step": 18820 + }, + { + "epoch": 1.4625242718446603, + "grad_norm": 0.41348567605018616, + "learning_rate": 9.486202620455857e-05, + "loss": 0.0114, + "step": 18830 + }, + { + "epoch": 1.4633009708737865, + "grad_norm": 0.31762513518333435, + "learning_rate": 9.485472298601704e-05, + "loss": 0.0093, + "step": 18840 + }, + { + "epoch": 1.4640776699029125, + "grad_norm": 0.37657129764556885, + "learning_rate": 9.484741486223043e-05, + "loss": 0.0126, + "step": 18850 + }, + { + "epoch": 1.4648543689320388, + "grad_norm": 0.22242870926856995, + "learning_rate": 9.484010183399797e-05, + "loss": 0.0116, + "step": 18860 + }, + { + "epoch": 1.465631067961165, + "grad_norm": 0.26894357800483704, + "learning_rate": 9.483278390211938e-05, + "loss": 0.0101, + "step": 18870 + }, + { + "epoch": 1.4664077669902913, + "grad_norm": 0.2502460777759552, + "learning_rate": 9.482546106739496e-05, + "loss": 0.0123, + "step": 18880 + }, + { + "epoch": 1.4671844660194175, + "grad_norm": 0.1992960274219513, + "learning_rate": 9.48181333306255e-05, + "loss": 0.012, + "step": 18890 + }, + { + "epoch": 1.4679611650485436, + "grad_norm": 0.17386280000209808, + "learning_rate": 9.481080069261237e-05, + "loss": 0.0147, + "step": 18900 + }, + { + "epoch": 1.4687378640776698, + "grad_norm": 0.2815113663673401, + "learning_rate": 9.480346315415745e-05, + "loss": 0.0119, + "step": 18910 + }, + { + "epoch": 1.469514563106796, + "grad_norm": 0.25870564579963684, + "learning_rate": 9.479612071606314e-05, + "loss": 0.0123, + "step": 18920 + }, + { + "epoch": 1.4702912621359223, + "grad_norm": 0.21534980833530426, + "learning_rate": 9.478877337913244e-05, + "loss": 0.0122, + "step": 18930 + }, + { + "epoch": 1.4710679611650486, + "grad_norm": 0.23489801585674286, + "learning_rate": 9.478142114416881e-05, + "loss": 0.0155, + "step": 18940 + }, + { + "epoch": 1.4718446601941748, + "grad_norm": 0.3154030740261078, + "learning_rate": 9.47740640119763e-05, + "loss": 0.0107, + "step": 18950 + }, + { + "epoch": 1.472621359223301, + "grad_norm": 0.1837935447692871, + "learning_rate": 9.476670198335947e-05, + "loss": 0.0119, + "step": 18960 + }, + { + "epoch": 1.4733980582524273, + "grad_norm": 0.26722967624664307, + "learning_rate": 9.47593350591234e-05, + "loss": 0.0093, + "step": 18970 + }, + { + "epoch": 1.4741747572815533, + "grad_norm": 0.18402956426143646, + "learning_rate": 9.475196324007376e-05, + "loss": 0.0117, + "step": 18980 + }, + { + "epoch": 1.4749514563106796, + "grad_norm": 0.19912244379520416, + "learning_rate": 9.474458652701669e-05, + "loss": 0.0087, + "step": 18990 + }, + { + "epoch": 1.4757281553398058, + "grad_norm": 0.1831018477678299, + "learning_rate": 9.473720492075892e-05, + "loss": 0.01, + "step": 19000 + }, + { + "epoch": 1.476504854368932, + "grad_norm": 0.21680769324302673, + "learning_rate": 9.472981842210768e-05, + "loss": 0.0085, + "step": 19010 + }, + { + "epoch": 1.4772815533980583, + "grad_norm": 0.2677271068096161, + "learning_rate": 9.472242703187074e-05, + "loss": 0.0115, + "step": 19020 + }, + { + "epoch": 1.4780582524271844, + "grad_norm": 0.2116953581571579, + "learning_rate": 9.471503075085643e-05, + "loss": 0.0123, + "step": 19030 + }, + { + "epoch": 1.4788349514563106, + "grad_norm": 0.5259804725646973, + "learning_rate": 9.470762957987359e-05, + "loss": 0.0108, + "step": 19040 + }, + { + "epoch": 1.4796116504854369, + "grad_norm": 0.26658788323402405, + "learning_rate": 9.470022351973158e-05, + "loss": 0.0131, + "step": 19050 + }, + { + "epoch": 1.480388349514563, + "grad_norm": 0.18177445232868195, + "learning_rate": 9.469281257124034e-05, + "loss": 0.0097, + "step": 19060 + }, + { + "epoch": 1.4811650485436894, + "grad_norm": 0.16992300748825073, + "learning_rate": 9.46853967352103e-05, + "loss": 0.0108, + "step": 19070 + }, + { + "epoch": 1.4819417475728156, + "grad_norm": 0.272699773311615, + "learning_rate": 9.467797601245246e-05, + "loss": 0.0118, + "step": 19080 + }, + { + "epoch": 1.4827184466019419, + "grad_norm": 0.36933261156082153, + "learning_rate": 9.467055040377834e-05, + "loss": 0.0124, + "step": 19090 + }, + { + "epoch": 1.4834951456310679, + "grad_norm": 0.31903642416000366, + "learning_rate": 9.466311990999999e-05, + "loss": 0.0107, + "step": 19100 + }, + { + "epoch": 1.4842718446601941, + "grad_norm": 0.3028613328933716, + "learning_rate": 9.465568453193e-05, + "loss": 0.0083, + "step": 19110 + }, + { + "epoch": 1.4850485436893204, + "grad_norm": 0.23350493609905243, + "learning_rate": 9.464824427038148e-05, + "loss": 0.0094, + "step": 19120 + }, + { + "epoch": 1.4858252427184466, + "grad_norm": 0.27606692910194397, + "learning_rate": 9.46407991261681e-05, + "loss": 0.0105, + "step": 19130 + }, + { + "epoch": 1.4866019417475729, + "grad_norm": 0.2742990553379059, + "learning_rate": 9.463334910010404e-05, + "loss": 0.01, + "step": 19140 + }, + { + "epoch": 1.487378640776699, + "grad_norm": 0.15537892282009125, + "learning_rate": 9.462589419300403e-05, + "loss": 0.0114, + "step": 19150 + }, + { + "epoch": 1.4881553398058252, + "grad_norm": 0.2264343649148941, + "learning_rate": 9.461843440568333e-05, + "loss": 0.0115, + "step": 19160 + }, + { + "epoch": 1.4889320388349514, + "grad_norm": 0.24126534163951874, + "learning_rate": 9.461096973895773e-05, + "loss": 0.0145, + "step": 19170 + }, + { + "epoch": 1.4897087378640776, + "grad_norm": 0.3324410617351532, + "learning_rate": 9.460350019364355e-05, + "loss": 0.0121, + "step": 19180 + }, + { + "epoch": 1.490485436893204, + "grad_norm": 0.3943367600440979, + "learning_rate": 9.459602577055764e-05, + "loss": 0.0116, + "step": 19190 + }, + { + "epoch": 1.4912621359223301, + "grad_norm": 0.2543356418609619, + "learning_rate": 9.45885464705174e-05, + "loss": 0.0097, + "step": 19200 + }, + { + "epoch": 1.4920388349514564, + "grad_norm": 0.42755618691444397, + "learning_rate": 9.458106229434076e-05, + "loss": 0.0101, + "step": 19210 + }, + { + "epoch": 1.4928155339805826, + "grad_norm": 0.3722035586833954, + "learning_rate": 9.457357324284617e-05, + "loss": 0.0113, + "step": 19220 + }, + { + "epoch": 1.4935922330097087, + "grad_norm": 0.521467387676239, + "learning_rate": 9.456607931685262e-05, + "loss": 0.0134, + "step": 19230 + }, + { + "epoch": 1.494368932038835, + "grad_norm": 0.42224013805389404, + "learning_rate": 9.455858051717965e-05, + "loss": 0.0116, + "step": 19240 + }, + { + "epoch": 1.4951456310679612, + "grad_norm": 0.32680898904800415, + "learning_rate": 9.45510768446473e-05, + "loss": 0.0117, + "step": 19250 + }, + { + "epoch": 1.4959223300970874, + "grad_norm": 0.1694612354040146, + "learning_rate": 9.454356830007618e-05, + "loss": 0.0106, + "step": 19260 + }, + { + "epoch": 1.4966990291262137, + "grad_norm": 0.32032904028892517, + "learning_rate": 9.45360548842874e-05, + "loss": 0.0106, + "step": 19270 + }, + { + "epoch": 1.4974757281553397, + "grad_norm": 0.33176732063293457, + "learning_rate": 9.452853659810261e-05, + "loss": 0.0112, + "step": 19280 + }, + { + "epoch": 1.498252427184466, + "grad_norm": 0.29915496706962585, + "learning_rate": 9.452101344234401e-05, + "loss": 0.0088, + "step": 19290 + }, + { + "epoch": 1.4990291262135922, + "grad_norm": 0.179878368973732, + "learning_rate": 9.451348541783431e-05, + "loss": 0.0086, + "step": 19300 + }, + { + "epoch": 1.4998058252427184, + "grad_norm": 0.415340781211853, + "learning_rate": 9.450595252539678e-05, + "loss": 0.0102, + "step": 19310 + }, + { + "epoch": 1.5005825242718447, + "grad_norm": 0.1870359480381012, + "learning_rate": 9.449841476585518e-05, + "loss": 0.0105, + "step": 19320 + }, + { + "epoch": 1.501359223300971, + "grad_norm": 0.17683514952659607, + "learning_rate": 9.449087214003384e-05, + "loss": 0.0107, + "step": 19330 + }, + { + "epoch": 1.5021359223300972, + "grad_norm": 0.15662778913974762, + "learning_rate": 9.448332464875765e-05, + "loss": 0.0108, + "step": 19340 + }, + { + "epoch": 1.5029126213592234, + "grad_norm": 0.3077988028526306, + "learning_rate": 9.447577229285192e-05, + "loss": 0.0104, + "step": 19350 + }, + { + "epoch": 1.5036893203883495, + "grad_norm": 0.1777542382478714, + "learning_rate": 9.446821507314261e-05, + "loss": 0.0108, + "step": 19360 + }, + { + "epoch": 1.5044660194174757, + "grad_norm": 0.2195350080728531, + "learning_rate": 9.446065299045617e-05, + "loss": 0.0085, + "step": 19370 + }, + { + "epoch": 1.505242718446602, + "grad_norm": 0.17341473698616028, + "learning_rate": 9.445308604561955e-05, + "loss": 0.0112, + "step": 19380 + }, + { + "epoch": 1.5060194174757282, + "grad_norm": 0.20728644728660583, + "learning_rate": 9.444551423946028e-05, + "loss": 0.0133, + "step": 19390 + }, + { + "epoch": 1.5067961165048542, + "grad_norm": 0.2780629098415375, + "learning_rate": 9.443793757280638e-05, + "loss": 0.0104, + "step": 19400 + }, + { + "epoch": 1.5075728155339805, + "grad_norm": 0.2700847387313843, + "learning_rate": 9.443035604648646e-05, + "loss": 0.0118, + "step": 19410 + }, + { + "epoch": 1.5083495145631067, + "grad_norm": 0.3459249436855316, + "learning_rate": 9.44227696613296e-05, + "loss": 0.0094, + "step": 19420 + }, + { + "epoch": 1.509126213592233, + "grad_norm": 0.32839131355285645, + "learning_rate": 9.441517841816542e-05, + "loss": 0.0139, + "step": 19430 + }, + { + "epoch": 1.5099029126213592, + "grad_norm": 0.24281065165996552, + "learning_rate": 9.440758231782413e-05, + "loss": 0.0095, + "step": 19440 + }, + { + "epoch": 1.5106796116504855, + "grad_norm": 0.29823043942451477, + "learning_rate": 9.439998136113639e-05, + "loss": 0.0105, + "step": 19450 + }, + { + "epoch": 1.5114563106796117, + "grad_norm": 0.19388704001903534, + "learning_rate": 9.439237554893344e-05, + "loss": 0.0092, + "step": 19460 + }, + { + "epoch": 1.512233009708738, + "grad_norm": 0.20366883277893066, + "learning_rate": 9.438476488204705e-05, + "loss": 0.0094, + "step": 19470 + }, + { + "epoch": 1.5130097087378642, + "grad_norm": 0.24761496484279633, + "learning_rate": 9.43771493613095e-05, + "loss": 0.012, + "step": 19480 + }, + { + "epoch": 1.5137864077669902, + "grad_norm": 0.3519802987575531, + "learning_rate": 9.436952898755362e-05, + "loss": 0.0121, + "step": 19490 + }, + { + "epoch": 1.5145631067961165, + "grad_norm": 0.32314440608024597, + "learning_rate": 9.436190376161276e-05, + "loss": 0.01, + "step": 19500 + }, + { + "epoch": 1.5153398058252427, + "grad_norm": 0.27540576457977295, + "learning_rate": 9.43542736843208e-05, + "loss": 0.0127, + "step": 19510 + }, + { + "epoch": 1.5161165048543688, + "grad_norm": 0.20272788405418396, + "learning_rate": 9.434663875651216e-05, + "loss": 0.014, + "step": 19520 + }, + { + "epoch": 1.516893203883495, + "grad_norm": 0.26643112301826477, + "learning_rate": 9.433899897902177e-05, + "loss": 0.0097, + "step": 19530 + }, + { + "epoch": 1.5176699029126213, + "grad_norm": 0.2633499503135681, + "learning_rate": 9.433135435268511e-05, + "loss": 0.0126, + "step": 19540 + }, + { + "epoch": 1.5184466019417475, + "grad_norm": 0.43590766191482544, + "learning_rate": 9.432370487833819e-05, + "loss": 0.0139, + "step": 19550 + }, + { + "epoch": 1.5192233009708738, + "grad_norm": 0.2875920832157135, + "learning_rate": 9.431605055681756e-05, + "loss": 0.0146, + "step": 19560 + }, + { + "epoch": 1.52, + "grad_norm": 0.4118689000606537, + "learning_rate": 9.430839138896026e-05, + "loss": 0.0123, + "step": 19570 + }, + { + "epoch": 1.5207766990291263, + "grad_norm": 0.2641434669494629, + "learning_rate": 9.43007273756039e-05, + "loss": 0.0136, + "step": 19580 + }, + { + "epoch": 1.5215533980582525, + "grad_norm": 0.33846548199653625, + "learning_rate": 9.429305851758658e-05, + "loss": 0.0104, + "step": 19590 + }, + { + "epoch": 1.5223300970873788, + "grad_norm": 0.24220015108585358, + "learning_rate": 9.428538481574699e-05, + "loss": 0.0093, + "step": 19600 + }, + { + "epoch": 1.523106796116505, + "grad_norm": 0.28415820002555847, + "learning_rate": 9.42777062709243e-05, + "loss": 0.0069, + "step": 19610 + }, + { + "epoch": 1.523883495145631, + "grad_norm": 0.4601912796497345, + "learning_rate": 9.427002288395821e-05, + "loss": 0.0126, + "step": 19620 + }, + { + "epoch": 1.5246601941747573, + "grad_norm": 0.533000648021698, + "learning_rate": 9.426233465568898e-05, + "loss": 0.0115, + "step": 19630 + }, + { + "epoch": 1.5254368932038835, + "grad_norm": 0.3108910024166107, + "learning_rate": 9.42546415869574e-05, + "loss": 0.0109, + "step": 19640 + }, + { + "epoch": 1.5262135922330096, + "grad_norm": 0.3118383288383484, + "learning_rate": 9.424694367860473e-05, + "loss": 0.0086, + "step": 19650 + }, + { + "epoch": 1.5269902912621358, + "grad_norm": 0.22571034729480743, + "learning_rate": 9.423924093147284e-05, + "loss": 0.0127, + "step": 19660 + }, + { + "epoch": 1.527766990291262, + "grad_norm": 0.36165308952331543, + "learning_rate": 9.423153334640407e-05, + "loss": 0.013, + "step": 19670 + }, + { + "epoch": 1.5285436893203883, + "grad_norm": 0.30098623037338257, + "learning_rate": 9.42238209242413e-05, + "loss": 0.0107, + "step": 19680 + }, + { + "epoch": 1.5293203883495146, + "grad_norm": 0.3649698793888092, + "learning_rate": 9.421610366582798e-05, + "loss": 0.0102, + "step": 19690 + }, + { + "epoch": 1.5300970873786408, + "grad_norm": 0.34021860361099243, + "learning_rate": 9.420838157200803e-05, + "loss": 0.0161, + "step": 19700 + }, + { + "epoch": 1.530873786407767, + "grad_norm": 0.4120229184627533, + "learning_rate": 9.420065464362594e-05, + "loss": 0.009, + "step": 19710 + }, + { + "epoch": 1.5316504854368933, + "grad_norm": 0.36690929532051086, + "learning_rate": 9.419292288152673e-05, + "loss": 0.0112, + "step": 19720 + }, + { + "epoch": 1.5324271844660196, + "grad_norm": 0.3114622235298157, + "learning_rate": 9.418518628655588e-05, + "loss": 0.0101, + "step": 19730 + }, + { + "epoch": 1.5332038834951456, + "grad_norm": 0.394896924495697, + "learning_rate": 9.417744485955951e-05, + "loss": 0.0166, + "step": 19740 + }, + { + "epoch": 1.5339805825242718, + "grad_norm": 0.1428711861371994, + "learning_rate": 9.41696986013842e-05, + "loss": 0.0113, + "step": 19750 + }, + { + "epoch": 1.534757281553398, + "grad_norm": 0.1608455628156662, + "learning_rate": 9.416194751287705e-05, + "loss": 0.0126, + "step": 19760 + }, + { + "epoch": 1.5355339805825243, + "grad_norm": 0.38728922605514526, + "learning_rate": 9.415419159488572e-05, + "loss": 0.0114, + "step": 19770 + }, + { + "epoch": 1.5363106796116504, + "grad_norm": 0.21650604903697968, + "learning_rate": 9.414643084825837e-05, + "loss": 0.014, + "step": 19780 + }, + { + "epoch": 1.5370873786407766, + "grad_norm": 0.18449589610099792, + "learning_rate": 9.413866527384372e-05, + "loss": 0.0145, + "step": 19790 + }, + { + "epoch": 1.5378640776699029, + "grad_norm": 0.33984288573265076, + "learning_rate": 9.4130894872491e-05, + "loss": 0.0114, + "step": 19800 + }, + { + "epoch": 1.538640776699029, + "grad_norm": 0.19352926313877106, + "learning_rate": 9.412311964504998e-05, + "loss": 0.0115, + "step": 19810 + }, + { + "epoch": 1.5394174757281553, + "grad_norm": 0.159114807844162, + "learning_rate": 9.411533959237091e-05, + "loss": 0.0129, + "step": 19820 + }, + { + "epoch": 1.5401941747572816, + "grad_norm": 0.30051377415657043, + "learning_rate": 9.410755471530464e-05, + "loss": 0.0134, + "step": 19830 + }, + { + "epoch": 1.5409708737864078, + "grad_norm": 0.47652390599250793, + "learning_rate": 9.40997650147025e-05, + "loss": 0.014, + "step": 19840 + }, + { + "epoch": 1.541747572815534, + "grad_norm": 0.3514070510864258, + "learning_rate": 9.409197049141637e-05, + "loss": 0.0164, + "step": 19850 + }, + { + "epoch": 1.5425242718446603, + "grad_norm": 0.27670636773109436, + "learning_rate": 9.408417114629863e-05, + "loss": 0.0137, + "step": 19860 + }, + { + "epoch": 1.5433009708737864, + "grad_norm": 0.25373151898384094, + "learning_rate": 9.40763669802022e-05, + "loss": 0.0112, + "step": 19870 + }, + { + "epoch": 1.5440776699029126, + "grad_norm": 0.31871625781059265, + "learning_rate": 9.406855799398056e-05, + "loss": 0.0103, + "step": 19880 + }, + { + "epoch": 1.5448543689320389, + "grad_norm": 0.340659499168396, + "learning_rate": 9.406074418848767e-05, + "loss": 0.0128, + "step": 19890 + }, + { + "epoch": 1.545631067961165, + "grad_norm": 0.36127740144729614, + "learning_rate": 9.405292556457805e-05, + "loss": 0.0125, + "step": 19900 + }, + { + "epoch": 1.5464077669902911, + "grad_norm": 0.22690503299236298, + "learning_rate": 9.404510212310671e-05, + "loss": 0.0125, + "step": 19910 + }, + { + "epoch": 1.5471844660194174, + "grad_norm": 0.26705342531204224, + "learning_rate": 9.403727386492924e-05, + "loss": 0.0112, + "step": 19920 + }, + { + "epoch": 1.5479611650485436, + "grad_norm": 0.24585935473442078, + "learning_rate": 9.40294407909017e-05, + "loss": 0.0106, + "step": 19930 + }, + { + "epoch": 1.54873786407767, + "grad_norm": 0.2893551290035248, + "learning_rate": 9.40216029018807e-05, + "loss": 0.0103, + "step": 19940 + }, + { + "epoch": 1.5495145631067961, + "grad_norm": 0.235193133354187, + "learning_rate": 9.401376019872338e-05, + "loss": 0.013, + "step": 19950 + }, + { + "epoch": 1.5502912621359224, + "grad_norm": 0.20903724431991577, + "learning_rate": 9.400591268228746e-05, + "loss": 0.0097, + "step": 19960 + }, + { + "epoch": 1.5510679611650486, + "grad_norm": 0.32513514161109924, + "learning_rate": 9.399806035343106e-05, + "loss": 0.0102, + "step": 19970 + }, + { + "epoch": 1.5518446601941749, + "grad_norm": 0.24852415919303894, + "learning_rate": 9.399020321301294e-05, + "loss": 0.01, + "step": 19980 + }, + { + "epoch": 1.5526213592233011, + "grad_norm": 0.2642050087451935, + "learning_rate": 9.398234126189234e-05, + "loss": 0.0089, + "step": 19990 + }, + { + "epoch": 1.5533980582524272, + "grad_norm": 0.28317031264305115, + "learning_rate": 9.397447450092902e-05, + "loss": 0.0086, + "step": 20000 + }, + { + "epoch": 1.5541747572815534, + "grad_norm": 0.2631814777851105, + "learning_rate": 9.39666029309833e-05, + "loss": 0.0107, + "step": 20010 + }, + { + "epoch": 1.5549514563106797, + "grad_norm": 0.22639253735542297, + "learning_rate": 9.395872655291596e-05, + "loss": 0.0137, + "step": 20020 + }, + { + "epoch": 1.5557281553398057, + "grad_norm": 0.19113343954086304, + "learning_rate": 9.395084536758838e-05, + "loss": 0.0101, + "step": 20030 + }, + { + "epoch": 1.556504854368932, + "grad_norm": 0.16811758279800415, + "learning_rate": 9.394295937586243e-05, + "loss": 0.0085, + "step": 20040 + }, + { + "epoch": 1.5572815533980582, + "grad_norm": 0.2410849630832672, + "learning_rate": 9.393506857860052e-05, + "loss": 0.0105, + "step": 20050 + }, + { + "epoch": 1.5580582524271844, + "grad_norm": 0.2370774745941162, + "learning_rate": 9.392717297666555e-05, + "loss": 0.0108, + "step": 20060 + }, + { + "epoch": 1.5588349514563107, + "grad_norm": 0.21098680794239044, + "learning_rate": 9.391927257092101e-05, + "loss": 0.0128, + "step": 20070 + }, + { + "epoch": 1.559611650485437, + "grad_norm": 0.15026837587356567, + "learning_rate": 9.391136736223085e-05, + "loss": 0.0094, + "step": 20080 + }, + { + "epoch": 1.5603883495145632, + "grad_norm": 0.3756979703903198, + "learning_rate": 9.390345735145956e-05, + "loss": 0.0088, + "step": 20090 + }, + { + "epoch": 1.5611650485436894, + "grad_norm": 0.279360830783844, + "learning_rate": 9.389554253947219e-05, + "loss": 0.0135, + "step": 20100 + }, + { + "epoch": 1.5619417475728157, + "grad_norm": 0.37404754757881165, + "learning_rate": 9.388762292713428e-05, + "loss": 0.0104, + "step": 20110 + }, + { + "epoch": 1.5627184466019417, + "grad_norm": 0.2021477222442627, + "learning_rate": 9.38796985153119e-05, + "loss": 0.0112, + "step": 20120 + }, + { + "epoch": 1.563495145631068, + "grad_norm": 0.24811951816082, + "learning_rate": 9.387176930487169e-05, + "loss": 0.0119, + "step": 20130 + }, + { + "epoch": 1.5642718446601942, + "grad_norm": 0.21817646920681, + "learning_rate": 9.386383529668072e-05, + "loss": 0.0144, + "step": 20140 + }, + { + "epoch": 1.5650485436893202, + "grad_norm": 0.8295303583145142, + "learning_rate": 9.385589649160669e-05, + "loss": 0.0134, + "step": 20150 + }, + { + "epoch": 1.5658252427184465, + "grad_norm": 0.3000243306159973, + "learning_rate": 9.384795289051775e-05, + "loss": 0.0104, + "step": 20160 + }, + { + "epoch": 1.5666019417475727, + "grad_norm": 0.36566832661628723, + "learning_rate": 9.384000449428261e-05, + "loss": 0.0082, + "step": 20170 + }, + { + "epoch": 1.567378640776699, + "grad_norm": 0.40981560945510864, + "learning_rate": 9.383205130377048e-05, + "loss": 0.0131, + "step": 20180 + }, + { + "epoch": 1.5681553398058252, + "grad_norm": 0.3162985146045685, + "learning_rate": 9.382409331985114e-05, + "loss": 0.0097, + "step": 20190 + }, + { + "epoch": 1.5689320388349515, + "grad_norm": 0.17418353259563446, + "learning_rate": 9.381613054339482e-05, + "loss": 0.0159, + "step": 20200 + }, + { + "epoch": 1.5697087378640777, + "grad_norm": 0.31391260027885437, + "learning_rate": 9.380816297527235e-05, + "loss": 0.0088, + "step": 20210 + }, + { + "epoch": 1.570485436893204, + "grad_norm": 0.9788041710853577, + "learning_rate": 9.380019061635506e-05, + "loss": 0.0125, + "step": 20220 + }, + { + "epoch": 1.5712621359223302, + "grad_norm": 0.26943981647491455, + "learning_rate": 9.379221346751474e-05, + "loss": 0.0111, + "step": 20230 + }, + { + "epoch": 1.5720388349514565, + "grad_norm": 0.30791181325912476, + "learning_rate": 9.378423152962382e-05, + "loss": 0.0125, + "step": 20240 + }, + { + "epoch": 1.5728155339805825, + "grad_norm": 0.15682272613048553, + "learning_rate": 9.377624480355517e-05, + "loss": 0.0104, + "step": 20250 + }, + { + "epoch": 1.5735922330097087, + "grad_norm": 0.19056950509548187, + "learning_rate": 9.376825329018219e-05, + "loss": 0.0083, + "step": 20260 + }, + { + "epoch": 1.574368932038835, + "grad_norm": 0.20776142179965973, + "learning_rate": 9.376025699037884e-05, + "loss": 0.0101, + "step": 20270 + }, + { + "epoch": 1.575145631067961, + "grad_norm": 0.21107524633407593, + "learning_rate": 9.37522559050196e-05, + "loss": 0.0095, + "step": 20280 + }, + { + "epoch": 1.5759223300970873, + "grad_norm": 0.2884169816970825, + "learning_rate": 9.37442500349794e-05, + "loss": 0.0083, + "step": 20290 + }, + { + "epoch": 1.5766990291262135, + "grad_norm": 0.7759711742401123, + "learning_rate": 9.373623938113381e-05, + "loss": 0.0097, + "step": 20300 + }, + { + "epoch": 1.5774757281553398, + "grad_norm": 0.3079848885536194, + "learning_rate": 9.372822394435883e-05, + "loss": 0.0104, + "step": 20310 + }, + { + "epoch": 1.578252427184466, + "grad_norm": 0.29594936966896057, + "learning_rate": 9.372020372553102e-05, + "loss": 0.0101, + "step": 20320 + }, + { + "epoch": 1.5790291262135923, + "grad_norm": 0.20919524133205414, + "learning_rate": 9.371217872552746e-05, + "loss": 0.0087, + "step": 20330 + }, + { + "epoch": 1.5798058252427185, + "grad_norm": 0.25720998644828796, + "learning_rate": 9.370414894522576e-05, + "loss": 0.014, + "step": 20340 + }, + { + "epoch": 1.5805825242718448, + "grad_norm": 0.35241010785102844, + "learning_rate": 9.369611438550406e-05, + "loss": 0.0127, + "step": 20350 + }, + { + "epoch": 1.581359223300971, + "grad_norm": 0.1716156303882599, + "learning_rate": 9.368807504724095e-05, + "loss": 0.012, + "step": 20360 + }, + { + "epoch": 1.582135922330097, + "grad_norm": 0.3438595235347748, + "learning_rate": 9.368003093131565e-05, + "loss": 0.0124, + "step": 20370 + }, + { + "epoch": 1.5829126213592233, + "grad_norm": 0.3592586815357208, + "learning_rate": 9.367198203860785e-05, + "loss": 0.0108, + "step": 20380 + }, + { + "epoch": 1.5836893203883495, + "grad_norm": 0.33132678270339966, + "learning_rate": 9.366392836999774e-05, + "loss": 0.0094, + "step": 20390 + }, + { + "epoch": 1.5844660194174758, + "grad_norm": 0.27901679277420044, + "learning_rate": 9.365586992636607e-05, + "loss": 0.0099, + "step": 20400 + }, + { + "epoch": 1.5852427184466018, + "grad_norm": 0.32587963342666626, + "learning_rate": 9.364780670859412e-05, + "loss": 0.01, + "step": 20410 + }, + { + "epoch": 1.586019417475728, + "grad_norm": 0.5034458041191101, + "learning_rate": 9.363973871756364e-05, + "loss": 0.0113, + "step": 20420 + }, + { + "epoch": 1.5867961165048543, + "grad_norm": 0.3229667842388153, + "learning_rate": 9.363166595415696e-05, + "loss": 0.0142, + "step": 20430 + }, + { + "epoch": 1.5875728155339806, + "grad_norm": 0.27907049655914307, + "learning_rate": 9.362358841925686e-05, + "loss": 0.0126, + "step": 20440 + }, + { + "epoch": 1.5883495145631068, + "grad_norm": 0.2910282611846924, + "learning_rate": 9.361550611374674e-05, + "loss": 0.0105, + "step": 20450 + }, + { + "epoch": 1.589126213592233, + "grad_norm": 0.24764585494995117, + "learning_rate": 9.360741903851043e-05, + "loss": 0.0148, + "step": 20460 + }, + { + "epoch": 1.5899029126213593, + "grad_norm": 0.3486481308937073, + "learning_rate": 9.359932719443236e-05, + "loss": 0.0136, + "step": 20470 + }, + { + "epoch": 1.5906796116504855, + "grad_norm": 0.1674918681383133, + "learning_rate": 9.35912305823974e-05, + "loss": 0.0128, + "step": 20480 + }, + { + "epoch": 1.5914563106796118, + "grad_norm": 0.1878303587436676, + "learning_rate": 9.358312920329101e-05, + "loss": 0.0118, + "step": 20490 + }, + { + "epoch": 1.5922330097087378, + "grad_norm": 0.33640244603157043, + "learning_rate": 9.357502305799914e-05, + "loss": 0.0082, + "step": 20500 + }, + { + "epoch": 1.593009708737864, + "grad_norm": 0.22431637346744537, + "learning_rate": 9.356691214740824e-05, + "loss": 0.0148, + "step": 20510 + }, + { + "epoch": 1.5937864077669903, + "grad_norm": 0.2996147572994232, + "learning_rate": 9.355879647240535e-05, + "loss": 0.0135, + "step": 20520 + }, + { + "epoch": 1.5945631067961163, + "grad_norm": 0.2877601385116577, + "learning_rate": 9.355067603387798e-05, + "loss": 0.0144, + "step": 20530 + }, + { + "epoch": 1.5953398058252426, + "grad_norm": 0.2538755238056183, + "learning_rate": 9.354255083271412e-05, + "loss": 0.0112, + "step": 20540 + }, + { + "epoch": 1.5961165048543688, + "grad_norm": 0.20042572915554047, + "learning_rate": 9.353442086980239e-05, + "loss": 0.0116, + "step": 20550 + }, + { + "epoch": 1.596893203883495, + "grad_norm": 0.15362051129341125, + "learning_rate": 9.352628614603185e-05, + "loss": 0.0104, + "step": 20560 + }, + { + "epoch": 1.5976699029126213, + "grad_norm": 0.20880557596683502, + "learning_rate": 9.351814666229209e-05, + "loss": 0.0093, + "step": 20570 + }, + { + "epoch": 1.5984466019417476, + "grad_norm": 0.30289021134376526, + "learning_rate": 9.351000241947324e-05, + "loss": 0.0114, + "step": 20580 + }, + { + "epoch": 1.5992233009708738, + "grad_norm": 0.2518482506275177, + "learning_rate": 9.350185341846594e-05, + "loss": 0.0084, + "step": 20590 + }, + { + "epoch": 1.6, + "grad_norm": 0.38364267349243164, + "learning_rate": 9.349369966016134e-05, + "loss": 0.0108, + "step": 20600 + }, + { + "epoch": 1.6007766990291263, + "grad_norm": 0.2148597091436386, + "learning_rate": 9.348554114545117e-05, + "loss": 0.0125, + "step": 20610 + }, + { + "epoch": 1.6015533980582526, + "grad_norm": 0.30340269207954407, + "learning_rate": 9.347737787522758e-05, + "loss": 0.0097, + "step": 20620 + }, + { + "epoch": 1.6023300970873786, + "grad_norm": 0.4410690367221832, + "learning_rate": 9.346920985038332e-05, + "loss": 0.0104, + "step": 20630 + }, + { + "epoch": 1.6031067961165049, + "grad_norm": 0.21607555449008942, + "learning_rate": 9.346103707181162e-05, + "loss": 0.0095, + "step": 20640 + }, + { + "epoch": 1.603883495145631, + "grad_norm": 0.1709173023700714, + "learning_rate": 9.345285954040626e-05, + "loss": 0.0101, + "step": 20650 + }, + { + "epoch": 1.6046601941747571, + "grad_norm": 0.313885360956192, + "learning_rate": 9.34446772570615e-05, + "loss": 0.0166, + "step": 20660 + }, + { + "epoch": 1.6054368932038834, + "grad_norm": 0.324199914932251, + "learning_rate": 9.343649022267214e-05, + "loss": 0.0109, + "step": 20670 + }, + { + "epoch": 1.6062135922330096, + "grad_norm": 0.6620644330978394, + "learning_rate": 9.342829843813353e-05, + "loss": 0.0109, + "step": 20680 + }, + { + "epoch": 1.6069902912621359, + "grad_norm": 0.35710954666137695, + "learning_rate": 9.342010190434149e-05, + "loss": 0.0117, + "step": 20690 + }, + { + "epoch": 1.6077669902912621, + "grad_norm": 0.20891214907169342, + "learning_rate": 9.34119006221924e-05, + "loss": 0.0081, + "step": 20700 + }, + { + "epoch": 1.6085436893203884, + "grad_norm": 0.3952854871749878, + "learning_rate": 9.340369459258313e-05, + "loss": 0.0136, + "step": 20710 + }, + { + "epoch": 1.6093203883495146, + "grad_norm": 0.23131537437438965, + "learning_rate": 9.339548381641106e-05, + "loss": 0.0088, + "step": 20720 + }, + { + "epoch": 1.6100970873786409, + "grad_norm": 0.22632676362991333, + "learning_rate": 9.338726829457413e-05, + "loss": 0.0104, + "step": 20730 + }, + { + "epoch": 1.6108737864077671, + "grad_norm": 0.2737612724304199, + "learning_rate": 9.337904802797078e-05, + "loss": 0.0113, + "step": 20740 + }, + { + "epoch": 1.6116504854368932, + "grad_norm": 0.22265750169754028, + "learning_rate": 9.337082301749993e-05, + "loss": 0.0099, + "step": 20750 + }, + { + "epoch": 1.6124271844660194, + "grad_norm": 0.3701304793357849, + "learning_rate": 9.336259326406109e-05, + "loss": 0.0106, + "step": 20760 + }, + { + "epoch": 1.6132038834951457, + "grad_norm": 0.25122252106666565, + "learning_rate": 9.335435876855427e-05, + "loss": 0.0092, + "step": 20770 + }, + { + "epoch": 1.6139805825242717, + "grad_norm": 0.8990773558616638, + "learning_rate": 9.334611953187994e-05, + "loss": 0.0111, + "step": 20780 + }, + { + "epoch": 1.614757281553398, + "grad_norm": 0.341387003660202, + "learning_rate": 9.333787555493914e-05, + "loss": 0.0114, + "step": 20790 + }, + { + "epoch": 1.6155339805825242, + "grad_norm": 0.2006390243768692, + "learning_rate": 9.332962683863345e-05, + "loss": 0.0135, + "step": 20800 + }, + { + "epoch": 1.6163106796116504, + "grad_norm": 0.18417446315288544, + "learning_rate": 9.332137338386489e-05, + "loss": 0.013, + "step": 20810 + }, + { + "epoch": 1.6170873786407767, + "grad_norm": 0.2977626621723175, + "learning_rate": 9.33131151915361e-05, + "loss": 0.0112, + "step": 20820 + }, + { + "epoch": 1.617864077669903, + "grad_norm": 0.32060906291007996, + "learning_rate": 9.330485226255012e-05, + "loss": 0.0092, + "step": 20830 + }, + { + "epoch": 1.6186407766990292, + "grad_norm": 0.3547264337539673, + "learning_rate": 9.329658459781061e-05, + "loss": 0.0107, + "step": 20840 + }, + { + "epoch": 1.6194174757281554, + "grad_norm": 0.33860933780670166, + "learning_rate": 9.328831219822172e-05, + "loss": 0.0109, + "step": 20850 + }, + { + "epoch": 1.6201941747572817, + "grad_norm": 0.22614768147468567, + "learning_rate": 9.328003506468808e-05, + "loss": 0.011, + "step": 20860 + }, + { + "epoch": 1.620970873786408, + "grad_norm": 0.3093767464160919, + "learning_rate": 9.327175319811488e-05, + "loss": 0.0088, + "step": 20870 + }, + { + "epoch": 1.621747572815534, + "grad_norm": 0.2645571827888489, + "learning_rate": 9.326346659940781e-05, + "loss": 0.0106, + "step": 20880 + }, + { + "epoch": 1.6225242718446602, + "grad_norm": 0.42147761583328247, + "learning_rate": 9.325517526947308e-05, + "loss": 0.0097, + "step": 20890 + }, + { + "epoch": 1.6233009708737864, + "grad_norm": 0.18790210783481598, + "learning_rate": 9.32468792092174e-05, + "loss": 0.0096, + "step": 20900 + }, + { + "epoch": 1.6240776699029125, + "grad_norm": 0.19265003502368927, + "learning_rate": 9.323857841954803e-05, + "loss": 0.011, + "step": 20910 + }, + { + "epoch": 1.6248543689320387, + "grad_norm": 0.27039584517478943, + "learning_rate": 9.323027290137276e-05, + "loss": 0.0114, + "step": 20920 + }, + { + "epoch": 1.625631067961165, + "grad_norm": 0.29134202003479004, + "learning_rate": 9.322196265559981e-05, + "loss": 0.0115, + "step": 20930 + }, + { + "epoch": 1.6264077669902912, + "grad_norm": 0.24879412353038788, + "learning_rate": 9.321364768313803e-05, + "loss": 0.0105, + "step": 20940 + }, + { + "epoch": 1.6271844660194175, + "grad_norm": 0.29567092657089233, + "learning_rate": 9.32053279848967e-05, + "loss": 0.0134, + "step": 20950 + }, + { + "epoch": 1.6279611650485437, + "grad_norm": 1.1938129663467407, + "learning_rate": 9.319700356178567e-05, + "loss": 0.0125, + "step": 20960 + }, + { + "epoch": 1.62873786407767, + "grad_norm": 0.5450142025947571, + "learning_rate": 9.318867441471527e-05, + "loss": 0.0128, + "step": 20970 + }, + { + "epoch": 1.6295145631067962, + "grad_norm": 0.252196729183197, + "learning_rate": 9.318034054459637e-05, + "loss": 0.0137, + "step": 20980 + }, + { + "epoch": 1.6302912621359225, + "grad_norm": 0.2899428606033325, + "learning_rate": 9.317200195234034e-05, + "loss": 0.0121, + "step": 20990 + }, + { + "epoch": 1.6310679611650487, + "grad_norm": 0.24450337886810303, + "learning_rate": 9.316365863885909e-05, + "loss": 0.0165, + "step": 21000 + }, + { + "epoch": 1.6318446601941747, + "grad_norm": 0.20614710450172424, + "learning_rate": 9.315531060506502e-05, + "loss": 0.0122, + "step": 21010 + }, + { + "epoch": 1.632621359223301, + "grad_norm": 0.34260889887809753, + "learning_rate": 9.314695785187108e-05, + "loss": 0.0089, + "step": 21020 + }, + { + "epoch": 1.6333980582524272, + "grad_norm": 0.40466561913490295, + "learning_rate": 9.313860038019069e-05, + "loss": 0.0133, + "step": 21030 + }, + { + "epoch": 1.6341747572815533, + "grad_norm": 0.3734077513217926, + "learning_rate": 9.313023819093782e-05, + "loss": 0.0104, + "step": 21040 + }, + { + "epoch": 1.6349514563106795, + "grad_norm": 0.2911650240421295, + "learning_rate": 9.312187128502695e-05, + "loss": 0.0094, + "step": 21050 + }, + { + "epoch": 1.6357281553398058, + "grad_norm": 0.27784547209739685, + "learning_rate": 9.311349966337307e-05, + "loss": 0.0102, + "step": 21060 + }, + { + "epoch": 1.636504854368932, + "grad_norm": 0.3459400236606598, + "learning_rate": 9.310512332689169e-05, + "loss": 0.0116, + "step": 21070 + }, + { + "epoch": 1.6372815533980583, + "grad_norm": 0.23356443643569946, + "learning_rate": 9.309674227649883e-05, + "loss": 0.0129, + "step": 21080 + }, + { + "epoch": 1.6380582524271845, + "grad_norm": 0.22238752245903015, + "learning_rate": 9.308835651311103e-05, + "loss": 0.0085, + "step": 21090 + }, + { + "epoch": 1.6388349514563108, + "grad_norm": 0.2675740718841553, + "learning_rate": 9.307996603764533e-05, + "loss": 0.0104, + "step": 21100 + }, + { + "epoch": 1.639611650485437, + "grad_norm": 0.17875702679157257, + "learning_rate": 9.307157085101932e-05, + "loss": 0.0116, + "step": 21110 + }, + { + "epoch": 1.6403883495145632, + "grad_norm": 0.2935802638530731, + "learning_rate": 9.306317095415109e-05, + "loss": 0.0083, + "step": 21120 + }, + { + "epoch": 1.6411650485436893, + "grad_norm": 0.19191938638687134, + "learning_rate": 9.305476634795922e-05, + "loss": 0.0108, + "step": 21130 + }, + { + "epoch": 1.6419417475728155, + "grad_norm": 0.30373939871788025, + "learning_rate": 9.304635703336284e-05, + "loss": 0.0104, + "step": 21140 + }, + { + "epoch": 1.6427184466019418, + "grad_norm": 0.3051883578300476, + "learning_rate": 9.303794301128157e-05, + "loss": 0.0117, + "step": 21150 + }, + { + "epoch": 1.6434951456310678, + "grad_norm": 0.1889467090368271, + "learning_rate": 9.302952428263555e-05, + "loss": 0.0093, + "step": 21160 + }, + { + "epoch": 1.644271844660194, + "grad_norm": 0.24335873126983643, + "learning_rate": 9.302110084834545e-05, + "loss": 0.0108, + "step": 21170 + }, + { + "epoch": 1.6450485436893203, + "grad_norm": 0.17080847918987274, + "learning_rate": 9.301267270933245e-05, + "loss": 0.0101, + "step": 21180 + }, + { + "epoch": 1.6458252427184465, + "grad_norm": 0.2578194737434387, + "learning_rate": 9.300423986651823e-05, + "loss": 0.0091, + "step": 21190 + }, + { + "epoch": 1.6466019417475728, + "grad_norm": 0.3832862675189972, + "learning_rate": 9.299580232082501e-05, + "loss": 0.0132, + "step": 21200 + }, + { + "epoch": 1.647378640776699, + "grad_norm": 0.1925412267446518, + "learning_rate": 9.298736007317547e-05, + "loss": 0.0093, + "step": 21210 + }, + { + "epoch": 1.6481553398058253, + "grad_norm": 0.24540984630584717, + "learning_rate": 9.297891312449288e-05, + "loss": 0.0118, + "step": 21220 + }, + { + "epoch": 1.6489320388349515, + "grad_norm": 0.1849033683538437, + "learning_rate": 9.297046147570094e-05, + "loss": 0.0115, + "step": 21230 + }, + { + "epoch": 1.6497087378640778, + "grad_norm": 0.29349634051322937, + "learning_rate": 9.296200512772396e-05, + "loss": 0.0094, + "step": 21240 + }, + { + "epoch": 1.650485436893204, + "grad_norm": 0.3127397894859314, + "learning_rate": 9.295354408148668e-05, + "loss": 0.0102, + "step": 21250 + }, + { + "epoch": 1.65126213592233, + "grad_norm": 0.16552802920341492, + "learning_rate": 9.294507833791441e-05, + "loss": 0.01, + "step": 21260 + }, + { + "epoch": 1.6520388349514563, + "grad_norm": 0.21877069771289825, + "learning_rate": 9.293660789793295e-05, + "loss": 0.0108, + "step": 21270 + }, + { + "epoch": 1.6528155339805826, + "grad_norm": 0.15889938175678253, + "learning_rate": 9.292813276246858e-05, + "loss": 0.0092, + "step": 21280 + }, + { + "epoch": 1.6535922330097086, + "grad_norm": 0.25939682126045227, + "learning_rate": 9.291965293244816e-05, + "loss": 0.0096, + "step": 21290 + }, + { + "epoch": 1.6543689320388348, + "grad_norm": 0.2728210687637329, + "learning_rate": 9.291116840879904e-05, + "loss": 0.0128, + "step": 21300 + }, + { + "epoch": 1.655145631067961, + "grad_norm": 0.3977471590042114, + "learning_rate": 9.290267919244904e-05, + "loss": 0.0092, + "step": 21310 + }, + { + "epoch": 1.6559223300970873, + "grad_norm": 0.22483551502227783, + "learning_rate": 9.289418528432655e-05, + "loss": 0.0127, + "step": 21320 + }, + { + "epoch": 1.6566990291262136, + "grad_norm": 0.5181578993797302, + "learning_rate": 9.288568668536045e-05, + "loss": 0.0102, + "step": 21330 + }, + { + "epoch": 1.6574757281553398, + "grad_norm": 0.27192094922065735, + "learning_rate": 9.287718339648013e-05, + "loss": 0.0111, + "step": 21340 + }, + { + "epoch": 1.658252427184466, + "grad_norm": 0.29946932196617126, + "learning_rate": 9.28686754186155e-05, + "loss": 0.0113, + "step": 21350 + }, + { + "epoch": 1.6590291262135923, + "grad_norm": 0.2964247763156891, + "learning_rate": 9.286016275269698e-05, + "loss": 0.0114, + "step": 21360 + }, + { + "epoch": 1.6598058252427186, + "grad_norm": 0.30150070786476135, + "learning_rate": 9.285164539965551e-05, + "loss": 0.0104, + "step": 21370 + }, + { + "epoch": 1.6605825242718446, + "grad_norm": 0.31943872570991516, + "learning_rate": 9.284312336042251e-05, + "loss": 0.0107, + "step": 21380 + }, + { + "epoch": 1.6613592233009709, + "grad_norm": 0.33073413372039795, + "learning_rate": 9.283459663592996e-05, + "loss": 0.0091, + "step": 21390 + }, + { + "epoch": 1.662135922330097, + "grad_norm": 0.2571268081665039, + "learning_rate": 9.282606522711033e-05, + "loss": 0.0092, + "step": 21400 + }, + { + "epoch": 1.6629126213592234, + "grad_norm": 0.33180779218673706, + "learning_rate": 9.281752913489657e-05, + "loss": 0.0105, + "step": 21410 + }, + { + "epoch": 1.6636893203883494, + "grad_norm": 0.13658711314201355, + "learning_rate": 9.280898836022222e-05, + "loss": 0.0089, + "step": 21420 + }, + { + "epoch": 1.6644660194174756, + "grad_norm": 0.30564022064208984, + "learning_rate": 9.280044290402126e-05, + "loss": 0.0124, + "step": 21430 + }, + { + "epoch": 1.6652427184466019, + "grad_norm": 0.5140872001647949, + "learning_rate": 9.279189276722821e-05, + "loss": 0.0117, + "step": 21440 + }, + { + "epoch": 1.6660194174757281, + "grad_norm": 0.32279807329177856, + "learning_rate": 9.278333795077812e-05, + "loss": 0.0144, + "step": 21450 + }, + { + "epoch": 1.6667961165048544, + "grad_norm": 0.26558586955070496, + "learning_rate": 9.27747784556065e-05, + "loss": 0.0104, + "step": 21460 + }, + { + "epoch": 1.6675728155339806, + "grad_norm": 0.3372723162174225, + "learning_rate": 9.276621428264942e-05, + "loss": 0.0125, + "step": 21470 + }, + { + "epoch": 1.6683495145631069, + "grad_norm": 0.3173283040523529, + "learning_rate": 9.275764543284345e-05, + "loss": 0.0114, + "step": 21480 + }, + { + "epoch": 1.6691262135922331, + "grad_norm": 0.26747047901153564, + "learning_rate": 9.274907190712566e-05, + "loss": 0.0134, + "step": 21490 + }, + { + "epoch": 1.6699029126213594, + "grad_norm": 0.2159215211868286, + "learning_rate": 9.274049370643363e-05, + "loss": 0.0157, + "step": 21500 + }, + { + "epoch": 1.6706796116504854, + "grad_norm": 0.2607375979423523, + "learning_rate": 9.273191083170547e-05, + "loss": 0.0107, + "step": 21510 + }, + { + "epoch": 1.6714563106796116, + "grad_norm": 0.30212876200675964, + "learning_rate": 9.27233232838798e-05, + "loss": 0.0117, + "step": 21520 + }, + { + "epoch": 1.672233009708738, + "grad_norm": 0.2667211592197418, + "learning_rate": 9.27147310638957e-05, + "loss": 0.0096, + "step": 21530 + }, + { + "epoch": 1.673009708737864, + "grad_norm": 0.27567386627197266, + "learning_rate": 9.270613417269286e-05, + "loss": 0.009, + "step": 21540 + }, + { + "epoch": 1.6737864077669902, + "grad_norm": 0.2715089023113251, + "learning_rate": 9.269753261121138e-05, + "loss": 0.0132, + "step": 21550 + }, + { + "epoch": 1.6745631067961164, + "grad_norm": 0.29494509100914, + "learning_rate": 9.268892638039194e-05, + "loss": 0.0102, + "step": 21560 + }, + { + "epoch": 1.6753398058252427, + "grad_norm": 0.2666895091533661, + "learning_rate": 9.268031548117569e-05, + "loss": 0.0111, + "step": 21570 + }, + { + "epoch": 1.676116504854369, + "grad_norm": 0.2705308794975281, + "learning_rate": 9.26716999145043e-05, + "loss": 0.0108, + "step": 21580 + }, + { + "epoch": 1.6768932038834952, + "grad_norm": 0.25681692361831665, + "learning_rate": 9.266307968131998e-05, + "loss": 0.0131, + "step": 21590 + }, + { + "epoch": 1.6776699029126214, + "grad_norm": 0.41161149740219116, + "learning_rate": 9.26544547825654e-05, + "loss": 0.0092, + "step": 21600 + }, + { + "epoch": 1.6784466019417477, + "grad_norm": 0.40868014097213745, + "learning_rate": 9.264582521918376e-05, + "loss": 0.0125, + "step": 21610 + }, + { + "epoch": 1.679223300970874, + "grad_norm": 0.2689008116722107, + "learning_rate": 9.263719099211881e-05, + "loss": 0.0124, + "step": 21620 + }, + { + "epoch": 1.6800000000000002, + "grad_norm": 0.38747185468673706, + "learning_rate": 9.262855210231476e-05, + "loss": 0.0153, + "step": 21630 + }, + { + "epoch": 1.6807766990291262, + "grad_norm": 0.21008875966072083, + "learning_rate": 9.261990855071633e-05, + "loss": 0.0104, + "step": 21640 + }, + { + "epoch": 1.6815533980582524, + "grad_norm": 0.2336691915988922, + "learning_rate": 9.261126033826878e-05, + "loss": 0.0111, + "step": 21650 + }, + { + "epoch": 1.6823300970873787, + "grad_norm": 0.27686193585395813, + "learning_rate": 9.260260746591786e-05, + "loss": 0.009, + "step": 21660 + }, + { + "epoch": 1.6831067961165047, + "grad_norm": 0.30384474992752075, + "learning_rate": 9.259394993460985e-05, + "loss": 0.0117, + "step": 21670 + }, + { + "epoch": 1.683883495145631, + "grad_norm": 0.289278119802475, + "learning_rate": 9.258528774529151e-05, + "loss": 0.008, + "step": 21680 + }, + { + "epoch": 1.6846601941747572, + "grad_norm": 0.29690659046173096, + "learning_rate": 9.257662089891013e-05, + "loss": 0.0113, + "step": 21690 + }, + { + "epoch": 1.6854368932038835, + "grad_norm": 0.18047396838665009, + "learning_rate": 9.25679493964135e-05, + "loss": 0.0114, + "step": 21700 + }, + { + "epoch": 1.6862135922330097, + "grad_norm": 0.22638681530952454, + "learning_rate": 9.255927323874994e-05, + "loss": 0.0109, + "step": 21710 + }, + { + "epoch": 1.686990291262136, + "grad_norm": 0.39089328050613403, + "learning_rate": 9.255059242686822e-05, + "loss": 0.0103, + "step": 21720 + }, + { + "epoch": 1.6877669902912622, + "grad_norm": 0.18415148556232452, + "learning_rate": 9.254190696171769e-05, + "loss": 0.0117, + "step": 21730 + }, + { + "epoch": 1.6885436893203885, + "grad_norm": 0.27938827872276306, + "learning_rate": 9.25332168442482e-05, + "loss": 0.0086, + "step": 21740 + }, + { + "epoch": 1.6893203883495147, + "grad_norm": 0.4098835289478302, + "learning_rate": 9.252452207541004e-05, + "loss": 0.0097, + "step": 21750 + }, + { + "epoch": 1.6900970873786407, + "grad_norm": 0.31900542974472046, + "learning_rate": 9.251582265615409e-05, + "loss": 0.0097, + "step": 21760 + }, + { + "epoch": 1.690873786407767, + "grad_norm": 0.1471593827009201, + "learning_rate": 9.250711858743169e-05, + "loss": 0.0095, + "step": 21770 + }, + { + "epoch": 1.6916504854368932, + "grad_norm": 0.36370569467544556, + "learning_rate": 9.24984098701947e-05, + "loss": 0.0112, + "step": 21780 + }, + { + "epoch": 1.6924271844660193, + "grad_norm": 0.28746476769447327, + "learning_rate": 9.248969650539552e-05, + "loss": 0.0116, + "step": 21790 + }, + { + "epoch": 1.6932038834951455, + "grad_norm": 0.6260988116264343, + "learning_rate": 9.2480978493987e-05, + "loss": 0.0149, + "step": 21800 + }, + { + "epoch": 1.6939805825242717, + "grad_norm": 0.41353824734687805, + "learning_rate": 9.247225583692256e-05, + "loss": 0.0123, + "step": 21810 + }, + { + "epoch": 1.694757281553398, + "grad_norm": 0.28687232732772827, + "learning_rate": 9.246352853515607e-05, + "loss": 0.0116, + "step": 21820 + }, + { + "epoch": 1.6955339805825242, + "grad_norm": 0.4378269910812378, + "learning_rate": 9.245479658964194e-05, + "loss": 0.0121, + "step": 21830 + }, + { + "epoch": 1.6963106796116505, + "grad_norm": 0.3668811023235321, + "learning_rate": 9.244606000133507e-05, + "loss": 0.0122, + "step": 21840 + }, + { + "epoch": 1.6970873786407767, + "grad_norm": 0.18341311812400818, + "learning_rate": 9.24373187711909e-05, + "loss": 0.0102, + "step": 21850 + }, + { + "epoch": 1.697864077669903, + "grad_norm": 0.23925131559371948, + "learning_rate": 9.242857290016537e-05, + "loss": 0.0113, + "step": 21860 + }, + { + "epoch": 1.6986407766990292, + "grad_norm": 0.20874875783920288, + "learning_rate": 9.241982238921488e-05, + "loss": 0.0091, + "step": 21870 + }, + { + "epoch": 1.6994174757281555, + "grad_norm": 0.20086370408535004, + "learning_rate": 9.24110672392964e-05, + "loss": 0.0086, + "step": 21880 + }, + { + "epoch": 1.7001941747572815, + "grad_norm": 0.186046302318573, + "learning_rate": 9.240230745136737e-05, + "loss": 0.011, + "step": 21890 + }, + { + "epoch": 1.7009708737864078, + "grad_norm": 0.36652764678001404, + "learning_rate": 9.239354302638575e-05, + "loss": 0.0079, + "step": 21900 + }, + { + "epoch": 1.701747572815534, + "grad_norm": 0.23884235322475433, + "learning_rate": 9.238477396531e-05, + "loss": 0.0108, + "step": 21910 + }, + { + "epoch": 1.70252427184466, + "grad_norm": 0.2065657526254654, + "learning_rate": 9.23760002690991e-05, + "loss": 0.0113, + "step": 21920 + }, + { + "epoch": 1.7033009708737863, + "grad_norm": 0.37286576628685, + "learning_rate": 9.236722193871252e-05, + "loss": 0.0125, + "step": 21930 + }, + { + "epoch": 1.7040776699029125, + "grad_norm": 0.41154009103775024, + "learning_rate": 9.235843897511023e-05, + "loss": 0.0127, + "step": 21940 + }, + { + "epoch": 1.7048543689320388, + "grad_norm": 0.2862393856048584, + "learning_rate": 9.234965137925276e-05, + "loss": 0.0118, + "step": 21950 + }, + { + "epoch": 1.705631067961165, + "grad_norm": 0.314123272895813, + "learning_rate": 9.234085915210108e-05, + "loss": 0.0111, + "step": 21960 + }, + { + "epoch": 1.7064077669902913, + "grad_norm": 0.19322055578231812, + "learning_rate": 9.23320622946167e-05, + "loss": 0.0093, + "step": 21970 + }, + { + "epoch": 1.7071844660194175, + "grad_norm": 0.4596744179725647, + "learning_rate": 9.232326080776163e-05, + "loss": 0.0154, + "step": 21980 + }, + { + "epoch": 1.7079611650485438, + "grad_norm": 0.2937363386154175, + "learning_rate": 9.23144546924984e-05, + "loss": 0.014, + "step": 21990 + }, + { + "epoch": 1.70873786407767, + "grad_norm": 0.4633618891239166, + "learning_rate": 9.230564394979e-05, + "loss": 0.0107, + "step": 22000 + }, + { + "epoch": 1.7095145631067963, + "grad_norm": 0.3383917808532715, + "learning_rate": 9.22968285806e-05, + "loss": 0.0111, + "step": 22010 + }, + { + "epoch": 1.7102912621359223, + "grad_norm": 0.23791998624801636, + "learning_rate": 9.228800858589242e-05, + "loss": 0.0107, + "step": 22020 + }, + { + "epoch": 1.7110679611650486, + "grad_norm": 0.44418030977249146, + "learning_rate": 9.227918396663179e-05, + "loss": 0.0138, + "step": 22030 + }, + { + "epoch": 1.7118446601941748, + "grad_norm": 0.2892305552959442, + "learning_rate": 9.227035472378319e-05, + "loss": 0.0123, + "step": 22040 + }, + { + "epoch": 1.7126213592233008, + "grad_norm": 0.20302794873714447, + "learning_rate": 9.226152085831213e-05, + "loss": 0.0126, + "step": 22050 + }, + { + "epoch": 1.713398058252427, + "grad_norm": 0.3043377101421356, + "learning_rate": 9.22526823711847e-05, + "loss": 0.0089, + "step": 22060 + }, + { + "epoch": 1.7141747572815533, + "grad_norm": 0.1623103767633438, + "learning_rate": 9.224383926336745e-05, + "loss": 0.0092, + "step": 22070 + }, + { + "epoch": 1.7149514563106796, + "grad_norm": 0.2797333002090454, + "learning_rate": 9.223499153582744e-05, + "loss": 0.009, + "step": 22080 + }, + { + "epoch": 1.7157281553398058, + "grad_norm": 0.37937286496162415, + "learning_rate": 9.222613918953226e-05, + "loss": 0.0094, + "step": 22090 + }, + { + "epoch": 1.716504854368932, + "grad_norm": 0.5430834293365479, + "learning_rate": 9.221728222544999e-05, + "loss": 0.0108, + "step": 22100 + }, + { + "epoch": 1.7172815533980583, + "grad_norm": 0.3345825672149658, + "learning_rate": 9.22084206445492e-05, + "loss": 0.0113, + "step": 22110 + }, + { + "epoch": 1.7180582524271846, + "grad_norm": 0.2367641031742096, + "learning_rate": 9.2199554447799e-05, + "loss": 0.0131, + "step": 22120 + }, + { + "epoch": 1.7188349514563108, + "grad_norm": 0.2538117468357086, + "learning_rate": 9.219068363616897e-05, + "loss": 0.0124, + "step": 22130 + }, + { + "epoch": 1.7196116504854368, + "grad_norm": 0.24683308601379395, + "learning_rate": 9.218180821062919e-05, + "loss": 0.0112, + "step": 22140 + }, + { + "epoch": 1.720388349514563, + "grad_norm": 0.25192034244537354, + "learning_rate": 9.21729281721503e-05, + "loss": 0.015, + "step": 22150 + }, + { + "epoch": 1.7211650485436893, + "grad_norm": 0.1604170948266983, + "learning_rate": 9.216404352170339e-05, + "loss": 0.0113, + "step": 22160 + }, + { + "epoch": 1.7219417475728154, + "grad_norm": 0.321911096572876, + "learning_rate": 9.215515426026007e-05, + "loss": 0.0081, + "step": 22170 + }, + { + "epoch": 1.7227184466019416, + "grad_norm": 0.2581811547279358, + "learning_rate": 9.214626038879246e-05, + "loss": 0.0095, + "step": 22180 + }, + { + "epoch": 1.7234951456310679, + "grad_norm": 0.3411617577075958, + "learning_rate": 9.21373619082732e-05, + "loss": 0.0085, + "step": 22190 + }, + { + "epoch": 1.7242718446601941, + "grad_norm": 0.17328494787216187, + "learning_rate": 9.212845881967535e-05, + "loss": 0.0116, + "step": 22200 + }, + { + "epoch": 1.7250485436893204, + "grad_norm": 0.2725125551223755, + "learning_rate": 9.211955112397262e-05, + "loss": 0.0098, + "step": 22210 + }, + { + "epoch": 1.7258252427184466, + "grad_norm": 0.6789297461509705, + "learning_rate": 9.211063882213909e-05, + "loss": 0.0119, + "step": 22220 + }, + { + "epoch": 1.7266019417475729, + "grad_norm": 0.1812463402748108, + "learning_rate": 9.210172191514942e-05, + "loss": 0.0097, + "step": 22230 + }, + { + "epoch": 1.7273786407766991, + "grad_norm": 0.2157520055770874, + "learning_rate": 9.209280040397874e-05, + "loss": 0.0111, + "step": 22240 + }, + { + "epoch": 1.7281553398058254, + "grad_norm": 0.19295984506607056, + "learning_rate": 9.208387428960268e-05, + "loss": 0.0104, + "step": 22250 + }, + { + "epoch": 1.7289320388349516, + "grad_norm": 0.12601812183856964, + "learning_rate": 9.20749435729974e-05, + "loss": 0.009, + "step": 22260 + }, + { + "epoch": 1.7297087378640776, + "grad_norm": 0.20508895814418793, + "learning_rate": 9.206600825513957e-05, + "loss": 0.0083, + "step": 22270 + }, + { + "epoch": 1.7304854368932039, + "grad_norm": 0.14752072095870972, + "learning_rate": 9.20570683370063e-05, + "loss": 0.0073, + "step": 22280 + }, + { + "epoch": 1.7312621359223301, + "grad_norm": 0.22629468142986298, + "learning_rate": 9.204812381957528e-05, + "loss": 0.0089, + "step": 22290 + }, + { + "epoch": 1.7320388349514562, + "grad_norm": 0.2771843671798706, + "learning_rate": 9.203917470382465e-05, + "loss": 0.009, + "step": 22300 + }, + { + "epoch": 1.7328155339805824, + "grad_norm": 0.3673083782196045, + "learning_rate": 9.203022099073309e-05, + "loss": 0.0088, + "step": 22310 + }, + { + "epoch": 1.7335922330097087, + "grad_norm": 0.3308833837509155, + "learning_rate": 9.202126268127976e-05, + "loss": 0.0105, + "step": 22320 + }, + { + "epoch": 1.734368932038835, + "grad_norm": 0.31536388397216797, + "learning_rate": 9.20122997764443e-05, + "loss": 0.0118, + "step": 22330 + }, + { + "epoch": 1.7351456310679612, + "grad_norm": 0.26012274622917175, + "learning_rate": 9.200333227720692e-05, + "loss": 0.0103, + "step": 22340 + }, + { + "epoch": 1.7359223300970874, + "grad_norm": 0.20520132780075073, + "learning_rate": 9.199436018454826e-05, + "loss": 0.0122, + "step": 22350 + }, + { + "epoch": 1.7366990291262137, + "grad_norm": 0.3648604154586792, + "learning_rate": 9.198538349944952e-05, + "loss": 0.011, + "step": 22360 + }, + { + "epoch": 1.73747572815534, + "grad_norm": 0.41454029083251953, + "learning_rate": 9.197640222289234e-05, + "loss": 0.0123, + "step": 22370 + }, + { + "epoch": 1.7382524271844662, + "grad_norm": 0.24068580567836761, + "learning_rate": 9.196741635585895e-05, + "loss": 0.013, + "step": 22380 + }, + { + "epoch": 1.7390291262135922, + "grad_norm": 0.22070711851119995, + "learning_rate": 9.195842589933199e-05, + "loss": 0.0127, + "step": 22390 + }, + { + "epoch": 1.7398058252427184, + "grad_norm": 0.4507305324077606, + "learning_rate": 9.194943085429466e-05, + "loss": 0.0206, + "step": 22400 + }, + { + "epoch": 1.7405825242718447, + "grad_norm": 0.3716111183166504, + "learning_rate": 9.194043122173065e-05, + "loss": 0.0141, + "step": 22410 + }, + { + "epoch": 1.741359223300971, + "grad_norm": 0.15416587889194489, + "learning_rate": 9.193142700262413e-05, + "loss": 0.0118, + "step": 22420 + }, + { + "epoch": 1.742135922330097, + "grad_norm": 0.1948457956314087, + "learning_rate": 9.192241819795979e-05, + "loss": 0.0094, + "step": 22430 + }, + { + "epoch": 1.7429126213592232, + "grad_norm": 0.21277369558811188, + "learning_rate": 9.191340480872284e-05, + "loss": 0.0072, + "step": 22440 + }, + { + "epoch": 1.7436893203883495, + "grad_norm": 0.20595566928386688, + "learning_rate": 9.190438683589895e-05, + "loss": 0.0167, + "step": 22450 + }, + { + "epoch": 1.7444660194174757, + "grad_norm": 0.2341671884059906, + "learning_rate": 9.189536428047432e-05, + "loss": 0.0116, + "step": 22460 + }, + { + "epoch": 1.745242718446602, + "grad_norm": 0.21656185388565063, + "learning_rate": 9.188633714343564e-05, + "loss": 0.0134, + "step": 22470 + }, + { + "epoch": 1.7460194174757282, + "grad_norm": 0.36205458641052246, + "learning_rate": 9.18773054257701e-05, + "loss": 0.0112, + "step": 22480 + }, + { + "epoch": 1.7467961165048544, + "grad_norm": 0.6697400212287903, + "learning_rate": 9.18682691284654e-05, + "loss": 0.0131, + "step": 22490 + }, + { + "epoch": 1.7475728155339807, + "grad_norm": 0.3023332953453064, + "learning_rate": 9.185922825250974e-05, + "loss": 0.0114, + "step": 22500 + }, + { + "epoch": 1.748349514563107, + "grad_norm": 0.319563090801239, + "learning_rate": 9.185018279889181e-05, + "loss": 0.011, + "step": 22510 + }, + { + "epoch": 1.749126213592233, + "grad_norm": 0.3145526051521301, + "learning_rate": 9.184113276860082e-05, + "loss": 0.0125, + "step": 22520 + }, + { + "epoch": 1.7499029126213592, + "grad_norm": 0.24870535731315613, + "learning_rate": 9.183207816262645e-05, + "loss": 0.0121, + "step": 22530 + }, + { + "epoch": 1.7506796116504855, + "grad_norm": 0.2665894627571106, + "learning_rate": 9.182301898195891e-05, + "loss": 0.0096, + "step": 22540 + }, + { + "epoch": 1.7514563106796115, + "grad_norm": 0.31405436992645264, + "learning_rate": 9.181395522758889e-05, + "loss": 0.0089, + "step": 22550 + }, + { + "epoch": 1.7522330097087377, + "grad_norm": 0.33948540687561035, + "learning_rate": 9.180488690050759e-05, + "loss": 0.0124, + "step": 22560 + }, + { + "epoch": 1.753009708737864, + "grad_norm": 0.2555635869503021, + "learning_rate": 9.179581400170671e-05, + "loss": 0.0113, + "step": 22570 + }, + { + "epoch": 1.7537864077669902, + "grad_norm": 0.32515716552734375, + "learning_rate": 9.178673653217845e-05, + "loss": 0.0076, + "step": 22580 + }, + { + "epoch": 1.7545631067961165, + "grad_norm": 0.22643202543258667, + "learning_rate": 9.177765449291551e-05, + "loss": 0.008, + "step": 22590 + }, + { + "epoch": 1.7553398058252427, + "grad_norm": 0.1845671534538269, + "learning_rate": 9.176856788491109e-05, + "loss": 0.0123, + "step": 22600 + }, + { + "epoch": 1.756116504854369, + "grad_norm": 0.4158249795436859, + "learning_rate": 9.175947670915887e-05, + "loss": 0.0117, + "step": 22610 + }, + { + "epoch": 1.7568932038834952, + "grad_norm": 0.14715394377708435, + "learning_rate": 9.175038096665309e-05, + "loss": 0.0104, + "step": 22620 + }, + { + "epoch": 1.7576699029126215, + "grad_norm": 0.21999432146549225, + "learning_rate": 9.17412806583884e-05, + "loss": 0.0093, + "step": 22630 + }, + { + "epoch": 1.7584466019417477, + "grad_norm": 0.24753892421722412, + "learning_rate": 9.173217578536002e-05, + "loss": 0.0136, + "step": 22640 + }, + { + "epoch": 1.7592233009708738, + "grad_norm": 0.29826587438583374, + "learning_rate": 9.172306634856362e-05, + "loss": 0.0123, + "step": 22650 + }, + { + "epoch": 1.76, + "grad_norm": 0.2992755174636841, + "learning_rate": 9.171395234899545e-05, + "loss": 0.0119, + "step": 22660 + }, + { + "epoch": 1.7607766990291263, + "grad_norm": 0.23134292662143707, + "learning_rate": 9.170483378765214e-05, + "loss": 0.0124, + "step": 22670 + }, + { + "epoch": 1.7615533980582523, + "grad_norm": 0.4687230587005615, + "learning_rate": 9.169571066553091e-05, + "loss": 0.0113, + "step": 22680 + }, + { + "epoch": 1.7623300970873785, + "grad_norm": 0.2933005690574646, + "learning_rate": 9.168658298362946e-05, + "loss": 0.0124, + "step": 22690 + }, + { + "epoch": 1.7631067961165048, + "grad_norm": 0.272453248500824, + "learning_rate": 9.167745074294598e-05, + "loss": 0.0135, + "step": 22700 + }, + { + "epoch": 1.763883495145631, + "grad_norm": 0.2680216133594513, + "learning_rate": 9.166831394447913e-05, + "loss": 0.0119, + "step": 22710 + }, + { + "epoch": 1.7646601941747573, + "grad_norm": 0.20377568900585175, + "learning_rate": 9.165917258922812e-05, + "loss": 0.0102, + "step": 22720 + }, + { + "epoch": 1.7654368932038835, + "grad_norm": 0.42206090688705444, + "learning_rate": 9.165002667819262e-05, + "loss": 0.0122, + "step": 22730 + }, + { + "epoch": 1.7662135922330098, + "grad_norm": 0.183634415268898, + "learning_rate": 9.164087621237282e-05, + "loss": 0.0151, + "step": 22740 + }, + { + "epoch": 1.766990291262136, + "grad_norm": 0.16829320788383484, + "learning_rate": 9.163172119276942e-05, + "loss": 0.0102, + "step": 22750 + }, + { + "epoch": 1.7677669902912623, + "grad_norm": 0.2288396805524826, + "learning_rate": 9.162256162038358e-05, + "loss": 0.0133, + "step": 22760 + }, + { + "epoch": 1.7685436893203883, + "grad_norm": 0.43918606638908386, + "learning_rate": 9.161339749621698e-05, + "loss": 0.0099, + "step": 22770 + }, + { + "epoch": 1.7693203883495146, + "grad_norm": 0.4384485185146332, + "learning_rate": 9.160422882127177e-05, + "loss": 0.0119, + "step": 22780 + }, + { + "epoch": 1.7700970873786408, + "grad_norm": 0.22862038016319275, + "learning_rate": 9.159505559655069e-05, + "loss": 0.0107, + "step": 22790 + }, + { + "epoch": 1.7708737864077668, + "grad_norm": 0.3202110230922699, + "learning_rate": 9.158587782305684e-05, + "loss": 0.0108, + "step": 22800 + }, + { + "epoch": 1.771650485436893, + "grad_norm": 0.23610533773899078, + "learning_rate": 9.157669550179391e-05, + "loss": 0.0122, + "step": 22810 + }, + { + "epoch": 1.7724271844660193, + "grad_norm": 0.311308890581131, + "learning_rate": 9.156750863376609e-05, + "loss": 0.0116, + "step": 22820 + }, + { + "epoch": 1.7732038834951456, + "grad_norm": 0.2046850621700287, + "learning_rate": 9.155831721997801e-05, + "loss": 0.0111, + "step": 22830 + }, + { + "epoch": 1.7739805825242718, + "grad_norm": 0.2567538619041443, + "learning_rate": 9.154912126143484e-05, + "loss": 0.0078, + "step": 22840 + }, + { + "epoch": 1.774757281553398, + "grad_norm": 0.40574976801872253, + "learning_rate": 9.153992075914224e-05, + "loss": 0.0135, + "step": 22850 + }, + { + "epoch": 1.7755339805825243, + "grad_norm": 0.7392147779464722, + "learning_rate": 9.153071571410635e-05, + "loss": 0.0116, + "step": 22860 + }, + { + "epoch": 1.7763106796116506, + "grad_norm": 0.278291791677475, + "learning_rate": 9.152150612733384e-05, + "loss": 0.0112, + "step": 22870 + }, + { + "epoch": 1.7770873786407768, + "grad_norm": 0.1573210060596466, + "learning_rate": 9.151229199983184e-05, + "loss": 0.009, + "step": 22880 + }, + { + "epoch": 1.777864077669903, + "grad_norm": 0.2339610755443573, + "learning_rate": 9.150307333260802e-05, + "loss": 0.0104, + "step": 22890 + }, + { + "epoch": 1.778640776699029, + "grad_norm": 0.35488736629486084, + "learning_rate": 9.149385012667048e-05, + "loss": 0.011, + "step": 22900 + }, + { + "epoch": 1.7794174757281553, + "grad_norm": 0.3858167231082916, + "learning_rate": 9.148462238302788e-05, + "loss": 0.0118, + "step": 22910 + }, + { + "epoch": 1.7801941747572816, + "grad_norm": 0.2599697709083557, + "learning_rate": 9.147539010268936e-05, + "loss": 0.0122, + "step": 22920 + }, + { + "epoch": 1.7809708737864076, + "grad_norm": 0.28689783811569214, + "learning_rate": 9.14661532866645e-05, + "loss": 0.0118, + "step": 22930 + }, + { + "epoch": 1.7817475728155339, + "grad_norm": 0.25008007884025574, + "learning_rate": 9.145691193596348e-05, + "loss": 0.0072, + "step": 22940 + }, + { + "epoch": 1.7825242718446601, + "grad_norm": 0.38383257389068604, + "learning_rate": 9.144766605159691e-05, + "loss": 0.009, + "step": 22950 + }, + { + "epoch": 1.7833009708737864, + "grad_norm": 0.4403362572193146, + "learning_rate": 9.14384156345759e-05, + "loss": 0.0131, + "step": 22960 + }, + { + "epoch": 1.7840776699029126, + "grad_norm": 0.22079336643218994, + "learning_rate": 9.142916068591204e-05, + "loss": 0.0073, + "step": 22970 + }, + { + "epoch": 1.7848543689320389, + "grad_norm": 0.3584171533584595, + "learning_rate": 9.141990120661746e-05, + "loss": 0.0118, + "step": 22980 + }, + { + "epoch": 1.785631067961165, + "grad_norm": 0.26921600103378296, + "learning_rate": 9.141063719770475e-05, + "loss": 0.0096, + "step": 22990 + }, + { + "epoch": 1.7864077669902914, + "grad_norm": 0.3458331525325775, + "learning_rate": 9.140136866018704e-05, + "loss": 0.0105, + "step": 23000 + }, + { + "epoch": 1.7871844660194176, + "grad_norm": 0.3116539716720581, + "learning_rate": 9.139209559507788e-05, + "loss": 0.0119, + "step": 23010 + }, + { + "epoch": 1.7879611650485439, + "grad_norm": 0.28020793199539185, + "learning_rate": 9.13828180033914e-05, + "loss": 0.0107, + "step": 23020 + }, + { + "epoch": 1.7887378640776699, + "grad_norm": 0.2835920751094818, + "learning_rate": 9.137353588614212e-05, + "loss": 0.0094, + "step": 23030 + }, + { + "epoch": 1.7895145631067961, + "grad_norm": 0.2768464684486389, + "learning_rate": 9.136424924434519e-05, + "loss": 0.011, + "step": 23040 + }, + { + "epoch": 1.7902912621359224, + "grad_norm": 0.3186975121498108, + "learning_rate": 9.135495807901615e-05, + "loss": 0.0148, + "step": 23050 + }, + { + "epoch": 1.7910679611650484, + "grad_norm": 0.38047948479652405, + "learning_rate": 9.134566239117108e-05, + "loss": 0.0095, + "step": 23060 + }, + { + "epoch": 1.7918446601941747, + "grad_norm": 0.32194995880126953, + "learning_rate": 9.13363621818265e-05, + "loss": 0.0118, + "step": 23070 + }, + { + "epoch": 1.792621359223301, + "grad_norm": 0.1708296686410904, + "learning_rate": 9.132705745199953e-05, + "loss": 0.0118, + "step": 23080 + }, + { + "epoch": 1.7933980582524272, + "grad_norm": 0.25163325667381287, + "learning_rate": 9.131774820270768e-05, + "loss": 0.01, + "step": 23090 + }, + { + "epoch": 1.7941747572815534, + "grad_norm": 0.6383483409881592, + "learning_rate": 9.130843443496901e-05, + "loss": 0.0133, + "step": 23100 + }, + { + "epoch": 1.7949514563106796, + "grad_norm": 0.2917691469192505, + "learning_rate": 9.129911614980206e-05, + "loss": 0.0112, + "step": 23110 + }, + { + "epoch": 1.795728155339806, + "grad_norm": 0.17967094480991364, + "learning_rate": 9.128979334822584e-05, + "loss": 0.0097, + "step": 23120 + }, + { + "epoch": 1.7965048543689321, + "grad_norm": 0.37649205327033997, + "learning_rate": 9.128046603125992e-05, + "loss": 0.0093, + "step": 23130 + }, + { + "epoch": 1.7972815533980584, + "grad_norm": 0.2622484564781189, + "learning_rate": 9.12711341999243e-05, + "loss": 0.0119, + "step": 23140 + }, + { + "epoch": 1.7980582524271844, + "grad_norm": 0.8351167440414429, + "learning_rate": 9.12617978552395e-05, + "loss": 0.0122, + "step": 23150 + }, + { + "epoch": 1.7988349514563107, + "grad_norm": 0.3127365708351135, + "learning_rate": 9.12524569982265e-05, + "loss": 0.0114, + "step": 23160 + }, + { + "epoch": 1.799611650485437, + "grad_norm": 0.19516964256763458, + "learning_rate": 9.124311162990684e-05, + "loss": 0.0094, + "step": 23170 + }, + { + "epoch": 1.800388349514563, + "grad_norm": 0.23104628920555115, + "learning_rate": 9.12337617513025e-05, + "loss": 0.0129, + "step": 23180 + }, + { + "epoch": 1.8011650485436892, + "grad_norm": 0.23390421271324158, + "learning_rate": 9.122440736343596e-05, + "loss": 0.0101, + "step": 23190 + }, + { + "epoch": 1.8019417475728154, + "grad_norm": 0.3423613905906677, + "learning_rate": 9.12150484673302e-05, + "loss": 0.0107, + "step": 23200 + }, + { + "epoch": 1.8027184466019417, + "grad_norm": 0.42388659715652466, + "learning_rate": 9.120568506400873e-05, + "loss": 0.0132, + "step": 23210 + }, + { + "epoch": 1.803495145631068, + "grad_norm": 0.24719583988189697, + "learning_rate": 9.119631715449548e-05, + "loss": 0.0092, + "step": 23220 + }, + { + "epoch": 1.8042718446601942, + "grad_norm": 0.22980646789073944, + "learning_rate": 9.118694473981493e-05, + "loss": 0.0144, + "step": 23230 + }, + { + "epoch": 1.8050485436893204, + "grad_norm": 0.465053528547287, + "learning_rate": 9.117756782099203e-05, + "loss": 0.0133, + "step": 23240 + }, + { + "epoch": 1.8058252427184467, + "grad_norm": 0.28826791048049927, + "learning_rate": 9.11681863990522e-05, + "loss": 0.0112, + "step": 23250 + }, + { + "epoch": 1.806601941747573, + "grad_norm": 0.6123590469360352, + "learning_rate": 9.115880047502142e-05, + "loss": 0.0131, + "step": 23260 + }, + { + "epoch": 1.8073786407766992, + "grad_norm": 0.3585430085659027, + "learning_rate": 9.114941004992609e-05, + "loss": 0.0107, + "step": 23270 + }, + { + "epoch": 1.8081553398058252, + "grad_norm": 0.29511773586273193, + "learning_rate": 9.114001512479317e-05, + "loss": 0.0106, + "step": 23280 + }, + { + "epoch": 1.8089320388349515, + "grad_norm": 0.25701504945755005, + "learning_rate": 9.113061570065003e-05, + "loss": 0.0115, + "step": 23290 + }, + { + "epoch": 1.8097087378640777, + "grad_norm": 0.2606219947338104, + "learning_rate": 9.112121177852459e-05, + "loss": 0.0106, + "step": 23300 + }, + { + "epoch": 1.8104854368932037, + "grad_norm": 0.2163548767566681, + "learning_rate": 9.111180335944527e-05, + "loss": 0.008, + "step": 23310 + }, + { + "epoch": 1.81126213592233, + "grad_norm": 0.27611634135246277, + "learning_rate": 9.110239044444093e-05, + "loss": 0.0121, + "step": 23320 + }, + { + "epoch": 1.8120388349514562, + "grad_norm": 0.11816506087779999, + "learning_rate": 9.109297303454099e-05, + "loss": 0.0096, + "step": 23330 + }, + { + "epoch": 1.8128155339805825, + "grad_norm": 0.3408297002315521, + "learning_rate": 9.108355113077526e-05, + "loss": 0.0106, + "step": 23340 + }, + { + "epoch": 1.8135922330097087, + "grad_norm": 0.25762349367141724, + "learning_rate": 9.107412473417419e-05, + "loss": 0.0095, + "step": 23350 + }, + { + "epoch": 1.814368932038835, + "grad_norm": 0.47922226786613464, + "learning_rate": 9.106469384576858e-05, + "loss": 0.0112, + "step": 23360 + }, + { + "epoch": 1.8151456310679612, + "grad_norm": 0.20292668044567108, + "learning_rate": 9.105525846658978e-05, + "loss": 0.0101, + "step": 23370 + }, + { + "epoch": 1.8159223300970875, + "grad_norm": 0.229364275932312, + "learning_rate": 9.104581859766965e-05, + "loss": 0.0091, + "step": 23380 + }, + { + "epoch": 1.8166990291262137, + "grad_norm": 0.5076597332954407, + "learning_rate": 9.10363742400405e-05, + "loss": 0.0123, + "step": 23390 + }, + { + "epoch": 1.8174757281553398, + "grad_norm": 0.2363334596157074, + "learning_rate": 9.102692539473518e-05, + "loss": 0.0117, + "step": 23400 + }, + { + "epoch": 1.818252427184466, + "grad_norm": 0.3867867887020111, + "learning_rate": 9.101747206278697e-05, + "loss": 0.0134, + "step": 23410 + }, + { + "epoch": 1.8190291262135923, + "grad_norm": 0.3056931793689728, + "learning_rate": 9.100801424522968e-05, + "loss": 0.0093, + "step": 23420 + }, + { + "epoch": 1.8198058252427185, + "grad_norm": 0.1661452353000641, + "learning_rate": 9.099855194309762e-05, + "loss": 0.0096, + "step": 23430 + }, + { + "epoch": 1.8205825242718445, + "grad_norm": 0.14182595908641815, + "learning_rate": 9.098908515742554e-05, + "loss": 0.0122, + "step": 23440 + }, + { + "epoch": 1.8213592233009708, + "grad_norm": 0.13615281879901886, + "learning_rate": 9.097961388924873e-05, + "loss": 0.012, + "step": 23450 + }, + { + "epoch": 1.822135922330097, + "grad_norm": 0.2369363009929657, + "learning_rate": 9.097013813960298e-05, + "loss": 0.0113, + "step": 23460 + }, + { + "epoch": 1.8229126213592233, + "grad_norm": 0.3631996214389801, + "learning_rate": 9.09606579095245e-05, + "loss": 0.0129, + "step": 23470 + }, + { + "epoch": 1.8236893203883495, + "grad_norm": 0.30536001920700073, + "learning_rate": 9.095117320005008e-05, + "loss": 0.0101, + "step": 23480 + }, + { + "epoch": 1.8244660194174758, + "grad_norm": 0.26969602704048157, + "learning_rate": 9.094168401221691e-05, + "loss": 0.0101, + "step": 23490 + }, + { + "epoch": 1.825242718446602, + "grad_norm": 0.22058086097240448, + "learning_rate": 9.093219034706273e-05, + "loss": 0.0111, + "step": 23500 + }, + { + "epoch": 1.8260194174757283, + "grad_norm": 0.3291822373867035, + "learning_rate": 9.092269220562577e-05, + "loss": 0.0095, + "step": 23510 + }, + { + "epoch": 1.8267961165048545, + "grad_norm": 0.2546491324901581, + "learning_rate": 9.09131895889447e-05, + "loss": 0.0123, + "step": 23520 + }, + { + "epoch": 1.8275728155339805, + "grad_norm": 0.3047405779361725, + "learning_rate": 9.090368249805873e-05, + "loss": 0.0107, + "step": 23530 + }, + { + "epoch": 1.8283495145631068, + "grad_norm": 0.3782944977283478, + "learning_rate": 9.089417093400754e-05, + "loss": 0.0092, + "step": 23540 + }, + { + "epoch": 1.829126213592233, + "grad_norm": 0.5129823684692383, + "learning_rate": 9.088465489783131e-05, + "loss": 0.008, + "step": 23550 + }, + { + "epoch": 1.829902912621359, + "grad_norm": 0.14599476754665375, + "learning_rate": 9.087513439057068e-05, + "loss": 0.0087, + "step": 23560 + }, + { + "epoch": 1.8306796116504853, + "grad_norm": 0.17537981271743774, + "learning_rate": 9.08656094132668e-05, + "loss": 0.0089, + "step": 23570 + }, + { + "epoch": 1.8314563106796116, + "grad_norm": 0.27938005328178406, + "learning_rate": 9.085607996696134e-05, + "loss": 0.0133, + "step": 23580 + }, + { + "epoch": 1.8322330097087378, + "grad_norm": 0.30890098214149475, + "learning_rate": 9.084654605269639e-05, + "loss": 0.0135, + "step": 23590 + }, + { + "epoch": 1.833009708737864, + "grad_norm": 0.32830801606178284, + "learning_rate": 9.083700767151457e-05, + "loss": 0.0123, + "step": 23600 + }, + { + "epoch": 1.8337864077669903, + "grad_norm": 0.3524421155452728, + "learning_rate": 9.082746482445898e-05, + "loss": 0.0143, + "step": 23610 + }, + { + "epoch": 1.8345631067961166, + "grad_norm": 0.2510101795196533, + "learning_rate": 9.081791751257325e-05, + "loss": 0.0128, + "step": 23620 + }, + { + "epoch": 1.8353398058252428, + "grad_norm": 0.18593400716781616, + "learning_rate": 9.080836573690142e-05, + "loss": 0.0116, + "step": 23630 + }, + { + "epoch": 1.836116504854369, + "grad_norm": 0.19347108900547028, + "learning_rate": 9.079880949848805e-05, + "loss": 0.0113, + "step": 23640 + }, + { + "epoch": 1.8368932038834953, + "grad_norm": 0.24695166945457458, + "learning_rate": 9.078924879837822e-05, + "loss": 0.0116, + "step": 23650 + }, + { + "epoch": 1.8376699029126213, + "grad_norm": 0.3190532326698303, + "learning_rate": 9.077968363761747e-05, + "loss": 0.0114, + "step": 23660 + }, + { + "epoch": 1.8384466019417476, + "grad_norm": 0.3323638141155243, + "learning_rate": 9.077011401725182e-05, + "loss": 0.0123, + "step": 23670 + }, + { + "epoch": 1.8392233009708738, + "grad_norm": 0.4385032653808594, + "learning_rate": 9.07605399383278e-05, + "loss": 0.0097, + "step": 23680 + }, + { + "epoch": 1.8399999999999999, + "grad_norm": 0.24414420127868652, + "learning_rate": 9.075096140189243e-05, + "loss": 0.0094, + "step": 23690 + }, + { + "epoch": 1.840776699029126, + "grad_norm": 0.5521604418754578, + "learning_rate": 9.074137840899318e-05, + "loss": 0.0115, + "step": 23700 + }, + { + "epoch": 1.8415533980582524, + "grad_norm": 0.23896367847919464, + "learning_rate": 9.073179096067804e-05, + "loss": 0.0103, + "step": 23710 + }, + { + "epoch": 1.8423300970873786, + "grad_norm": 0.32184138894081116, + "learning_rate": 9.072219905799549e-05, + "loss": 0.0126, + "step": 23720 + }, + { + "epoch": 1.8431067961165049, + "grad_norm": 0.21912729740142822, + "learning_rate": 9.071260270199447e-05, + "loss": 0.0166, + "step": 23730 + }, + { + "epoch": 1.843883495145631, + "grad_norm": 0.2740424871444702, + "learning_rate": 9.070300189372441e-05, + "loss": 0.0084, + "step": 23740 + }, + { + "epoch": 1.8446601941747574, + "grad_norm": 0.22607213258743286, + "learning_rate": 9.069339663423528e-05, + "loss": 0.0116, + "step": 23750 + }, + { + "epoch": 1.8454368932038836, + "grad_norm": 0.2376834899187088, + "learning_rate": 9.068378692457747e-05, + "loss": 0.0115, + "step": 23760 + }, + { + "epoch": 1.8462135922330098, + "grad_norm": 0.17286567389965057, + "learning_rate": 9.067417276580189e-05, + "loss": 0.0106, + "step": 23770 + }, + { + "epoch": 1.8469902912621359, + "grad_norm": 0.4927041828632355, + "learning_rate": 9.066455415895993e-05, + "loss": 0.0108, + "step": 23780 + }, + { + "epoch": 1.8477669902912621, + "grad_norm": 0.4370918869972229, + "learning_rate": 9.065493110510346e-05, + "loss": 0.0108, + "step": 23790 + }, + { + "epoch": 1.8485436893203884, + "grad_norm": 0.37082675099372864, + "learning_rate": 9.064530360528484e-05, + "loss": 0.0114, + "step": 23800 + }, + { + "epoch": 1.8493203883495144, + "grad_norm": 0.38875406980514526, + "learning_rate": 9.063567166055695e-05, + "loss": 0.0102, + "step": 23810 + }, + { + "epoch": 1.8500970873786406, + "grad_norm": 0.3548334836959839, + "learning_rate": 9.062603527197308e-05, + "loss": 0.0098, + "step": 23820 + }, + { + "epoch": 1.850873786407767, + "grad_norm": 0.2255648970603943, + "learning_rate": 9.06163944405871e-05, + "loss": 0.0099, + "step": 23830 + }, + { + "epoch": 1.8516504854368931, + "grad_norm": 0.25740954279899597, + "learning_rate": 9.060674916745327e-05, + "loss": 0.0089, + "step": 23840 + }, + { + "epoch": 1.8524271844660194, + "grad_norm": 0.30771636962890625, + "learning_rate": 9.05970994536264e-05, + "loss": 0.0122, + "step": 23850 + }, + { + "epoch": 1.8532038834951456, + "grad_norm": 0.2984945476055145, + "learning_rate": 9.05874453001618e-05, + "loss": 0.0091, + "step": 23860 + }, + { + "epoch": 1.853980582524272, + "grad_norm": 0.3711586594581604, + "learning_rate": 9.057778670811517e-05, + "loss": 0.0093, + "step": 23870 + }, + { + "epoch": 1.8547572815533981, + "grad_norm": 0.2585851550102234, + "learning_rate": 9.056812367854281e-05, + "loss": 0.0125, + "step": 23880 + }, + { + "epoch": 1.8555339805825244, + "grad_norm": 0.22224102914333344, + "learning_rate": 9.055845621250143e-05, + "loss": 0.0102, + "step": 23890 + }, + { + "epoch": 1.8563106796116506, + "grad_norm": 0.24563992023468018, + "learning_rate": 9.054878431104825e-05, + "loss": 0.011, + "step": 23900 + }, + { + "epoch": 1.8570873786407767, + "grad_norm": 0.27054935693740845, + "learning_rate": 9.0539107975241e-05, + "loss": 0.0114, + "step": 23910 + }, + { + "epoch": 1.857864077669903, + "grad_norm": 0.2599163353443146, + "learning_rate": 9.052942720613784e-05, + "loss": 0.0146, + "step": 23920 + }, + { + "epoch": 1.8586407766990292, + "grad_norm": 0.3527103662490845, + "learning_rate": 9.051974200479745e-05, + "loss": 0.0125, + "step": 23930 + }, + { + "epoch": 1.8594174757281552, + "grad_norm": 0.29773446917533875, + "learning_rate": 9.051005237227901e-05, + "loss": 0.0109, + "step": 23940 + }, + { + "epoch": 1.8601941747572814, + "grad_norm": 0.8040980100631714, + "learning_rate": 9.050035830964215e-05, + "loss": 0.011, + "step": 23950 + }, + { + "epoch": 1.8609708737864077, + "grad_norm": 0.25013795495033264, + "learning_rate": 9.049065981794698e-05, + "loss": 0.0103, + "step": 23960 + }, + { + "epoch": 1.861747572815534, + "grad_norm": 0.3174832761287689, + "learning_rate": 9.048095689825414e-05, + "loss": 0.0083, + "step": 23970 + }, + { + "epoch": 1.8625242718446602, + "grad_norm": 0.24536262452602386, + "learning_rate": 9.047124955162472e-05, + "loss": 0.0088, + "step": 23980 + }, + { + "epoch": 1.8633009708737864, + "grad_norm": 0.2458130568265915, + "learning_rate": 9.046153777912028e-05, + "loss": 0.0123, + "step": 23990 + }, + { + "epoch": 1.8640776699029127, + "grad_norm": 0.5490935444831848, + "learning_rate": 9.045182158180292e-05, + "loss": 0.0121, + "step": 24000 + }, + { + "epoch": 1.864854368932039, + "grad_norm": 0.253305584192276, + "learning_rate": 9.044210096073516e-05, + "loss": 0.01, + "step": 24010 + }, + { + "epoch": 1.8656310679611652, + "grad_norm": 0.27363574504852295, + "learning_rate": 9.043237591698004e-05, + "loss": 0.0101, + "step": 24020 + }, + { + "epoch": 1.8664077669902914, + "grad_norm": 0.2932552397251129, + "learning_rate": 9.04226464516011e-05, + "loss": 0.0117, + "step": 24030 + }, + { + "epoch": 1.8671844660194175, + "grad_norm": 0.20899543166160583, + "learning_rate": 9.041291256566229e-05, + "loss": 0.0093, + "step": 24040 + }, + { + "epoch": 1.8679611650485437, + "grad_norm": 0.4257926940917969, + "learning_rate": 9.040317426022814e-05, + "loss": 0.0092, + "step": 24050 + }, + { + "epoch": 1.86873786407767, + "grad_norm": 0.4243113100528717, + "learning_rate": 9.03934315363636e-05, + "loss": 0.0085, + "step": 24060 + }, + { + "epoch": 1.869514563106796, + "grad_norm": 0.18580417335033417, + "learning_rate": 9.038368439513409e-05, + "loss": 0.0091, + "step": 24070 + }, + { + "epoch": 1.8702912621359222, + "grad_norm": 0.17787331342697144, + "learning_rate": 9.03739328376056e-05, + "loss": 0.0078, + "step": 24080 + }, + { + "epoch": 1.8710679611650485, + "grad_norm": 0.38382115960121155, + "learning_rate": 9.036417686484451e-05, + "loss": 0.0121, + "step": 24090 + }, + { + "epoch": 1.8718446601941747, + "grad_norm": 0.23837901651859283, + "learning_rate": 9.035441647791773e-05, + "loss": 0.0091, + "step": 24100 + }, + { + "epoch": 1.872621359223301, + "grad_norm": 0.2504710853099823, + "learning_rate": 9.034465167789263e-05, + "loss": 0.012, + "step": 24110 + }, + { + "epoch": 1.8733980582524272, + "grad_norm": 0.21944214403629303, + "learning_rate": 9.033488246583706e-05, + "loss": 0.0098, + "step": 24120 + }, + { + "epoch": 1.8741747572815535, + "grad_norm": 0.21448610723018646, + "learning_rate": 9.032510884281941e-05, + "loss": 0.0095, + "step": 24130 + }, + { + "epoch": 1.8749514563106797, + "grad_norm": 0.1985643357038498, + "learning_rate": 9.031533080990848e-05, + "loss": 0.0102, + "step": 24140 + }, + { + "epoch": 1.875728155339806, + "grad_norm": 0.3456476032733917, + "learning_rate": 9.030554836817358e-05, + "loss": 0.0081, + "step": 24150 + }, + { + "epoch": 1.876504854368932, + "grad_norm": 0.20296484231948853, + "learning_rate": 9.029576151868451e-05, + "loss": 0.0112, + "step": 24160 + }, + { + "epoch": 1.8772815533980582, + "grad_norm": 0.3550910949707031, + "learning_rate": 9.028597026251155e-05, + "loss": 0.0108, + "step": 24170 + }, + { + "epoch": 1.8780582524271845, + "grad_norm": 0.4802013337612152, + "learning_rate": 9.027617460072547e-05, + "loss": 0.0096, + "step": 24180 + }, + { + "epoch": 1.8788349514563105, + "grad_norm": 0.32149091362953186, + "learning_rate": 9.026637453439745e-05, + "loss": 0.0093, + "step": 24190 + }, + { + "epoch": 1.8796116504854368, + "grad_norm": 0.2031804770231247, + "learning_rate": 9.025657006459927e-05, + "loss": 0.0089, + "step": 24200 + }, + { + "epoch": 1.880388349514563, + "grad_norm": 0.175175741314888, + "learning_rate": 9.024676119240311e-05, + "loss": 0.0073, + "step": 24210 + }, + { + "epoch": 1.8811650485436893, + "grad_norm": 0.3109890818595886, + "learning_rate": 9.023694791888166e-05, + "loss": 0.0084, + "step": 24220 + }, + { + "epoch": 1.8819417475728155, + "grad_norm": 0.26898953318595886, + "learning_rate": 9.022713024510808e-05, + "loss": 0.0092, + "step": 24230 + }, + { + "epoch": 1.8827184466019418, + "grad_norm": 0.35818180441856384, + "learning_rate": 9.021730817215601e-05, + "loss": 0.0118, + "step": 24240 + }, + { + "epoch": 1.883495145631068, + "grad_norm": 0.18773244321346283, + "learning_rate": 9.02074817010996e-05, + "loss": 0.0099, + "step": 24250 + }, + { + "epoch": 1.8842718446601943, + "grad_norm": 0.24171724915504456, + "learning_rate": 9.019765083301342e-05, + "loss": 0.0125, + "step": 24260 + }, + { + "epoch": 1.8850485436893205, + "grad_norm": 0.16721276938915253, + "learning_rate": 9.01878155689726e-05, + "loss": 0.0103, + "step": 24270 + }, + { + "epoch": 1.8858252427184468, + "grad_norm": 0.27394548058509827, + "learning_rate": 9.017797591005268e-05, + "loss": 0.0105, + "step": 24280 + }, + { + "epoch": 1.8866019417475728, + "grad_norm": 0.47194963693618774, + "learning_rate": 9.016813185732972e-05, + "loss": 0.01, + "step": 24290 + }, + { + "epoch": 1.887378640776699, + "grad_norm": 0.30758360028266907, + "learning_rate": 9.015828341188027e-05, + "loss": 0.0106, + "step": 24300 + }, + { + "epoch": 1.8881553398058253, + "grad_norm": 0.2955356538295746, + "learning_rate": 9.01484305747813e-05, + "loss": 0.0094, + "step": 24310 + }, + { + "epoch": 1.8889320388349513, + "grad_norm": 0.24199821054935455, + "learning_rate": 9.013857334711033e-05, + "loss": 0.0104, + "step": 24320 + }, + { + "epoch": 1.8897087378640776, + "grad_norm": 0.2515270709991455, + "learning_rate": 9.012871172994534e-05, + "loss": 0.0073, + "step": 24330 + }, + { + "epoch": 1.8904854368932038, + "grad_norm": 0.21196117997169495, + "learning_rate": 9.011884572436476e-05, + "loss": 0.0108, + "step": 24340 + }, + { + "epoch": 1.89126213592233, + "grad_norm": 0.18872804939746857, + "learning_rate": 9.010897533144754e-05, + "loss": 0.0096, + "step": 24350 + }, + { + "epoch": 1.8920388349514563, + "grad_norm": 0.38968461751937866, + "learning_rate": 9.009910055227306e-05, + "loss": 0.0105, + "step": 24360 + }, + { + "epoch": 1.8928155339805826, + "grad_norm": 0.24989347159862518, + "learning_rate": 9.008922138792124e-05, + "loss": 0.0077, + "step": 24370 + }, + { + "epoch": 1.8935922330097088, + "grad_norm": 0.2411191761493683, + "learning_rate": 9.007933783947244e-05, + "loss": 0.0084, + "step": 24380 + }, + { + "epoch": 1.894368932038835, + "grad_norm": 0.24339668452739716, + "learning_rate": 9.006944990800752e-05, + "loss": 0.0148, + "step": 24390 + }, + { + "epoch": 1.8951456310679613, + "grad_norm": 0.27159640192985535, + "learning_rate": 9.005955759460779e-05, + "loss": 0.0092, + "step": 24400 + }, + { + "epoch": 1.8959223300970873, + "grad_norm": 0.5513491034507751, + "learning_rate": 9.004966090035508e-05, + "loss": 0.0112, + "step": 24410 + }, + { + "epoch": 1.8966990291262136, + "grad_norm": 0.7910110950469971, + "learning_rate": 9.003975982633166e-05, + "loss": 0.015, + "step": 24420 + }, + { + "epoch": 1.8974757281553398, + "grad_norm": 0.2569427788257599, + "learning_rate": 9.00298543736203e-05, + "loss": 0.0095, + "step": 24430 + }, + { + "epoch": 1.898252427184466, + "grad_norm": 0.24996568262577057, + "learning_rate": 9.001994454330427e-05, + "loss": 0.0088, + "step": 24440 + }, + { + "epoch": 1.899029126213592, + "grad_norm": 0.2329368144273758, + "learning_rate": 9.001003033646727e-05, + "loss": 0.0115, + "step": 24450 + }, + { + "epoch": 1.8998058252427183, + "grad_norm": 0.16572806239128113, + "learning_rate": 9.00001117541935e-05, + "loss": 0.0103, + "step": 24460 + }, + { + "epoch": 1.9005825242718446, + "grad_norm": 0.41291263699531555, + "learning_rate": 8.999018879756764e-05, + "loss": 0.0118, + "step": 24470 + }, + { + "epoch": 1.9013592233009708, + "grad_norm": 0.23406653106212616, + "learning_rate": 8.998026146767487e-05, + "loss": 0.0095, + "step": 24480 + }, + { + "epoch": 1.902135922330097, + "grad_norm": 0.2221105545759201, + "learning_rate": 8.99703297656008e-05, + "loss": 0.0086, + "step": 24490 + }, + { + "epoch": 1.9029126213592233, + "grad_norm": 0.21892380714416504, + "learning_rate": 8.996039369243156e-05, + "loss": 0.0101, + "step": 24500 + }, + { + "epoch": 1.9036893203883496, + "grad_norm": 0.5681636929512024, + "learning_rate": 8.995045324925378e-05, + "loss": 0.0076, + "step": 24510 + }, + { + "epoch": 1.9044660194174758, + "grad_norm": 0.12799863517284393, + "learning_rate": 8.994050843715448e-05, + "loss": 0.008, + "step": 24520 + }, + { + "epoch": 1.905242718446602, + "grad_norm": 0.6063812971115112, + "learning_rate": 8.993055925722121e-05, + "loss": 0.0115, + "step": 24530 + }, + { + "epoch": 1.9060194174757281, + "grad_norm": 0.41086146235466003, + "learning_rate": 8.992060571054202e-05, + "loss": 0.014, + "step": 24540 + }, + { + "epoch": 1.9067961165048544, + "grad_norm": 0.2269793450832367, + "learning_rate": 8.991064779820542e-05, + "loss": 0.0105, + "step": 24550 + }, + { + "epoch": 1.9075728155339806, + "grad_norm": 0.23183397948741913, + "learning_rate": 8.990068552130036e-05, + "loss": 0.0112, + "step": 24560 + }, + { + "epoch": 1.9083495145631066, + "grad_norm": 0.23440879583358765, + "learning_rate": 8.989071888091634e-05, + "loss": 0.0105, + "step": 24570 + }, + { + "epoch": 1.909126213592233, + "grad_norm": 0.2445647120475769, + "learning_rate": 8.988074787814329e-05, + "loss": 0.0108, + "step": 24580 + }, + { + "epoch": 1.9099029126213591, + "grad_norm": 0.16711175441741943, + "learning_rate": 8.987077251407158e-05, + "loss": 0.0097, + "step": 24590 + }, + { + "epoch": 1.9106796116504854, + "grad_norm": 0.36956676840782166, + "learning_rate": 8.986079278979216e-05, + "loss": 0.0111, + "step": 24600 + }, + { + "epoch": 1.9114563106796116, + "grad_norm": 0.14931683242321014, + "learning_rate": 8.985080870639635e-05, + "loss": 0.0109, + "step": 24610 + }, + { + "epoch": 1.9122330097087379, + "grad_norm": 0.34268006682395935, + "learning_rate": 8.984082026497603e-05, + "loss": 0.0132, + "step": 24620 + }, + { + "epoch": 1.9130097087378641, + "grad_norm": 0.21891778707504272, + "learning_rate": 8.98308274666235e-05, + "loss": 0.0101, + "step": 24630 + }, + { + "epoch": 1.9137864077669904, + "grad_norm": 0.3406601548194885, + "learning_rate": 8.982083031243155e-05, + "loss": 0.0091, + "step": 24640 + }, + { + "epoch": 1.9145631067961166, + "grad_norm": 0.23556014895439148, + "learning_rate": 8.98108288034935e-05, + "loss": 0.0083, + "step": 24650 + }, + { + "epoch": 1.9153398058252429, + "grad_norm": 0.3041257858276367, + "learning_rate": 8.980082294090305e-05, + "loss": 0.0117, + "step": 24660 + }, + { + "epoch": 1.916116504854369, + "grad_norm": 0.21751345694065094, + "learning_rate": 8.979081272575443e-05, + "loss": 0.0127, + "step": 24670 + }, + { + "epoch": 1.9168932038834952, + "grad_norm": 0.2625728249549866, + "learning_rate": 8.978079815914236e-05, + "loss": 0.0133, + "step": 24680 + }, + { + "epoch": 1.9176699029126214, + "grad_norm": 0.35791918635368347, + "learning_rate": 8.977077924216202e-05, + "loss": 0.014, + "step": 24690 + }, + { + "epoch": 1.9184466019417474, + "grad_norm": 0.18931543827056885, + "learning_rate": 8.976075597590905e-05, + "loss": 0.0095, + "step": 24700 + }, + { + "epoch": 1.9192233009708737, + "grad_norm": 0.3889313340187073, + "learning_rate": 8.975072836147958e-05, + "loss": 0.0116, + "step": 24710 + }, + { + "epoch": 1.92, + "grad_norm": 0.29314056038856506, + "learning_rate": 8.974069639997025e-05, + "loss": 0.0146, + "step": 24720 + }, + { + "epoch": 1.9207766990291262, + "grad_norm": 0.23205384612083435, + "learning_rate": 8.973066009247808e-05, + "loss": 0.0088, + "step": 24730 + }, + { + "epoch": 1.9215533980582524, + "grad_norm": 0.23826105892658234, + "learning_rate": 8.972061944010066e-05, + "loss": 0.0083, + "step": 24740 + }, + { + "epoch": 1.9223300970873787, + "grad_norm": 0.19725842773914337, + "learning_rate": 8.971057444393603e-05, + "loss": 0.0092, + "step": 24750 + }, + { + "epoch": 1.923106796116505, + "grad_norm": 0.19362252950668335, + "learning_rate": 8.970052510508268e-05, + "loss": 0.0066, + "step": 24760 + }, + { + "epoch": 1.9238834951456312, + "grad_norm": 0.17156580090522766, + "learning_rate": 8.969047142463959e-05, + "loss": 0.0082, + "step": 24770 + }, + { + "epoch": 1.9246601941747574, + "grad_norm": 0.28662851452827454, + "learning_rate": 8.968041340370621e-05, + "loss": 0.0102, + "step": 24780 + }, + { + "epoch": 1.9254368932038834, + "grad_norm": 0.23702706396579742, + "learning_rate": 8.96703510433825e-05, + "loss": 0.0083, + "step": 24790 + }, + { + "epoch": 1.9262135922330097, + "grad_norm": 0.2764281630516052, + "learning_rate": 8.966028434476883e-05, + "loss": 0.0079, + "step": 24800 + }, + { + "epoch": 1.926990291262136, + "grad_norm": 0.43214869499206543, + "learning_rate": 8.96502133089661e-05, + "loss": 0.0084, + "step": 24810 + }, + { + "epoch": 1.927766990291262, + "grad_norm": 0.22674016654491425, + "learning_rate": 8.964013793707564e-05, + "loss": 0.0087, + "step": 24820 + }, + { + "epoch": 1.9285436893203882, + "grad_norm": 0.32721197605133057, + "learning_rate": 8.963005823019932e-05, + "loss": 0.0083, + "step": 24830 + }, + { + "epoch": 1.9293203883495145, + "grad_norm": 0.48149412870407104, + "learning_rate": 8.961997418943939e-05, + "loss": 0.0085, + "step": 24840 + }, + { + "epoch": 1.9300970873786407, + "grad_norm": 0.29451608657836914, + "learning_rate": 8.960988581589865e-05, + "loss": 0.0098, + "step": 24850 + }, + { + "epoch": 1.930873786407767, + "grad_norm": 1.31987726688385, + "learning_rate": 8.959979311068037e-05, + "loss": 0.0085, + "step": 24860 + }, + { + "epoch": 1.9316504854368932, + "grad_norm": 0.22894135117530823, + "learning_rate": 8.958969607488823e-05, + "loss": 0.009, + "step": 24870 + }, + { + "epoch": 1.9324271844660195, + "grad_norm": 0.2518688440322876, + "learning_rate": 8.957959470962647e-05, + "loss": 0.0109, + "step": 24880 + }, + { + "epoch": 1.9332038834951457, + "grad_norm": 0.2435186356306076, + "learning_rate": 8.956948901599971e-05, + "loss": 0.0084, + "step": 24890 + }, + { + "epoch": 1.933980582524272, + "grad_norm": 0.44692128896713257, + "learning_rate": 8.955937899511315e-05, + "loss": 0.0094, + "step": 24900 + }, + { + "epoch": 1.9347572815533982, + "grad_norm": 0.2279718518257141, + "learning_rate": 8.954926464807238e-05, + "loss": 0.011, + "step": 24910 + }, + { + "epoch": 1.9355339805825242, + "grad_norm": 0.15589946508407593, + "learning_rate": 8.953914597598347e-05, + "loss": 0.0096, + "step": 24920 + }, + { + "epoch": 1.9363106796116505, + "grad_norm": 0.2334069162607193, + "learning_rate": 8.952902297995303e-05, + "loss": 0.0098, + "step": 24930 + }, + { + "epoch": 1.9370873786407767, + "grad_norm": 0.21970222890377045, + "learning_rate": 8.951889566108804e-05, + "loss": 0.012, + "step": 24940 + }, + { + "epoch": 1.9378640776699028, + "grad_norm": 0.30755746364593506, + "learning_rate": 8.950876402049606e-05, + "loss": 0.0121, + "step": 24950 + }, + { + "epoch": 1.938640776699029, + "grad_norm": 0.22506771981716156, + "learning_rate": 8.949862805928504e-05, + "loss": 0.0087, + "step": 24960 + }, + { + "epoch": 1.9394174757281553, + "grad_norm": 0.24096764624118805, + "learning_rate": 8.948848777856343e-05, + "loss": 0.0165, + "step": 24970 + }, + { + "epoch": 1.9401941747572815, + "grad_norm": 0.30084428191185, + "learning_rate": 8.947834317944017e-05, + "loss": 0.0117, + "step": 24980 + }, + { + "epoch": 1.9409708737864078, + "grad_norm": 0.3110416829586029, + "learning_rate": 8.946819426302466e-05, + "loss": 0.0087, + "step": 24990 + }, + { + "epoch": 1.941747572815534, + "grad_norm": 0.25158095359802246, + "learning_rate": 8.945804103042676e-05, + "loss": 0.0093, + "step": 25000 + }, + { + "epoch": 1.9425242718446603, + "grad_norm": 0.5961136221885681, + "learning_rate": 8.944788348275681e-05, + "loss": 0.0136, + "step": 25010 + }, + { + "epoch": 1.9433009708737865, + "grad_norm": 0.32038572430610657, + "learning_rate": 8.943772162112565e-05, + "loss": 0.0081, + "step": 25020 + }, + { + "epoch": 1.9440776699029128, + "grad_norm": 0.3786918818950653, + "learning_rate": 8.942755544664454e-05, + "loss": 0.0099, + "step": 25030 + }, + { + "epoch": 1.9448543689320388, + "grad_norm": 0.2405775636434555, + "learning_rate": 8.941738496042525e-05, + "loss": 0.0135, + "step": 25040 + }, + { + "epoch": 1.945631067961165, + "grad_norm": 0.5464943051338196, + "learning_rate": 8.940721016357999e-05, + "loss": 0.0127, + "step": 25050 + }, + { + "epoch": 1.9464077669902913, + "grad_norm": 0.2548221945762634, + "learning_rate": 8.939703105722148e-05, + "loss": 0.0104, + "step": 25060 + }, + { + "epoch": 1.9471844660194175, + "grad_norm": 0.19346153736114502, + "learning_rate": 8.93868476424629e-05, + "loss": 0.0095, + "step": 25070 + }, + { + "epoch": 1.9479611650485436, + "grad_norm": 0.25373053550720215, + "learning_rate": 8.937665992041786e-05, + "loss": 0.0093, + "step": 25080 + }, + { + "epoch": 1.9487378640776698, + "grad_norm": 0.2683076560497284, + "learning_rate": 8.93664678922005e-05, + "loss": 0.0093, + "step": 25090 + }, + { + "epoch": 1.949514563106796, + "grad_norm": 0.43705153465270996, + "learning_rate": 8.93562715589254e-05, + "loss": 0.0127, + "step": 25100 + }, + { + "epoch": 1.9502912621359223, + "grad_norm": 0.24162530899047852, + "learning_rate": 8.934607092170762e-05, + "loss": 0.0112, + "step": 25110 + }, + { + "epoch": 1.9510679611650485, + "grad_norm": 0.3522210121154785, + "learning_rate": 8.933586598166266e-05, + "loss": 0.0106, + "step": 25120 + }, + { + "epoch": 1.9518446601941748, + "grad_norm": 0.4260611832141876, + "learning_rate": 8.932565673990655e-05, + "loss": 0.0112, + "step": 25130 + }, + { + "epoch": 1.952621359223301, + "grad_norm": 0.3170967102050781, + "learning_rate": 8.931544319755574e-05, + "loss": 0.0127, + "step": 25140 + }, + { + "epoch": 1.9533980582524273, + "grad_norm": 0.4107160270214081, + "learning_rate": 8.930522535572718e-05, + "loss": 0.0125, + "step": 25150 + }, + { + "epoch": 1.9541747572815535, + "grad_norm": 0.29937395453453064, + "learning_rate": 8.929500321553826e-05, + "loss": 0.0127, + "step": 25160 + }, + { + "epoch": 1.9549514563106796, + "grad_norm": 0.2743000090122223, + "learning_rate": 8.928477677810686e-05, + "loss": 0.01, + "step": 25170 + }, + { + "epoch": 1.9557281553398058, + "grad_norm": 0.6999452710151672, + "learning_rate": 8.927454604455137e-05, + "loss": 0.008, + "step": 25180 + }, + { + "epoch": 1.956504854368932, + "grad_norm": 0.4728536903858185, + "learning_rate": 8.926431101599053e-05, + "loss": 0.0099, + "step": 25190 + }, + { + "epoch": 1.957281553398058, + "grad_norm": 0.3196612596511841, + "learning_rate": 8.925407169354369e-05, + "loss": 0.0087, + "step": 25200 + }, + { + "epoch": 1.9580582524271843, + "grad_norm": 0.24019098281860352, + "learning_rate": 8.92438280783306e-05, + "loss": 0.0085, + "step": 25210 + }, + { + "epoch": 1.9588349514563106, + "grad_norm": 0.33533599972724915, + "learning_rate": 8.923358017147146e-05, + "loss": 0.0086, + "step": 25220 + }, + { + "epoch": 1.9596116504854368, + "grad_norm": 0.279203325510025, + "learning_rate": 8.922332797408697e-05, + "loss": 0.0084, + "step": 25230 + }, + { + "epoch": 1.960388349514563, + "grad_norm": 0.3880096673965454, + "learning_rate": 8.921307148729831e-05, + "loss": 0.011, + "step": 25240 + }, + { + "epoch": 1.9611650485436893, + "grad_norm": 0.45331913232803345, + "learning_rate": 8.920281071222712e-05, + "loss": 0.013, + "step": 25250 + }, + { + "epoch": 1.9619417475728156, + "grad_norm": 0.6693563461303711, + "learning_rate": 8.919254564999548e-05, + "loss": 0.0093, + "step": 25260 + }, + { + "epoch": 1.9627184466019418, + "grad_norm": 0.3069773018360138, + "learning_rate": 8.918227630172598e-05, + "loss": 0.0091, + "step": 25270 + }, + { + "epoch": 1.963495145631068, + "grad_norm": 0.2801685035228729, + "learning_rate": 8.917200266854165e-05, + "loss": 0.0084, + "step": 25280 + }, + { + "epoch": 1.9642718446601943, + "grad_norm": 0.28048834204673767, + "learning_rate": 8.9161724751566e-05, + "loss": 0.0081, + "step": 25290 + }, + { + "epoch": 1.9650485436893204, + "grad_norm": 0.23335999250411987, + "learning_rate": 8.915144255192302e-05, + "loss": 0.0087, + "step": 25300 + }, + { + "epoch": 1.9658252427184466, + "grad_norm": 0.22390495240688324, + "learning_rate": 8.914115607073714e-05, + "loss": 0.0098, + "step": 25310 + }, + { + "epoch": 1.9666019417475729, + "grad_norm": 0.2845410406589508, + "learning_rate": 8.913086530913327e-05, + "loss": 0.01, + "step": 25320 + }, + { + "epoch": 1.9673786407766989, + "grad_norm": 0.34402137994766235, + "learning_rate": 8.912057026823681e-05, + "loss": 0.0132, + "step": 25330 + }, + { + "epoch": 1.9681553398058251, + "grad_norm": 0.30076852440834045, + "learning_rate": 8.91102709491736e-05, + "loss": 0.012, + "step": 25340 + }, + { + "epoch": 1.9689320388349514, + "grad_norm": 0.21421772241592407, + "learning_rate": 8.909996735306996e-05, + "loss": 0.012, + "step": 25350 + }, + { + "epoch": 1.9697087378640776, + "grad_norm": 0.16789229214191437, + "learning_rate": 8.908965948105268e-05, + "loss": 0.0097, + "step": 25360 + }, + { + "epoch": 1.9704854368932039, + "grad_norm": 0.24587883055210114, + "learning_rate": 8.907934733424901e-05, + "loss": 0.0111, + "step": 25370 + }, + { + "epoch": 1.9712621359223301, + "grad_norm": 0.20621439814567566, + "learning_rate": 8.906903091378666e-05, + "loss": 0.0102, + "step": 25380 + }, + { + "epoch": 1.9720388349514564, + "grad_norm": 0.7951043844223022, + "learning_rate": 8.905871022079384e-05, + "loss": 0.0093, + "step": 25390 + }, + { + "epoch": 1.9728155339805826, + "grad_norm": 0.2097780555486679, + "learning_rate": 8.90483852563992e-05, + "loss": 0.009, + "step": 25400 + }, + { + "epoch": 1.9735922330097089, + "grad_norm": 0.20771747827529907, + "learning_rate": 8.903805602173185e-05, + "loss": 0.0105, + "step": 25410 + }, + { + "epoch": 1.974368932038835, + "grad_norm": 0.19166171550750732, + "learning_rate": 8.902772251792137e-05, + "loss": 0.0109, + "step": 25420 + }, + { + "epoch": 1.9751456310679611, + "grad_norm": 0.21086691319942474, + "learning_rate": 8.901738474609786e-05, + "loss": 0.0109, + "step": 25430 + }, + { + "epoch": 1.9759223300970874, + "grad_norm": 0.38631999492645264, + "learning_rate": 8.900704270739179e-05, + "loss": 0.0116, + "step": 25440 + }, + { + "epoch": 1.9766990291262136, + "grad_norm": 0.4686657786369324, + "learning_rate": 8.89966964029342e-05, + "loss": 0.0096, + "step": 25450 + }, + { + "epoch": 1.9774757281553397, + "grad_norm": 0.18823473155498505, + "learning_rate": 8.898634583385652e-05, + "loss": 0.0129, + "step": 25460 + }, + { + "epoch": 1.978252427184466, + "grad_norm": 0.33728769421577454, + "learning_rate": 8.897599100129065e-05, + "loss": 0.0083, + "step": 25470 + }, + { + "epoch": 1.9790291262135922, + "grad_norm": 0.3295867443084717, + "learning_rate": 8.896563190636903e-05, + "loss": 0.0097, + "step": 25480 + }, + { + "epoch": 1.9798058252427184, + "grad_norm": 0.29112064838409424, + "learning_rate": 8.895526855022448e-05, + "loss": 0.0091, + "step": 25490 + }, + { + "epoch": 1.9805825242718447, + "grad_norm": 0.29285329580307007, + "learning_rate": 8.894490093399033e-05, + "loss": 0.0108, + "step": 25500 + }, + { + "epoch": 1.981359223300971, + "grad_norm": 0.24603575468063354, + "learning_rate": 8.893452905880035e-05, + "loss": 0.0123, + "step": 25510 + }, + { + "epoch": 1.9821359223300972, + "grad_norm": 0.25666549801826477, + "learning_rate": 8.892415292578883e-05, + "loss": 0.009, + "step": 25520 + }, + { + "epoch": 1.9829126213592234, + "grad_norm": 0.2912655770778656, + "learning_rate": 8.891377253609046e-05, + "loss": 0.0121, + "step": 25530 + }, + { + "epoch": 1.9836893203883497, + "grad_norm": 0.3097323775291443, + "learning_rate": 8.890338789084043e-05, + "loss": 0.0086, + "step": 25540 + }, + { + "epoch": 1.9844660194174757, + "grad_norm": 0.2728452980518341, + "learning_rate": 8.88929989911744e-05, + "loss": 0.0102, + "step": 25550 + }, + { + "epoch": 1.985242718446602, + "grad_norm": 0.4667101204395294, + "learning_rate": 8.888260583822847e-05, + "loss": 0.0109, + "step": 25560 + }, + { + "epoch": 1.9860194174757282, + "grad_norm": 0.2156187742948532, + "learning_rate": 8.887220843313921e-05, + "loss": 0.0089, + "step": 25570 + }, + { + "epoch": 1.9867961165048542, + "grad_norm": 0.17711329460144043, + "learning_rate": 8.88618067770437e-05, + "loss": 0.0063, + "step": 25580 + }, + { + "epoch": 1.9875728155339805, + "grad_norm": 0.21006780862808228, + "learning_rate": 8.885140087107942e-05, + "loss": 0.0105, + "step": 25590 + }, + { + "epoch": 1.9883495145631067, + "grad_norm": 0.2906184792518616, + "learning_rate": 8.884099071638436e-05, + "loss": 0.012, + "step": 25600 + }, + { + "epoch": 1.989126213592233, + "grad_norm": 0.967646062374115, + "learning_rate": 8.883057631409695e-05, + "loss": 0.0115, + "step": 25610 + }, + { + "epoch": 1.9899029126213592, + "grad_norm": 0.3093593716621399, + "learning_rate": 8.882015766535608e-05, + "loss": 0.0081, + "step": 25620 + }, + { + "epoch": 1.9906796116504855, + "grad_norm": 0.2026798129081726, + "learning_rate": 8.880973477130115e-05, + "loss": 0.0103, + "step": 25630 + }, + { + "epoch": 1.9914563106796117, + "grad_norm": 0.17328740656375885, + "learning_rate": 8.879930763307197e-05, + "loss": 0.0099, + "step": 25640 + }, + { + "epoch": 1.992233009708738, + "grad_norm": 0.16256333887577057, + "learning_rate": 8.878887625180884e-05, + "loss": 0.0114, + "step": 25650 + }, + { + "epoch": 1.9930097087378642, + "grad_norm": 0.26584890484809875, + "learning_rate": 8.877844062865253e-05, + "loss": 0.0105, + "step": 25660 + }, + { + "epoch": 1.9937864077669905, + "grad_norm": 0.3334271013736725, + "learning_rate": 8.876800076474424e-05, + "loss": 0.0107, + "step": 25670 + }, + { + "epoch": 1.9945631067961165, + "grad_norm": 0.22593723237514496, + "learning_rate": 8.875755666122568e-05, + "loss": 0.0112, + "step": 25680 + }, + { + "epoch": 1.9953398058252427, + "grad_norm": 0.45445889234542847, + "learning_rate": 8.8747108319239e-05, + "loss": 0.0111, + "step": 25690 + }, + { + "epoch": 1.996116504854369, + "grad_norm": 0.2449665069580078, + "learning_rate": 8.87366557399268e-05, + "loss": 0.0122, + "step": 25700 + }, + { + "epoch": 1.996893203883495, + "grad_norm": 0.22817136347293854, + "learning_rate": 8.872619892443217e-05, + "loss": 0.0086, + "step": 25710 + }, + { + "epoch": 1.9976699029126213, + "grad_norm": 0.287654310464859, + "learning_rate": 8.871573787389865e-05, + "loss": 0.0103, + "step": 25720 + }, + { + "epoch": 1.9984466019417475, + "grad_norm": 0.3307591676712036, + "learning_rate": 8.870527258947024e-05, + "loss": 0.0104, + "step": 25730 + }, + { + "epoch": 1.9992233009708738, + "grad_norm": 0.15839941799640656, + "learning_rate": 8.869480307229143e-05, + "loss": 0.0123, + "step": 25740 + }, + { + "epoch": 2.0, + "grad_norm": 0.3330371081829071, + "learning_rate": 8.868432932350712e-05, + "loss": 0.0097, + "step": 25750 + }, + { + "epoch": 2.0007766990291262, + "grad_norm": 0.8655486702919006, + "learning_rate": 8.867385134426272e-05, + "loss": 0.0118, + "step": 25760 + }, + { + "epoch": 2.0015533980582525, + "grad_norm": 0.3326025605201721, + "learning_rate": 8.866336913570407e-05, + "loss": 0.0101, + "step": 25770 + }, + { + "epoch": 2.0023300970873787, + "grad_norm": 0.226521834731102, + "learning_rate": 8.865288269897751e-05, + "loss": 0.0125, + "step": 25780 + }, + { + "epoch": 2.003106796116505, + "grad_norm": 0.244436576962471, + "learning_rate": 8.864239203522981e-05, + "loss": 0.0084, + "step": 25790 + }, + { + "epoch": 2.0038834951456312, + "grad_norm": 0.33748000860214233, + "learning_rate": 8.863189714560822e-05, + "loss": 0.0096, + "step": 25800 + }, + { + "epoch": 2.0046601941747575, + "grad_norm": 0.2909768223762512, + "learning_rate": 8.862139803126043e-05, + "loss": 0.0091, + "step": 25810 + }, + { + "epoch": 2.0054368932038833, + "grad_norm": 0.26858770847320557, + "learning_rate": 8.861089469333463e-05, + "loss": 0.0097, + "step": 25820 + }, + { + "epoch": 2.0062135922330095, + "grad_norm": 0.4998282492160797, + "learning_rate": 8.860038713297944e-05, + "loss": 0.0107, + "step": 25830 + }, + { + "epoch": 2.006990291262136, + "grad_norm": 0.39736172556877136, + "learning_rate": 8.858987535134394e-05, + "loss": 0.012, + "step": 25840 + }, + { + "epoch": 2.007766990291262, + "grad_norm": 0.5087403655052185, + "learning_rate": 8.857935934957769e-05, + "loss": 0.0124, + "step": 25850 + }, + { + "epoch": 2.0085436893203883, + "grad_norm": 0.29262638092041016, + "learning_rate": 8.856883912883071e-05, + "loss": 0.0098, + "step": 25860 + }, + { + "epoch": 2.0093203883495145, + "grad_norm": 0.36906981468200684, + "learning_rate": 8.855831469025346e-05, + "loss": 0.01, + "step": 25870 + }, + { + "epoch": 2.010097087378641, + "grad_norm": 0.2324717789888382, + "learning_rate": 8.854778603499689e-05, + "loss": 0.009, + "step": 25880 + }, + { + "epoch": 2.010873786407767, + "grad_norm": 0.23415371775627136, + "learning_rate": 8.85372531642124e-05, + "loss": 0.0117, + "step": 25890 + }, + { + "epoch": 2.0116504854368933, + "grad_norm": 0.32003217935562134, + "learning_rate": 8.852671607905185e-05, + "loss": 0.0096, + "step": 25900 + }, + { + "epoch": 2.0124271844660195, + "grad_norm": 0.1559273898601532, + "learning_rate": 8.851617478066754e-05, + "loss": 0.01, + "step": 25910 + }, + { + "epoch": 2.013203883495146, + "grad_norm": 0.17121383547782898, + "learning_rate": 8.850562927021227e-05, + "loss": 0.0095, + "step": 25920 + }, + { + "epoch": 2.013980582524272, + "grad_norm": 0.1762336790561676, + "learning_rate": 8.849507954883928e-05, + "loss": 0.0106, + "step": 25930 + }, + { + "epoch": 2.014757281553398, + "grad_norm": 0.39683112502098083, + "learning_rate": 8.848452561770226e-05, + "loss": 0.0083, + "step": 25940 + }, + { + "epoch": 2.015533980582524, + "grad_norm": 0.4877319633960724, + "learning_rate": 8.847396747795538e-05, + "loss": 0.0108, + "step": 25950 + }, + { + "epoch": 2.0163106796116503, + "grad_norm": 0.3053849935531616, + "learning_rate": 8.846340513075327e-05, + "loss": 0.0119, + "step": 25960 + }, + { + "epoch": 2.0170873786407766, + "grad_norm": 0.3509213328361511, + "learning_rate": 8.845283857725099e-05, + "loss": 0.012, + "step": 25970 + }, + { + "epoch": 2.017864077669903, + "grad_norm": 0.35338473320007324, + "learning_rate": 8.844226781860409e-05, + "loss": 0.0121, + "step": 25980 + }, + { + "epoch": 2.018640776699029, + "grad_norm": 0.21119998395442963, + "learning_rate": 8.84316928559686e-05, + "loss": 0.0125, + "step": 25990 + }, + { + "epoch": 2.0194174757281553, + "grad_norm": 0.42717644572257996, + "learning_rate": 8.842111369050094e-05, + "loss": 0.0079, + "step": 26000 + }, + { + "epoch": 2.0201941747572816, + "grad_norm": 0.2820080518722534, + "learning_rate": 8.841053032335808e-05, + "loss": 0.012, + "step": 26010 + }, + { + "epoch": 2.020970873786408, + "grad_norm": 0.3426004648208618, + "learning_rate": 8.839994275569735e-05, + "loss": 0.0091, + "step": 26020 + }, + { + "epoch": 2.021747572815534, + "grad_norm": 0.9124481678009033, + "learning_rate": 8.838935098867662e-05, + "loss": 0.0126, + "step": 26030 + }, + { + "epoch": 2.0225242718446603, + "grad_norm": 0.4599023759365082, + "learning_rate": 8.837875502345418e-05, + "loss": 0.011, + "step": 26040 + }, + { + "epoch": 2.0233009708737866, + "grad_norm": 0.306731253862381, + "learning_rate": 8.83681548611888e-05, + "loss": 0.0091, + "step": 26050 + }, + { + "epoch": 2.024077669902913, + "grad_norm": 0.4248434901237488, + "learning_rate": 8.835755050303969e-05, + "loss": 0.0143, + "step": 26060 + }, + { + "epoch": 2.0248543689320386, + "grad_norm": 0.3465054929256439, + "learning_rate": 8.834694195016653e-05, + "loss": 0.0083, + "step": 26070 + }, + { + "epoch": 2.025631067961165, + "grad_norm": 0.23080940544605255, + "learning_rate": 8.833632920372942e-05, + "loss": 0.0108, + "step": 26080 + }, + { + "epoch": 2.026407766990291, + "grad_norm": 0.29702433943748474, + "learning_rate": 8.832571226488903e-05, + "loss": 0.0105, + "step": 26090 + }, + { + "epoch": 2.0271844660194174, + "grad_norm": 0.16922031342983246, + "learning_rate": 8.831509113480634e-05, + "loss": 0.0109, + "step": 26100 + }, + { + "epoch": 2.0279611650485436, + "grad_norm": 0.2401813119649887, + "learning_rate": 8.83044658146429e-05, + "loss": 0.0096, + "step": 26110 + }, + { + "epoch": 2.02873786407767, + "grad_norm": 0.15932728350162506, + "learning_rate": 8.829383630556067e-05, + "loss": 0.0086, + "step": 26120 + }, + { + "epoch": 2.029514563106796, + "grad_norm": 0.2417396903038025, + "learning_rate": 8.828320260872207e-05, + "loss": 0.0102, + "step": 26130 + }, + { + "epoch": 2.0302912621359224, + "grad_norm": 0.35507839918136597, + "learning_rate": 8.827256472529e-05, + "loss": 0.0112, + "step": 26140 + }, + { + "epoch": 2.0310679611650486, + "grad_norm": 0.35317716002464294, + "learning_rate": 8.826192265642778e-05, + "loss": 0.0112, + "step": 26150 + }, + { + "epoch": 2.031844660194175, + "grad_norm": 0.2193492203950882, + "learning_rate": 8.825127640329923e-05, + "loss": 0.008, + "step": 26160 + }, + { + "epoch": 2.032621359223301, + "grad_norm": 0.2101997584104538, + "learning_rate": 8.824062596706861e-05, + "loss": 0.0114, + "step": 26170 + }, + { + "epoch": 2.0333980582524274, + "grad_norm": 0.26057693362236023, + "learning_rate": 8.822997134890062e-05, + "loss": 0.0098, + "step": 26180 + }, + { + "epoch": 2.0341747572815536, + "grad_norm": 0.23473291099071503, + "learning_rate": 8.821931254996044e-05, + "loss": 0.0094, + "step": 26190 + }, + { + "epoch": 2.0349514563106794, + "grad_norm": 0.32664602994918823, + "learning_rate": 8.82086495714137e-05, + "loss": 0.0102, + "step": 26200 + }, + { + "epoch": 2.0357281553398057, + "grad_norm": 0.20993676781654358, + "learning_rate": 8.81979824144265e-05, + "loss": 0.0099, + "step": 26210 + }, + { + "epoch": 2.036504854368932, + "grad_norm": 0.2620920240879059, + "learning_rate": 8.818731108016536e-05, + "loss": 0.0086, + "step": 26220 + }, + { + "epoch": 2.037281553398058, + "grad_norm": 0.16866694390773773, + "learning_rate": 8.81766355697973e-05, + "loss": 0.011, + "step": 26230 + }, + { + "epoch": 2.0380582524271844, + "grad_norm": 0.20653703808784485, + "learning_rate": 8.816595588448977e-05, + "loss": 0.0104, + "step": 26240 + }, + { + "epoch": 2.0388349514563107, + "grad_norm": 0.19779467582702637, + "learning_rate": 8.81552720254107e-05, + "loss": 0.0114, + "step": 26250 + }, + { + "epoch": 2.039611650485437, + "grad_norm": 0.7528218626976013, + "learning_rate": 8.814458399372842e-05, + "loss": 0.01, + "step": 26260 + }, + { + "epoch": 2.040388349514563, + "grad_norm": 0.2388061285018921, + "learning_rate": 8.813389179061181e-05, + "loss": 0.0093, + "step": 26270 + }, + { + "epoch": 2.0411650485436894, + "grad_norm": 0.3618921935558319, + "learning_rate": 8.812319541723012e-05, + "loss": 0.0081, + "step": 26280 + }, + { + "epoch": 2.0419417475728157, + "grad_norm": 0.4222402572631836, + "learning_rate": 8.811249487475309e-05, + "loss": 0.0121, + "step": 26290 + }, + { + "epoch": 2.042718446601942, + "grad_norm": 0.39422014355659485, + "learning_rate": 8.810179016435092e-05, + "loss": 0.0086, + "step": 26300 + }, + { + "epoch": 2.043495145631068, + "grad_norm": 0.2976618707180023, + "learning_rate": 8.809108128719428e-05, + "loss": 0.008, + "step": 26310 + }, + { + "epoch": 2.044271844660194, + "grad_norm": 0.18888849020004272, + "learning_rate": 8.808036824445424e-05, + "loss": 0.0099, + "step": 26320 + }, + { + "epoch": 2.04504854368932, + "grad_norm": 0.20306001603603363, + "learning_rate": 8.806965103730238e-05, + "loss": 0.0113, + "step": 26330 + }, + { + "epoch": 2.0458252427184465, + "grad_norm": 0.32885611057281494, + "learning_rate": 8.805892966691074e-05, + "loss": 0.0118, + "step": 26340 + }, + { + "epoch": 2.0466019417475727, + "grad_norm": 0.3941752016544342, + "learning_rate": 8.804820413445175e-05, + "loss": 0.0139, + "step": 26350 + }, + { + "epoch": 2.047378640776699, + "grad_norm": 0.3086930811405182, + "learning_rate": 8.803747444109837e-05, + "loss": 0.0123, + "step": 26360 + }, + { + "epoch": 2.048155339805825, + "grad_norm": 0.3382115364074707, + "learning_rate": 8.802674058802399e-05, + "loss": 0.0124, + "step": 26370 + }, + { + "epoch": 2.0489320388349515, + "grad_norm": 0.27042800188064575, + "learning_rate": 8.801600257640241e-05, + "loss": 0.0127, + "step": 26380 + }, + { + "epoch": 2.0497087378640777, + "grad_norm": 0.30346232652664185, + "learning_rate": 8.800526040740795e-05, + "loss": 0.0107, + "step": 26390 + }, + { + "epoch": 2.050485436893204, + "grad_norm": 0.26836585998535156, + "learning_rate": 8.799451408221535e-05, + "loss": 0.0112, + "step": 26400 + }, + { + "epoch": 2.05126213592233, + "grad_norm": 0.2833973169326782, + "learning_rate": 8.798376360199982e-05, + "loss": 0.0094, + "step": 26410 + }, + { + "epoch": 2.0520388349514564, + "grad_norm": 0.20374715328216553, + "learning_rate": 8.797300896793701e-05, + "loss": 0.0082, + "step": 26420 + }, + { + "epoch": 2.0528155339805827, + "grad_norm": 0.28640905022621155, + "learning_rate": 8.796225018120302e-05, + "loss": 0.011, + "step": 26430 + }, + { + "epoch": 2.053592233009709, + "grad_norm": 0.24600860476493835, + "learning_rate": 8.795148724297444e-05, + "loss": 0.008, + "step": 26440 + }, + { + "epoch": 2.0543689320388347, + "grad_norm": 0.2098691761493683, + "learning_rate": 8.794072015442825e-05, + "loss": 0.0133, + "step": 26450 + }, + { + "epoch": 2.055145631067961, + "grad_norm": 0.19284644722938538, + "learning_rate": 8.792994891674198e-05, + "loss": 0.0118, + "step": 26460 + }, + { + "epoch": 2.0559223300970872, + "grad_norm": 0.1683252453804016, + "learning_rate": 8.79191735310935e-05, + "loss": 0.0088, + "step": 26470 + }, + { + "epoch": 2.0566990291262135, + "grad_norm": 0.2446446418762207, + "learning_rate": 8.790839399866122e-05, + "loss": 0.0116, + "step": 26480 + }, + { + "epoch": 2.0574757281553397, + "grad_norm": 0.19525524973869324, + "learning_rate": 8.789761032062397e-05, + "loss": 0.0093, + "step": 26490 + }, + { + "epoch": 2.058252427184466, + "grad_norm": 0.28249654173851013, + "learning_rate": 8.788682249816103e-05, + "loss": 0.0107, + "step": 26500 + }, + { + "epoch": 2.0590291262135922, + "grad_norm": 0.2917088568210602, + "learning_rate": 8.787603053245215e-05, + "loss": 0.013, + "step": 26510 + }, + { + "epoch": 2.0598058252427185, + "grad_norm": 0.33616310358047485, + "learning_rate": 8.78652344246775e-05, + "loss": 0.0131, + "step": 26520 + }, + { + "epoch": 2.0605825242718447, + "grad_norm": 0.41611969470977783, + "learning_rate": 8.785443417601776e-05, + "loss": 0.0083, + "step": 26530 + }, + { + "epoch": 2.061359223300971, + "grad_norm": 0.22880424559116364, + "learning_rate": 8.784362978765401e-05, + "loss": 0.0121, + "step": 26540 + }, + { + "epoch": 2.0621359223300972, + "grad_norm": 0.25020021200180054, + "learning_rate": 8.783282126076779e-05, + "loss": 0.0137, + "step": 26550 + }, + { + "epoch": 2.0629126213592235, + "grad_norm": 0.19580639898777008, + "learning_rate": 8.782200859654112e-05, + "loss": 0.0084, + "step": 26560 + }, + { + "epoch": 2.0636893203883497, + "grad_norm": 0.23694510757923126, + "learning_rate": 8.781119179615646e-05, + "loss": 0.0119, + "step": 26570 + }, + { + "epoch": 2.0644660194174755, + "grad_norm": 0.25917884707450867, + "learning_rate": 8.780037086079674e-05, + "loss": 0.0113, + "step": 26580 + }, + { + "epoch": 2.065242718446602, + "grad_norm": 0.29604440927505493, + "learning_rate": 8.778954579164527e-05, + "loss": 0.0108, + "step": 26590 + }, + { + "epoch": 2.066019417475728, + "grad_norm": 0.21829672157764435, + "learning_rate": 8.777871658988588e-05, + "loss": 0.0116, + "step": 26600 + }, + { + "epoch": 2.0667961165048543, + "grad_norm": 0.24866285920143127, + "learning_rate": 8.776788325670285e-05, + "loss": 0.0117, + "step": 26610 + }, + { + "epoch": 2.0675728155339805, + "grad_norm": 0.18554557859897614, + "learning_rate": 8.775704579328089e-05, + "loss": 0.0103, + "step": 26620 + }, + { + "epoch": 2.068349514563107, + "grad_norm": 0.22283433377742767, + "learning_rate": 8.774620420080517e-05, + "loss": 0.0103, + "step": 26630 + }, + { + "epoch": 2.069126213592233, + "grad_norm": 0.301207572221756, + "learning_rate": 8.773535848046131e-05, + "loss": 0.0139, + "step": 26640 + }, + { + "epoch": 2.0699029126213593, + "grad_norm": 0.33335405588150024, + "learning_rate": 8.772450863343538e-05, + "loss": 0.0127, + "step": 26650 + }, + { + "epoch": 2.0706796116504855, + "grad_norm": 0.286865234375, + "learning_rate": 8.77136546609139e-05, + "loss": 0.0098, + "step": 26660 + }, + { + "epoch": 2.071456310679612, + "grad_norm": 0.2774697542190552, + "learning_rate": 8.770279656408385e-05, + "loss": 0.0105, + "step": 26670 + }, + { + "epoch": 2.072233009708738, + "grad_norm": 0.23086364567279816, + "learning_rate": 8.769193434413265e-05, + "loss": 0.0092, + "step": 26680 + }, + { + "epoch": 2.0730097087378643, + "grad_norm": 0.2664088010787964, + "learning_rate": 8.76810680022482e-05, + "loss": 0.0089, + "step": 26690 + }, + { + "epoch": 2.07378640776699, + "grad_norm": 0.24288155138492584, + "learning_rate": 8.767019753961878e-05, + "loss": 0.0088, + "step": 26700 + }, + { + "epoch": 2.0745631067961163, + "grad_norm": 0.25371643900871277, + "learning_rate": 8.765932295743321e-05, + "loss": 0.0128, + "step": 26710 + }, + { + "epoch": 2.0753398058252426, + "grad_norm": 0.1864880472421646, + "learning_rate": 8.764844425688068e-05, + "loss": 0.0124, + "step": 26720 + }, + { + "epoch": 2.076116504854369, + "grad_norm": 0.4515987038612366, + "learning_rate": 8.763756143915092e-05, + "loss": 0.0141, + "step": 26730 + }, + { + "epoch": 2.076893203883495, + "grad_norm": 0.36724841594696045, + "learning_rate": 8.7626674505434e-05, + "loss": 0.0116, + "step": 26740 + }, + { + "epoch": 2.0776699029126213, + "grad_norm": 0.21332351863384247, + "learning_rate": 8.761578345692053e-05, + "loss": 0.0091, + "step": 26750 + }, + { + "epoch": 2.0784466019417476, + "grad_norm": 0.5818511843681335, + "learning_rate": 8.760488829480156e-05, + "loss": 0.0097, + "step": 26760 + }, + { + "epoch": 2.079223300970874, + "grad_norm": 0.20983514189720154, + "learning_rate": 8.759398902026854e-05, + "loss": 0.0096, + "step": 26770 + }, + { + "epoch": 2.08, + "grad_norm": 0.24876774847507477, + "learning_rate": 8.758308563451339e-05, + "loss": 0.01, + "step": 26780 + }, + { + "epoch": 2.0807766990291263, + "grad_norm": 0.28654634952545166, + "learning_rate": 8.75721781387285e-05, + "loss": 0.0102, + "step": 26790 + }, + { + "epoch": 2.0815533980582526, + "grad_norm": 0.20360247790813446, + "learning_rate": 8.75612665341067e-05, + "loss": 0.0094, + "step": 26800 + }, + { + "epoch": 2.082330097087379, + "grad_norm": 0.2834315299987793, + "learning_rate": 8.755035082184126e-05, + "loss": 0.0097, + "step": 26810 + }, + { + "epoch": 2.083106796116505, + "grad_norm": 0.29335933923721313, + "learning_rate": 8.753943100312592e-05, + "loss": 0.0094, + "step": 26820 + }, + { + "epoch": 2.083883495145631, + "grad_norm": 0.27878129482269287, + "learning_rate": 8.752850707915484e-05, + "loss": 0.0113, + "step": 26830 + }, + { + "epoch": 2.084660194174757, + "grad_norm": 0.25845786929130554, + "learning_rate": 8.751757905112264e-05, + "loss": 0.0124, + "step": 26840 + }, + { + "epoch": 2.0854368932038834, + "grad_norm": 0.348503053188324, + "learning_rate": 8.75066469202244e-05, + "loss": 0.0106, + "step": 26850 + }, + { + "epoch": 2.0862135922330096, + "grad_norm": 0.32364094257354736, + "learning_rate": 8.749571068765567e-05, + "loss": 0.011, + "step": 26860 + }, + { + "epoch": 2.086990291262136, + "grad_norm": 0.2500811219215393, + "learning_rate": 8.748477035461238e-05, + "loss": 0.0105, + "step": 26870 + }, + { + "epoch": 2.087766990291262, + "grad_norm": 0.29453006386756897, + "learning_rate": 8.747382592229095e-05, + "loss": 0.0117, + "step": 26880 + }, + { + "epoch": 2.0885436893203884, + "grad_norm": 0.3557249903678894, + "learning_rate": 8.746287739188828e-05, + "loss": 0.0101, + "step": 26890 + }, + { + "epoch": 2.0893203883495146, + "grad_norm": 0.16051267087459564, + "learning_rate": 8.745192476460165e-05, + "loss": 0.0107, + "step": 26900 + }, + { + "epoch": 2.090097087378641, + "grad_norm": 0.23714977502822876, + "learning_rate": 8.744096804162882e-05, + "loss": 0.0107, + "step": 26910 + }, + { + "epoch": 2.090873786407767, + "grad_norm": 0.3624429702758789, + "learning_rate": 8.743000722416804e-05, + "loss": 0.0092, + "step": 26920 + }, + { + "epoch": 2.0916504854368934, + "grad_norm": 0.2772027850151062, + "learning_rate": 8.741904231341793e-05, + "loss": 0.0102, + "step": 26930 + }, + { + "epoch": 2.0924271844660196, + "grad_norm": 0.4357546865940094, + "learning_rate": 8.740807331057762e-05, + "loss": 0.0155, + "step": 26940 + }, + { + "epoch": 2.0932038834951454, + "grad_norm": 0.28510040044784546, + "learning_rate": 8.739710021684667e-05, + "loss": 0.0116, + "step": 26950 + }, + { + "epoch": 2.0939805825242717, + "grad_norm": 0.260579377412796, + "learning_rate": 8.738612303342503e-05, + "loss": 0.0126, + "step": 26960 + }, + { + "epoch": 2.094757281553398, + "grad_norm": 0.29062512516975403, + "learning_rate": 8.73751417615132e-05, + "loss": 0.0096, + "step": 26970 + }, + { + "epoch": 2.095533980582524, + "grad_norm": 0.7106029391288757, + "learning_rate": 8.736415640231208e-05, + "loss": 0.0112, + "step": 26980 + }, + { + "epoch": 2.0963106796116504, + "grad_norm": 0.37009575963020325, + "learning_rate": 8.735316695702297e-05, + "loss": 0.0159, + "step": 26990 + }, + { + "epoch": 2.0970873786407767, + "grad_norm": 0.22133243083953857, + "learning_rate": 8.734217342684769e-05, + "loss": 0.0112, + "step": 27000 + }, + { + "epoch": 2.097864077669903, + "grad_norm": 0.3324931859970093, + "learning_rate": 8.733117581298847e-05, + "loss": 0.01, + "step": 27010 + }, + { + "epoch": 2.098640776699029, + "grad_norm": 0.3007976710796356, + "learning_rate": 8.732017411664796e-05, + "loss": 0.01, + "step": 27020 + }, + { + "epoch": 2.0994174757281554, + "grad_norm": 0.23545752465724945, + "learning_rate": 8.730916833902936e-05, + "loss": 0.0082, + "step": 27030 + }, + { + "epoch": 2.1001941747572817, + "grad_norm": 0.3735288381576538, + "learning_rate": 8.729815848133618e-05, + "loss": 0.0097, + "step": 27040 + }, + { + "epoch": 2.100970873786408, + "grad_norm": 0.1872875988483429, + "learning_rate": 8.728714454477247e-05, + "loss": 0.008, + "step": 27050 + }, + { + "epoch": 2.101747572815534, + "grad_norm": 0.1904950886964798, + "learning_rate": 8.727612653054269e-05, + "loss": 0.0128, + "step": 27060 + }, + { + "epoch": 2.1025242718446604, + "grad_norm": 0.31322792172431946, + "learning_rate": 8.726510443985176e-05, + "loss": 0.0111, + "step": 27070 + }, + { + "epoch": 2.103300970873786, + "grad_norm": 0.39726412296295166, + "learning_rate": 8.725407827390503e-05, + "loss": 0.0113, + "step": 27080 + }, + { + "epoch": 2.1040776699029125, + "grad_norm": 0.33558252453804016, + "learning_rate": 8.724304803390833e-05, + "loss": 0.0102, + "step": 27090 + }, + { + "epoch": 2.1048543689320387, + "grad_norm": 0.2727872133255005, + "learning_rate": 8.723201372106788e-05, + "loss": 0.0125, + "step": 27100 + }, + { + "epoch": 2.105631067961165, + "grad_norm": 0.34416520595550537, + "learning_rate": 8.722097533659038e-05, + "loss": 0.0084, + "step": 27110 + }, + { + "epoch": 2.106407766990291, + "grad_norm": 0.20320375263690948, + "learning_rate": 8.720993288168299e-05, + "loss": 0.0084, + "step": 27120 + }, + { + "epoch": 2.1071844660194174, + "grad_norm": 0.37382933497428894, + "learning_rate": 8.719888635755327e-05, + "loss": 0.0108, + "step": 27130 + }, + { + "epoch": 2.1079611650485437, + "grad_norm": 0.23292802274227142, + "learning_rate": 8.718783576540928e-05, + "loss": 0.0088, + "step": 27140 + }, + { + "epoch": 2.10873786407767, + "grad_norm": 0.21761563420295715, + "learning_rate": 8.717678110645948e-05, + "loss": 0.0106, + "step": 27150 + }, + { + "epoch": 2.109514563106796, + "grad_norm": 0.2450675666332245, + "learning_rate": 8.716572238191279e-05, + "loss": 0.0094, + "step": 27160 + }, + { + "epoch": 2.1102912621359224, + "grad_norm": 0.23909831047058105, + "learning_rate": 8.715465959297857e-05, + "loss": 0.0052, + "step": 27170 + }, + { + "epoch": 2.1110679611650487, + "grad_norm": 0.25634488463401794, + "learning_rate": 8.714359274086665e-05, + "loss": 0.0109, + "step": 27180 + }, + { + "epoch": 2.111844660194175, + "grad_norm": 0.2873077988624573, + "learning_rate": 8.713252182678726e-05, + "loss": 0.0111, + "step": 27190 + }, + { + "epoch": 2.1126213592233007, + "grad_norm": 0.23519060015678406, + "learning_rate": 8.712144685195112e-05, + "loss": 0.0125, + "step": 27200 + }, + { + "epoch": 2.113398058252427, + "grad_norm": 0.20737214386463165, + "learning_rate": 8.711036781756936e-05, + "loss": 0.0087, + "step": 27210 + }, + { + "epoch": 2.1141747572815532, + "grad_norm": 0.522699236869812, + "learning_rate": 8.709928472485357e-05, + "loss": 0.0101, + "step": 27220 + }, + { + "epoch": 2.1149514563106795, + "grad_norm": 0.14313991367816925, + "learning_rate": 8.708819757501579e-05, + "loss": 0.0069, + "step": 27230 + }, + { + "epoch": 2.1157281553398057, + "grad_norm": 0.2999138832092285, + "learning_rate": 8.707710636926846e-05, + "loss": 0.0107, + "step": 27240 + }, + { + "epoch": 2.116504854368932, + "grad_norm": 0.3313700556755066, + "learning_rate": 8.706601110882455e-05, + "loss": 0.0075, + "step": 27250 + }, + { + "epoch": 2.1172815533980582, + "grad_norm": 0.18990039825439453, + "learning_rate": 8.705491179489738e-05, + "loss": 0.0097, + "step": 27260 + }, + { + "epoch": 2.1180582524271845, + "grad_norm": 0.2538468837738037, + "learning_rate": 8.704380842870077e-05, + "loss": 0.0091, + "step": 27270 + }, + { + "epoch": 2.1188349514563107, + "grad_norm": 0.3337843120098114, + "learning_rate": 8.703270101144895e-05, + "loss": 0.0104, + "step": 27280 + }, + { + "epoch": 2.119611650485437, + "grad_norm": 0.22751203179359436, + "learning_rate": 8.702158954435664e-05, + "loss": 0.0098, + "step": 27290 + }, + { + "epoch": 2.1203883495145632, + "grad_norm": 0.369710236787796, + "learning_rate": 8.701047402863896e-05, + "loss": 0.0147, + "step": 27300 + }, + { + "epoch": 2.1211650485436895, + "grad_norm": 0.2259179651737213, + "learning_rate": 8.699935446551148e-05, + "loss": 0.0077, + "step": 27310 + }, + { + "epoch": 2.1219417475728157, + "grad_norm": 0.1956859678030014, + "learning_rate": 8.698823085619022e-05, + "loss": 0.0101, + "step": 27320 + }, + { + "epoch": 2.1227184466019415, + "grad_norm": 0.8482722043991089, + "learning_rate": 8.697710320189166e-05, + "loss": 0.0104, + "step": 27330 + }, + { + "epoch": 2.123495145631068, + "grad_norm": 0.3411474823951721, + "learning_rate": 8.696597150383268e-05, + "loss": 0.0108, + "step": 27340 + }, + { + "epoch": 2.124271844660194, + "grad_norm": 0.1569032073020935, + "learning_rate": 8.695483576323063e-05, + "loss": 0.0092, + "step": 27350 + }, + { + "epoch": 2.1250485436893203, + "grad_norm": 0.27753207087516785, + "learning_rate": 8.69436959813033e-05, + "loss": 0.0094, + "step": 27360 + }, + { + "epoch": 2.1258252427184465, + "grad_norm": 0.2851411700248718, + "learning_rate": 8.693255215926892e-05, + "loss": 0.0117, + "step": 27370 + }, + { + "epoch": 2.1266019417475728, + "grad_norm": 0.26051875948905945, + "learning_rate": 8.692140429834617e-05, + "loss": 0.0101, + "step": 27380 + }, + { + "epoch": 2.127378640776699, + "grad_norm": 0.26166093349456787, + "learning_rate": 8.691025239975415e-05, + "loss": 0.009, + "step": 27390 + }, + { + "epoch": 2.1281553398058253, + "grad_norm": 0.28022557497024536, + "learning_rate": 8.689909646471243e-05, + "loss": 0.0113, + "step": 27400 + }, + { + "epoch": 2.1289320388349515, + "grad_norm": 0.2568403482437134, + "learning_rate": 8.688793649444099e-05, + "loss": 0.009, + "step": 27410 + }, + { + "epoch": 2.1297087378640778, + "grad_norm": 0.1962161660194397, + "learning_rate": 8.687677249016029e-05, + "loss": 0.0089, + "step": 27420 + }, + { + "epoch": 2.130485436893204, + "grad_norm": 0.23708808422088623, + "learning_rate": 8.686560445309118e-05, + "loss": 0.0072, + "step": 27430 + }, + { + "epoch": 2.1312621359223303, + "grad_norm": 0.5707203149795532, + "learning_rate": 8.685443238445499e-05, + "loss": 0.0096, + "step": 27440 + }, + { + "epoch": 2.1320388349514565, + "grad_norm": 0.39403632283210754, + "learning_rate": 8.68432562854735e-05, + "loss": 0.0093, + "step": 27450 + }, + { + "epoch": 2.1328155339805823, + "grad_norm": 0.19390107691287994, + "learning_rate": 8.683207615736887e-05, + "loss": 0.0117, + "step": 27460 + }, + { + "epoch": 2.1335922330097086, + "grad_norm": 0.3963286280632019, + "learning_rate": 8.682089200136379e-05, + "loss": 0.0099, + "step": 27470 + }, + { + "epoch": 2.134368932038835, + "grad_norm": 0.27800503373146057, + "learning_rate": 8.680970381868132e-05, + "loss": 0.0097, + "step": 27480 + }, + { + "epoch": 2.135145631067961, + "grad_norm": 0.34551236033439636, + "learning_rate": 8.679851161054498e-05, + "loss": 0.0116, + "step": 27490 + }, + { + "epoch": 2.1359223300970873, + "grad_norm": 0.3509752154350281, + "learning_rate": 8.678731537817873e-05, + "loss": 0.0123, + "step": 27500 + }, + { + "epoch": 2.1366990291262136, + "grad_norm": 0.31517836451530457, + "learning_rate": 8.677611512280697e-05, + "loss": 0.0081, + "step": 27510 + }, + { + "epoch": 2.13747572815534, + "grad_norm": 0.34958985447883606, + "learning_rate": 8.676491084565457e-05, + "loss": 0.0102, + "step": 27520 + }, + { + "epoch": 2.138252427184466, + "grad_norm": 0.278145432472229, + "learning_rate": 8.675370254794678e-05, + "loss": 0.0093, + "step": 27530 + }, + { + "epoch": 2.1390291262135923, + "grad_norm": 0.3079357147216797, + "learning_rate": 8.674249023090935e-05, + "loss": 0.0073, + "step": 27540 + }, + { + "epoch": 2.1398058252427186, + "grad_norm": 0.2918591797351837, + "learning_rate": 8.673127389576843e-05, + "loss": 0.0068, + "step": 27550 + }, + { + "epoch": 2.140582524271845, + "grad_norm": 0.24414248764514923, + "learning_rate": 8.67200535437506e-05, + "loss": 0.0097, + "step": 27560 + }, + { + "epoch": 2.141359223300971, + "grad_norm": 0.5366660356521606, + "learning_rate": 8.670882917608296e-05, + "loss": 0.0079, + "step": 27570 + }, + { + "epoch": 2.1421359223300973, + "grad_norm": 0.2550643980503082, + "learning_rate": 8.669760079399292e-05, + "loss": 0.0086, + "step": 27580 + }, + { + "epoch": 2.142912621359223, + "grad_norm": 0.43162858486175537, + "learning_rate": 8.668636839870845e-05, + "loss": 0.0079, + "step": 27590 + }, + { + "epoch": 2.1436893203883494, + "grad_norm": 0.32624509930610657, + "learning_rate": 8.667513199145789e-05, + "loss": 0.0128, + "step": 27600 + }, + { + "epoch": 2.1444660194174756, + "grad_norm": 0.2927224338054657, + "learning_rate": 8.666389157347002e-05, + "loss": 0.0097, + "step": 27610 + }, + { + "epoch": 2.145242718446602, + "grad_norm": 0.2527632713317871, + "learning_rate": 8.66526471459741e-05, + "loss": 0.0121, + "step": 27620 + }, + { + "epoch": 2.146019417475728, + "grad_norm": 0.3914009630680084, + "learning_rate": 8.66413987101998e-05, + "loss": 0.0105, + "step": 27630 + }, + { + "epoch": 2.1467961165048544, + "grad_norm": 0.39487704634666443, + "learning_rate": 8.663014626737723e-05, + "loss": 0.009, + "step": 27640 + }, + { + "epoch": 2.1475728155339806, + "grad_norm": 0.24259595572948456, + "learning_rate": 8.661888981873691e-05, + "loss": 0.0125, + "step": 27650 + }, + { + "epoch": 2.148349514563107, + "grad_norm": 0.21398942172527313, + "learning_rate": 8.660762936550988e-05, + "loss": 0.01, + "step": 27660 + }, + { + "epoch": 2.149126213592233, + "grad_norm": 0.21898649632930756, + "learning_rate": 8.659636490892753e-05, + "loss": 0.0085, + "step": 27670 + }, + { + "epoch": 2.1499029126213594, + "grad_norm": 0.3611460030078888, + "learning_rate": 8.658509645022174e-05, + "loss": 0.0113, + "step": 27680 + }, + { + "epoch": 2.1506796116504856, + "grad_norm": 0.26224491000175476, + "learning_rate": 8.657382399062481e-05, + "loss": 0.0079, + "step": 27690 + }, + { + "epoch": 2.151456310679612, + "grad_norm": 0.33444052934646606, + "learning_rate": 8.656254753136946e-05, + "loss": 0.0085, + "step": 27700 + }, + { + "epoch": 2.1522330097087377, + "grad_norm": 0.14802174270153046, + "learning_rate": 8.655126707368891e-05, + "loss": 0.0086, + "step": 27710 + }, + { + "epoch": 2.153009708737864, + "grad_norm": 0.21017338335514069, + "learning_rate": 8.653998261881672e-05, + "loss": 0.0075, + "step": 27720 + }, + { + "epoch": 2.15378640776699, + "grad_norm": 0.19220803678035736, + "learning_rate": 8.652869416798699e-05, + "loss": 0.0104, + "step": 27730 + }, + { + "epoch": 2.1545631067961164, + "grad_norm": 0.2407376617193222, + "learning_rate": 8.651740172243417e-05, + "loss": 0.0077, + "step": 27740 + }, + { + "epoch": 2.1553398058252426, + "grad_norm": 0.23842860758304596, + "learning_rate": 8.65061052833932e-05, + "loss": 0.0123, + "step": 27750 + }, + { + "epoch": 2.156116504854369, + "grad_norm": 0.26046231389045715, + "learning_rate": 8.649480485209945e-05, + "loss": 0.0087, + "step": 27760 + }, + { + "epoch": 2.156893203883495, + "grad_norm": 0.36321699619293213, + "learning_rate": 8.64835004297887e-05, + "loss": 0.0121, + "step": 27770 + }, + { + "epoch": 2.1576699029126214, + "grad_norm": 0.28222161531448364, + "learning_rate": 8.64721920176972e-05, + "loss": 0.0113, + "step": 27780 + }, + { + "epoch": 2.1584466019417476, + "grad_norm": 0.3145509660243988, + "learning_rate": 8.646087961706164e-05, + "loss": 0.009, + "step": 27790 + }, + { + "epoch": 2.159223300970874, + "grad_norm": 0.22150106728076935, + "learning_rate": 8.644956322911908e-05, + "loss": 0.0103, + "step": 27800 + }, + { + "epoch": 2.16, + "grad_norm": 0.23474468290805817, + "learning_rate": 8.643824285510709e-05, + "loss": 0.0083, + "step": 27810 + }, + { + "epoch": 2.1607766990291264, + "grad_norm": 0.19980204105377197, + "learning_rate": 8.642691849626364e-05, + "loss": 0.0099, + "step": 27820 + }, + { + "epoch": 2.1615533980582526, + "grad_norm": 0.1985950618982315, + "learning_rate": 8.641559015382717e-05, + "loss": 0.009, + "step": 27830 + }, + { + "epoch": 2.1623300970873784, + "grad_norm": 0.2657061517238617, + "learning_rate": 8.640425782903649e-05, + "loss": 0.0092, + "step": 27840 + }, + { + "epoch": 2.1631067961165047, + "grad_norm": 0.524271547794342, + "learning_rate": 8.639292152313091e-05, + "loss": 0.0129, + "step": 27850 + }, + { + "epoch": 2.163883495145631, + "grad_norm": 0.23962101340293884, + "learning_rate": 8.638158123735015e-05, + "loss": 0.0091, + "step": 27860 + }, + { + "epoch": 2.164660194174757, + "grad_norm": 0.47195157408714294, + "learning_rate": 8.637023697293436e-05, + "loss": 0.0101, + "step": 27870 + }, + { + "epoch": 2.1654368932038834, + "grad_norm": 0.16014014184474945, + "learning_rate": 8.635888873112414e-05, + "loss": 0.0072, + "step": 27880 + }, + { + "epoch": 2.1662135922330097, + "grad_norm": 0.2170243114233017, + "learning_rate": 8.634753651316052e-05, + "loss": 0.0116, + "step": 27890 + }, + { + "epoch": 2.166990291262136, + "grad_norm": 0.2580581605434418, + "learning_rate": 8.633618032028496e-05, + "loss": 0.0084, + "step": 27900 + }, + { + "epoch": 2.167766990291262, + "grad_norm": 0.3031270503997803, + "learning_rate": 8.632482015373934e-05, + "loss": 0.0106, + "step": 27910 + }, + { + "epoch": 2.1685436893203884, + "grad_norm": 0.2071293592453003, + "learning_rate": 8.6313456014766e-05, + "loss": 0.0104, + "step": 27920 + }, + { + "epoch": 2.1693203883495147, + "grad_norm": 0.4052988886833191, + "learning_rate": 8.630208790460771e-05, + "loss": 0.0135, + "step": 27930 + }, + { + "epoch": 2.170097087378641, + "grad_norm": 0.328755259513855, + "learning_rate": 8.629071582450768e-05, + "loss": 0.0103, + "step": 27940 + }, + { + "epoch": 2.170873786407767, + "grad_norm": 0.20915603637695312, + "learning_rate": 8.62793397757095e-05, + "loss": 0.0117, + "step": 27950 + }, + { + "epoch": 2.171650485436893, + "grad_norm": 0.1860688179731369, + "learning_rate": 8.626795975945729e-05, + "loss": 0.0089, + "step": 27960 + }, + { + "epoch": 2.1724271844660192, + "grad_norm": 0.24792106449604034, + "learning_rate": 8.625657577699551e-05, + "loss": 0.0089, + "step": 27970 + }, + { + "epoch": 2.1732038834951455, + "grad_norm": 0.30432799458503723, + "learning_rate": 8.624518782956914e-05, + "loss": 0.0099, + "step": 27980 + }, + { + "epoch": 2.1739805825242717, + "grad_norm": 0.1742774397134781, + "learning_rate": 8.62337959184235e-05, + "loss": 0.0082, + "step": 27990 + }, + { + "epoch": 2.174757281553398, + "grad_norm": 0.12936753034591675, + "learning_rate": 8.622240004480441e-05, + "loss": 0.0072, + "step": 28000 + }, + { + "epoch": 2.1755339805825242, + "grad_norm": 0.33484554290771484, + "learning_rate": 8.621100020995814e-05, + "loss": 0.0095, + "step": 28010 + }, + { + "epoch": 2.1763106796116505, + "grad_norm": 0.7081558108329773, + "learning_rate": 8.619959641513132e-05, + "loss": 0.0125, + "step": 28020 + }, + { + "epoch": 2.1770873786407767, + "grad_norm": 0.21148259937763214, + "learning_rate": 8.618818866157105e-05, + "loss": 0.0114, + "step": 28030 + }, + { + "epoch": 2.177864077669903, + "grad_norm": 0.1804848164319992, + "learning_rate": 8.617677695052487e-05, + "loss": 0.0086, + "step": 28040 + }, + { + "epoch": 2.1786407766990292, + "grad_norm": 0.181147038936615, + "learning_rate": 8.616536128324078e-05, + "loss": 0.0097, + "step": 28050 + }, + { + "epoch": 2.1794174757281555, + "grad_norm": 0.3149910271167755, + "learning_rate": 8.615394166096712e-05, + "loss": 0.009, + "step": 28060 + }, + { + "epoch": 2.1801941747572817, + "grad_norm": 0.26539066433906555, + "learning_rate": 8.614251808495279e-05, + "loss": 0.0085, + "step": 28070 + }, + { + "epoch": 2.180970873786408, + "grad_norm": 0.17117393016815186, + "learning_rate": 8.6131090556447e-05, + "loss": 0.0107, + "step": 28080 + }, + { + "epoch": 2.1817475728155338, + "grad_norm": 0.16519391536712646, + "learning_rate": 8.611965907669947e-05, + "loss": 0.0101, + "step": 28090 + }, + { + "epoch": 2.18252427184466, + "grad_norm": 0.25103163719177246, + "learning_rate": 8.610822364696034e-05, + "loss": 0.0107, + "step": 28100 + }, + { + "epoch": 2.1833009708737863, + "grad_norm": 0.24916191399097443, + "learning_rate": 8.609678426848015e-05, + "loss": 0.0102, + "step": 28110 + }, + { + "epoch": 2.1840776699029125, + "grad_norm": 0.5019299387931824, + "learning_rate": 8.60853409425099e-05, + "loss": 0.0082, + "step": 28120 + }, + { + "epoch": 2.1848543689320388, + "grad_norm": 0.19695144891738892, + "learning_rate": 8.607389367030104e-05, + "loss": 0.0098, + "step": 28130 + }, + { + "epoch": 2.185631067961165, + "grad_norm": 0.30278724431991577, + "learning_rate": 8.606244245310538e-05, + "loss": 0.0076, + "step": 28140 + }, + { + "epoch": 2.1864077669902913, + "grad_norm": 0.7663271427154541, + "learning_rate": 8.605098729217525e-05, + "loss": 0.0078, + "step": 28150 + }, + { + "epoch": 2.1871844660194175, + "grad_norm": 0.20453065633773804, + "learning_rate": 8.603952818876335e-05, + "loss": 0.0095, + "step": 28160 + }, + { + "epoch": 2.1879611650485438, + "grad_norm": 0.22945354878902435, + "learning_rate": 8.602806514412281e-05, + "loss": 0.0081, + "step": 28170 + }, + { + "epoch": 2.18873786407767, + "grad_norm": 0.1749470829963684, + "learning_rate": 8.601659815950726e-05, + "loss": 0.0103, + "step": 28180 + }, + { + "epoch": 2.1895145631067963, + "grad_norm": 0.4017576277256012, + "learning_rate": 8.600512723617067e-05, + "loss": 0.0074, + "step": 28190 + }, + { + "epoch": 2.1902912621359225, + "grad_norm": 0.1383991688489914, + "learning_rate": 8.59936523753675e-05, + "loss": 0.0101, + "step": 28200 + }, + { + "epoch": 2.1910679611650483, + "grad_norm": 0.2952671945095062, + "learning_rate": 8.598217357835264e-05, + "loss": 0.008, + "step": 28210 + }, + { + "epoch": 2.1918446601941746, + "grad_norm": 0.5990915894508362, + "learning_rate": 8.597069084638135e-05, + "loss": 0.0119, + "step": 28220 + }, + { + "epoch": 2.192621359223301, + "grad_norm": 0.2866552174091339, + "learning_rate": 8.595920418070939e-05, + "loss": 0.0099, + "step": 28230 + }, + { + "epoch": 2.193398058252427, + "grad_norm": 0.20923613011837006, + "learning_rate": 8.594771358259295e-05, + "loss": 0.0098, + "step": 28240 + }, + { + "epoch": 2.1941747572815533, + "grad_norm": 0.6196771264076233, + "learning_rate": 8.593621905328858e-05, + "loss": 0.0104, + "step": 28250 + }, + { + "epoch": 2.1949514563106796, + "grad_norm": 0.4010656774044037, + "learning_rate": 8.592472059405333e-05, + "loss": 0.0107, + "step": 28260 + }, + { + "epoch": 2.195728155339806, + "grad_norm": 0.2791481018066406, + "learning_rate": 8.591321820614464e-05, + "loss": 0.011, + "step": 28270 + }, + { + "epoch": 2.196504854368932, + "grad_norm": 0.30249491333961487, + "learning_rate": 8.590171189082041e-05, + "loss": 0.0128, + "step": 28280 + }, + { + "epoch": 2.1972815533980583, + "grad_norm": 0.3165256977081299, + "learning_rate": 8.589020164933894e-05, + "loss": 0.0091, + "step": 28290 + }, + { + "epoch": 2.1980582524271846, + "grad_norm": 0.2907841205596924, + "learning_rate": 8.587868748295898e-05, + "loss": 0.007, + "step": 28300 + }, + { + "epoch": 2.198834951456311, + "grad_norm": 0.23033270239830017, + "learning_rate": 8.586716939293971e-05, + "loss": 0.0073, + "step": 28310 + }, + { + "epoch": 2.199611650485437, + "grad_norm": 0.2795586884021759, + "learning_rate": 8.58556473805407e-05, + "loss": 0.0109, + "step": 28320 + }, + { + "epoch": 2.2003883495145633, + "grad_norm": 0.18085913360118866, + "learning_rate": 8.584412144702202e-05, + "loss": 0.0073, + "step": 28330 + }, + { + "epoch": 2.2011650485436896, + "grad_norm": 0.20078428089618683, + "learning_rate": 8.58325915936441e-05, + "loss": 0.0065, + "step": 28340 + }, + { + "epoch": 2.2019417475728154, + "grad_norm": 0.20729182660579681, + "learning_rate": 8.582105782166783e-05, + "loss": 0.0096, + "step": 28350 + }, + { + "epoch": 2.2027184466019416, + "grad_norm": 0.300108402967453, + "learning_rate": 8.580952013235455e-05, + "loss": 0.0075, + "step": 28360 + }, + { + "epoch": 2.203495145631068, + "grad_norm": 0.358894407749176, + "learning_rate": 8.579797852696596e-05, + "loss": 0.0112, + "step": 28370 + }, + { + "epoch": 2.204271844660194, + "grad_norm": 0.292537122964859, + "learning_rate": 8.578643300676428e-05, + "loss": 0.0099, + "step": 28380 + }, + { + "epoch": 2.2050485436893204, + "grad_norm": 0.2889319956302643, + "learning_rate": 8.577488357301209e-05, + "loss": 0.011, + "step": 28390 + }, + { + "epoch": 2.2058252427184466, + "grad_norm": 0.19336657226085663, + "learning_rate": 8.576333022697242e-05, + "loss": 0.0095, + "step": 28400 + }, + { + "epoch": 2.206601941747573, + "grad_norm": 0.4039100706577301, + "learning_rate": 8.575177296990873e-05, + "loss": 0.0082, + "step": 28410 + }, + { + "epoch": 2.207378640776699, + "grad_norm": 0.2133830338716507, + "learning_rate": 8.574021180308489e-05, + "loss": 0.0113, + "step": 28420 + }, + { + "epoch": 2.2081553398058253, + "grad_norm": 0.19446609914302826, + "learning_rate": 8.572864672776523e-05, + "loss": 0.0104, + "step": 28430 + }, + { + "epoch": 2.2089320388349516, + "grad_norm": 0.25125449895858765, + "learning_rate": 8.571707774521447e-05, + "loss": 0.0133, + "step": 28440 + }, + { + "epoch": 2.209708737864078, + "grad_norm": 0.34735211730003357, + "learning_rate": 8.57055048566978e-05, + "loss": 0.0127, + "step": 28450 + }, + { + "epoch": 2.2104854368932036, + "grad_norm": 0.394175261259079, + "learning_rate": 8.569392806348078e-05, + "loss": 0.0157, + "step": 28460 + }, + { + "epoch": 2.21126213592233, + "grad_norm": 0.2883295714855194, + "learning_rate": 8.568234736682947e-05, + "loss": 0.0074, + "step": 28470 + }, + { + "epoch": 2.212038834951456, + "grad_norm": 0.5581421256065369, + "learning_rate": 8.567076276801029e-05, + "loss": 0.0101, + "step": 28480 + }, + { + "epoch": 2.2128155339805824, + "grad_norm": 0.3521933853626251, + "learning_rate": 8.565917426829013e-05, + "loss": 0.0095, + "step": 28490 + }, + { + "epoch": 2.2135922330097086, + "grad_norm": 0.27735841274261475, + "learning_rate": 8.564758186893628e-05, + "loss": 0.0082, + "step": 28500 + }, + { + "epoch": 2.214368932038835, + "grad_norm": 0.15727989375591278, + "learning_rate": 8.563598557121649e-05, + "loss": 0.0085, + "step": 28510 + }, + { + "epoch": 2.215145631067961, + "grad_norm": 0.15307796001434326, + "learning_rate": 8.562438537639888e-05, + "loss": 0.0108, + "step": 28520 + }, + { + "epoch": 2.2159223300970874, + "grad_norm": 0.47517192363739014, + "learning_rate": 8.561278128575206e-05, + "loss": 0.0092, + "step": 28530 + }, + { + "epoch": 2.2166990291262136, + "grad_norm": 0.216777041554451, + "learning_rate": 8.5601173300545e-05, + "loss": 0.0128, + "step": 28540 + }, + { + "epoch": 2.21747572815534, + "grad_norm": 0.5249974131584167, + "learning_rate": 8.558956142204717e-05, + "loss": 0.0108, + "step": 28550 + }, + { + "epoch": 2.218252427184466, + "grad_norm": 0.2507629692554474, + "learning_rate": 8.55779456515284e-05, + "loss": 0.0065, + "step": 28560 + }, + { + "epoch": 2.2190291262135924, + "grad_norm": 0.20326243340969086, + "learning_rate": 8.556632599025898e-05, + "loss": 0.0071, + "step": 28570 + }, + { + "epoch": 2.2198058252427186, + "grad_norm": 0.2856482267379761, + "learning_rate": 8.555470243950964e-05, + "loss": 0.0102, + "step": 28580 + }, + { + "epoch": 2.220582524271845, + "grad_norm": 0.1490890383720398, + "learning_rate": 8.554307500055148e-05, + "loss": 0.0084, + "step": 28590 + }, + { + "epoch": 2.2213592233009707, + "grad_norm": 0.219789519906044, + "learning_rate": 8.553144367465609e-05, + "loss": 0.0105, + "step": 28600 + }, + { + "epoch": 2.222135922330097, + "grad_norm": 0.23631040751934052, + "learning_rate": 8.551980846309544e-05, + "loss": 0.0105, + "step": 28610 + }, + { + "epoch": 2.222912621359223, + "grad_norm": 0.30333080887794495, + "learning_rate": 8.550816936714193e-05, + "loss": 0.0133, + "step": 28620 + }, + { + "epoch": 2.2236893203883494, + "grad_norm": 0.21700070798397064, + "learning_rate": 8.549652638806841e-05, + "loss": 0.0096, + "step": 28630 + }, + { + "epoch": 2.2244660194174757, + "grad_norm": 0.26130425930023193, + "learning_rate": 8.548487952714812e-05, + "loss": 0.009, + "step": 28640 + }, + { + "epoch": 2.225242718446602, + "grad_norm": 0.4009096026420593, + "learning_rate": 8.547322878565478e-05, + "loss": 0.009, + "step": 28650 + }, + { + "epoch": 2.226019417475728, + "grad_norm": 0.16493816673755646, + "learning_rate": 8.546157416486245e-05, + "loss": 0.0094, + "step": 28660 + }, + { + "epoch": 2.2267961165048544, + "grad_norm": 0.40393999218940735, + "learning_rate": 8.54499156660457e-05, + "loss": 0.0112, + "step": 28670 + }, + { + "epoch": 2.2275728155339807, + "grad_norm": 0.24784229695796967, + "learning_rate": 8.543825329047947e-05, + "loss": 0.0108, + "step": 28680 + }, + { + "epoch": 2.228349514563107, + "grad_norm": 0.21557176113128662, + "learning_rate": 8.542658703943913e-05, + "loss": 0.0102, + "step": 28690 + }, + { + "epoch": 2.229126213592233, + "grad_norm": 0.5215991735458374, + "learning_rate": 8.541491691420051e-05, + "loss": 0.0105, + "step": 28700 + }, + { + "epoch": 2.2299029126213594, + "grad_norm": 0.19448915123939514, + "learning_rate": 8.54032429160398e-05, + "loss": 0.0089, + "step": 28710 + }, + { + "epoch": 2.2306796116504852, + "grad_norm": 0.3144998550415039, + "learning_rate": 8.539156504623369e-05, + "loss": 0.0085, + "step": 28720 + }, + { + "epoch": 2.2314563106796115, + "grad_norm": 0.1484372615814209, + "learning_rate": 8.537988330605923e-05, + "loss": 0.0103, + "step": 28730 + }, + { + "epoch": 2.2322330097087377, + "grad_norm": 0.2479926496744156, + "learning_rate": 8.536819769679393e-05, + "loss": 0.008, + "step": 28740 + }, + { + "epoch": 2.233009708737864, + "grad_norm": 0.5515297651290894, + "learning_rate": 8.53565082197157e-05, + "loss": 0.0119, + "step": 28750 + }, + { + "epoch": 2.2337864077669902, + "grad_norm": 0.25696250796318054, + "learning_rate": 8.534481487610289e-05, + "loss": 0.0107, + "step": 28760 + }, + { + "epoch": 2.2345631067961165, + "grad_norm": 0.24265210330486298, + "learning_rate": 8.533311766723428e-05, + "loss": 0.0116, + "step": 28770 + }, + { + "epoch": 2.2353398058252427, + "grad_norm": 0.3417203724384308, + "learning_rate": 8.532141659438901e-05, + "loss": 0.0084, + "step": 28780 + }, + { + "epoch": 2.236116504854369, + "grad_norm": 0.34780624508857727, + "learning_rate": 8.530971165884675e-05, + "loss": 0.0128, + "step": 28790 + }, + { + "epoch": 2.236893203883495, + "grad_norm": 0.27717292308807373, + "learning_rate": 8.529800286188752e-05, + "loss": 0.0127, + "step": 28800 + }, + { + "epoch": 2.2376699029126215, + "grad_norm": 0.3699066936969757, + "learning_rate": 8.528629020479175e-05, + "loss": 0.0123, + "step": 28810 + }, + { + "epoch": 2.2384466019417477, + "grad_norm": 0.19311712682247162, + "learning_rate": 8.527457368884033e-05, + "loss": 0.0079, + "step": 28820 + }, + { + "epoch": 2.239223300970874, + "grad_norm": 0.5376977324485779, + "learning_rate": 8.526285331531458e-05, + "loss": 0.0086, + "step": 28830 + }, + { + "epoch": 2.24, + "grad_norm": 0.29700931906700134, + "learning_rate": 8.525112908549621e-05, + "loss": 0.0074, + "step": 28840 + }, + { + "epoch": 2.240776699029126, + "grad_norm": 0.38394856452941895, + "learning_rate": 8.523940100066735e-05, + "loss": 0.0116, + "step": 28850 + }, + { + "epoch": 2.2415533980582523, + "grad_norm": 0.21649663150310516, + "learning_rate": 8.52276690621106e-05, + "loss": 0.0077, + "step": 28860 + }, + { + "epoch": 2.2423300970873785, + "grad_norm": 0.2977299392223358, + "learning_rate": 8.521593327110889e-05, + "loss": 0.0101, + "step": 28870 + }, + { + "epoch": 2.2431067961165048, + "grad_norm": 0.2955136299133301, + "learning_rate": 8.520419362894569e-05, + "loss": 0.0091, + "step": 28880 + }, + { + "epoch": 2.243883495145631, + "grad_norm": 0.23324622213840485, + "learning_rate": 8.51924501369048e-05, + "loss": 0.0077, + "step": 28890 + }, + { + "epoch": 2.2446601941747573, + "grad_norm": 0.46555227041244507, + "learning_rate": 8.518070279627047e-05, + "loss": 0.0089, + "step": 28900 + }, + { + "epoch": 2.2454368932038835, + "grad_norm": 0.30199623107910156, + "learning_rate": 8.516895160832737e-05, + "loss": 0.0088, + "step": 28910 + }, + { + "epoch": 2.2462135922330098, + "grad_norm": 0.4767347276210785, + "learning_rate": 8.515719657436061e-05, + "loss": 0.0111, + "step": 28920 + }, + { + "epoch": 2.246990291262136, + "grad_norm": 0.18043695390224457, + "learning_rate": 8.514543769565568e-05, + "loss": 0.0097, + "step": 28930 + }, + { + "epoch": 2.2477669902912623, + "grad_norm": 0.1505870521068573, + "learning_rate": 8.513367497349853e-05, + "loss": 0.0086, + "step": 28940 + }, + { + "epoch": 2.2485436893203885, + "grad_norm": 0.21264630556106567, + "learning_rate": 8.51219084091755e-05, + "loss": 0.0075, + "step": 28950 + }, + { + "epoch": 2.2493203883495148, + "grad_norm": 0.19062717258930206, + "learning_rate": 8.511013800397338e-05, + "loss": 0.0077, + "step": 28960 + }, + { + "epoch": 2.2500970873786406, + "grad_norm": 0.1433190554380417, + "learning_rate": 8.509836375917937e-05, + "loss": 0.0099, + "step": 28970 + }, + { + "epoch": 2.250873786407767, + "grad_norm": 0.2226121723651886, + "learning_rate": 8.508658567608104e-05, + "loss": 0.0071, + "step": 28980 + }, + { + "epoch": 2.251650485436893, + "grad_norm": 0.28309887647628784, + "learning_rate": 8.507480375596647e-05, + "loss": 0.0115, + "step": 28990 + }, + { + "epoch": 2.2524271844660193, + "grad_norm": 0.2810218036174774, + "learning_rate": 8.506301800012408e-05, + "loss": 0.0087, + "step": 29000 + }, + { + "epoch": 2.2532038834951456, + "grad_norm": 0.1473521590232849, + "learning_rate": 8.505122840984278e-05, + "loss": 0.0086, + "step": 29010 + }, + { + "epoch": 2.253980582524272, + "grad_norm": 0.39863121509552, + "learning_rate": 8.503943498641182e-05, + "loss": 0.0105, + "step": 29020 + }, + { + "epoch": 2.254757281553398, + "grad_norm": 0.23546937108039856, + "learning_rate": 8.502763773112095e-05, + "loss": 0.0078, + "step": 29030 + }, + { + "epoch": 2.2555339805825243, + "grad_norm": 0.22627361118793488, + "learning_rate": 8.501583664526026e-05, + "loss": 0.0085, + "step": 29040 + }, + { + "epoch": 2.2563106796116505, + "grad_norm": 0.40364259481430054, + "learning_rate": 8.500403173012032e-05, + "loss": 0.0077, + "step": 29050 + }, + { + "epoch": 2.257087378640777, + "grad_norm": 0.25498512387275696, + "learning_rate": 8.499222298699211e-05, + "loss": 0.0103, + "step": 29060 + }, + { + "epoch": 2.257864077669903, + "grad_norm": 0.2633298635482788, + "learning_rate": 8.498041041716701e-05, + "loss": 0.0079, + "step": 29070 + }, + { + "epoch": 2.2586407766990293, + "grad_norm": 0.15738631784915924, + "learning_rate": 8.496859402193681e-05, + "loss": 0.007, + "step": 29080 + }, + { + "epoch": 2.2594174757281555, + "grad_norm": 0.25918078422546387, + "learning_rate": 8.495677380259374e-05, + "loss": 0.0129, + "step": 29090 + }, + { + "epoch": 2.260194174757282, + "grad_norm": 0.3149375915527344, + "learning_rate": 8.494494976043045e-05, + "loss": 0.0095, + "step": 29100 + }, + { + "epoch": 2.2609708737864076, + "grad_norm": 0.4106540083885193, + "learning_rate": 8.493312189673998e-05, + "loss": 0.0123, + "step": 29110 + }, + { + "epoch": 2.261747572815534, + "grad_norm": 0.24440962076187134, + "learning_rate": 8.492129021281584e-05, + "loss": 0.0106, + "step": 29120 + }, + { + "epoch": 2.26252427184466, + "grad_norm": 0.16000671684741974, + "learning_rate": 8.490945470995188e-05, + "loss": 0.0099, + "step": 29130 + }, + { + "epoch": 2.2633009708737863, + "grad_norm": 0.26336047053337097, + "learning_rate": 8.489761538944247e-05, + "loss": 0.0118, + "step": 29140 + }, + { + "epoch": 2.2640776699029126, + "grad_norm": 0.2861344814300537, + "learning_rate": 8.48857722525823e-05, + "loss": 0.0086, + "step": 29150 + }, + { + "epoch": 2.264854368932039, + "grad_norm": 0.18076561391353607, + "learning_rate": 8.487392530066652e-05, + "loss": 0.0102, + "step": 29160 + }, + { + "epoch": 2.265631067961165, + "grad_norm": 0.29958972334861755, + "learning_rate": 8.486207453499069e-05, + "loss": 0.0086, + "step": 29170 + }, + { + "epoch": 2.2664077669902913, + "grad_norm": 0.2185526192188263, + "learning_rate": 8.485021995685082e-05, + "loss": 0.0126, + "step": 29180 + }, + { + "epoch": 2.2671844660194176, + "grad_norm": 0.3685406446456909, + "learning_rate": 8.483836156754328e-05, + "loss": 0.0076, + "step": 29190 + }, + { + "epoch": 2.267961165048544, + "grad_norm": 0.19634585082530975, + "learning_rate": 8.482649936836491e-05, + "loss": 0.0116, + "step": 29200 + }, + { + "epoch": 2.26873786407767, + "grad_norm": 0.28301477432250977, + "learning_rate": 8.481463336061293e-05, + "loss": 0.0077, + "step": 29210 + }, + { + "epoch": 2.269514563106796, + "grad_norm": 0.2877519130706787, + "learning_rate": 8.480276354558496e-05, + "loss": 0.0097, + "step": 29220 + }, + { + "epoch": 2.270291262135922, + "grad_norm": 0.3010934293270111, + "learning_rate": 8.479088992457913e-05, + "loss": 0.0103, + "step": 29230 + }, + { + "epoch": 2.2710679611650484, + "grad_norm": 0.1515607386827469, + "learning_rate": 8.477901249889387e-05, + "loss": 0.0076, + "step": 29240 + }, + { + "epoch": 2.2718446601941746, + "grad_norm": 0.30888351798057556, + "learning_rate": 8.47671312698281e-05, + "loss": 0.0093, + "step": 29250 + }, + { + "epoch": 2.272621359223301, + "grad_norm": 0.3935754597187042, + "learning_rate": 8.475524623868112e-05, + "loss": 0.0102, + "step": 29260 + }, + { + "epoch": 2.273398058252427, + "grad_norm": 0.32198187708854675, + "learning_rate": 8.474335740675266e-05, + "loss": 0.0123, + "step": 29270 + }, + { + "epoch": 2.2741747572815534, + "grad_norm": 0.4598105549812317, + "learning_rate": 8.473146477534289e-05, + "loss": 0.0091, + "step": 29280 + }, + { + "epoch": 2.2749514563106796, + "grad_norm": 0.2585817873477936, + "learning_rate": 8.471956834575232e-05, + "loss": 0.0116, + "step": 29290 + }, + { + "epoch": 2.275728155339806, + "grad_norm": 0.23053885996341705, + "learning_rate": 8.470766811928197e-05, + "loss": 0.007, + "step": 29300 + }, + { + "epoch": 2.276504854368932, + "grad_norm": 0.16380882263183594, + "learning_rate": 8.469576409723323e-05, + "loss": 0.0104, + "step": 29310 + }, + { + "epoch": 2.2772815533980584, + "grad_norm": 0.4292449355125427, + "learning_rate": 8.468385628090788e-05, + "loss": 0.0096, + "step": 29320 + }, + { + "epoch": 2.2780582524271846, + "grad_norm": 0.1626473069190979, + "learning_rate": 8.467194467160815e-05, + "loss": 0.0103, + "step": 29330 + }, + { + "epoch": 2.278834951456311, + "grad_norm": 0.37962427735328674, + "learning_rate": 8.466002927063667e-05, + "loss": 0.0108, + "step": 29340 + }, + { + "epoch": 2.279611650485437, + "grad_norm": 0.26483672857284546, + "learning_rate": 8.464811007929651e-05, + "loss": 0.0112, + "step": 29350 + }, + { + "epoch": 2.280388349514563, + "grad_norm": 0.5634516477584839, + "learning_rate": 8.463618709889114e-05, + "loss": 0.0089, + "step": 29360 + }, + { + "epoch": 2.281165048543689, + "grad_norm": 0.18568454682826996, + "learning_rate": 8.462426033072442e-05, + "loss": 0.0068, + "step": 29370 + }, + { + "epoch": 2.2819417475728154, + "grad_norm": 0.3288404941558838, + "learning_rate": 8.461232977610061e-05, + "loss": 0.0096, + "step": 29380 + }, + { + "epoch": 2.2827184466019417, + "grad_norm": 0.24852092564105988, + "learning_rate": 8.46003954363245e-05, + "loss": 0.0088, + "step": 29390 + }, + { + "epoch": 2.283495145631068, + "grad_norm": 0.37524473667144775, + "learning_rate": 8.458845731270115e-05, + "loss": 0.0122, + "step": 29400 + }, + { + "epoch": 2.284271844660194, + "grad_norm": 0.37863677740097046, + "learning_rate": 8.45765154065361e-05, + "loss": 0.0095, + "step": 29410 + }, + { + "epoch": 2.2850485436893204, + "grad_norm": 0.4089648425579071, + "learning_rate": 8.456456971913532e-05, + "loss": 0.0117, + "step": 29420 + }, + { + "epoch": 2.2858252427184467, + "grad_norm": 0.34244146943092346, + "learning_rate": 8.455262025180517e-05, + "loss": 0.0124, + "step": 29430 + }, + { + "epoch": 2.286601941747573, + "grad_norm": 0.22421570122241974, + "learning_rate": 8.454066700585242e-05, + "loss": 0.0096, + "step": 29440 + }, + { + "epoch": 2.287378640776699, + "grad_norm": 0.416823148727417, + "learning_rate": 8.452870998258423e-05, + "loss": 0.0096, + "step": 29450 + }, + { + "epoch": 2.2881553398058254, + "grad_norm": 0.23442937433719635, + "learning_rate": 8.451674918330825e-05, + "loss": 0.0109, + "step": 29460 + }, + { + "epoch": 2.288932038834951, + "grad_norm": 0.22435978055000305, + "learning_rate": 8.450478460933246e-05, + "loss": 0.0091, + "step": 29470 + }, + { + "epoch": 2.2897087378640775, + "grad_norm": 0.3164006471633911, + "learning_rate": 8.449281626196532e-05, + "loss": 0.0091, + "step": 29480 + }, + { + "epoch": 2.2904854368932037, + "grad_norm": 0.2908742427825928, + "learning_rate": 8.448084414251564e-05, + "loss": 0.0074, + "step": 29490 + }, + { + "epoch": 2.29126213592233, + "grad_norm": 0.2507326006889343, + "learning_rate": 8.446886825229271e-05, + "loss": 0.0067, + "step": 29500 + }, + { + "epoch": 2.292038834951456, + "grad_norm": 0.22040560841560364, + "learning_rate": 8.445688859260615e-05, + "loss": 0.0142, + "step": 29510 + }, + { + "epoch": 2.2928155339805825, + "grad_norm": 1.0713036060333252, + "learning_rate": 8.444490516476606e-05, + "loss": 0.0088, + "step": 29520 + }, + { + "epoch": 2.2935922330097087, + "grad_norm": 0.14739404618740082, + "learning_rate": 8.443291797008293e-05, + "loss": 0.0071, + "step": 29530 + }, + { + "epoch": 2.294368932038835, + "grad_norm": 0.21152527630329132, + "learning_rate": 8.442092700986765e-05, + "loss": 0.0083, + "step": 29540 + }, + { + "epoch": 2.295145631067961, + "grad_norm": 0.45223093032836914, + "learning_rate": 8.440893228543156e-05, + "loss": 0.0064, + "step": 29550 + }, + { + "epoch": 2.2959223300970875, + "grad_norm": 0.26329609751701355, + "learning_rate": 8.439693379808638e-05, + "loss": 0.0104, + "step": 29560 + }, + { + "epoch": 2.2966990291262137, + "grad_norm": 0.7136750221252441, + "learning_rate": 8.43849315491442e-05, + "loss": 0.0103, + "step": 29570 + }, + { + "epoch": 2.29747572815534, + "grad_norm": 0.36205917596817017, + "learning_rate": 8.437292553991763e-05, + "loss": 0.0095, + "step": 29580 + }, + { + "epoch": 2.298252427184466, + "grad_norm": 0.2262820303440094, + "learning_rate": 8.436091577171959e-05, + "loss": 0.0128, + "step": 29590 + }, + { + "epoch": 2.2990291262135925, + "grad_norm": 0.3818943500518799, + "learning_rate": 8.434890224586347e-05, + "loss": 0.0117, + "step": 29600 + }, + { + "epoch": 2.2998058252427183, + "grad_norm": 0.24168163537979126, + "learning_rate": 8.433688496366303e-05, + "loss": 0.0096, + "step": 29610 + }, + { + "epoch": 2.3005825242718445, + "grad_norm": 0.315337210893631, + "learning_rate": 8.432486392643248e-05, + "loss": 0.0081, + "step": 29620 + }, + { + "epoch": 2.3013592233009708, + "grad_norm": 0.18543006479740143, + "learning_rate": 8.431283913548643e-05, + "loss": 0.0082, + "step": 29630 + }, + { + "epoch": 2.302135922330097, + "grad_norm": 0.30911752581596375, + "learning_rate": 8.430081059213985e-05, + "loss": 0.0097, + "step": 29640 + }, + { + "epoch": 2.3029126213592233, + "grad_norm": 0.2827785611152649, + "learning_rate": 8.428877829770823e-05, + "loss": 0.0104, + "step": 29650 + }, + { + "epoch": 2.3036893203883495, + "grad_norm": 0.3111993372440338, + "learning_rate": 8.427674225350735e-05, + "loss": 0.0125, + "step": 29660 + }, + { + "epoch": 2.3044660194174758, + "grad_norm": 0.2203007936477661, + "learning_rate": 8.426470246085347e-05, + "loss": 0.0095, + "step": 29670 + }, + { + "epoch": 2.305242718446602, + "grad_norm": 0.5444714426994324, + "learning_rate": 8.425265892106324e-05, + "loss": 0.0083, + "step": 29680 + }, + { + "epoch": 2.3060194174757283, + "grad_norm": 0.27932044863700867, + "learning_rate": 8.424061163545374e-05, + "loss": 0.01, + "step": 29690 + }, + { + "epoch": 2.3067961165048545, + "grad_norm": 0.11906617879867554, + "learning_rate": 8.422856060534243e-05, + "loss": 0.0096, + "step": 29700 + }, + { + "epoch": 2.3075728155339807, + "grad_norm": 0.303242027759552, + "learning_rate": 8.421650583204718e-05, + "loss": 0.0085, + "step": 29710 + }, + { + "epoch": 2.3083495145631066, + "grad_norm": 0.2886945605278015, + "learning_rate": 8.420444731688633e-05, + "loss": 0.0096, + "step": 29720 + }, + { + "epoch": 2.309126213592233, + "grad_norm": 0.3589329123497009, + "learning_rate": 8.419238506117852e-05, + "loss": 0.0108, + "step": 29730 + }, + { + "epoch": 2.309902912621359, + "grad_norm": 0.18540576100349426, + "learning_rate": 8.418031906624289e-05, + "loss": 0.0082, + "step": 29740 + }, + { + "epoch": 2.3106796116504853, + "grad_norm": 0.47706079483032227, + "learning_rate": 8.416824933339898e-05, + "loss": 0.0117, + "step": 29750 + }, + { + "epoch": 2.3114563106796115, + "grad_norm": 0.330640971660614, + "learning_rate": 8.415617586396667e-05, + "loss": 0.0122, + "step": 29760 + }, + { + "epoch": 2.312233009708738, + "grad_norm": 0.9275242686271667, + "learning_rate": 8.414409865926632e-05, + "loss": 0.0151, + "step": 29770 + }, + { + "epoch": 2.313009708737864, + "grad_norm": 0.34688904881477356, + "learning_rate": 8.413201772061867e-05, + "loss": 0.009, + "step": 29780 + }, + { + "epoch": 2.3137864077669903, + "grad_norm": 0.3365571200847626, + "learning_rate": 8.411993304934488e-05, + "loss": 0.008, + "step": 29790 + }, + { + "epoch": 2.3145631067961165, + "grad_norm": 0.3360922932624817, + "learning_rate": 8.410784464676654e-05, + "loss": 0.0095, + "step": 29800 + }, + { + "epoch": 2.315339805825243, + "grad_norm": 0.1768513172864914, + "learning_rate": 8.409575251420556e-05, + "loss": 0.0095, + "step": 29810 + }, + { + "epoch": 2.316116504854369, + "grad_norm": 0.21453295648097992, + "learning_rate": 8.408365665298435e-05, + "loss": 0.0115, + "step": 29820 + }, + { + "epoch": 2.3168932038834953, + "grad_norm": 0.33720555901527405, + "learning_rate": 8.40715570644257e-05, + "loss": 0.0104, + "step": 29830 + }, + { + "epoch": 2.3176699029126215, + "grad_norm": 0.23783722519874573, + "learning_rate": 8.40594537498528e-05, + "loss": 0.0138, + "step": 29840 + }, + { + "epoch": 2.318446601941748, + "grad_norm": 0.2970128655433655, + "learning_rate": 8.404734671058924e-05, + "loss": 0.0109, + "step": 29850 + }, + { + "epoch": 2.319223300970874, + "grad_norm": 0.5049226880073547, + "learning_rate": 8.403523594795902e-05, + "loss": 0.011, + "step": 29860 + }, + { + "epoch": 2.32, + "grad_norm": 0.4836846888065338, + "learning_rate": 8.402312146328659e-05, + "loss": 0.0132, + "step": 29870 + }, + { + "epoch": 2.320776699029126, + "grad_norm": 0.2950120270252228, + "learning_rate": 8.401100325789675e-05, + "loss": 0.0111, + "step": 29880 + }, + { + "epoch": 2.3215533980582523, + "grad_norm": 0.7185761332511902, + "learning_rate": 8.399888133311472e-05, + "loss": 0.0127, + "step": 29890 + }, + { + "epoch": 2.3223300970873786, + "grad_norm": 0.24834813177585602, + "learning_rate": 8.398675569026613e-05, + "loss": 0.0086, + "step": 29900 + }, + { + "epoch": 2.323106796116505, + "grad_norm": 0.15090256929397583, + "learning_rate": 8.397462633067705e-05, + "loss": 0.0096, + "step": 29910 + }, + { + "epoch": 2.323883495145631, + "grad_norm": 0.21080993115901947, + "learning_rate": 8.396249325567392e-05, + "loss": 0.0067, + "step": 29920 + }, + { + "epoch": 2.3246601941747573, + "grad_norm": 0.2524457573890686, + "learning_rate": 8.395035646658357e-05, + "loss": 0.0112, + "step": 29930 + }, + { + "epoch": 2.3254368932038836, + "grad_norm": 0.1588609218597412, + "learning_rate": 8.39382159647333e-05, + "loss": 0.0084, + "step": 29940 + }, + { + "epoch": 2.32621359223301, + "grad_norm": 0.3289870619773865, + "learning_rate": 8.392607175145075e-05, + "loss": 0.0092, + "step": 29950 + }, + { + "epoch": 2.326990291262136, + "grad_norm": 0.21457475423812866, + "learning_rate": 8.3913923828064e-05, + "loss": 0.01, + "step": 29960 + }, + { + "epoch": 2.3277669902912623, + "grad_norm": 0.36048662662506104, + "learning_rate": 8.390177219590152e-05, + "loss": 0.0106, + "step": 29970 + }, + { + "epoch": 2.328543689320388, + "grad_norm": 0.39216676354408264, + "learning_rate": 8.388961685629222e-05, + "loss": 0.0081, + "step": 29980 + }, + { + "epoch": 2.3293203883495144, + "grad_norm": 0.24558646976947784, + "learning_rate": 8.387745781056536e-05, + "loss": 0.0097, + "step": 29990 + }, + { + "epoch": 2.3300970873786406, + "grad_norm": 0.19542734324932098, + "learning_rate": 8.386529506005065e-05, + "loss": 0.0114, + "step": 30000 + }, + { + "epoch": 2.330873786407767, + "grad_norm": 0.23404166102409363, + "learning_rate": 8.38531286060782e-05, + "loss": 0.0089, + "step": 30010 + }, + { + "epoch": 2.331650485436893, + "grad_norm": 0.21129478514194489, + "learning_rate": 8.384095844997849e-05, + "loss": 0.0103, + "step": 30020 + }, + { + "epoch": 2.3324271844660194, + "grad_norm": 0.1679915487766266, + "learning_rate": 8.382878459308245e-05, + "loss": 0.0084, + "step": 30030 + }, + { + "epoch": 2.3332038834951456, + "grad_norm": 0.18840953707695007, + "learning_rate": 8.381660703672138e-05, + "loss": 0.0084, + "step": 30040 + }, + { + "epoch": 2.333980582524272, + "grad_norm": 0.25785961747169495, + "learning_rate": 8.380442578222702e-05, + "loss": 0.0066, + "step": 30050 + }, + { + "epoch": 2.334757281553398, + "grad_norm": 0.18759380280971527, + "learning_rate": 8.379224083093146e-05, + "loss": 0.0107, + "step": 30060 + }, + { + "epoch": 2.3355339805825244, + "grad_norm": 0.19741055369377136, + "learning_rate": 8.378005218416727e-05, + "loss": 0.0087, + "step": 30070 + }, + { + "epoch": 2.3363106796116506, + "grad_norm": 0.2509554922580719, + "learning_rate": 8.376785984326735e-05, + "loss": 0.0078, + "step": 30080 + }, + { + "epoch": 2.337087378640777, + "grad_norm": 0.1914428323507309, + "learning_rate": 8.375566380956506e-05, + "loss": 0.0097, + "step": 30090 + }, + { + "epoch": 2.337864077669903, + "grad_norm": 0.23438695073127747, + "learning_rate": 8.374346408439411e-05, + "loss": 0.0107, + "step": 30100 + }, + { + "epoch": 2.3386407766990294, + "grad_norm": 0.2578018307685852, + "learning_rate": 8.373126066908868e-05, + "loss": 0.0095, + "step": 30110 + }, + { + "epoch": 2.339417475728155, + "grad_norm": 0.3055989444255829, + "learning_rate": 8.371905356498326e-05, + "loss": 0.0122, + "step": 30120 + }, + { + "epoch": 2.3401941747572814, + "grad_norm": 0.2710946202278137, + "learning_rate": 8.370684277341288e-05, + "loss": 0.0139, + "step": 30130 + }, + { + "epoch": 2.3409708737864077, + "grad_norm": 0.36536598205566406, + "learning_rate": 8.369462829571282e-05, + "loss": 0.0097, + "step": 30140 + }, + { + "epoch": 2.341747572815534, + "grad_norm": 0.46609413623809814, + "learning_rate": 8.36824101332189e-05, + "loss": 0.011, + "step": 30150 + }, + { + "epoch": 2.34252427184466, + "grad_norm": 0.35855260491371155, + "learning_rate": 8.367018828726721e-05, + "loss": 0.0098, + "step": 30160 + }, + { + "epoch": 2.3433009708737864, + "grad_norm": 0.4082709848880768, + "learning_rate": 8.365796275919438e-05, + "loss": 0.0117, + "step": 30170 + }, + { + "epoch": 2.3440776699029127, + "grad_norm": 0.2889266014099121, + "learning_rate": 8.364573355033734e-05, + "loss": 0.0093, + "step": 30180 + }, + { + "epoch": 2.344854368932039, + "grad_norm": 0.24624553322792053, + "learning_rate": 8.363350066203346e-05, + "loss": 0.0082, + "step": 30190 + }, + { + "epoch": 2.345631067961165, + "grad_norm": 0.29866302013397217, + "learning_rate": 8.362126409562053e-05, + "loss": 0.011, + "step": 30200 + }, + { + "epoch": 2.3464077669902914, + "grad_norm": 0.32844507694244385, + "learning_rate": 8.360902385243667e-05, + "loss": 0.0104, + "step": 30210 + }, + { + "epoch": 2.3471844660194177, + "grad_norm": 0.24857617914676666, + "learning_rate": 8.359677993382052e-05, + "loss": 0.0086, + "step": 30220 + }, + { + "epoch": 2.3479611650485435, + "grad_norm": 0.3114752769470215, + "learning_rate": 8.358453234111103e-05, + "loss": 0.0112, + "step": 30230 + }, + { + "epoch": 2.3487378640776697, + "grad_norm": 0.4296402633190155, + "learning_rate": 8.357228107564756e-05, + "loss": 0.0093, + "step": 30240 + }, + { + "epoch": 2.349514563106796, + "grad_norm": 0.17906232178211212, + "learning_rate": 8.356002613876993e-05, + "loss": 0.0084, + "step": 30250 + }, + { + "epoch": 2.350291262135922, + "grad_norm": 0.2287811040878296, + "learning_rate": 8.35477675318183e-05, + "loss": 0.0078, + "step": 30260 + }, + { + "epoch": 2.3510679611650485, + "grad_norm": 0.23874159157276154, + "learning_rate": 8.353550525613323e-05, + "loss": 0.0077, + "step": 30270 + }, + { + "epoch": 2.3518446601941747, + "grad_norm": 0.24335704743862152, + "learning_rate": 8.352323931305572e-05, + "loss": 0.0107, + "step": 30280 + }, + { + "epoch": 2.352621359223301, + "grad_norm": 0.1995450258255005, + "learning_rate": 8.351096970392717e-05, + "loss": 0.0071, + "step": 30290 + }, + { + "epoch": 2.353398058252427, + "grad_norm": 0.22260253131389618, + "learning_rate": 8.349869643008937e-05, + "loss": 0.01, + "step": 30300 + }, + { + "epoch": 2.3541747572815535, + "grad_norm": 0.23325948417186737, + "learning_rate": 8.348641949288449e-05, + "loss": 0.0104, + "step": 30310 + }, + { + "epoch": 2.3549514563106797, + "grad_norm": 0.33266010880470276, + "learning_rate": 8.34741388936551e-05, + "loss": 0.0083, + "step": 30320 + }, + { + "epoch": 2.355728155339806, + "grad_norm": 0.2830928862094879, + "learning_rate": 8.346185463374423e-05, + "loss": 0.0114, + "step": 30330 + }, + { + "epoch": 2.356504854368932, + "grad_norm": 0.2681994140148163, + "learning_rate": 8.344956671449524e-05, + "loss": 0.009, + "step": 30340 + }, + { + "epoch": 2.3572815533980584, + "grad_norm": 0.30104056000709534, + "learning_rate": 8.343727513725192e-05, + "loss": 0.0103, + "step": 30350 + }, + { + "epoch": 2.3580582524271847, + "grad_norm": 0.32235679030418396, + "learning_rate": 8.342497990335847e-05, + "loss": 0.0106, + "step": 30360 + }, + { + "epoch": 2.3588349514563105, + "grad_norm": 0.31946465373039246, + "learning_rate": 8.341268101415946e-05, + "loss": 0.0091, + "step": 30370 + }, + { + "epoch": 2.3596116504854368, + "grad_norm": 0.3369065225124359, + "learning_rate": 8.34003784709999e-05, + "loss": 0.0119, + "step": 30380 + }, + { + "epoch": 2.360388349514563, + "grad_norm": 0.17520824074745178, + "learning_rate": 8.338807227522516e-05, + "loss": 0.0097, + "step": 30390 + }, + { + "epoch": 2.3611650485436892, + "grad_norm": 0.5096744894981384, + "learning_rate": 8.337576242818103e-05, + "loss": 0.0092, + "step": 30400 + }, + { + "epoch": 2.3619417475728155, + "grad_norm": 0.4483618140220642, + "learning_rate": 8.336344893121372e-05, + "loss": 0.0087, + "step": 30410 + }, + { + "epoch": 2.3627184466019417, + "grad_norm": 0.35839542746543884, + "learning_rate": 8.335113178566977e-05, + "loss": 0.0099, + "step": 30420 + }, + { + "epoch": 2.363495145631068, + "grad_norm": 0.28388452529907227, + "learning_rate": 8.33388109928962e-05, + "loss": 0.0077, + "step": 30430 + }, + { + "epoch": 2.3642718446601942, + "grad_norm": 0.2963455617427826, + "learning_rate": 8.33264865542404e-05, + "loss": 0.0087, + "step": 30440 + }, + { + "epoch": 2.3650485436893205, + "grad_norm": 0.4181365370750427, + "learning_rate": 8.331415847105013e-05, + "loss": 0.0094, + "step": 30450 + }, + { + "epoch": 2.3658252427184467, + "grad_norm": 0.2412540316581726, + "learning_rate": 8.330182674467357e-05, + "loss": 0.0086, + "step": 30460 + }, + { + "epoch": 2.366601941747573, + "grad_norm": 0.19715820252895355, + "learning_rate": 8.32894913764593e-05, + "loss": 0.0112, + "step": 30470 + }, + { + "epoch": 2.367378640776699, + "grad_norm": 0.20842310786247253, + "learning_rate": 8.327715236775633e-05, + "loss": 0.0088, + "step": 30480 + }, + { + "epoch": 2.368155339805825, + "grad_norm": 0.2542135417461395, + "learning_rate": 8.326480971991398e-05, + "loss": 0.0095, + "step": 30490 + }, + { + "epoch": 2.3689320388349513, + "grad_norm": 0.4968356490135193, + "learning_rate": 8.325246343428206e-05, + "loss": 0.0119, + "step": 30500 + }, + { + "epoch": 2.3697087378640775, + "grad_norm": 0.1958872228860855, + "learning_rate": 8.324011351221072e-05, + "loss": 0.0112, + "step": 30510 + }, + { + "epoch": 2.370485436893204, + "grad_norm": 0.23622477054595947, + "learning_rate": 8.322775995505057e-05, + "loss": 0.0086, + "step": 30520 + }, + { + "epoch": 2.37126213592233, + "grad_norm": 0.25754040479660034, + "learning_rate": 8.321540276415254e-05, + "loss": 0.0134, + "step": 30530 + }, + { + "epoch": 2.3720388349514563, + "grad_norm": 0.30077576637268066, + "learning_rate": 8.320304194086798e-05, + "loss": 0.007, + "step": 30540 + }, + { + "epoch": 2.3728155339805825, + "grad_norm": 0.3897506594657898, + "learning_rate": 8.31906774865487e-05, + "loss": 0.0116, + "step": 30550 + }, + { + "epoch": 2.373592233009709, + "grad_norm": 0.24630610644817352, + "learning_rate": 8.317830940254682e-05, + "loss": 0.0075, + "step": 30560 + }, + { + "epoch": 2.374368932038835, + "grad_norm": 0.38953086733818054, + "learning_rate": 8.316593769021491e-05, + "loss": 0.0119, + "step": 30570 + }, + { + "epoch": 2.3751456310679613, + "grad_norm": 0.2652002274990082, + "learning_rate": 8.315356235090592e-05, + "loss": 0.0083, + "step": 30580 + }, + { + "epoch": 2.3759223300970875, + "grad_norm": 0.24446968734264374, + "learning_rate": 8.314118338597319e-05, + "loss": 0.012, + "step": 30590 + }, + { + "epoch": 2.376699029126214, + "grad_norm": 0.13073121011257172, + "learning_rate": 8.312880079677048e-05, + "loss": 0.0114, + "step": 30600 + }, + { + "epoch": 2.37747572815534, + "grad_norm": 0.22208330035209656, + "learning_rate": 8.311641458465191e-05, + "loss": 0.0107, + "step": 30610 + }, + { + "epoch": 2.378252427184466, + "grad_norm": 0.2638915181159973, + "learning_rate": 8.310402475097205e-05, + "loss": 0.0089, + "step": 30620 + }, + { + "epoch": 2.379029126213592, + "grad_norm": 0.15964384377002716, + "learning_rate": 8.309163129708581e-05, + "loss": 0.0078, + "step": 30630 + }, + { + "epoch": 2.3798058252427183, + "grad_norm": 0.1474626660346985, + "learning_rate": 8.307923422434852e-05, + "loss": 0.0093, + "step": 30640 + }, + { + "epoch": 2.3805825242718446, + "grad_norm": 0.15897077322006226, + "learning_rate": 8.30668335341159e-05, + "loss": 0.0076, + "step": 30650 + }, + { + "epoch": 2.381359223300971, + "grad_norm": 0.24975688755512238, + "learning_rate": 8.305442922774408e-05, + "loss": 0.0139, + "step": 30660 + }, + { + "epoch": 2.382135922330097, + "grad_norm": 0.21926604211330414, + "learning_rate": 8.304202130658959e-05, + "loss": 0.0103, + "step": 30670 + }, + { + "epoch": 2.3829126213592233, + "grad_norm": 0.17819324135780334, + "learning_rate": 8.302960977200931e-05, + "loss": 0.0084, + "step": 30680 + }, + { + "epoch": 2.3836893203883496, + "grad_norm": 0.12160328775644302, + "learning_rate": 8.301719462536058e-05, + "loss": 0.0073, + "step": 30690 + }, + { + "epoch": 2.384466019417476, + "grad_norm": 0.20225770771503448, + "learning_rate": 8.300477586800108e-05, + "loss": 0.0081, + "step": 30700 + }, + { + "epoch": 2.385242718446602, + "grad_norm": 0.2903963327407837, + "learning_rate": 8.299235350128892e-05, + "loss": 0.0119, + "step": 30710 + }, + { + "epoch": 2.3860194174757283, + "grad_norm": 0.25164473056793213, + "learning_rate": 8.297992752658258e-05, + "loss": 0.0102, + "step": 30720 + }, + { + "epoch": 2.386796116504854, + "grad_norm": 0.2067314088344574, + "learning_rate": 8.296749794524094e-05, + "loss": 0.0106, + "step": 30730 + }, + { + "epoch": 2.3875728155339804, + "grad_norm": 0.21080702543258667, + "learning_rate": 8.295506475862332e-05, + "loss": 0.0096, + "step": 30740 + }, + { + "epoch": 2.3883495145631066, + "grad_norm": 0.1510646641254425, + "learning_rate": 8.294262796808933e-05, + "loss": 0.007, + "step": 30750 + }, + { + "epoch": 2.389126213592233, + "grad_norm": 0.2605535387992859, + "learning_rate": 8.293018757499909e-05, + "loss": 0.0056, + "step": 30760 + }, + { + "epoch": 2.389902912621359, + "grad_norm": 0.49299830198287964, + "learning_rate": 8.291774358071305e-05, + "loss": 0.0088, + "step": 30770 + }, + { + "epoch": 2.3906796116504854, + "grad_norm": 0.19101133942604065, + "learning_rate": 8.290529598659205e-05, + "loss": 0.01, + "step": 30780 + }, + { + "epoch": 2.3914563106796116, + "grad_norm": 0.2556106746196747, + "learning_rate": 8.289284479399738e-05, + "loss": 0.013, + "step": 30790 + }, + { + "epoch": 2.392233009708738, + "grad_norm": 0.3678683936595917, + "learning_rate": 8.288039000429064e-05, + "loss": 0.0093, + "step": 30800 + }, + { + "epoch": 2.393009708737864, + "grad_norm": 0.14799098670482635, + "learning_rate": 8.286793161883388e-05, + "loss": 0.0077, + "step": 30810 + }, + { + "epoch": 2.3937864077669904, + "grad_norm": 0.24527689814567566, + "learning_rate": 8.285546963898954e-05, + "loss": 0.0077, + "step": 30820 + }, + { + "epoch": 2.3945631067961166, + "grad_norm": 0.26020243763923645, + "learning_rate": 8.284300406612044e-05, + "loss": 0.0108, + "step": 30830 + }, + { + "epoch": 2.395339805825243, + "grad_norm": 0.17371907830238342, + "learning_rate": 8.283053490158978e-05, + "loss": 0.01, + "step": 30840 + }, + { + "epoch": 2.396116504854369, + "grad_norm": 0.7195031642913818, + "learning_rate": 8.28180621467612e-05, + "loss": 0.01, + "step": 30850 + }, + { + "epoch": 2.3968932038834954, + "grad_norm": 0.2933526933193207, + "learning_rate": 8.280558580299867e-05, + "loss": 0.0089, + "step": 30860 + }, + { + "epoch": 2.3976699029126216, + "grad_norm": 0.2216930091381073, + "learning_rate": 8.279310587166661e-05, + "loss": 0.0097, + "step": 30870 + }, + { + "epoch": 2.3984466019417474, + "grad_norm": 0.23118750751018524, + "learning_rate": 8.278062235412978e-05, + "loss": 0.0089, + "step": 30880 + }, + { + "epoch": 2.3992233009708737, + "grad_norm": 0.1634669005870819, + "learning_rate": 8.276813525175339e-05, + "loss": 0.0082, + "step": 30890 + }, + { + "epoch": 2.4, + "grad_norm": 0.31719258427619934, + "learning_rate": 8.2755644565903e-05, + "loss": 0.009, + "step": 30900 + }, + { + "epoch": 2.400776699029126, + "grad_norm": 0.185744047164917, + "learning_rate": 8.274315029794454e-05, + "loss": 0.0083, + "step": 30910 + }, + { + "epoch": 2.4015533980582524, + "grad_norm": 0.253418505191803, + "learning_rate": 8.273065244924443e-05, + "loss": 0.0078, + "step": 30920 + }, + { + "epoch": 2.4023300970873787, + "grad_norm": 0.2987808585166931, + "learning_rate": 8.271815102116936e-05, + "loss": 0.0079, + "step": 30930 + }, + { + "epoch": 2.403106796116505, + "grad_norm": 0.19147485494613647, + "learning_rate": 8.270564601508648e-05, + "loss": 0.0098, + "step": 30940 + }, + { + "epoch": 2.403883495145631, + "grad_norm": 0.28625357151031494, + "learning_rate": 8.269313743236333e-05, + "loss": 0.0125, + "step": 30950 + }, + { + "epoch": 2.4046601941747574, + "grad_norm": 0.19365404546260834, + "learning_rate": 8.268062527436783e-05, + "loss": 0.0087, + "step": 30960 + }, + { + "epoch": 2.4054368932038837, + "grad_norm": 0.3410350978374481, + "learning_rate": 8.266810954246828e-05, + "loss": 0.011, + "step": 30970 + }, + { + "epoch": 2.40621359223301, + "grad_norm": 0.668620765209198, + "learning_rate": 8.265559023803338e-05, + "loss": 0.0095, + "step": 30980 + }, + { + "epoch": 2.4069902912621357, + "grad_norm": 0.21338090300559998, + "learning_rate": 8.264306736243223e-05, + "loss": 0.0122, + "step": 30990 + }, + { + "epoch": 2.407766990291262, + "grad_norm": 0.23741622269153595, + "learning_rate": 8.263054091703432e-05, + "loss": 0.0086, + "step": 31000 + }, + { + "epoch": 2.408543689320388, + "grad_norm": 0.18499812483787537, + "learning_rate": 8.26180109032095e-05, + "loss": 0.0099, + "step": 31010 + }, + { + "epoch": 2.4093203883495145, + "grad_norm": 0.3189586400985718, + "learning_rate": 8.260547732232807e-05, + "loss": 0.0123, + "step": 31020 + }, + { + "epoch": 2.4100970873786407, + "grad_norm": 0.32763367891311646, + "learning_rate": 8.259294017576063e-05, + "loss": 0.0096, + "step": 31030 + }, + { + "epoch": 2.410873786407767, + "grad_norm": 0.1948084980249405, + "learning_rate": 8.258039946487828e-05, + "loss": 0.0132, + "step": 31040 + }, + { + "epoch": 2.411650485436893, + "grad_norm": 0.2992154061794281, + "learning_rate": 8.256785519105241e-05, + "loss": 0.0115, + "step": 31050 + }, + { + "epoch": 2.4124271844660194, + "grad_norm": 0.42021048069000244, + "learning_rate": 8.255530735565488e-05, + "loss": 0.0117, + "step": 31060 + }, + { + "epoch": 2.4132038834951457, + "grad_norm": 0.35128092765808105, + "learning_rate": 8.254275596005787e-05, + "loss": 0.0096, + "step": 31070 + }, + { + "epoch": 2.413980582524272, + "grad_norm": 0.37216150760650635, + "learning_rate": 8.2530201005634e-05, + "loss": 0.0085, + "step": 31080 + }, + { + "epoch": 2.414757281553398, + "grad_norm": 0.4222675561904907, + "learning_rate": 8.251764249375626e-05, + "loss": 0.0136, + "step": 31090 + }, + { + "epoch": 2.4155339805825244, + "grad_norm": 0.27118194103240967, + "learning_rate": 8.250508042579803e-05, + "loss": 0.0092, + "step": 31100 + }, + { + "epoch": 2.4163106796116507, + "grad_norm": 0.3195841908454895, + "learning_rate": 8.249251480313307e-05, + "loss": 0.0112, + "step": 31110 + }, + { + "epoch": 2.417087378640777, + "grad_norm": 0.3171212375164032, + "learning_rate": 8.247994562713555e-05, + "loss": 0.0095, + "step": 31120 + }, + { + "epoch": 2.4178640776699027, + "grad_norm": 0.4423663020133972, + "learning_rate": 8.246737289917998e-05, + "loss": 0.009, + "step": 31130 + }, + { + "epoch": 2.418640776699029, + "grad_norm": 0.267473429441452, + "learning_rate": 8.245479662064135e-05, + "loss": 0.0097, + "step": 31140 + }, + { + "epoch": 2.4194174757281552, + "grad_norm": 0.2289944589138031, + "learning_rate": 8.244221679289496e-05, + "loss": 0.0125, + "step": 31150 + }, + { + "epoch": 2.4201941747572815, + "grad_norm": 0.19665056467056274, + "learning_rate": 8.242963341731652e-05, + "loss": 0.01, + "step": 31160 + }, + { + "epoch": 2.4209708737864077, + "grad_norm": 0.32902324199676514, + "learning_rate": 8.24170464952821e-05, + "loss": 0.0086, + "step": 31170 + }, + { + "epoch": 2.421747572815534, + "grad_norm": 0.3978017568588257, + "learning_rate": 8.240445602816824e-05, + "loss": 0.0079, + "step": 31180 + }, + { + "epoch": 2.4225242718446602, + "grad_norm": 0.20274251699447632, + "learning_rate": 8.239186201735179e-05, + "loss": 0.0084, + "step": 31190 + }, + { + "epoch": 2.4233009708737865, + "grad_norm": 0.3943507969379425, + "learning_rate": 8.237926446420998e-05, + "loss": 0.0108, + "step": 31200 + }, + { + "epoch": 2.4240776699029127, + "grad_norm": 0.3575437366962433, + "learning_rate": 8.236666337012052e-05, + "loss": 0.0089, + "step": 31210 + }, + { + "epoch": 2.424854368932039, + "grad_norm": 0.43548035621643066, + "learning_rate": 8.23540587364614e-05, + "loss": 0.0079, + "step": 31220 + }, + { + "epoch": 2.4256310679611652, + "grad_norm": 0.5959183573722839, + "learning_rate": 8.234145056461107e-05, + "loss": 0.0103, + "step": 31230 + }, + { + "epoch": 2.426407766990291, + "grad_norm": 0.2447260171175003, + "learning_rate": 8.232883885594832e-05, + "loss": 0.008, + "step": 31240 + }, + { + "epoch": 2.4271844660194173, + "grad_norm": 0.27538827061653137, + "learning_rate": 8.231622361185236e-05, + "loss": 0.0085, + "step": 31250 + }, + { + "epoch": 2.4279611650485435, + "grad_norm": 0.3041943609714508, + "learning_rate": 8.230360483370278e-05, + "loss": 0.0085, + "step": 31260 + }, + { + "epoch": 2.42873786407767, + "grad_norm": 0.26087409257888794, + "learning_rate": 8.229098252287953e-05, + "loss": 0.01, + "step": 31270 + }, + { + "epoch": 2.429514563106796, + "grad_norm": 0.23456817865371704, + "learning_rate": 8.2278356680763e-05, + "loss": 0.0089, + "step": 31280 + }, + { + "epoch": 2.4302912621359223, + "grad_norm": 0.20356965065002441, + "learning_rate": 8.22657273087339e-05, + "loss": 0.0118, + "step": 31290 + }, + { + "epoch": 2.4310679611650485, + "grad_norm": 0.3377608358860016, + "learning_rate": 8.225309440817336e-05, + "loss": 0.0099, + "step": 31300 + }, + { + "epoch": 2.431844660194175, + "grad_norm": 0.2678244411945343, + "learning_rate": 8.224045798046293e-05, + "loss": 0.0077, + "step": 31310 + }, + { + "epoch": 2.432621359223301, + "grad_norm": 0.26631903648376465, + "learning_rate": 8.22278180269845e-05, + "loss": 0.0108, + "step": 31320 + }, + { + "epoch": 2.4333980582524273, + "grad_norm": 0.28007304668426514, + "learning_rate": 8.221517454912031e-05, + "loss": 0.0091, + "step": 31330 + }, + { + "epoch": 2.4341747572815535, + "grad_norm": 0.20592497289180756, + "learning_rate": 8.220252754825308e-05, + "loss": 0.0126, + "step": 31340 + }, + { + "epoch": 2.4349514563106798, + "grad_norm": 0.23438234627246857, + "learning_rate": 8.218987702576586e-05, + "loss": 0.0066, + "step": 31350 + }, + { + "epoch": 2.435728155339806, + "grad_norm": 0.21972887217998505, + "learning_rate": 8.217722298304207e-05, + "loss": 0.0092, + "step": 31360 + }, + { + "epoch": 2.4365048543689323, + "grad_norm": 0.3705545663833618, + "learning_rate": 8.216456542146557e-05, + "loss": 0.0087, + "step": 31370 + }, + { + "epoch": 2.437281553398058, + "grad_norm": 0.6648123860359192, + "learning_rate": 8.215190434242055e-05, + "loss": 0.0092, + "step": 31380 + }, + { + "epoch": 2.4380582524271843, + "grad_norm": 0.25205352902412415, + "learning_rate": 8.213923974729161e-05, + "loss": 0.0089, + "step": 31390 + }, + { + "epoch": 2.4388349514563106, + "grad_norm": 0.1531076282262802, + "learning_rate": 8.212657163746373e-05, + "loss": 0.0084, + "step": 31400 + }, + { + "epoch": 2.439611650485437, + "grad_norm": 0.14339621365070343, + "learning_rate": 8.211390001432227e-05, + "loss": 0.0117, + "step": 31410 + }, + { + "epoch": 2.440388349514563, + "grad_norm": 0.24164047837257385, + "learning_rate": 8.210122487925297e-05, + "loss": 0.0094, + "step": 31420 + }, + { + "epoch": 2.4411650485436893, + "grad_norm": 0.14294371008872986, + "learning_rate": 8.208854623364202e-05, + "loss": 0.0086, + "step": 31430 + }, + { + "epoch": 2.4419417475728156, + "grad_norm": 0.3160654604434967, + "learning_rate": 8.207586407887589e-05, + "loss": 0.0091, + "step": 31440 + }, + { + "epoch": 2.442718446601942, + "grad_norm": 0.16487258672714233, + "learning_rate": 8.206317841634148e-05, + "loss": 0.0114, + "step": 31450 + }, + { + "epoch": 2.443495145631068, + "grad_norm": 0.22495615482330322, + "learning_rate": 8.205048924742609e-05, + "loss": 0.0093, + "step": 31460 + }, + { + "epoch": 2.4442718446601943, + "grad_norm": 0.21728910505771637, + "learning_rate": 8.203779657351738e-05, + "loss": 0.0089, + "step": 31470 + }, + { + "epoch": 2.4450485436893206, + "grad_norm": 0.2575297951698303, + "learning_rate": 8.20251003960034e-05, + "loss": 0.0129, + "step": 31480 + }, + { + "epoch": 2.4458252427184464, + "grad_norm": 0.615207850933075, + "learning_rate": 8.201240071627258e-05, + "loss": 0.0108, + "step": 31490 + }, + { + "epoch": 2.4466019417475726, + "grad_norm": 0.4022900462150574, + "learning_rate": 8.199969753571377e-05, + "loss": 0.0089, + "step": 31500 + }, + { + "epoch": 2.447378640776699, + "grad_norm": 0.20992223918437958, + "learning_rate": 8.198699085571615e-05, + "loss": 0.0084, + "step": 31510 + }, + { + "epoch": 2.448155339805825, + "grad_norm": 0.17653287947177887, + "learning_rate": 8.197428067766928e-05, + "loss": 0.0077, + "step": 31520 + }, + { + "epoch": 2.4489320388349514, + "grad_norm": 0.23858977854251862, + "learning_rate": 8.196156700296316e-05, + "loss": 0.0115, + "step": 31530 + }, + { + "epoch": 2.4497087378640776, + "grad_norm": 0.2560061812400818, + "learning_rate": 8.194884983298814e-05, + "loss": 0.0092, + "step": 31540 + }, + { + "epoch": 2.450485436893204, + "grad_norm": 0.33911654353141785, + "learning_rate": 8.193612916913491e-05, + "loss": 0.0067, + "step": 31550 + }, + { + "epoch": 2.45126213592233, + "grad_norm": 0.24436406791210175, + "learning_rate": 8.192340501279463e-05, + "loss": 0.0102, + "step": 31560 + }, + { + "epoch": 2.4520388349514564, + "grad_norm": 0.2652478814125061, + "learning_rate": 8.191067736535876e-05, + "loss": 0.0099, + "step": 31570 + }, + { + "epoch": 2.4528155339805826, + "grad_norm": 0.5468562245368958, + "learning_rate": 8.18979462282192e-05, + "loss": 0.0128, + "step": 31580 + }, + { + "epoch": 2.453592233009709, + "grad_norm": 0.29931649565696716, + "learning_rate": 8.188521160276819e-05, + "loss": 0.0092, + "step": 31590 + }, + { + "epoch": 2.454368932038835, + "grad_norm": 0.2408657670021057, + "learning_rate": 8.187247349039837e-05, + "loss": 0.0094, + "step": 31600 + }, + { + "epoch": 2.4551456310679614, + "grad_norm": 0.21426790952682495, + "learning_rate": 8.185973189250278e-05, + "loss": 0.0104, + "step": 31610 + }, + { + "epoch": 2.4559223300970876, + "grad_norm": 0.16726185381412506, + "learning_rate": 8.184698681047482e-05, + "loss": 0.0097, + "step": 31620 + }, + { + "epoch": 2.4566990291262134, + "grad_norm": 0.2116374969482422, + "learning_rate": 8.183423824570827e-05, + "loss": 0.0072, + "step": 31630 + }, + { + "epoch": 2.4574757281553397, + "grad_norm": 0.3274523913860321, + "learning_rate": 8.182148619959725e-05, + "loss": 0.0098, + "step": 31640 + }, + { + "epoch": 2.458252427184466, + "grad_norm": 0.31491944193840027, + "learning_rate": 8.180873067353636e-05, + "loss": 0.0059, + "step": 31650 + }, + { + "epoch": 2.459029126213592, + "grad_norm": 0.1934959590435028, + "learning_rate": 8.179597166892052e-05, + "loss": 0.0106, + "step": 31660 + }, + { + "epoch": 2.4598058252427184, + "grad_norm": 0.4060825705528259, + "learning_rate": 8.178320918714501e-05, + "loss": 0.0128, + "step": 31670 + }, + { + "epoch": 2.4605825242718447, + "grad_norm": 0.22897806763648987, + "learning_rate": 8.177044322960554e-05, + "loss": 0.0082, + "step": 31680 + }, + { + "epoch": 2.461359223300971, + "grad_norm": 0.33754944801330566, + "learning_rate": 8.175767379769816e-05, + "loss": 0.0108, + "step": 31690 + }, + { + "epoch": 2.462135922330097, + "grad_norm": 0.2804928123950958, + "learning_rate": 8.174490089281932e-05, + "loss": 0.0113, + "step": 31700 + }, + { + "epoch": 2.4629126213592234, + "grad_norm": 0.32950085401535034, + "learning_rate": 8.173212451636584e-05, + "loss": 0.0117, + "step": 31710 + }, + { + "epoch": 2.4636893203883496, + "grad_norm": 0.19024738669395447, + "learning_rate": 8.171934466973493e-05, + "loss": 0.0089, + "step": 31720 + }, + { + "epoch": 2.464466019417476, + "grad_norm": 0.23381240665912628, + "learning_rate": 8.170656135432418e-05, + "loss": 0.009, + "step": 31730 + }, + { + "epoch": 2.4652427184466017, + "grad_norm": 0.31158292293548584, + "learning_rate": 8.169377457153155e-05, + "loss": 0.0093, + "step": 31740 + }, + { + "epoch": 2.466019417475728, + "grad_norm": 0.37471216917037964, + "learning_rate": 8.168098432275539e-05, + "loss": 0.0096, + "step": 31750 + }, + { + "epoch": 2.466796116504854, + "grad_norm": 0.17226547002792358, + "learning_rate": 8.166819060939442e-05, + "loss": 0.0109, + "step": 31760 + }, + { + "epoch": 2.4675728155339804, + "grad_norm": 0.24889278411865234, + "learning_rate": 8.165539343284772e-05, + "loss": 0.0085, + "step": 31770 + }, + { + "epoch": 2.4683495145631067, + "grad_norm": 0.5102967023849487, + "learning_rate": 8.16425927945148e-05, + "loss": 0.0093, + "step": 31780 + }, + { + "epoch": 2.469126213592233, + "grad_norm": 0.3023306727409363, + "learning_rate": 8.162978869579551e-05, + "loss": 0.0105, + "step": 31790 + }, + { + "epoch": 2.469902912621359, + "grad_norm": 0.26859521865844727, + "learning_rate": 8.161698113809007e-05, + "loss": 0.0112, + "step": 31800 + }, + { + "epoch": 2.4706796116504854, + "grad_norm": 0.3324812650680542, + "learning_rate": 8.160417012279911e-05, + "loss": 0.0107, + "step": 31810 + }, + { + "epoch": 2.4714563106796117, + "grad_norm": 0.33653372526168823, + "learning_rate": 8.159135565132363e-05, + "loss": 0.0101, + "step": 31820 + }, + { + "epoch": 2.472233009708738, + "grad_norm": 0.26491740345954895, + "learning_rate": 8.157853772506498e-05, + "loss": 0.0118, + "step": 31830 + }, + { + "epoch": 2.473009708737864, + "grad_norm": 0.27279895544052124, + "learning_rate": 8.156571634542494e-05, + "loss": 0.0089, + "step": 31840 + }, + { + "epoch": 2.4737864077669904, + "grad_norm": 0.21443438529968262, + "learning_rate": 8.15528915138056e-05, + "loss": 0.0099, + "step": 31850 + }, + { + "epoch": 2.4745631067961167, + "grad_norm": 0.21623903512954712, + "learning_rate": 8.154006323160949e-05, + "loss": 0.0098, + "step": 31860 + }, + { + "epoch": 2.475339805825243, + "grad_norm": 0.25142917037010193, + "learning_rate": 8.152723150023949e-05, + "loss": 0.0089, + "step": 31870 + }, + { + "epoch": 2.4761165048543687, + "grad_norm": 0.30883294343948364, + "learning_rate": 8.151439632109886e-05, + "loss": 0.0087, + "step": 31880 + }, + { + "epoch": 2.476893203883495, + "grad_norm": 0.2093716561794281, + "learning_rate": 8.150155769559122e-05, + "loss": 0.0089, + "step": 31890 + }, + { + "epoch": 2.4776699029126212, + "grad_norm": 0.4413011968135834, + "learning_rate": 8.148871562512058e-05, + "loss": 0.0144, + "step": 31900 + }, + { + "epoch": 2.4784466019417475, + "grad_norm": 0.19292831420898438, + "learning_rate": 8.147587011109136e-05, + "loss": 0.0109, + "step": 31910 + }, + { + "epoch": 2.4792233009708737, + "grad_norm": 0.3477630019187927, + "learning_rate": 8.14630211549083e-05, + "loss": 0.0086, + "step": 31920 + }, + { + "epoch": 2.48, + "grad_norm": 0.28484413027763367, + "learning_rate": 8.145016875797655e-05, + "loss": 0.0088, + "step": 31930 + }, + { + "epoch": 2.4807766990291262, + "grad_norm": 0.16401483118534088, + "learning_rate": 8.143731292170164e-05, + "loss": 0.0119, + "step": 31940 + }, + { + "epoch": 2.4815533980582525, + "grad_norm": 0.40042561292648315, + "learning_rate": 8.142445364748944e-05, + "loss": 0.0118, + "step": 31950 + }, + { + "epoch": 2.4823300970873787, + "grad_norm": 0.2694847881793976, + "learning_rate": 8.141159093674624e-05, + "loss": 0.01, + "step": 31960 + }, + { + "epoch": 2.483106796116505, + "grad_norm": 0.2681480944156647, + "learning_rate": 8.139872479087869e-05, + "loss": 0.0087, + "step": 31970 + }, + { + "epoch": 2.4838834951456312, + "grad_norm": 0.20851776003837585, + "learning_rate": 8.13858552112938e-05, + "loss": 0.0103, + "step": 31980 + }, + { + "epoch": 2.4846601941747575, + "grad_norm": 0.25956952571868896, + "learning_rate": 8.137298219939895e-05, + "loss": 0.0082, + "step": 31990 + }, + { + "epoch": 2.4854368932038833, + "grad_norm": 0.23260152339935303, + "learning_rate": 8.136010575660196e-05, + "loss": 0.0114, + "step": 32000 + }, + { + "epoch": 2.4862135922330095, + "grad_norm": 0.27697697281837463, + "learning_rate": 8.134722588431095e-05, + "loss": 0.0088, + "step": 32010 + }, + { + "epoch": 2.4869902912621358, + "grad_norm": 0.26332616806030273, + "learning_rate": 8.133434258393444e-05, + "loss": 0.0098, + "step": 32020 + }, + { + "epoch": 2.487766990291262, + "grad_norm": 0.28564438223838806, + "learning_rate": 8.132145585688134e-05, + "loss": 0.0094, + "step": 32030 + }, + { + "epoch": 2.4885436893203883, + "grad_norm": 0.3544200360774994, + "learning_rate": 8.130856570456093e-05, + "loss": 0.0067, + "step": 32040 + }, + { + "epoch": 2.4893203883495145, + "grad_norm": 0.4293207824230194, + "learning_rate": 8.129567212838283e-05, + "loss": 0.0096, + "step": 32050 + }, + { + "epoch": 2.4900970873786408, + "grad_norm": 0.4109458923339844, + "learning_rate": 8.128277512975708e-05, + "loss": 0.0087, + "step": 32060 + }, + { + "epoch": 2.490873786407767, + "grad_norm": 0.4214215874671936, + "learning_rate": 8.126987471009408e-05, + "loss": 0.0062, + "step": 32070 + }, + { + "epoch": 2.4916504854368933, + "grad_norm": 0.5627671480178833, + "learning_rate": 8.125697087080459e-05, + "loss": 0.0114, + "step": 32080 + }, + { + "epoch": 2.4924271844660195, + "grad_norm": 0.32981783151626587, + "learning_rate": 8.124406361329976e-05, + "loss": 0.0072, + "step": 32090 + }, + { + "epoch": 2.4932038834951458, + "grad_norm": 0.1800670176744461, + "learning_rate": 8.123115293899111e-05, + "loss": 0.0081, + "step": 32100 + }, + { + "epoch": 2.493980582524272, + "grad_norm": 0.4606460928916931, + "learning_rate": 8.121823884929055e-05, + "loss": 0.009, + "step": 32110 + }, + { + "epoch": 2.4947572815533983, + "grad_norm": 0.21335957944393158, + "learning_rate": 8.12053213456103e-05, + "loss": 0.0074, + "step": 32120 + }, + { + "epoch": 2.4955339805825245, + "grad_norm": 0.2281060665845871, + "learning_rate": 8.119240042936303e-05, + "loss": 0.0081, + "step": 32130 + }, + { + "epoch": 2.4963106796116503, + "grad_norm": 0.19069506227970123, + "learning_rate": 8.117947610196175e-05, + "loss": 0.0068, + "step": 32140 + }, + { + "epoch": 2.4970873786407766, + "grad_norm": 0.41245853900909424, + "learning_rate": 8.116654836481982e-05, + "loss": 0.0088, + "step": 32150 + }, + { + "epoch": 2.497864077669903, + "grad_norm": 0.22853510081768036, + "learning_rate": 8.115361721935106e-05, + "loss": 0.0099, + "step": 32160 + }, + { + "epoch": 2.498640776699029, + "grad_norm": 0.15818633139133453, + "learning_rate": 8.114068266696953e-05, + "loss": 0.0109, + "step": 32170 + }, + { + "epoch": 2.4994174757281553, + "grad_norm": 0.31254100799560547, + "learning_rate": 8.112774470908978e-05, + "loss": 0.0112, + "step": 32180 + }, + { + "epoch": 2.5001941747572816, + "grad_norm": 0.25660598278045654, + "learning_rate": 8.111480334712665e-05, + "loss": 0.0073, + "step": 32190 + }, + { + "epoch": 2.500970873786408, + "grad_norm": 0.42954379320144653, + "learning_rate": 8.110185858249542e-05, + "loss": 0.0098, + "step": 32200 + }, + { + "epoch": 2.501747572815534, + "grad_norm": 0.38190317153930664, + "learning_rate": 8.108891041661168e-05, + "loss": 0.0096, + "step": 32210 + }, + { + "epoch": 2.5025242718446603, + "grad_norm": 0.34002792835235596, + "learning_rate": 8.107595885089146e-05, + "loss": 0.0091, + "step": 32220 + }, + { + "epoch": 2.5033009708737866, + "grad_norm": 0.39349091053009033, + "learning_rate": 8.106300388675108e-05, + "loss": 0.0088, + "step": 32230 + }, + { + "epoch": 2.5040776699029124, + "grad_norm": 0.2082548439502716, + "learning_rate": 8.105004552560731e-05, + "loss": 0.0085, + "step": 32240 + }, + { + "epoch": 2.5048543689320386, + "grad_norm": 0.3507069945335388, + "learning_rate": 8.103708376887724e-05, + "loss": 0.011, + "step": 32250 + }, + { + "epoch": 2.505631067961165, + "grad_norm": 0.2139807939529419, + "learning_rate": 8.102411861797836e-05, + "loss": 0.0076, + "step": 32260 + }, + { + "epoch": 2.506407766990291, + "grad_norm": 0.18353818356990814, + "learning_rate": 8.101115007432851e-05, + "loss": 0.0078, + "step": 32270 + }, + { + "epoch": 2.5071844660194174, + "grad_norm": 0.2165270447731018, + "learning_rate": 8.09981781393459e-05, + "loss": 0.0078, + "step": 32280 + }, + { + "epoch": 2.5079611650485436, + "grad_norm": 0.2500276267528534, + "learning_rate": 8.098520281444915e-05, + "loss": 0.0095, + "step": 32290 + }, + { + "epoch": 2.50873786407767, + "grad_norm": 0.5092483162879944, + "learning_rate": 8.09722241010572e-05, + "loss": 0.0083, + "step": 32300 + }, + { + "epoch": 2.509514563106796, + "grad_norm": 0.2860886752605438, + "learning_rate": 8.095924200058939e-05, + "loss": 0.0079, + "step": 32310 + }, + { + "epoch": 2.5102912621359224, + "grad_norm": 0.20776139199733734, + "learning_rate": 8.094625651446541e-05, + "loss": 0.0088, + "step": 32320 + }, + { + "epoch": 2.5110679611650486, + "grad_norm": 1.45559823513031, + "learning_rate": 8.093326764410536e-05, + "loss": 0.0123, + "step": 32330 + }, + { + "epoch": 2.511844660194175, + "grad_norm": 0.266028493642807, + "learning_rate": 8.092027539092966e-05, + "loss": 0.0104, + "step": 32340 + }, + { + "epoch": 2.512621359223301, + "grad_norm": 0.28847789764404297, + "learning_rate": 8.090727975635913e-05, + "loss": 0.0089, + "step": 32350 + }, + { + "epoch": 2.5133980582524273, + "grad_norm": 0.21778492629528046, + "learning_rate": 8.089428074181497e-05, + "loss": 0.0094, + "step": 32360 + }, + { + "epoch": 2.5141747572815536, + "grad_norm": 0.45029622316360474, + "learning_rate": 8.088127834871871e-05, + "loss": 0.0093, + "step": 32370 + }, + { + "epoch": 2.51495145631068, + "grad_norm": 0.34132492542266846, + "learning_rate": 8.086827257849226e-05, + "loss": 0.0095, + "step": 32380 + }, + { + "epoch": 2.515728155339806, + "grad_norm": 0.3598629832267761, + "learning_rate": 8.085526343255795e-05, + "loss": 0.0114, + "step": 32390 + }, + { + "epoch": 2.516504854368932, + "grad_norm": 0.5130255222320557, + "learning_rate": 8.084225091233842e-05, + "loss": 0.0132, + "step": 32400 + }, + { + "epoch": 2.517281553398058, + "grad_norm": 0.24028435349464417, + "learning_rate": 8.082923501925668e-05, + "loss": 0.0066, + "step": 32410 + }, + { + "epoch": 2.5180582524271844, + "grad_norm": 0.22876372933387756, + "learning_rate": 8.081621575473617e-05, + "loss": 0.0099, + "step": 32420 + }, + { + "epoch": 2.5188349514563106, + "grad_norm": 0.24132081866264343, + "learning_rate": 8.080319312020064e-05, + "loss": 0.0091, + "step": 32430 + }, + { + "epoch": 2.519611650485437, + "grad_norm": 0.5648980736732483, + "learning_rate": 8.079016711707421e-05, + "loss": 0.0098, + "step": 32440 + }, + { + "epoch": 2.520388349514563, + "grad_norm": 0.2513822615146637, + "learning_rate": 8.077713774678139e-05, + "loss": 0.0107, + "step": 32450 + }, + { + "epoch": 2.5211650485436894, + "grad_norm": 0.2371026575565338, + "learning_rate": 8.076410501074707e-05, + "loss": 0.0102, + "step": 32460 + }, + { + "epoch": 2.5219417475728156, + "grad_norm": 0.23717695474624634, + "learning_rate": 8.075106891039647e-05, + "loss": 0.0086, + "step": 32470 + }, + { + "epoch": 2.522718446601942, + "grad_norm": 0.2736029624938965, + "learning_rate": 8.073802944715523e-05, + "loss": 0.0067, + "step": 32480 + }, + { + "epoch": 2.5234951456310677, + "grad_norm": 0.37426793575286865, + "learning_rate": 8.072498662244929e-05, + "loss": 0.0091, + "step": 32490 + }, + { + "epoch": 2.524271844660194, + "grad_norm": 0.2622626721858978, + "learning_rate": 8.0711940437705e-05, + "loss": 0.0123, + "step": 32500 + }, + { + "epoch": 2.52504854368932, + "grad_norm": 0.3397826850414276, + "learning_rate": 8.06988908943491e-05, + "loss": 0.0082, + "step": 32510 + }, + { + "epoch": 2.5258252427184464, + "grad_norm": 0.45657432079315186, + "learning_rate": 8.068583799380863e-05, + "loss": 0.01, + "step": 32520 + }, + { + "epoch": 2.5266019417475727, + "grad_norm": 0.1897309571504593, + "learning_rate": 8.067278173751104e-05, + "loss": 0.008, + "step": 32530 + }, + { + "epoch": 2.527378640776699, + "grad_norm": 0.2227896749973297, + "learning_rate": 8.065972212688417e-05, + "loss": 0.0114, + "step": 32540 + }, + { + "epoch": 2.528155339805825, + "grad_norm": 0.2379530370235443, + "learning_rate": 8.064665916335618e-05, + "loss": 0.0095, + "step": 32550 + }, + { + "epoch": 2.5289320388349514, + "grad_norm": 0.18046027421951294, + "learning_rate": 8.063359284835564e-05, + "loss": 0.0067, + "step": 32560 + }, + { + "epoch": 2.5297087378640777, + "grad_norm": 0.19705353677272797, + "learning_rate": 8.062052318331142e-05, + "loss": 0.0067, + "step": 32570 + }, + { + "epoch": 2.530485436893204, + "grad_norm": 0.21136023104190826, + "learning_rate": 8.060745016965283e-05, + "loss": 0.0125, + "step": 32580 + }, + { + "epoch": 2.53126213592233, + "grad_norm": 0.32433488965034485, + "learning_rate": 8.059437380880952e-05, + "loss": 0.01, + "step": 32590 + }, + { + "epoch": 2.5320388349514564, + "grad_norm": 0.26820990443229675, + "learning_rate": 8.058129410221146e-05, + "loss": 0.0098, + "step": 32600 + }, + { + "epoch": 2.5328155339805827, + "grad_norm": 0.2527337074279785, + "learning_rate": 8.056821105128908e-05, + "loss": 0.0081, + "step": 32610 + }, + { + "epoch": 2.533592233009709, + "grad_norm": 0.2632618248462677, + "learning_rate": 8.05551246574731e-05, + "loss": 0.0095, + "step": 32620 + }, + { + "epoch": 2.534368932038835, + "grad_norm": 0.36227554082870483, + "learning_rate": 8.05420349221946e-05, + "loss": 0.0086, + "step": 32630 + }, + { + "epoch": 2.5351456310679614, + "grad_norm": 0.2578650116920471, + "learning_rate": 8.05289418468851e-05, + "loss": 0.0091, + "step": 32640 + }, + { + "epoch": 2.5359223300970872, + "grad_norm": 0.31243473291397095, + "learning_rate": 8.051584543297642e-05, + "loss": 0.0116, + "step": 32650 + }, + { + "epoch": 2.5366990291262135, + "grad_norm": 0.14885200560092926, + "learning_rate": 8.050274568190074e-05, + "loss": 0.0099, + "step": 32660 + }, + { + "epoch": 2.5374757281553397, + "grad_norm": 0.5063350200653076, + "learning_rate": 8.048964259509067e-05, + "loss": 0.0093, + "step": 32670 + }, + { + "epoch": 2.538252427184466, + "grad_norm": 0.17810551822185516, + "learning_rate": 8.047653617397914e-05, + "loss": 0.0078, + "step": 32680 + }, + { + "epoch": 2.5390291262135922, + "grad_norm": 0.3062770962715149, + "learning_rate": 8.046342641999941e-05, + "loss": 0.012, + "step": 32690 + }, + { + "epoch": 2.5398058252427185, + "grad_norm": 0.2592880129814148, + "learning_rate": 8.045031333458517e-05, + "loss": 0.0095, + "step": 32700 + }, + { + "epoch": 2.5405825242718447, + "grad_norm": 0.30289193987846375, + "learning_rate": 8.043719691917047e-05, + "loss": 0.0088, + "step": 32710 + }, + { + "epoch": 2.541359223300971, + "grad_norm": 0.22925230860710144, + "learning_rate": 8.042407717518966e-05, + "loss": 0.01, + "step": 32720 + }, + { + "epoch": 2.542135922330097, + "grad_norm": 0.327730655670166, + "learning_rate": 8.041095410407751e-05, + "loss": 0.0074, + "step": 32730 + }, + { + "epoch": 2.542912621359223, + "grad_norm": 0.3517782986164093, + "learning_rate": 8.039782770726913e-05, + "loss": 0.01, + "step": 32740 + }, + { + "epoch": 2.5436893203883493, + "grad_norm": 0.293504923582077, + "learning_rate": 8.038469798620004e-05, + "loss": 0.0084, + "step": 32750 + }, + { + "epoch": 2.5444660194174755, + "grad_norm": 0.28352969884872437, + "learning_rate": 8.037156494230604e-05, + "loss": 0.0061, + "step": 32760 + }, + { + "epoch": 2.5452427184466018, + "grad_norm": 0.3500165045261383, + "learning_rate": 8.035842857702338e-05, + "loss": 0.0105, + "step": 32770 + }, + { + "epoch": 2.546019417475728, + "grad_norm": 0.22249512374401093, + "learning_rate": 8.03452888917886e-05, + "loss": 0.0093, + "step": 32780 + }, + { + "epoch": 2.5467961165048543, + "grad_norm": 0.2407541275024414, + "learning_rate": 8.033214588803866e-05, + "loss": 0.0094, + "step": 32790 + }, + { + "epoch": 2.5475728155339805, + "grad_norm": 0.3084067106246948, + "learning_rate": 8.031899956721083e-05, + "loss": 0.0095, + "step": 32800 + }, + { + "epoch": 2.5483495145631068, + "grad_norm": 0.4067210853099823, + "learning_rate": 8.030584993074282e-05, + "loss": 0.0097, + "step": 32810 + }, + { + "epoch": 2.549126213592233, + "grad_norm": 0.38938474655151367, + "learning_rate": 8.02926969800726e-05, + "loss": 0.0067, + "step": 32820 + }, + { + "epoch": 2.5499029126213593, + "grad_norm": 0.30144134163856506, + "learning_rate": 8.027954071663859e-05, + "loss": 0.0096, + "step": 32830 + }, + { + "epoch": 2.5506796116504855, + "grad_norm": 0.25291526317596436, + "learning_rate": 8.026638114187954e-05, + "loss": 0.0074, + "step": 32840 + }, + { + "epoch": 2.5514563106796118, + "grad_norm": 0.3356281816959381, + "learning_rate": 8.025321825723456e-05, + "loss": 0.01, + "step": 32850 + }, + { + "epoch": 2.552233009708738, + "grad_norm": 0.6567737460136414, + "learning_rate": 8.02400520641431e-05, + "loss": 0.0087, + "step": 32860 + }, + { + "epoch": 2.5530097087378643, + "grad_norm": 0.23608516156673431, + "learning_rate": 8.022688256404501e-05, + "loss": 0.0065, + "step": 32870 + }, + { + "epoch": 2.5537864077669905, + "grad_norm": 0.13868838548660278, + "learning_rate": 8.02137097583805e-05, + "loss": 0.0084, + "step": 32880 + }, + { + "epoch": 2.5545631067961168, + "grad_norm": 0.2547271251678467, + "learning_rate": 8.02005336485901e-05, + "loss": 0.0076, + "step": 32890 + }, + { + "epoch": 2.5553398058252426, + "grad_norm": 0.23031190037727356, + "learning_rate": 8.018735423611476e-05, + "loss": 0.0094, + "step": 32900 + }, + { + "epoch": 2.556116504854369, + "grad_norm": 0.11130394041538239, + "learning_rate": 8.017417152239574e-05, + "loss": 0.0074, + "step": 32910 + }, + { + "epoch": 2.556893203883495, + "grad_norm": 0.16910330951213837, + "learning_rate": 8.01609855088747e-05, + "loss": 0.0072, + "step": 32920 + }, + { + "epoch": 2.5576699029126213, + "grad_norm": 0.28392481803894043, + "learning_rate": 8.014779619699362e-05, + "loss": 0.0088, + "step": 32930 + }, + { + "epoch": 2.5584466019417476, + "grad_norm": 0.21546722948551178, + "learning_rate": 8.013460358819489e-05, + "loss": 0.0066, + "step": 32940 + }, + { + "epoch": 2.559223300970874, + "grad_norm": 0.2325804978609085, + "learning_rate": 8.01214076839212e-05, + "loss": 0.0093, + "step": 32950 + }, + { + "epoch": 2.56, + "grad_norm": 0.2703475058078766, + "learning_rate": 8.010820848561565e-05, + "loss": 0.0086, + "step": 32960 + }, + { + "epoch": 2.5607766990291263, + "grad_norm": 0.28202733397483826, + "learning_rate": 8.009500599472171e-05, + "loss": 0.0093, + "step": 32970 + }, + { + "epoch": 2.5615533980582526, + "grad_norm": 0.3525589108467102, + "learning_rate": 8.008180021268314e-05, + "loss": 0.0113, + "step": 32980 + }, + { + "epoch": 2.562330097087379, + "grad_norm": 0.20883336663246155, + "learning_rate": 8.006859114094414e-05, + "loss": 0.0066, + "step": 32990 + }, + { + "epoch": 2.5631067961165046, + "grad_norm": 0.41271957755088806, + "learning_rate": 8.005537878094921e-05, + "loss": 0.0077, + "step": 33000 + }, + { + "epoch": 2.563883495145631, + "grad_norm": 0.17208552360534668, + "learning_rate": 8.004216313414323e-05, + "loss": 0.0088, + "step": 33010 + }, + { + "epoch": 2.564660194174757, + "grad_norm": 0.31599387526512146, + "learning_rate": 8.002894420197149e-05, + "loss": 0.0065, + "step": 33020 + }, + { + "epoch": 2.5654368932038834, + "grad_norm": 0.33322587609291077, + "learning_rate": 8.001572198587954e-05, + "loss": 0.01, + "step": 33030 + }, + { + "epoch": 2.5662135922330096, + "grad_norm": 0.19848190248012543, + "learning_rate": 8.000249648731338e-05, + "loss": 0.0089, + "step": 33040 + }, + { + "epoch": 2.566990291262136, + "grad_norm": 0.18209025263786316, + "learning_rate": 7.998926770771928e-05, + "loss": 0.0088, + "step": 33050 + }, + { + "epoch": 2.567766990291262, + "grad_norm": 0.2601945102214813, + "learning_rate": 7.997603564854397e-05, + "loss": 0.0087, + "step": 33060 + }, + { + "epoch": 2.5685436893203883, + "grad_norm": 0.45831015706062317, + "learning_rate": 7.996280031123448e-05, + "loss": 0.008, + "step": 33070 + }, + { + "epoch": 2.5693203883495146, + "grad_norm": 0.24203629791736603, + "learning_rate": 7.994956169723818e-05, + "loss": 0.0093, + "step": 33080 + }, + { + "epoch": 2.570097087378641, + "grad_norm": 0.34298256039619446, + "learning_rate": 7.993631980800285e-05, + "loss": 0.0078, + "step": 33090 + }, + { + "epoch": 2.570873786407767, + "grad_norm": 0.24056237936019897, + "learning_rate": 7.992307464497659e-05, + "loss": 0.0102, + "step": 33100 + }, + { + "epoch": 2.5716504854368933, + "grad_norm": 0.21631745994091034, + "learning_rate": 7.990982620960787e-05, + "loss": 0.0077, + "step": 33110 + }, + { + "epoch": 2.5724271844660196, + "grad_norm": 0.3564737141132355, + "learning_rate": 7.989657450334554e-05, + "loss": 0.0107, + "step": 33120 + }, + { + "epoch": 2.573203883495146, + "grad_norm": 0.22749410569667816, + "learning_rate": 7.988331952763877e-05, + "loss": 0.0094, + "step": 33130 + }, + { + "epoch": 2.573980582524272, + "grad_norm": 0.22968889772891998, + "learning_rate": 7.987006128393709e-05, + "loss": 0.0088, + "step": 33140 + }, + { + "epoch": 2.574757281553398, + "grad_norm": 0.21150153875350952, + "learning_rate": 7.985679977369043e-05, + "loss": 0.0128, + "step": 33150 + }, + { + "epoch": 2.575533980582524, + "grad_norm": 0.218857541680336, + "learning_rate": 7.984353499834902e-05, + "loss": 0.0081, + "step": 33160 + }, + { + "epoch": 2.5763106796116504, + "grad_norm": 0.2093677967786789, + "learning_rate": 7.983026695936351e-05, + "loss": 0.0094, + "step": 33170 + }, + { + "epoch": 2.5770873786407766, + "grad_norm": 0.3894709646701813, + "learning_rate": 7.981699565818486e-05, + "loss": 0.008, + "step": 33180 + }, + { + "epoch": 2.577864077669903, + "grad_norm": 0.3377281129360199, + "learning_rate": 7.980372109626437e-05, + "loss": 0.0081, + "step": 33190 + }, + { + "epoch": 2.578640776699029, + "grad_norm": 0.2501084506511688, + "learning_rate": 7.979044327505375e-05, + "loss": 0.0102, + "step": 33200 + }, + { + "epoch": 2.5794174757281554, + "grad_norm": 0.5459915399551392, + "learning_rate": 7.977716219600506e-05, + "loss": 0.0097, + "step": 33210 + }, + { + "epoch": 2.5801941747572816, + "grad_norm": 0.14519917964935303, + "learning_rate": 7.97638778605707e-05, + "loss": 0.0089, + "step": 33220 + }, + { + "epoch": 2.580970873786408, + "grad_norm": 0.2410793900489807, + "learning_rate": 7.975059027020338e-05, + "loss": 0.01, + "step": 33230 + }, + { + "epoch": 2.581747572815534, + "grad_norm": 0.5008218884468079, + "learning_rate": 7.973729942635623e-05, + "loss": 0.0123, + "step": 33240 + }, + { + "epoch": 2.58252427184466, + "grad_norm": 0.18385988473892212, + "learning_rate": 7.972400533048273e-05, + "loss": 0.0104, + "step": 33250 + }, + { + "epoch": 2.583300970873786, + "grad_norm": 0.22211644053459167, + "learning_rate": 7.97107079840367e-05, + "loss": 0.0129, + "step": 33260 + }, + { + "epoch": 2.5840776699029124, + "grad_norm": 0.24059271812438965, + "learning_rate": 7.969740738847231e-05, + "loss": 0.0111, + "step": 33270 + }, + { + "epoch": 2.5848543689320387, + "grad_norm": 0.18878741562366486, + "learning_rate": 7.968410354524411e-05, + "loss": 0.0081, + "step": 33280 + }, + { + "epoch": 2.585631067961165, + "grad_norm": 0.1479063630104065, + "learning_rate": 7.967079645580697e-05, + "loss": 0.0063, + "step": 33290 + }, + { + "epoch": 2.586407766990291, + "grad_norm": 0.23345200717449188, + "learning_rate": 7.965748612161612e-05, + "loss": 0.0105, + "step": 33300 + }, + { + "epoch": 2.5871844660194174, + "grad_norm": 0.2993038594722748, + "learning_rate": 7.96441725441272e-05, + "loss": 0.0095, + "step": 33310 + }, + { + "epoch": 2.5879611650485437, + "grad_norm": 0.24695469439029694, + "learning_rate": 7.963085572479614e-05, + "loss": 0.009, + "step": 33320 + }, + { + "epoch": 2.58873786407767, + "grad_norm": 0.19599291682243347, + "learning_rate": 7.961753566507924e-05, + "loss": 0.0092, + "step": 33330 + }, + { + "epoch": 2.589514563106796, + "grad_norm": 0.1979759782552719, + "learning_rate": 7.960421236643316e-05, + "loss": 0.0086, + "step": 33340 + }, + { + "epoch": 2.5902912621359224, + "grad_norm": 0.2621246576309204, + "learning_rate": 7.959088583031496e-05, + "loss": 0.0117, + "step": 33350 + }, + { + "epoch": 2.5910679611650487, + "grad_norm": 0.3845173120498657, + "learning_rate": 7.957755605818194e-05, + "loss": 0.0095, + "step": 33360 + }, + { + "epoch": 2.591844660194175, + "grad_norm": 0.4263865053653717, + "learning_rate": 7.956422305149185e-05, + "loss": 0.0118, + "step": 33370 + }, + { + "epoch": 2.592621359223301, + "grad_norm": 0.29643514752388, + "learning_rate": 7.95508868117028e-05, + "loss": 0.012, + "step": 33380 + }, + { + "epoch": 2.5933980582524274, + "grad_norm": 0.36128780245780945, + "learning_rate": 7.953754734027318e-05, + "loss": 0.0109, + "step": 33390 + }, + { + "epoch": 2.5941747572815537, + "grad_norm": 0.19847166538238525, + "learning_rate": 7.952420463866182e-05, + "loss": 0.0104, + "step": 33400 + }, + { + "epoch": 2.5949514563106795, + "grad_norm": 0.40893346071243286, + "learning_rate": 7.951085870832782e-05, + "loss": 0.0088, + "step": 33410 + }, + { + "epoch": 2.5957281553398057, + "grad_norm": 0.1902025043964386, + "learning_rate": 7.949750955073067e-05, + "loss": 0.0063, + "step": 33420 + }, + { + "epoch": 2.596504854368932, + "grad_norm": 1.3255982398986816, + "learning_rate": 7.948415716733022e-05, + "loss": 0.0108, + "step": 33430 + }, + { + "epoch": 2.597281553398058, + "grad_norm": 0.1944120079278946, + "learning_rate": 7.947080155958669e-05, + "loss": 0.0079, + "step": 33440 + }, + { + "epoch": 2.5980582524271845, + "grad_norm": 0.1994962841272354, + "learning_rate": 7.94574427289606e-05, + "loss": 0.0092, + "step": 33450 + }, + { + "epoch": 2.5988349514563107, + "grad_norm": 0.27018776535987854, + "learning_rate": 7.944408067691284e-05, + "loss": 0.0094, + "step": 33460 + }, + { + "epoch": 2.599611650485437, + "grad_norm": 0.5076410174369812, + "learning_rate": 7.943071540490473e-05, + "loss": 0.0097, + "step": 33470 + }, + { + "epoch": 2.600388349514563, + "grad_norm": 0.37221217155456543, + "learning_rate": 7.94173469143978e-05, + "loss": 0.0105, + "step": 33480 + }, + { + "epoch": 2.6011650485436895, + "grad_norm": 0.22132326662540436, + "learning_rate": 7.940397520685406e-05, + "loss": 0.0106, + "step": 33490 + }, + { + "epoch": 2.6019417475728153, + "grad_norm": 0.35250124335289, + "learning_rate": 7.939060028373577e-05, + "loss": 0.0122, + "step": 33500 + }, + { + "epoch": 2.6027184466019415, + "grad_norm": 0.3242928385734558, + "learning_rate": 7.937722214650565e-05, + "loss": 0.0083, + "step": 33510 + }, + { + "epoch": 2.6034951456310678, + "grad_norm": 0.23831790685653687, + "learning_rate": 7.936384079662666e-05, + "loss": 0.0056, + "step": 33520 + }, + { + "epoch": 2.604271844660194, + "grad_norm": 0.27061963081359863, + "learning_rate": 7.93504562355622e-05, + "loss": 0.008, + "step": 33530 + }, + { + "epoch": 2.6050485436893203, + "grad_norm": 0.16086247563362122, + "learning_rate": 7.933706846477599e-05, + "loss": 0.0087, + "step": 33540 + }, + { + "epoch": 2.6058252427184465, + "grad_norm": 0.2202489972114563, + "learning_rate": 7.932367748573206e-05, + "loss": 0.0107, + "step": 33550 + }, + { + "epoch": 2.6066019417475728, + "grad_norm": 0.22887039184570312, + "learning_rate": 7.931028329989485e-05, + "loss": 0.0093, + "step": 33560 + }, + { + "epoch": 2.607378640776699, + "grad_norm": 0.22922217845916748, + "learning_rate": 7.929688590872913e-05, + "loss": 0.0073, + "step": 33570 + }, + { + "epoch": 2.6081553398058253, + "grad_norm": 0.19872453808784485, + "learning_rate": 7.928348531370003e-05, + "loss": 0.0069, + "step": 33580 + }, + { + "epoch": 2.6089320388349515, + "grad_norm": 0.4451902508735657, + "learning_rate": 7.927008151627297e-05, + "loss": 0.0108, + "step": 33590 + }, + { + "epoch": 2.6097087378640778, + "grad_norm": 0.4460749924182892, + "learning_rate": 7.925667451791383e-05, + "loss": 0.0083, + "step": 33600 + }, + { + "epoch": 2.610485436893204, + "grad_norm": 0.1982658952474594, + "learning_rate": 7.924326432008874e-05, + "loss": 0.0082, + "step": 33610 + }, + { + "epoch": 2.6112621359223303, + "grad_norm": 0.30412817001342773, + "learning_rate": 7.922985092426422e-05, + "loss": 0.0122, + "step": 33620 + }, + { + "epoch": 2.6120388349514565, + "grad_norm": 0.2672972083091736, + "learning_rate": 7.921643433190717e-05, + "loss": 0.0117, + "step": 33630 + }, + { + "epoch": 2.6128155339805827, + "grad_norm": 0.3669448792934418, + "learning_rate": 7.920301454448478e-05, + "loss": 0.0134, + "step": 33640 + }, + { + "epoch": 2.613592233009709, + "grad_norm": 0.26256653666496277, + "learning_rate": 7.918959156346461e-05, + "loss": 0.0151, + "step": 33650 + }, + { + "epoch": 2.614368932038835, + "grad_norm": 0.36174699664115906, + "learning_rate": 7.91761653903146e-05, + "loss": 0.01, + "step": 33660 + }, + { + "epoch": 2.615145631067961, + "grad_norm": 0.19157296419143677, + "learning_rate": 7.916273602650302e-05, + "loss": 0.0102, + "step": 33670 + }, + { + "epoch": 2.6159223300970873, + "grad_norm": 0.5143858790397644, + "learning_rate": 7.914930347349847e-05, + "loss": 0.0112, + "step": 33680 + }, + { + "epoch": 2.6166990291262135, + "grad_norm": 0.20809124410152435, + "learning_rate": 7.913586773276992e-05, + "loss": 0.0073, + "step": 33690 + }, + { + "epoch": 2.61747572815534, + "grad_norm": 0.3705143332481384, + "learning_rate": 7.912242880578667e-05, + "loss": 0.0088, + "step": 33700 + }, + { + "epoch": 2.618252427184466, + "grad_norm": 0.16912801563739777, + "learning_rate": 7.910898669401839e-05, + "loss": 0.0097, + "step": 33710 + }, + { + "epoch": 2.6190291262135923, + "grad_norm": 0.26120784878730774, + "learning_rate": 7.909554139893511e-05, + "loss": 0.009, + "step": 33720 + }, + { + "epoch": 2.6198058252427185, + "grad_norm": 0.3761865794658661, + "learning_rate": 7.908209292200715e-05, + "loss": 0.0095, + "step": 33730 + }, + { + "epoch": 2.620582524271845, + "grad_norm": 0.3075152039527893, + "learning_rate": 7.906864126470523e-05, + "loss": 0.0129, + "step": 33740 + }, + { + "epoch": 2.6213592233009706, + "grad_norm": 0.10972011834383011, + "learning_rate": 7.905518642850041e-05, + "loss": 0.0094, + "step": 33750 + }, + { + "epoch": 2.622135922330097, + "grad_norm": 0.9097667336463928, + "learning_rate": 7.904172841486409e-05, + "loss": 0.0077, + "step": 33760 + }, + { + "epoch": 2.622912621359223, + "grad_norm": 0.33143699169158936, + "learning_rate": 7.902826722526801e-05, + "loss": 0.0096, + "step": 33770 + }, + { + "epoch": 2.6236893203883493, + "grad_norm": 0.1728900671005249, + "learning_rate": 7.901480286118427e-05, + "loss": 0.0088, + "step": 33780 + }, + { + "epoch": 2.6244660194174756, + "grad_norm": 0.251126229763031, + "learning_rate": 7.900133532408531e-05, + "loss": 0.0091, + "step": 33790 + }, + { + "epoch": 2.625242718446602, + "grad_norm": 0.6286218762397766, + "learning_rate": 7.898786461544395e-05, + "loss": 0.0097, + "step": 33800 + }, + { + "epoch": 2.626019417475728, + "grad_norm": 0.2081255167722702, + "learning_rate": 7.897439073673325e-05, + "loss": 0.0069, + "step": 33810 + }, + { + "epoch": 2.6267961165048543, + "grad_norm": 0.31479185819625854, + "learning_rate": 7.896091368942677e-05, + "loss": 0.0094, + "step": 33820 + }, + { + "epoch": 2.6275728155339806, + "grad_norm": 0.5535783171653748, + "learning_rate": 7.894743347499832e-05, + "loss": 0.01, + "step": 33830 + }, + { + "epoch": 2.628349514563107, + "grad_norm": 0.3638274371623993, + "learning_rate": 7.893395009492203e-05, + "loss": 0.0083, + "step": 33840 + }, + { + "epoch": 2.629126213592233, + "grad_norm": 0.20329166948795319, + "learning_rate": 7.892046355067248e-05, + "loss": 0.0111, + "step": 33850 + }, + { + "epoch": 2.6299029126213593, + "grad_norm": 0.32154402136802673, + "learning_rate": 7.890697384372451e-05, + "loss": 0.0102, + "step": 33860 + }, + { + "epoch": 2.6306796116504856, + "grad_norm": 0.2702656686306, + "learning_rate": 7.889348097555336e-05, + "loss": 0.0087, + "step": 33870 + }, + { + "epoch": 2.631456310679612, + "grad_norm": 0.5149534940719604, + "learning_rate": 7.887998494763455e-05, + "loss": 0.0094, + "step": 33880 + }, + { + "epoch": 2.632233009708738, + "grad_norm": 0.32127904891967773, + "learning_rate": 7.886648576144404e-05, + "loss": 0.0091, + "step": 33890 + }, + { + "epoch": 2.6330097087378643, + "grad_norm": 0.35123443603515625, + "learning_rate": 7.885298341845802e-05, + "loss": 0.0076, + "step": 33900 + }, + { + "epoch": 2.63378640776699, + "grad_norm": 0.30238768458366394, + "learning_rate": 7.883947792015311e-05, + "loss": 0.0099, + "step": 33910 + }, + { + "epoch": 2.6345631067961164, + "grad_norm": 0.309237539768219, + "learning_rate": 7.882596926800628e-05, + "loss": 0.0084, + "step": 33920 + }, + { + "epoch": 2.6353398058252426, + "grad_norm": 0.22828176617622375, + "learning_rate": 7.881245746349477e-05, + "loss": 0.0073, + "step": 33930 + }, + { + "epoch": 2.636116504854369, + "grad_norm": 0.16847480833530426, + "learning_rate": 7.879894250809623e-05, + "loss": 0.0096, + "step": 33940 + }, + { + "epoch": 2.636893203883495, + "grad_norm": 0.46894389390945435, + "learning_rate": 7.878542440328865e-05, + "loss": 0.0108, + "step": 33950 + }, + { + "epoch": 2.6376699029126214, + "grad_norm": 0.2064306139945984, + "learning_rate": 7.877190315055031e-05, + "loss": 0.0109, + "step": 33960 + }, + { + "epoch": 2.6384466019417476, + "grad_norm": 0.3188708424568176, + "learning_rate": 7.875837875135991e-05, + "loss": 0.0118, + "step": 33970 + }, + { + "epoch": 2.639223300970874, + "grad_norm": 0.3641189932823181, + "learning_rate": 7.874485120719646e-05, + "loss": 0.0089, + "step": 33980 + }, + { + "epoch": 2.64, + "grad_norm": 0.20251014828681946, + "learning_rate": 7.873132051953928e-05, + "loss": 0.0083, + "step": 33990 + }, + { + "epoch": 2.6407766990291264, + "grad_norm": 0.4078265130519867, + "learning_rate": 7.87177866898681e-05, + "loss": 0.0149, + "step": 34000 + }, + { + "epoch": 2.641553398058252, + "grad_norm": 0.34470701217651367, + "learning_rate": 7.870424971966294e-05, + "loss": 0.0109, + "step": 34010 + }, + { + "epoch": 2.6423300970873784, + "grad_norm": 0.2925356924533844, + "learning_rate": 7.869070961040419e-05, + "loss": 0.0093, + "step": 34020 + }, + { + "epoch": 2.6431067961165047, + "grad_norm": 0.16082829236984253, + "learning_rate": 7.867716636357257e-05, + "loss": 0.0108, + "step": 34030 + }, + { + "epoch": 2.643883495145631, + "grad_norm": 0.3048584759235382, + "learning_rate": 7.866361998064915e-05, + "loss": 0.0105, + "step": 34040 + }, + { + "epoch": 2.644660194174757, + "grad_norm": 0.26284152269363403, + "learning_rate": 7.865007046311534e-05, + "loss": 0.0057, + "step": 34050 + }, + { + "epoch": 2.6454368932038834, + "grad_norm": 0.31908825039863586, + "learning_rate": 7.86365178124529e-05, + "loss": 0.0074, + "step": 34060 + }, + { + "epoch": 2.6462135922330097, + "grad_norm": 0.33053386211395264, + "learning_rate": 7.862296203014394e-05, + "loss": 0.0098, + "step": 34070 + }, + { + "epoch": 2.646990291262136, + "grad_norm": 0.42964592576026917, + "learning_rate": 7.860940311767088e-05, + "loss": 0.0077, + "step": 34080 + }, + { + "epoch": 2.647766990291262, + "grad_norm": 0.25783011317253113, + "learning_rate": 7.85958410765165e-05, + "loss": 0.0089, + "step": 34090 + }, + { + "epoch": 2.6485436893203884, + "grad_norm": 0.2339617908000946, + "learning_rate": 7.858227590816394e-05, + "loss": 0.0086, + "step": 34100 + }, + { + "epoch": 2.6493203883495147, + "grad_norm": 0.3143027424812317, + "learning_rate": 7.856870761409664e-05, + "loss": 0.0095, + "step": 34110 + }, + { + "epoch": 2.650097087378641, + "grad_norm": 0.24702656269073486, + "learning_rate": 7.855513619579846e-05, + "loss": 0.0078, + "step": 34120 + }, + { + "epoch": 2.650873786407767, + "grad_norm": 0.1865878701210022, + "learning_rate": 7.85415616547535e-05, + "loss": 0.0077, + "step": 34130 + }, + { + "epoch": 2.6516504854368934, + "grad_norm": 0.24841415882110596, + "learning_rate": 7.852798399244627e-05, + "loss": 0.009, + "step": 34140 + }, + { + "epoch": 2.6524271844660197, + "grad_norm": 0.2845427989959717, + "learning_rate": 7.851440321036161e-05, + "loss": 0.0108, + "step": 34150 + }, + { + "epoch": 2.6532038834951455, + "grad_norm": 0.21230922639369965, + "learning_rate": 7.850081930998468e-05, + "loss": 0.008, + "step": 34160 + }, + { + "epoch": 2.6539805825242717, + "grad_norm": 0.1835821568965912, + "learning_rate": 7.8487232292801e-05, + "loss": 0.0086, + "step": 34170 + }, + { + "epoch": 2.654757281553398, + "grad_norm": 0.24129123985767365, + "learning_rate": 7.847364216029642e-05, + "loss": 0.0111, + "step": 34180 + }, + { + "epoch": 2.655533980582524, + "grad_norm": 0.29433658719062805, + "learning_rate": 7.846004891395716e-05, + "loss": 0.0093, + "step": 34190 + }, + { + "epoch": 2.6563106796116505, + "grad_norm": 0.25312724709510803, + "learning_rate": 7.844645255526972e-05, + "loss": 0.009, + "step": 34200 + }, + { + "epoch": 2.6570873786407767, + "grad_norm": 0.2241377830505371, + "learning_rate": 7.843285308572101e-05, + "loss": 0.0103, + "step": 34210 + }, + { + "epoch": 2.657864077669903, + "grad_norm": 0.39336878061294556, + "learning_rate": 7.841925050679823e-05, + "loss": 0.0091, + "step": 34220 + }, + { + "epoch": 2.658640776699029, + "grad_norm": 0.2363830953836441, + "learning_rate": 7.840564481998895e-05, + "loss": 0.0067, + "step": 34230 + }, + { + "epoch": 2.6594174757281555, + "grad_norm": 0.3389822542667389, + "learning_rate": 7.839203602678105e-05, + "loss": 0.009, + "step": 34240 + }, + { + "epoch": 2.6601941747572817, + "grad_norm": 0.18180891871452332, + "learning_rate": 7.837842412866279e-05, + "loss": 0.0073, + "step": 34250 + }, + { + "epoch": 2.6609708737864075, + "grad_norm": 0.1728755235671997, + "learning_rate": 7.836480912712272e-05, + "loss": 0.0074, + "step": 34260 + }, + { + "epoch": 2.6617475728155338, + "grad_norm": 0.2258961796760559, + "learning_rate": 7.835119102364976e-05, + "loss": 0.0087, + "step": 34270 + }, + { + "epoch": 2.66252427184466, + "grad_norm": 0.24575649201869965, + "learning_rate": 7.83375698197332e-05, + "loss": 0.0079, + "step": 34280 + }, + { + "epoch": 2.6633009708737863, + "grad_norm": 0.405940979719162, + "learning_rate": 7.83239455168626e-05, + "loss": 0.0102, + "step": 34290 + }, + { + "epoch": 2.6640776699029125, + "grad_norm": 0.20656795799732208, + "learning_rate": 7.83103181165279e-05, + "loss": 0.0069, + "step": 34300 + }, + { + "epoch": 2.6648543689320388, + "grad_norm": 0.22317340970039368, + "learning_rate": 7.829668762021937e-05, + "loss": 0.0115, + "step": 34310 + }, + { + "epoch": 2.665631067961165, + "grad_norm": 0.32094013690948486, + "learning_rate": 7.828305402942764e-05, + "loss": 0.0096, + "step": 34320 + }, + { + "epoch": 2.6664077669902913, + "grad_norm": 0.2518889009952545, + "learning_rate": 7.826941734564363e-05, + "loss": 0.0096, + "step": 34330 + }, + { + "epoch": 2.6671844660194175, + "grad_norm": 0.17968307435512543, + "learning_rate": 7.825577757035865e-05, + "loss": 0.008, + "step": 34340 + }, + { + "epoch": 2.6679611650485437, + "grad_norm": 0.21422652900218964, + "learning_rate": 7.824213470506431e-05, + "loss": 0.008, + "step": 34350 + }, + { + "epoch": 2.66873786407767, + "grad_norm": 0.24274249374866486, + "learning_rate": 7.822848875125257e-05, + "loss": 0.0083, + "step": 34360 + }, + { + "epoch": 2.6695145631067962, + "grad_norm": 0.20929571986198425, + "learning_rate": 7.821483971041576e-05, + "loss": 0.0096, + "step": 34370 + }, + { + "epoch": 2.6702912621359225, + "grad_norm": 0.260433554649353, + "learning_rate": 7.820118758404649e-05, + "loss": 0.0107, + "step": 34380 + }, + { + "epoch": 2.6710679611650487, + "grad_norm": 0.27437928318977356, + "learning_rate": 7.818753237363776e-05, + "loss": 0.0081, + "step": 34390 + }, + { + "epoch": 2.671844660194175, + "grad_norm": 0.24625422060489655, + "learning_rate": 7.817387408068286e-05, + "loss": 0.0083, + "step": 34400 + }, + { + "epoch": 2.6726213592233012, + "grad_norm": 0.2834695279598236, + "learning_rate": 7.816021270667544e-05, + "loss": 0.0108, + "step": 34410 + }, + { + "epoch": 2.673398058252427, + "grad_norm": 0.30415046215057373, + "learning_rate": 7.81465482531095e-05, + "loss": 0.009, + "step": 34420 + }, + { + "epoch": 2.6741747572815533, + "grad_norm": 0.22857354581356049, + "learning_rate": 7.813288072147938e-05, + "loss": 0.0099, + "step": 34430 + }, + { + "epoch": 2.6749514563106795, + "grad_norm": 0.1727994978427887, + "learning_rate": 7.811921011327972e-05, + "loss": 0.0077, + "step": 34440 + }, + { + "epoch": 2.675728155339806, + "grad_norm": 0.3364970088005066, + "learning_rate": 7.810553643000549e-05, + "loss": 0.0096, + "step": 34450 + }, + { + "epoch": 2.676504854368932, + "grad_norm": 0.38920772075653076, + "learning_rate": 7.809185967315206e-05, + "loss": 0.0087, + "step": 34460 + }, + { + "epoch": 2.6772815533980583, + "grad_norm": 0.4926975667476654, + "learning_rate": 7.80781798442151e-05, + "loss": 0.0078, + "step": 34470 + }, + { + "epoch": 2.6780582524271845, + "grad_norm": 0.22291821241378784, + "learning_rate": 7.806449694469061e-05, + "loss": 0.0074, + "step": 34480 + }, + { + "epoch": 2.678834951456311, + "grad_norm": 0.37095674872398376, + "learning_rate": 7.805081097607492e-05, + "loss": 0.0105, + "step": 34490 + }, + { + "epoch": 2.679611650485437, + "grad_norm": 0.1804320514202118, + "learning_rate": 7.803712193986474e-05, + "loss": 0.0064, + "step": 34500 + }, + { + "epoch": 2.680388349514563, + "grad_norm": 0.1993637979030609, + "learning_rate": 7.802342983755702e-05, + "loss": 0.0091, + "step": 34510 + }, + { + "epoch": 2.681165048543689, + "grad_norm": 0.4107533097267151, + "learning_rate": 7.800973467064918e-05, + "loss": 0.0086, + "step": 34520 + }, + { + "epoch": 2.6819417475728153, + "grad_norm": 0.15461032092571259, + "learning_rate": 7.799603644063884e-05, + "loss": 0.0081, + "step": 34530 + }, + { + "epoch": 2.6827184466019416, + "grad_norm": 0.26436173915863037, + "learning_rate": 7.798233514902405e-05, + "loss": 0.0075, + "step": 34540 + }, + { + "epoch": 2.683495145631068, + "grad_norm": 0.16450650990009308, + "learning_rate": 7.796863079730318e-05, + "loss": 0.0078, + "step": 34550 + }, + { + "epoch": 2.684271844660194, + "grad_norm": 0.18931496143341064, + "learning_rate": 7.795492338697488e-05, + "loss": 0.0083, + "step": 34560 + }, + { + "epoch": 2.6850485436893203, + "grad_norm": 0.2026437669992447, + "learning_rate": 7.794121291953819e-05, + "loss": 0.0093, + "step": 34570 + }, + { + "epoch": 2.6858252427184466, + "grad_norm": 0.295364111661911, + "learning_rate": 7.792749939649246e-05, + "loss": 0.0095, + "step": 34580 + }, + { + "epoch": 2.686601941747573, + "grad_norm": 0.3293825685977936, + "learning_rate": 7.79137828193374e-05, + "loss": 0.0084, + "step": 34590 + }, + { + "epoch": 2.687378640776699, + "grad_norm": 0.11800755560398102, + "learning_rate": 7.790006318957301e-05, + "loss": 0.0099, + "step": 34600 + }, + { + "epoch": 2.6881553398058253, + "grad_norm": 0.18871060013771057, + "learning_rate": 7.788634050869965e-05, + "loss": 0.0078, + "step": 34610 + }, + { + "epoch": 2.6889320388349516, + "grad_norm": 0.178923562169075, + "learning_rate": 7.787261477821803e-05, + "loss": 0.0108, + "step": 34620 + }, + { + "epoch": 2.689708737864078, + "grad_norm": 0.36912813782691956, + "learning_rate": 7.785888599962916e-05, + "loss": 0.0105, + "step": 34630 + }, + { + "epoch": 2.690485436893204, + "grad_norm": 0.26782986521720886, + "learning_rate": 7.784515417443439e-05, + "loss": 0.0111, + "step": 34640 + }, + { + "epoch": 2.6912621359223303, + "grad_norm": 0.3033566474914551, + "learning_rate": 7.783141930413545e-05, + "loss": 0.0112, + "step": 34650 + }, + { + "epoch": 2.6920388349514566, + "grad_norm": 0.20033203065395355, + "learning_rate": 7.78176813902343e-05, + "loss": 0.0098, + "step": 34660 + }, + { + "epoch": 2.6928155339805824, + "grad_norm": 0.28593096137046814, + "learning_rate": 7.780394043423336e-05, + "loss": 0.0087, + "step": 34670 + }, + { + "epoch": 2.6935922330097086, + "grad_norm": 0.21803420782089233, + "learning_rate": 7.77901964376353e-05, + "loss": 0.0075, + "step": 34680 + }, + { + "epoch": 2.694368932038835, + "grad_norm": 0.2787322998046875, + "learning_rate": 7.777644940194316e-05, + "loss": 0.0074, + "step": 34690 + }, + { + "epoch": 2.695145631067961, + "grad_norm": 0.36242738366127014, + "learning_rate": 7.776269932866023e-05, + "loss": 0.0077, + "step": 34700 + }, + { + "epoch": 2.6959223300970874, + "grad_norm": 0.19080805778503418, + "learning_rate": 7.774894621929026e-05, + "loss": 0.0071, + "step": 34710 + }, + { + "epoch": 2.6966990291262136, + "grad_norm": 0.6590007543563843, + "learning_rate": 7.773519007533725e-05, + "loss": 0.01, + "step": 34720 + }, + { + "epoch": 2.69747572815534, + "grad_norm": 0.25538456439971924, + "learning_rate": 7.772143089830556e-05, + "loss": 0.0085, + "step": 34730 + }, + { + "epoch": 2.698252427184466, + "grad_norm": 0.6373974084854126, + "learning_rate": 7.770766868969985e-05, + "loss": 0.0076, + "step": 34740 + }, + { + "epoch": 2.6990291262135924, + "grad_norm": 0.1425027698278427, + "learning_rate": 7.769390345102518e-05, + "loss": 0.0072, + "step": 34750 + }, + { + "epoch": 2.699805825242718, + "grad_norm": 0.4120965301990509, + "learning_rate": 7.768013518378683e-05, + "loss": 0.0092, + "step": 34760 + }, + { + "epoch": 2.7005825242718444, + "grad_norm": 0.29125136137008667, + "learning_rate": 7.766636388949053e-05, + "loss": 0.0078, + "step": 34770 + }, + { + "epoch": 2.7013592233009707, + "grad_norm": 0.3036712408065796, + "learning_rate": 7.765258956964229e-05, + "loss": 0.0105, + "step": 34780 + }, + { + "epoch": 2.702135922330097, + "grad_norm": 0.3538512885570526, + "learning_rate": 7.76388122257484e-05, + "loss": 0.0092, + "step": 34790 + }, + { + "epoch": 2.702912621359223, + "grad_norm": 0.28584274649620056, + "learning_rate": 7.762503185931558e-05, + "loss": 0.0116, + "step": 34800 + }, + { + "epoch": 2.7036893203883494, + "grad_norm": 0.1690511256456375, + "learning_rate": 7.76112484718508e-05, + "loss": 0.0092, + "step": 34810 + }, + { + "epoch": 2.7044660194174757, + "grad_norm": 0.28643786907196045, + "learning_rate": 7.75974620648614e-05, + "loss": 0.0094, + "step": 34820 + }, + { + "epoch": 2.705242718446602, + "grad_norm": 0.28367936611175537, + "learning_rate": 7.758367263985503e-05, + "loss": 0.0097, + "step": 34830 + }, + { + "epoch": 2.706019417475728, + "grad_norm": 0.2523626685142517, + "learning_rate": 7.75698801983397e-05, + "loss": 0.0083, + "step": 34840 + }, + { + "epoch": 2.7067961165048544, + "grad_norm": 0.2483253926038742, + "learning_rate": 7.755608474182372e-05, + "loss": 0.0088, + "step": 34850 + }, + { + "epoch": 2.7075728155339807, + "grad_norm": 0.2744215726852417, + "learning_rate": 7.754228627181574e-05, + "loss": 0.009, + "step": 34860 + }, + { + "epoch": 2.708349514563107, + "grad_norm": 0.4512402415275574, + "learning_rate": 7.752848478982476e-05, + "loss": 0.0078, + "step": 34870 + }, + { + "epoch": 2.709126213592233, + "grad_norm": 0.18823838233947754, + "learning_rate": 7.751468029736006e-05, + "loss": 0.0086, + "step": 34880 + }, + { + "epoch": 2.7099029126213594, + "grad_norm": 0.21626649796962738, + "learning_rate": 7.750087279593129e-05, + "loss": 0.0083, + "step": 34890 + }, + { + "epoch": 2.7106796116504857, + "grad_norm": 0.2561870515346527, + "learning_rate": 7.748706228704843e-05, + "loss": 0.0061, + "step": 34900 + }, + { + "epoch": 2.711456310679612, + "grad_norm": 0.2880573570728302, + "learning_rate": 7.747324877222176e-05, + "loss": 0.0084, + "step": 34910 + }, + { + "epoch": 2.7122330097087377, + "grad_norm": 0.46330124139785767, + "learning_rate": 7.745943225296188e-05, + "loss": 0.0119, + "step": 34920 + }, + { + "epoch": 2.713009708737864, + "grad_norm": 0.47645607590675354, + "learning_rate": 7.744561273077981e-05, + "loss": 0.0125, + "step": 34930 + }, + { + "epoch": 2.71378640776699, + "grad_norm": 0.3670835494995117, + "learning_rate": 7.743179020718678e-05, + "loss": 0.0097, + "step": 34940 + }, + { + "epoch": 2.7145631067961165, + "grad_norm": 0.401012122631073, + "learning_rate": 7.741796468369443e-05, + "loss": 0.0096, + "step": 34950 + }, + { + "epoch": 2.7153398058252427, + "grad_norm": 0.330078125, + "learning_rate": 7.740413616181466e-05, + "loss": 0.0093, + "step": 34960 + }, + { + "epoch": 2.716116504854369, + "grad_norm": 0.5054227113723755, + "learning_rate": 7.739030464305978e-05, + "loss": 0.0079, + "step": 34970 + }, + { + "epoch": 2.716893203883495, + "grad_norm": 0.5723316073417664, + "learning_rate": 7.737647012894235e-05, + "loss": 0.0121, + "step": 34980 + }, + { + "epoch": 2.7176699029126214, + "grad_norm": 0.18369348347187042, + "learning_rate": 7.736263262097532e-05, + "loss": 0.007, + "step": 34990 + }, + { + "epoch": 2.7184466019417477, + "grad_norm": 0.2994880676269531, + "learning_rate": 7.734879212067192e-05, + "loss": 0.0102, + "step": 35000 + }, + { + "epoch": 2.719223300970874, + "grad_norm": 0.4226432144641876, + "learning_rate": 7.733494862954573e-05, + "loss": 0.0109, + "step": 35010 + }, + { + "epoch": 2.7199999999999998, + "grad_norm": 0.278887003660202, + "learning_rate": 7.732110214911066e-05, + "loss": 0.0071, + "step": 35020 + }, + { + "epoch": 2.720776699029126, + "grad_norm": 0.21477265655994415, + "learning_rate": 7.730725268088094e-05, + "loss": 0.0088, + "step": 35030 + }, + { + "epoch": 2.7215533980582522, + "grad_norm": 0.2649634778499603, + "learning_rate": 7.729340022637112e-05, + "loss": 0.0095, + "step": 35040 + }, + { + "epoch": 2.7223300970873785, + "grad_norm": 0.14170978963375092, + "learning_rate": 7.727954478709607e-05, + "loss": 0.0108, + "step": 35050 + }, + { + "epoch": 2.7231067961165047, + "grad_norm": 0.19799751043319702, + "learning_rate": 7.726568636457103e-05, + "loss": 0.0092, + "step": 35060 + }, + { + "epoch": 2.723883495145631, + "grad_norm": 0.21315066516399384, + "learning_rate": 7.725182496031153e-05, + "loss": 0.0078, + "step": 35070 + }, + { + "epoch": 2.7246601941747572, + "grad_norm": 0.2381603717803955, + "learning_rate": 7.723796057583341e-05, + "loss": 0.0071, + "step": 35080 + }, + { + "epoch": 2.7254368932038835, + "grad_norm": 0.2999923825263977, + "learning_rate": 7.722409321265287e-05, + "loss": 0.0078, + "step": 35090 + }, + { + "epoch": 2.7262135922330097, + "grad_norm": 0.37068620324134827, + "learning_rate": 7.721022287228645e-05, + "loss": 0.0109, + "step": 35100 + }, + { + "epoch": 2.726990291262136, + "grad_norm": 0.11851242929697037, + "learning_rate": 7.719634955625094e-05, + "loss": 0.0092, + "step": 35110 + }, + { + "epoch": 2.7277669902912622, + "grad_norm": 0.1883537471294403, + "learning_rate": 7.718247326606355e-05, + "loss": 0.0083, + "step": 35120 + }, + { + "epoch": 2.7285436893203885, + "grad_norm": 0.24460063874721527, + "learning_rate": 7.716859400324175e-05, + "loss": 0.0142, + "step": 35130 + }, + { + "epoch": 2.7293203883495147, + "grad_norm": 0.4474724531173706, + "learning_rate": 7.715471176930335e-05, + "loss": 0.0085, + "step": 35140 + }, + { + "epoch": 2.730097087378641, + "grad_norm": 0.18488474190235138, + "learning_rate": 7.714082656576651e-05, + "loss": 0.0093, + "step": 35150 + }, + { + "epoch": 2.7308737864077672, + "grad_norm": 0.6164620518684387, + "learning_rate": 7.712693839414968e-05, + "loss": 0.0079, + "step": 35160 + }, + { + "epoch": 2.731650485436893, + "grad_norm": 0.3449232280254364, + "learning_rate": 7.711304725597164e-05, + "loss": 0.0099, + "step": 35170 + }, + { + "epoch": 2.7324271844660193, + "grad_norm": 0.347209095954895, + "learning_rate": 7.709915315275151e-05, + "loss": 0.0096, + "step": 35180 + }, + { + "epoch": 2.7332038834951455, + "grad_norm": 0.3897068202495575, + "learning_rate": 7.708525608600876e-05, + "loss": 0.0095, + "step": 35190 + }, + { + "epoch": 2.733980582524272, + "grad_norm": 0.2644908130168915, + "learning_rate": 7.707135605726311e-05, + "loss": 0.0099, + "step": 35200 + }, + { + "epoch": 2.734757281553398, + "grad_norm": 0.20075683295726776, + "learning_rate": 7.705745306803466e-05, + "loss": 0.0089, + "step": 35210 + }, + { + "epoch": 2.7355339805825243, + "grad_norm": 0.2137058526277542, + "learning_rate": 7.704354711984383e-05, + "loss": 0.0106, + "step": 35220 + }, + { + "epoch": 2.7363106796116505, + "grad_norm": 0.2822689712047577, + "learning_rate": 7.702963821421133e-05, + "loss": 0.0077, + "step": 35230 + }, + { + "epoch": 2.737087378640777, + "grad_norm": 0.27235203981399536, + "learning_rate": 7.701572635265826e-05, + "loss": 0.0084, + "step": 35240 + }, + { + "epoch": 2.737864077669903, + "grad_norm": 0.31467440724372864, + "learning_rate": 7.700181153670596e-05, + "loss": 0.0091, + "step": 35250 + }, + { + "epoch": 2.7386407766990293, + "grad_norm": 0.5211147665977478, + "learning_rate": 7.698789376787616e-05, + "loss": 0.0118, + "step": 35260 + }, + { + "epoch": 2.739417475728155, + "grad_norm": 0.36517614126205444, + "learning_rate": 7.697397304769085e-05, + "loss": 0.01, + "step": 35270 + }, + { + "epoch": 2.7401941747572813, + "grad_norm": 0.3686685264110565, + "learning_rate": 7.696004937767241e-05, + "loss": 0.0106, + "step": 35280 + }, + { + "epoch": 2.7409708737864076, + "grad_norm": 0.29751119017601013, + "learning_rate": 7.694612275934352e-05, + "loss": 0.0084, + "step": 35290 + }, + { + "epoch": 2.741747572815534, + "grad_norm": 0.23448322713375092, + "learning_rate": 7.693219319422714e-05, + "loss": 0.0097, + "step": 35300 + }, + { + "epoch": 2.74252427184466, + "grad_norm": 0.17281928658485413, + "learning_rate": 7.69182606838466e-05, + "loss": 0.0088, + "step": 35310 + }, + { + "epoch": 2.7433009708737863, + "grad_norm": 0.3875148892402649, + "learning_rate": 7.690432522972558e-05, + "loss": 0.0095, + "step": 35320 + }, + { + "epoch": 2.7440776699029126, + "grad_norm": 0.44230714440345764, + "learning_rate": 7.689038683338796e-05, + "loss": 0.0069, + "step": 35330 + }, + { + "epoch": 2.744854368932039, + "grad_norm": 0.2551564574241638, + "learning_rate": 7.687644549635808e-05, + "loss": 0.0088, + "step": 35340 + }, + { + "epoch": 2.745631067961165, + "grad_norm": 0.24827489256858826, + "learning_rate": 7.686250122016053e-05, + "loss": 0.006, + "step": 35350 + }, + { + "epoch": 2.7464077669902913, + "grad_norm": 0.4761821925640106, + "learning_rate": 7.684855400632022e-05, + "loss": 0.0064, + "step": 35360 + }, + { + "epoch": 2.7471844660194176, + "grad_norm": 0.3755880296230316, + "learning_rate": 7.683460385636243e-05, + "loss": 0.0122, + "step": 35370 + }, + { + "epoch": 2.747961165048544, + "grad_norm": 0.1978040486574173, + "learning_rate": 7.68206507718127e-05, + "loss": 0.0095, + "step": 35380 + }, + { + "epoch": 2.74873786407767, + "grad_norm": 0.1820240467786789, + "learning_rate": 7.680669475419693e-05, + "loss": 0.0095, + "step": 35390 + }, + { + "epoch": 2.7495145631067963, + "grad_norm": 0.22580300271511078, + "learning_rate": 7.679273580504132e-05, + "loss": 0.0065, + "step": 35400 + }, + { + "epoch": 2.7502912621359226, + "grad_norm": 0.19293361902236938, + "learning_rate": 7.67787739258724e-05, + "loss": 0.009, + "step": 35410 + }, + { + "epoch": 2.751067961165049, + "grad_norm": 0.2774483561515808, + "learning_rate": 7.676480911821705e-05, + "loss": 0.0085, + "step": 35420 + }, + { + "epoch": 2.7518446601941746, + "grad_norm": 0.24459494650363922, + "learning_rate": 7.675084138360238e-05, + "loss": 0.0088, + "step": 35430 + }, + { + "epoch": 2.752621359223301, + "grad_norm": 0.220210000872612, + "learning_rate": 7.673687072355592e-05, + "loss": 0.0084, + "step": 35440 + }, + { + "epoch": 2.753398058252427, + "grad_norm": 0.2938345670700073, + "learning_rate": 7.67228971396055e-05, + "loss": 0.0071, + "step": 35450 + }, + { + "epoch": 2.7541747572815534, + "grad_norm": 0.29633447527885437, + "learning_rate": 7.670892063327922e-05, + "loss": 0.0061, + "step": 35460 + }, + { + "epoch": 2.7549514563106796, + "grad_norm": 0.42961063981056213, + "learning_rate": 7.669494120610552e-05, + "loss": 0.0074, + "step": 35470 + }, + { + "epoch": 2.755728155339806, + "grad_norm": 0.393551230430603, + "learning_rate": 7.66809588596132e-05, + "loss": 0.0066, + "step": 35480 + }, + { + "epoch": 2.756504854368932, + "grad_norm": 0.24117547273635864, + "learning_rate": 7.666697359533132e-05, + "loss": 0.007, + "step": 35490 + }, + { + "epoch": 2.7572815533980584, + "grad_norm": 0.17566588521003723, + "learning_rate": 7.665298541478932e-05, + "loss": 0.0102, + "step": 35500 + }, + { + "epoch": 2.7580582524271846, + "grad_norm": 0.3758625090122223, + "learning_rate": 7.66389943195169e-05, + "loss": 0.0098, + "step": 35510 + }, + { + "epoch": 2.7588349514563104, + "grad_norm": 0.25492724776268005, + "learning_rate": 7.66250003110441e-05, + "loss": 0.0105, + "step": 35520 + }, + { + "epoch": 2.7596116504854367, + "grad_norm": 0.31475475430488586, + "learning_rate": 7.66110033909013e-05, + "loss": 0.01, + "step": 35530 + }, + { + "epoch": 2.760388349514563, + "grad_norm": 0.2582232654094696, + "learning_rate": 7.659700356061918e-05, + "loss": 0.0084, + "step": 35540 + }, + { + "epoch": 2.761165048543689, + "grad_norm": 0.4568030536174774, + "learning_rate": 7.658300082172875e-05, + "loss": 0.0072, + "step": 35550 + }, + { + "epoch": 2.7619417475728154, + "grad_norm": 0.19697722792625427, + "learning_rate": 7.65689951757613e-05, + "loss": 0.0087, + "step": 35560 + }, + { + "epoch": 2.7627184466019417, + "grad_norm": 0.13693369925022125, + "learning_rate": 7.655498662424849e-05, + "loss": 0.0065, + "step": 35570 + }, + { + "epoch": 2.763495145631068, + "grad_norm": 0.2334088236093521, + "learning_rate": 7.654097516872227e-05, + "loss": 0.0075, + "step": 35580 + }, + { + "epoch": 2.764271844660194, + "grad_norm": 0.24066823720932007, + "learning_rate": 7.65269608107149e-05, + "loss": 0.0099, + "step": 35590 + }, + { + "epoch": 2.7650485436893204, + "grad_norm": 0.2496543824672699, + "learning_rate": 7.6512943551759e-05, + "loss": 0.0118, + "step": 35600 + }, + { + "epoch": 2.7658252427184467, + "grad_norm": 0.49443385004997253, + "learning_rate": 7.649892339338743e-05, + "loss": 0.0098, + "step": 35610 + }, + { + "epoch": 2.766601941747573, + "grad_norm": 0.20061221718788147, + "learning_rate": 7.648490033713344e-05, + "loss": 0.01, + "step": 35620 + }, + { + "epoch": 2.767378640776699, + "grad_norm": 0.27307504415512085, + "learning_rate": 7.647087438453058e-05, + "loss": 0.0087, + "step": 35630 + }, + { + "epoch": 2.7681553398058254, + "grad_norm": 0.2643120586872101, + "learning_rate": 7.64568455371127e-05, + "loss": 0.0113, + "step": 35640 + }, + { + "epoch": 2.7689320388349516, + "grad_norm": 0.5940812230110168, + "learning_rate": 7.644281379641396e-05, + "loss": 0.0102, + "step": 35650 + }, + { + "epoch": 2.769708737864078, + "grad_norm": 0.24393735826015472, + "learning_rate": 7.642877916396887e-05, + "loss": 0.0077, + "step": 35660 + }, + { + "epoch": 2.770485436893204, + "grad_norm": 0.18203891813755035, + "learning_rate": 7.641474164131221e-05, + "loss": 0.0084, + "step": 35670 + }, + { + "epoch": 2.77126213592233, + "grad_norm": 0.1638091504573822, + "learning_rate": 7.640070122997913e-05, + "loss": 0.0079, + "step": 35680 + }, + { + "epoch": 2.772038834951456, + "grad_norm": 0.44658228754997253, + "learning_rate": 7.638665793150505e-05, + "loss": 0.0102, + "step": 35690 + }, + { + "epoch": 2.7728155339805824, + "grad_norm": 0.27928030490875244, + "learning_rate": 7.637261174742574e-05, + "loss": 0.0091, + "step": 35700 + }, + { + "epoch": 2.7735922330097087, + "grad_norm": 0.46688106656074524, + "learning_rate": 7.635856267927725e-05, + "loss": 0.0074, + "step": 35710 + }, + { + "epoch": 2.774368932038835, + "grad_norm": 0.6723981499671936, + "learning_rate": 7.634451072859597e-05, + "loss": 0.0093, + "step": 35720 + }, + { + "epoch": 2.775145631067961, + "grad_norm": 0.3197277784347534, + "learning_rate": 7.633045589691863e-05, + "loss": 0.0087, + "step": 35730 + }, + { + "epoch": 2.7759223300970874, + "grad_norm": 0.2033558487892151, + "learning_rate": 7.63163981857822e-05, + "loss": 0.0068, + "step": 35740 + }, + { + "epoch": 2.7766990291262137, + "grad_norm": 0.16293147206306458, + "learning_rate": 7.630233759672403e-05, + "loss": 0.008, + "step": 35750 + }, + { + "epoch": 2.77747572815534, + "grad_norm": 0.3294324278831482, + "learning_rate": 7.628827413128177e-05, + "loss": 0.0093, + "step": 35760 + }, + { + "epoch": 2.7782524271844657, + "grad_norm": 0.37669336795806885, + "learning_rate": 7.627420779099338e-05, + "loss": 0.0082, + "step": 35770 + }, + { + "epoch": 2.779029126213592, + "grad_norm": 0.1947929710149765, + "learning_rate": 7.626013857739711e-05, + "loss": 0.0086, + "step": 35780 + }, + { + "epoch": 2.7798058252427182, + "grad_norm": 0.4392220079898834, + "learning_rate": 7.624606649203158e-05, + "loss": 0.0099, + "step": 35790 + }, + { + "epoch": 2.7805825242718445, + "grad_norm": 0.26801812648773193, + "learning_rate": 7.623199153643569e-05, + "loss": 0.0096, + "step": 35800 + }, + { + "epoch": 2.7813592233009707, + "grad_norm": 0.32677480578422546, + "learning_rate": 7.621791371214863e-05, + "loss": 0.0106, + "step": 35810 + }, + { + "epoch": 2.782135922330097, + "grad_norm": 0.19355209171772003, + "learning_rate": 7.620383302070995e-05, + "loss": 0.0085, + "step": 35820 + }, + { + "epoch": 2.7829126213592232, + "grad_norm": 0.32921820878982544, + "learning_rate": 7.61897494636595e-05, + "loss": 0.0098, + "step": 35830 + }, + { + "epoch": 2.7836893203883495, + "grad_norm": 0.31883710622787476, + "learning_rate": 7.617566304253739e-05, + "loss": 0.0116, + "step": 35840 + }, + { + "epoch": 2.7844660194174757, + "grad_norm": 0.13857287168502808, + "learning_rate": 7.616157375888416e-05, + "loss": 0.0067, + "step": 35850 + }, + { + "epoch": 2.785242718446602, + "grad_norm": 0.2595357894897461, + "learning_rate": 7.614748161424053e-05, + "loss": 0.0102, + "step": 35860 + }, + { + "epoch": 2.7860194174757282, + "grad_norm": 0.3528435528278351, + "learning_rate": 7.613338661014763e-05, + "loss": 0.0074, + "step": 35870 + }, + { + "epoch": 2.7867961165048545, + "grad_norm": 0.29862654209136963, + "learning_rate": 7.611928874814686e-05, + "loss": 0.0118, + "step": 35880 + }, + { + "epoch": 2.7875728155339807, + "grad_norm": 0.16619178652763367, + "learning_rate": 7.610518802977996e-05, + "loss": 0.0073, + "step": 35890 + }, + { + "epoch": 2.788349514563107, + "grad_norm": 0.3844333291053772, + "learning_rate": 7.609108445658893e-05, + "loss": 0.0086, + "step": 35900 + }, + { + "epoch": 2.7891262135922332, + "grad_norm": 0.167180597782135, + "learning_rate": 7.607697803011612e-05, + "loss": 0.0118, + "step": 35910 + }, + { + "epoch": 2.7899029126213595, + "grad_norm": 0.30743634700775146, + "learning_rate": 7.606286875190421e-05, + "loss": 0.0077, + "step": 35920 + }, + { + "epoch": 2.7906796116504853, + "grad_norm": 0.2725575864315033, + "learning_rate": 7.604875662349617e-05, + "loss": 0.0094, + "step": 35930 + }, + { + "epoch": 2.7914563106796115, + "grad_norm": 0.2505688965320587, + "learning_rate": 7.603464164643525e-05, + "loss": 0.0095, + "step": 35940 + }, + { + "epoch": 2.792233009708738, + "grad_norm": 0.3933985233306885, + "learning_rate": 7.602052382226507e-05, + "loss": 0.0082, + "step": 35950 + }, + { + "epoch": 2.793009708737864, + "grad_norm": 0.6164013743400574, + "learning_rate": 7.600640315252954e-05, + "loss": 0.0086, + "step": 35960 + }, + { + "epoch": 2.7937864077669903, + "grad_norm": 0.1476677507162094, + "learning_rate": 7.599227963877284e-05, + "loss": 0.0067, + "step": 35970 + }, + { + "epoch": 2.7945631067961165, + "grad_norm": 0.2890162169933319, + "learning_rate": 7.597815328253954e-05, + "loss": 0.0083, + "step": 35980 + }, + { + "epoch": 2.7953398058252428, + "grad_norm": 0.3445001244544983, + "learning_rate": 7.596402408537443e-05, + "loss": 0.0082, + "step": 35990 + }, + { + "epoch": 2.796116504854369, + "grad_norm": 0.2556093633174896, + "learning_rate": 7.59498920488227e-05, + "loss": 0.0096, + "step": 36000 + }, + { + "epoch": 2.7968932038834953, + "grad_norm": 0.2629258632659912, + "learning_rate": 7.593575717442979e-05, + "loss": 0.0107, + "step": 36010 + }, + { + "epoch": 2.7976699029126215, + "grad_norm": 0.26754674315452576, + "learning_rate": 7.592161946374147e-05, + "loss": 0.0132, + "step": 36020 + }, + { + "epoch": 2.7984466019417473, + "grad_norm": 0.2174038589000702, + "learning_rate": 7.590747891830381e-05, + "loss": 0.0097, + "step": 36030 + }, + { + "epoch": 2.7992233009708736, + "grad_norm": 0.33219996094703674, + "learning_rate": 7.58933355396632e-05, + "loss": 0.0099, + "step": 36040 + }, + { + "epoch": 2.8, + "grad_norm": 0.28781184554100037, + "learning_rate": 7.587918932936636e-05, + "loss": 0.0076, + "step": 36050 + }, + { + "epoch": 2.800776699029126, + "grad_norm": 0.23263731598854065, + "learning_rate": 7.586504028896028e-05, + "loss": 0.0089, + "step": 36060 + }, + { + "epoch": 2.8015533980582523, + "grad_norm": 0.2948753237724304, + "learning_rate": 7.585088841999228e-05, + "loss": 0.0097, + "step": 36070 + }, + { + "epoch": 2.8023300970873786, + "grad_norm": 0.3161655366420746, + "learning_rate": 7.583673372400999e-05, + "loss": 0.0097, + "step": 36080 + }, + { + "epoch": 2.803106796116505, + "grad_norm": 0.2203778326511383, + "learning_rate": 7.582257620256134e-05, + "loss": 0.0106, + "step": 36090 + }, + { + "epoch": 2.803883495145631, + "grad_norm": 0.2798476219177246, + "learning_rate": 7.580841585719458e-05, + "loss": 0.0085, + "step": 36100 + }, + { + "epoch": 2.8046601941747573, + "grad_norm": 0.3066023886203766, + "learning_rate": 7.579425268945825e-05, + "loss": 0.0091, + "step": 36110 + }, + { + "epoch": 2.8054368932038836, + "grad_norm": 0.12792952358722687, + "learning_rate": 7.578008670090127e-05, + "loss": 0.0076, + "step": 36120 + }, + { + "epoch": 2.80621359223301, + "grad_norm": 0.15828628838062286, + "learning_rate": 7.576591789307272e-05, + "loss": 0.0093, + "step": 36130 + }, + { + "epoch": 2.806990291262136, + "grad_norm": 0.4177095293998718, + "learning_rate": 7.575174626752216e-05, + "loss": 0.008, + "step": 36140 + }, + { + "epoch": 2.8077669902912623, + "grad_norm": 0.2706030309200287, + "learning_rate": 7.573757182579934e-05, + "loss": 0.0091, + "step": 36150 + }, + { + "epoch": 2.8085436893203886, + "grad_norm": 0.3178713321685791, + "learning_rate": 7.572339456945435e-05, + "loss": 0.0078, + "step": 36160 + }, + { + "epoch": 2.809320388349515, + "grad_norm": 0.5926786065101624, + "learning_rate": 7.570921450003762e-05, + "loss": 0.009, + "step": 36170 + }, + { + "epoch": 2.8100970873786406, + "grad_norm": 0.46950387954711914, + "learning_rate": 7.569503161909983e-05, + "loss": 0.009, + "step": 36180 + }, + { + "epoch": 2.810873786407767, + "grad_norm": 0.18132276833057404, + "learning_rate": 7.568084592819202e-05, + "loss": 0.0062, + "step": 36190 + }, + { + "epoch": 2.811650485436893, + "grad_norm": 0.26120641827583313, + "learning_rate": 7.566665742886551e-05, + "loss": 0.0077, + "step": 36200 + }, + { + "epoch": 2.8124271844660194, + "grad_norm": 0.18289056420326233, + "learning_rate": 7.565246612267194e-05, + "loss": 0.008, + "step": 36210 + }, + { + "epoch": 2.8132038834951456, + "grad_norm": 0.23650729656219482, + "learning_rate": 7.563827201116325e-05, + "loss": 0.0071, + "step": 36220 + }, + { + "epoch": 2.813980582524272, + "grad_norm": 0.5309130549430847, + "learning_rate": 7.562407509589164e-05, + "loss": 0.0108, + "step": 36230 + }, + { + "epoch": 2.814757281553398, + "grad_norm": 0.2612965703010559, + "learning_rate": 7.560987537840973e-05, + "loss": 0.0084, + "step": 36240 + }, + { + "epoch": 2.8155339805825244, + "grad_norm": 0.21210142970085144, + "learning_rate": 7.559567286027036e-05, + "loss": 0.0064, + "step": 36250 + }, + { + "epoch": 2.8163106796116506, + "grad_norm": 0.22180667519569397, + "learning_rate": 7.558146754302668e-05, + "loss": 0.0111, + "step": 36260 + }, + { + "epoch": 2.817087378640777, + "grad_norm": 0.3286290466785431, + "learning_rate": 7.556725942823217e-05, + "loss": 0.0088, + "step": 36270 + }, + { + "epoch": 2.8178640776699027, + "grad_norm": 0.5280014276504517, + "learning_rate": 7.555304851744061e-05, + "loss": 0.0112, + "step": 36280 + }, + { + "epoch": 2.818640776699029, + "grad_norm": 0.1849006563425064, + "learning_rate": 7.553883481220608e-05, + "loss": 0.0066, + "step": 36290 + }, + { + "epoch": 2.819417475728155, + "grad_norm": 0.25473156571388245, + "learning_rate": 7.552461831408298e-05, + "loss": 0.0083, + "step": 36300 + }, + { + "epoch": 2.8201941747572814, + "grad_norm": 0.18492472171783447, + "learning_rate": 7.551039902462599e-05, + "loss": 0.0082, + "step": 36310 + }, + { + "epoch": 2.8209708737864077, + "grad_norm": 0.12105652689933777, + "learning_rate": 7.549617694539011e-05, + "loss": 0.0081, + "step": 36320 + }, + { + "epoch": 2.821747572815534, + "grad_norm": 0.32500433921813965, + "learning_rate": 7.548195207793066e-05, + "loss": 0.0128, + "step": 36330 + }, + { + "epoch": 2.82252427184466, + "grad_norm": 0.5458679795265198, + "learning_rate": 7.546772442380323e-05, + "loss": 0.0099, + "step": 36340 + }, + { + "epoch": 2.8233009708737864, + "grad_norm": 0.2707365155220032, + "learning_rate": 7.545349398456376e-05, + "loss": 0.0095, + "step": 36350 + }, + { + "epoch": 2.8240776699029126, + "grad_norm": 0.3144366443157196, + "learning_rate": 7.543926076176845e-05, + "loss": 0.0097, + "step": 36360 + }, + { + "epoch": 2.824854368932039, + "grad_norm": 0.24453841149806976, + "learning_rate": 7.542502475697385e-05, + "loss": 0.0098, + "step": 36370 + }, + { + "epoch": 2.825631067961165, + "grad_norm": 0.27243202924728394, + "learning_rate": 7.541078597173675e-05, + "loss": 0.0117, + "step": 36380 + }, + { + "epoch": 2.8264077669902914, + "grad_norm": 0.2156183272600174, + "learning_rate": 7.539654440761431e-05, + "loss": 0.0079, + "step": 36390 + }, + { + "epoch": 2.8271844660194176, + "grad_norm": 0.16529788076877594, + "learning_rate": 7.538230006616395e-05, + "loss": 0.0072, + "step": 36400 + }, + { + "epoch": 2.827961165048544, + "grad_norm": 0.7916640043258667, + "learning_rate": 7.536805294894342e-05, + "loss": 0.009, + "step": 36410 + }, + { + "epoch": 2.82873786407767, + "grad_norm": 0.2141803354024887, + "learning_rate": 7.535380305751076e-05, + "loss": 0.007, + "step": 36420 + }, + { + "epoch": 2.8295145631067964, + "grad_norm": 0.36209169030189514, + "learning_rate": 7.533955039342431e-05, + "loss": 0.0109, + "step": 36430 + }, + { + "epoch": 2.830291262135922, + "grad_norm": 0.140488862991333, + "learning_rate": 7.532529495824274e-05, + "loss": 0.0086, + "step": 36440 + }, + { + "epoch": 2.8310679611650484, + "grad_norm": 0.16760650277137756, + "learning_rate": 7.531103675352497e-05, + "loss": 0.0083, + "step": 36450 + }, + { + "epoch": 2.8318446601941747, + "grad_norm": 0.2190031260251999, + "learning_rate": 7.529677578083029e-05, + "loss": 0.0086, + "step": 36460 + }, + { + "epoch": 2.832621359223301, + "grad_norm": 0.23905044794082642, + "learning_rate": 7.528251204171823e-05, + "loss": 0.0082, + "step": 36470 + }, + { + "epoch": 2.833398058252427, + "grad_norm": 0.8668004870414734, + "learning_rate": 7.526824553774867e-05, + "loss": 0.0095, + "step": 36480 + }, + { + "epoch": 2.8341747572815534, + "grad_norm": 0.24368993937969208, + "learning_rate": 7.525397627048178e-05, + "loss": 0.0079, + "step": 36490 + }, + { + "epoch": 2.8349514563106797, + "grad_norm": 0.19566261768341064, + "learning_rate": 7.523970424147802e-05, + "loss": 0.0083, + "step": 36500 + }, + { + "epoch": 2.835728155339806, + "grad_norm": 0.2479712963104248, + "learning_rate": 7.522542945229813e-05, + "loss": 0.0083, + "step": 36510 + }, + { + "epoch": 2.836504854368932, + "grad_norm": 0.22717449069023132, + "learning_rate": 7.521115190450321e-05, + "loss": 0.008, + "step": 36520 + }, + { + "epoch": 2.837281553398058, + "grad_norm": 0.3291579484939575, + "learning_rate": 7.519687159965462e-05, + "loss": 0.0083, + "step": 36530 + }, + { + "epoch": 2.8380582524271842, + "grad_norm": 0.2248668074607849, + "learning_rate": 7.518258853931403e-05, + "loss": 0.0075, + "step": 36540 + }, + { + "epoch": 2.8388349514563105, + "grad_norm": 0.3457690477371216, + "learning_rate": 7.516830272504342e-05, + "loss": 0.0104, + "step": 36550 + }, + { + "epoch": 2.8396116504854367, + "grad_norm": 0.2984236180782318, + "learning_rate": 7.515401415840509e-05, + "loss": 0.0081, + "step": 36560 + }, + { + "epoch": 2.840388349514563, + "grad_norm": 0.4495633840560913, + "learning_rate": 7.513972284096155e-05, + "loss": 0.011, + "step": 36570 + }, + { + "epoch": 2.8411650485436892, + "grad_norm": 1.0354080200195312, + "learning_rate": 7.512542877427574e-05, + "loss": 0.0077, + "step": 36580 + }, + { + "epoch": 2.8419417475728155, + "grad_norm": 0.31687116622924805, + "learning_rate": 7.51111319599108e-05, + "loss": 0.0089, + "step": 36590 + }, + { + "epoch": 2.8427184466019417, + "grad_norm": 0.24401642382144928, + "learning_rate": 7.509683239943024e-05, + "loss": 0.0097, + "step": 36600 + }, + { + "epoch": 2.843495145631068, + "grad_norm": 0.31116974353790283, + "learning_rate": 7.50825300943978e-05, + "loss": 0.0109, + "step": 36610 + }, + { + "epoch": 2.8442718446601942, + "grad_norm": 0.2577105760574341, + "learning_rate": 7.50682250463776e-05, + "loss": 0.0113, + "step": 36620 + }, + { + "epoch": 2.8450485436893205, + "grad_norm": 0.2553131580352783, + "learning_rate": 7.5053917256934e-05, + "loss": 0.007, + "step": 36630 + }, + { + "epoch": 2.8458252427184467, + "grad_norm": 0.1547163724899292, + "learning_rate": 7.503960672763166e-05, + "loss": 0.0091, + "step": 36640 + }, + { + "epoch": 2.846601941747573, + "grad_norm": 0.3752181828022003, + "learning_rate": 7.502529346003559e-05, + "loss": 0.0084, + "step": 36650 + }, + { + "epoch": 2.847378640776699, + "grad_norm": 0.25979316234588623, + "learning_rate": 7.501097745571106e-05, + "loss": 0.0098, + "step": 36660 + }, + { + "epoch": 2.8481553398058255, + "grad_norm": 0.28662291169166565, + "learning_rate": 7.499665871622361e-05, + "loss": 0.0101, + "step": 36670 + }, + { + "epoch": 2.8489320388349517, + "grad_norm": 0.19564040005207062, + "learning_rate": 7.498233724313917e-05, + "loss": 0.0088, + "step": 36680 + }, + { + "epoch": 2.8497087378640775, + "grad_norm": 0.2657186686992645, + "learning_rate": 7.496801303802387e-05, + "loss": 0.0075, + "step": 36690 + }, + { + "epoch": 2.8504854368932038, + "grad_norm": 0.294119656085968, + "learning_rate": 7.495368610244423e-05, + "loss": 0.0087, + "step": 36700 + }, + { + "epoch": 2.85126213592233, + "grad_norm": 0.1946844756603241, + "learning_rate": 7.493935643796697e-05, + "loss": 0.0089, + "step": 36710 + }, + { + "epoch": 2.8520388349514563, + "grad_norm": 0.24988344311714172, + "learning_rate": 7.49250240461592e-05, + "loss": 0.0091, + "step": 36720 + }, + { + "epoch": 2.8528155339805825, + "grad_norm": 0.19058650732040405, + "learning_rate": 7.491068892858826e-05, + "loss": 0.0067, + "step": 36730 + }, + { + "epoch": 2.8535922330097088, + "grad_norm": 0.2990257740020752, + "learning_rate": 7.489635108682184e-05, + "loss": 0.0078, + "step": 36740 + }, + { + "epoch": 2.854368932038835, + "grad_norm": 0.262846440076828, + "learning_rate": 7.48820105224279e-05, + "loss": 0.0104, + "step": 36750 + }, + { + "epoch": 2.8551456310679613, + "grad_norm": 0.252003014087677, + "learning_rate": 7.48676672369747e-05, + "loss": 0.0061, + "step": 36760 + }, + { + "epoch": 2.8559223300970875, + "grad_norm": 0.16165700554847717, + "learning_rate": 7.485332123203079e-05, + "loss": 0.0067, + "step": 36770 + }, + { + "epoch": 2.8566990291262133, + "grad_norm": 0.32087570428848267, + "learning_rate": 7.483897250916504e-05, + "loss": 0.0076, + "step": 36780 + }, + { + "epoch": 2.8574757281553396, + "grad_norm": 0.30613255500793457, + "learning_rate": 7.48246210699466e-05, + "loss": 0.0073, + "step": 36790 + }, + { + "epoch": 2.858252427184466, + "grad_norm": 0.22761870920658112, + "learning_rate": 7.481026691594492e-05, + "loss": 0.0064, + "step": 36800 + }, + { + "epoch": 2.859029126213592, + "grad_norm": 0.2781147360801697, + "learning_rate": 7.479591004872975e-05, + "loss": 0.0131, + "step": 36810 + }, + { + "epoch": 2.8598058252427183, + "grad_norm": 0.20959512889385223, + "learning_rate": 7.478155046987114e-05, + "loss": 0.0088, + "step": 36820 + }, + { + "epoch": 2.8605825242718446, + "grad_norm": 0.20407000184059143, + "learning_rate": 7.476718818093944e-05, + "loss": 0.0089, + "step": 36830 + }, + { + "epoch": 2.861359223300971, + "grad_norm": 0.18305113911628723, + "learning_rate": 7.475282318350527e-05, + "loss": 0.009, + "step": 36840 + }, + { + "epoch": 2.862135922330097, + "grad_norm": 0.23636409640312195, + "learning_rate": 7.473845547913957e-05, + "loss": 0.0127, + "step": 36850 + }, + { + "epoch": 2.8629126213592233, + "grad_norm": 0.7145712375640869, + "learning_rate": 7.472408506941357e-05, + "loss": 0.0099, + "step": 36860 + }, + { + "epoch": 2.8636893203883496, + "grad_norm": 0.39595115184783936, + "learning_rate": 7.470971195589881e-05, + "loss": 0.0072, + "step": 36870 + }, + { + "epoch": 2.864466019417476, + "grad_norm": 0.23815493285655975, + "learning_rate": 7.46953361401671e-05, + "loss": 0.0075, + "step": 36880 + }, + { + "epoch": 2.865242718446602, + "grad_norm": 0.5501708984375, + "learning_rate": 7.468095762379055e-05, + "loss": 0.0129, + "step": 36890 + }, + { + "epoch": 2.8660194174757283, + "grad_norm": 0.2070484459400177, + "learning_rate": 7.466657640834158e-05, + "loss": 0.009, + "step": 36900 + }, + { + "epoch": 2.8667961165048546, + "grad_norm": 0.21802902221679688, + "learning_rate": 7.46521924953929e-05, + "loss": 0.0081, + "step": 36910 + }, + { + "epoch": 2.867572815533981, + "grad_norm": 0.16186782717704773, + "learning_rate": 7.463780588651752e-05, + "loss": 0.0089, + "step": 36920 + }, + { + "epoch": 2.868349514563107, + "grad_norm": 0.24871322512626648, + "learning_rate": 7.462341658328873e-05, + "loss": 0.0112, + "step": 36930 + }, + { + "epoch": 2.869126213592233, + "grad_norm": 0.23215007781982422, + "learning_rate": 7.460902458728012e-05, + "loss": 0.0078, + "step": 36940 + }, + { + "epoch": 2.869902912621359, + "grad_norm": 0.21700119972229004, + "learning_rate": 7.459462990006558e-05, + "loss": 0.0066, + "step": 36950 + }, + { + "epoch": 2.8706796116504854, + "grad_norm": 0.24981383979320526, + "learning_rate": 7.458023252321929e-05, + "loss": 0.0095, + "step": 36960 + }, + { + "epoch": 2.8714563106796116, + "grad_norm": 0.2945190370082855, + "learning_rate": 7.456583245831573e-05, + "loss": 0.0092, + "step": 36970 + }, + { + "epoch": 2.872233009708738, + "grad_norm": 0.19977720081806183, + "learning_rate": 7.455142970692968e-05, + "loss": 0.0092, + "step": 36980 + }, + { + "epoch": 2.873009708737864, + "grad_norm": 0.20910139381885529, + "learning_rate": 7.453702427063615e-05, + "loss": 0.0086, + "step": 36990 + }, + { + "epoch": 2.8737864077669903, + "grad_norm": 0.18347042798995972, + "learning_rate": 7.452261615101057e-05, + "loss": 0.0086, + "step": 37000 + }, + { + "epoch": 2.8745631067961166, + "grad_norm": 0.3882339596748352, + "learning_rate": 7.450820534962856e-05, + "loss": 0.0083, + "step": 37010 + }, + { + "epoch": 2.875339805825243, + "grad_norm": 0.3993628919124603, + "learning_rate": 7.449379186806604e-05, + "loss": 0.0074, + "step": 37020 + }, + { + "epoch": 2.876116504854369, + "grad_norm": 0.11334159970283508, + "learning_rate": 7.44793757078993e-05, + "loss": 0.0096, + "step": 37030 + }, + { + "epoch": 2.876893203883495, + "grad_norm": 0.578117311000824, + "learning_rate": 7.44649568707048e-05, + "loss": 0.0121, + "step": 37040 + }, + { + "epoch": 2.877669902912621, + "grad_norm": 0.523746132850647, + "learning_rate": 7.445053535805942e-05, + "loss": 0.0083, + "step": 37050 + }, + { + "epoch": 2.8784466019417474, + "grad_norm": 0.9494485259056091, + "learning_rate": 7.443611117154024e-05, + "loss": 0.0092, + "step": 37060 + }, + { + "epoch": 2.8792233009708736, + "grad_norm": 0.3874399662017822, + "learning_rate": 7.442168431272469e-05, + "loss": 0.0096, + "step": 37070 + }, + { + "epoch": 2.88, + "grad_norm": 0.33323046565055847, + "learning_rate": 7.440725478319044e-05, + "loss": 0.0094, + "step": 37080 + }, + { + "epoch": 2.880776699029126, + "grad_norm": 0.20913174748420715, + "learning_rate": 7.439282258451553e-05, + "loss": 0.0095, + "step": 37090 + }, + { + "epoch": 2.8815533980582524, + "grad_norm": 0.33804211020469666, + "learning_rate": 7.43783877182782e-05, + "loss": 0.009, + "step": 37100 + }, + { + "epoch": 2.8823300970873786, + "grad_norm": 0.20540180802345276, + "learning_rate": 7.436395018605707e-05, + "loss": 0.0085, + "step": 37110 + }, + { + "epoch": 2.883106796116505, + "grad_norm": 0.21145959198474884, + "learning_rate": 7.434950998943094e-05, + "loss": 0.0078, + "step": 37120 + }, + { + "epoch": 2.883883495145631, + "grad_norm": 0.2108081728219986, + "learning_rate": 7.433506712997902e-05, + "loss": 0.0102, + "step": 37130 + }, + { + "epoch": 2.8846601941747574, + "grad_norm": 0.22837188839912415, + "learning_rate": 7.432062160928077e-05, + "loss": 0.0134, + "step": 37140 + }, + { + "epoch": 2.8854368932038836, + "grad_norm": 0.2687339782714844, + "learning_rate": 7.430617342891588e-05, + "loss": 0.0077, + "step": 37150 + }, + { + "epoch": 2.88621359223301, + "grad_norm": 0.2018287628889084, + "learning_rate": 7.429172259046442e-05, + "loss": 0.0089, + "step": 37160 + }, + { + "epoch": 2.886990291262136, + "grad_norm": 0.646565854549408, + "learning_rate": 7.42772690955067e-05, + "loss": 0.0077, + "step": 37170 + }, + { + "epoch": 2.8877669902912624, + "grad_norm": 0.1870768517255783, + "learning_rate": 7.426281294562332e-05, + "loss": 0.0087, + "step": 37180 + }, + { + "epoch": 2.888543689320388, + "grad_norm": 0.2002519965171814, + "learning_rate": 7.424835414239519e-05, + "loss": 0.007, + "step": 37190 + }, + { + "epoch": 2.8893203883495144, + "grad_norm": 0.22629249095916748, + "learning_rate": 7.423389268740352e-05, + "loss": 0.0062, + "step": 37200 + }, + { + "epoch": 2.8900970873786407, + "grad_norm": 0.15628278255462646, + "learning_rate": 7.42194285822298e-05, + "loss": 0.01, + "step": 37210 + }, + { + "epoch": 2.890873786407767, + "grad_norm": 0.24304834008216858, + "learning_rate": 7.420496182845576e-05, + "loss": 0.0068, + "step": 37220 + }, + { + "epoch": 2.891650485436893, + "grad_norm": 0.672275722026825, + "learning_rate": 7.419049242766348e-05, + "loss": 0.0099, + "step": 37230 + }, + { + "epoch": 2.8924271844660194, + "grad_norm": 0.7069689035415649, + "learning_rate": 7.417602038143533e-05, + "loss": 0.0097, + "step": 37240 + }, + { + "epoch": 2.8932038834951457, + "grad_norm": 0.28242552280426025, + "learning_rate": 7.416154569135393e-05, + "loss": 0.0101, + "step": 37250 + }, + { + "epoch": 2.893980582524272, + "grad_norm": 0.20501968264579773, + "learning_rate": 7.414706835900222e-05, + "loss": 0.0092, + "step": 37260 + }, + { + "epoch": 2.894757281553398, + "grad_norm": 0.23595073819160461, + "learning_rate": 7.413258838596345e-05, + "loss": 0.0079, + "step": 37270 + }, + { + "epoch": 2.8955339805825244, + "grad_norm": 0.369900107383728, + "learning_rate": 7.411810577382106e-05, + "loss": 0.011, + "step": 37280 + }, + { + "epoch": 2.8963106796116502, + "grad_norm": 0.20635677874088287, + "learning_rate": 7.410362052415889e-05, + "loss": 0.0079, + "step": 37290 + }, + { + "epoch": 2.8970873786407765, + "grad_norm": 0.1881965696811676, + "learning_rate": 7.408913263856102e-05, + "loss": 0.0075, + "step": 37300 + }, + { + "epoch": 2.8978640776699027, + "grad_norm": 0.22469472885131836, + "learning_rate": 7.407464211861183e-05, + "loss": 0.0101, + "step": 37310 + }, + { + "epoch": 2.898640776699029, + "grad_norm": 0.28833451867103577, + "learning_rate": 7.406014896589597e-05, + "loss": 0.0083, + "step": 37320 + }, + { + "epoch": 2.8994174757281552, + "grad_norm": 0.7725570201873779, + "learning_rate": 7.404565318199838e-05, + "loss": 0.0083, + "step": 37330 + }, + { + "epoch": 2.9001941747572815, + "grad_norm": 0.19505400955677032, + "learning_rate": 7.403115476850432e-05, + "loss": 0.0066, + "step": 37340 + }, + { + "epoch": 2.9009708737864077, + "grad_norm": 0.3422318994998932, + "learning_rate": 7.401665372699932e-05, + "loss": 0.0101, + "step": 37350 + }, + { + "epoch": 2.901747572815534, + "grad_norm": 0.35197946429252625, + "learning_rate": 7.400215005906916e-05, + "loss": 0.0076, + "step": 37360 + }, + { + "epoch": 2.90252427184466, + "grad_norm": 0.28273168206214905, + "learning_rate": 7.398764376629996e-05, + "loss": 0.0088, + "step": 37370 + }, + { + "epoch": 2.9033009708737865, + "grad_norm": 0.28187286853790283, + "learning_rate": 7.39731348502781e-05, + "loss": 0.0089, + "step": 37380 + }, + { + "epoch": 2.9040776699029127, + "grad_norm": 0.315491646528244, + "learning_rate": 7.395862331259028e-05, + "loss": 0.0081, + "step": 37390 + }, + { + "epoch": 2.904854368932039, + "grad_norm": 0.5143993496894836, + "learning_rate": 7.394410915482341e-05, + "loss": 0.0085, + "step": 37400 + }, + { + "epoch": 2.905631067961165, + "grad_norm": 0.33126363158226013, + "learning_rate": 7.392959237856476e-05, + "loss": 0.0071, + "step": 37410 + }, + { + "epoch": 2.9064077669902915, + "grad_norm": 0.3435538113117218, + "learning_rate": 7.391507298540187e-05, + "loss": 0.0093, + "step": 37420 + }, + { + "epoch": 2.9071844660194177, + "grad_norm": 0.6924113631248474, + "learning_rate": 7.390055097692257e-05, + "loss": 0.008, + "step": 37430 + }, + { + "epoch": 2.907961165048544, + "grad_norm": 0.2076479196548462, + "learning_rate": 7.388602635471494e-05, + "loss": 0.0082, + "step": 37440 + }, + { + "epoch": 2.9087378640776698, + "grad_norm": 0.27476027607917786, + "learning_rate": 7.387149912036738e-05, + "loss": 0.0069, + "step": 37450 + }, + { + "epoch": 2.909514563106796, + "grad_norm": 0.5104552507400513, + "learning_rate": 7.385696927546854e-05, + "loss": 0.0074, + "step": 37460 + }, + { + "epoch": 2.9102912621359223, + "grad_norm": 0.4574578106403351, + "learning_rate": 7.384243682160743e-05, + "loss": 0.0105, + "step": 37470 + }, + { + "epoch": 2.9110679611650485, + "grad_norm": 0.3090344965457916, + "learning_rate": 7.382790176037329e-05, + "loss": 0.0089, + "step": 37480 + }, + { + "epoch": 2.9118446601941748, + "grad_norm": 0.21601852774620056, + "learning_rate": 7.381336409335562e-05, + "loss": 0.0076, + "step": 37490 + }, + { + "epoch": 2.912621359223301, + "grad_norm": 0.19484342634677887, + "learning_rate": 7.379882382214426e-05, + "loss": 0.0102, + "step": 37500 + }, + { + "epoch": 2.9133980582524273, + "grad_norm": 0.29342105984687805, + "learning_rate": 7.378428094832931e-05, + "loss": 0.0068, + "step": 37510 + }, + { + "epoch": 2.9141747572815535, + "grad_norm": 0.18665426969528198, + "learning_rate": 7.376973547350114e-05, + "loss": 0.0071, + "step": 37520 + }, + { + "epoch": 2.9149514563106798, + "grad_norm": 0.37949684262275696, + "learning_rate": 7.375518739925046e-05, + "loss": 0.007, + "step": 37530 + }, + { + "epoch": 2.9157281553398056, + "grad_norm": 0.25459036231040955, + "learning_rate": 7.374063672716818e-05, + "loss": 0.0069, + "step": 37540 + }, + { + "epoch": 2.916504854368932, + "grad_norm": 0.21772845089435577, + "learning_rate": 7.372608345884558e-05, + "loss": 0.008, + "step": 37550 + }, + { + "epoch": 2.917281553398058, + "grad_norm": 0.15874241292476654, + "learning_rate": 7.371152759587415e-05, + "loss": 0.0063, + "step": 37560 + }, + { + "epoch": 2.9180582524271843, + "grad_norm": 0.23098178207874298, + "learning_rate": 7.369696913984572e-05, + "loss": 0.012, + "step": 37570 + }, + { + "epoch": 2.9188349514563106, + "grad_norm": 0.21918566524982452, + "learning_rate": 7.368240809235237e-05, + "loss": 0.0079, + "step": 37580 + }, + { + "epoch": 2.919611650485437, + "grad_norm": 0.2220563441514969, + "learning_rate": 7.366784445498649e-05, + "loss": 0.0081, + "step": 37590 + }, + { + "epoch": 2.920388349514563, + "grad_norm": 0.1683976948261261, + "learning_rate": 7.36532782293407e-05, + "loss": 0.0087, + "step": 37600 + }, + { + "epoch": 2.9211650485436893, + "grad_norm": 0.4236260950565338, + "learning_rate": 7.363870941700797e-05, + "loss": 0.0095, + "step": 37610 + }, + { + "epoch": 2.9219417475728156, + "grad_norm": 0.2370765656232834, + "learning_rate": 7.362413801958152e-05, + "loss": 0.0084, + "step": 37620 + }, + { + "epoch": 2.922718446601942, + "grad_norm": 0.1648528277873993, + "learning_rate": 7.360956403865486e-05, + "loss": 0.0074, + "step": 37630 + }, + { + "epoch": 2.923495145631068, + "grad_norm": 0.6250187158584595, + "learning_rate": 7.359498747582177e-05, + "loss": 0.0081, + "step": 37640 + }, + { + "epoch": 2.9242718446601943, + "grad_norm": 0.16078118979930878, + "learning_rate": 7.358040833267634e-05, + "loss": 0.0068, + "step": 37650 + }, + { + "epoch": 2.9250485436893205, + "grad_norm": 0.31987810134887695, + "learning_rate": 7.356582661081288e-05, + "loss": 0.0071, + "step": 37660 + }, + { + "epoch": 2.925825242718447, + "grad_norm": 0.40545833110809326, + "learning_rate": 7.355124231182607e-05, + "loss": 0.0085, + "step": 37670 + }, + { + "epoch": 2.926601941747573, + "grad_norm": 0.21679037809371948, + "learning_rate": 7.353665543731083e-05, + "loss": 0.0077, + "step": 37680 + }, + { + "epoch": 2.9273786407766993, + "grad_norm": 0.28255006670951843, + "learning_rate": 7.352206598886229e-05, + "loss": 0.0082, + "step": 37690 + }, + { + "epoch": 2.928155339805825, + "grad_norm": 0.4250715374946594, + "learning_rate": 7.350747396807601e-05, + "loss": 0.0089, + "step": 37700 + }, + { + "epoch": 2.9289320388349513, + "grad_norm": 0.20261211693286896, + "learning_rate": 7.349287937654772e-05, + "loss": 0.0078, + "step": 37710 + }, + { + "epoch": 2.9297087378640776, + "grad_norm": 0.3033435046672821, + "learning_rate": 7.347828221587345e-05, + "loss": 0.0104, + "step": 37720 + }, + { + "epoch": 2.930485436893204, + "grad_norm": 0.40414801239967346, + "learning_rate": 7.346368248764955e-05, + "loss": 0.0066, + "step": 37730 + }, + { + "epoch": 2.93126213592233, + "grad_norm": 0.6554610729217529, + "learning_rate": 7.34490801934726e-05, + "loss": 0.0087, + "step": 37740 + }, + { + "epoch": 2.9320388349514563, + "grad_norm": 0.17830680310726166, + "learning_rate": 7.343447533493947e-05, + "loss": 0.0071, + "step": 37750 + }, + { + "epoch": 2.9328155339805826, + "grad_norm": 0.22125239670276642, + "learning_rate": 7.341986791364739e-05, + "loss": 0.0059, + "step": 37760 + }, + { + "epoch": 2.933592233009709, + "grad_norm": 0.14642781019210815, + "learning_rate": 7.340525793119374e-05, + "loss": 0.0098, + "step": 37770 + }, + { + "epoch": 2.934368932038835, + "grad_norm": 0.41601425409317017, + "learning_rate": 7.33906453891763e-05, + "loss": 0.0086, + "step": 37780 + }, + { + "epoch": 2.935145631067961, + "grad_norm": 0.16680605709552765, + "learning_rate": 7.337603028919303e-05, + "loss": 0.0069, + "step": 37790 + }, + { + "epoch": 2.935922330097087, + "grad_norm": 0.3882351517677307, + "learning_rate": 7.336141263284225e-05, + "loss": 0.007, + "step": 37800 + }, + { + "epoch": 2.9366990291262134, + "grad_norm": 0.1451406329870224, + "learning_rate": 7.334679242172249e-05, + "loss": 0.0047, + "step": 37810 + }, + { + "epoch": 2.9374757281553396, + "grad_norm": 0.28242069482803345, + "learning_rate": 7.33321696574326e-05, + "loss": 0.0097, + "step": 37820 + }, + { + "epoch": 2.938252427184466, + "grad_norm": 0.23719896376132965, + "learning_rate": 7.331754434157174e-05, + "loss": 0.0076, + "step": 37830 + }, + { + "epoch": 2.939029126213592, + "grad_norm": 0.40142935514450073, + "learning_rate": 7.330291647573928e-05, + "loss": 0.0103, + "step": 37840 + }, + { + "epoch": 2.9398058252427184, + "grad_norm": 0.14988282322883606, + "learning_rate": 7.32882860615349e-05, + "loss": 0.0088, + "step": 37850 + }, + { + "epoch": 2.9405825242718446, + "grad_norm": 0.30288735032081604, + "learning_rate": 7.327365310055858e-05, + "loss": 0.0084, + "step": 37860 + }, + { + "epoch": 2.941359223300971, + "grad_norm": 0.19227224588394165, + "learning_rate": 7.325901759441055e-05, + "loss": 0.0071, + "step": 37870 + }, + { + "epoch": 2.942135922330097, + "grad_norm": 0.21701323986053467, + "learning_rate": 7.324437954469132e-05, + "loss": 0.0071, + "step": 37880 + }, + { + "epoch": 2.9429126213592234, + "grad_norm": 0.14870251715183258, + "learning_rate": 7.32297389530017e-05, + "loss": 0.01, + "step": 37890 + }, + { + "epoch": 2.9436893203883496, + "grad_norm": 0.23570899665355682, + "learning_rate": 7.321509582094274e-05, + "loss": 0.0075, + "step": 37900 + }, + { + "epoch": 2.944466019417476, + "grad_norm": 0.41186535358428955, + "learning_rate": 7.320045015011582e-05, + "loss": 0.0074, + "step": 37910 + }, + { + "epoch": 2.945242718446602, + "grad_norm": 0.16797210276126862, + "learning_rate": 7.318580194212253e-05, + "loss": 0.0057, + "step": 37920 + }, + { + "epoch": 2.9460194174757284, + "grad_norm": 0.3093582093715668, + "learning_rate": 7.317115119856483e-05, + "loss": 0.0093, + "step": 37930 + }, + { + "epoch": 2.9467961165048546, + "grad_norm": 0.17265872657299042, + "learning_rate": 7.315649792104485e-05, + "loss": 0.0105, + "step": 37940 + }, + { + "epoch": 2.9475728155339804, + "grad_norm": 0.21754087507724762, + "learning_rate": 7.314184211116507e-05, + "loss": 0.01, + "step": 37950 + }, + { + "epoch": 2.9483495145631067, + "grad_norm": 0.3348013460636139, + "learning_rate": 7.312718377052823e-05, + "loss": 0.009, + "step": 37960 + }, + { + "epoch": 2.949126213592233, + "grad_norm": 0.3544814884662628, + "learning_rate": 7.311252290073737e-05, + "loss": 0.0105, + "step": 37970 + }, + { + "epoch": 2.949902912621359, + "grad_norm": 0.36550912261009216, + "learning_rate": 7.309785950339572e-05, + "loss": 0.0101, + "step": 37980 + }, + { + "epoch": 2.9506796116504854, + "grad_norm": 0.3428972065448761, + "learning_rate": 7.308319358010692e-05, + "loss": 0.0107, + "step": 37990 + }, + { + "epoch": 2.9514563106796117, + "grad_norm": 0.348423570394516, + "learning_rate": 7.306852513247474e-05, + "loss": 0.0082, + "step": 38000 + }, + { + "epoch": 2.952233009708738, + "grad_norm": 0.2857747972011566, + "learning_rate": 7.305385416210336e-05, + "loss": 0.0079, + "step": 38010 + }, + { + "epoch": 2.953009708737864, + "grad_norm": 0.14394892752170563, + "learning_rate": 7.303918067059711e-05, + "loss": 0.0083, + "step": 38020 + }, + { + "epoch": 2.9537864077669904, + "grad_norm": 0.3640015125274658, + "learning_rate": 7.302450465956075e-05, + "loss": 0.0078, + "step": 38030 + }, + { + "epoch": 2.9545631067961167, + "grad_norm": 0.26694783568382263, + "learning_rate": 7.300982613059914e-05, + "loss": 0.008, + "step": 38040 + }, + { + "epoch": 2.9553398058252425, + "grad_norm": 0.20970934629440308, + "learning_rate": 7.299514508531757e-05, + "loss": 0.0107, + "step": 38050 + }, + { + "epoch": 2.9561165048543687, + "grad_norm": 0.24720212817192078, + "learning_rate": 7.298046152532148e-05, + "loss": 0.0094, + "step": 38060 + }, + { + "epoch": 2.956893203883495, + "grad_norm": 0.17805178463459015, + "learning_rate": 7.296577545221668e-05, + "loss": 0.0092, + "step": 38070 + }, + { + "epoch": 2.957669902912621, + "grad_norm": 0.21750889718532562, + "learning_rate": 7.29510868676092e-05, + "loss": 0.0107, + "step": 38080 + }, + { + "epoch": 2.9584466019417475, + "grad_norm": 0.40489619970321655, + "learning_rate": 7.293639577310538e-05, + "loss": 0.0117, + "step": 38090 + }, + { + "epoch": 2.9592233009708737, + "grad_norm": 0.1905594766139984, + "learning_rate": 7.292170217031179e-05, + "loss": 0.0097, + "step": 38100 + }, + { + "epoch": 2.96, + "grad_norm": 0.214363694190979, + "learning_rate": 7.290700606083532e-05, + "loss": 0.0105, + "step": 38110 + }, + { + "epoch": 2.960776699029126, + "grad_norm": 0.35597196221351624, + "learning_rate": 7.28923074462831e-05, + "loss": 0.0099, + "step": 38120 + }, + { + "epoch": 2.9615533980582525, + "grad_norm": 0.1591271460056305, + "learning_rate": 7.287760632826255e-05, + "loss": 0.01, + "step": 38130 + }, + { + "epoch": 2.9623300970873787, + "grad_norm": 0.2513035237789154, + "learning_rate": 7.286290270838139e-05, + "loss": 0.0078, + "step": 38140 + }, + { + "epoch": 2.963106796116505, + "grad_norm": 0.19612133502960205, + "learning_rate": 7.284819658824756e-05, + "loss": 0.0095, + "step": 38150 + }, + { + "epoch": 2.963883495145631, + "grad_norm": 0.15526387095451355, + "learning_rate": 7.283348796946929e-05, + "loss": 0.0083, + "step": 38160 + }, + { + "epoch": 2.9646601941747575, + "grad_norm": 0.24459274113178253, + "learning_rate": 7.28187768536551e-05, + "loss": 0.008, + "step": 38170 + }, + { + "epoch": 2.9654368932038837, + "grad_norm": 0.19083550572395325, + "learning_rate": 7.280406324241379e-05, + "loss": 0.0104, + "step": 38180 + }, + { + "epoch": 2.96621359223301, + "grad_norm": 0.3151388168334961, + "learning_rate": 7.278934713735441e-05, + "loss": 0.0065, + "step": 38190 + }, + { + "epoch": 2.9669902912621358, + "grad_norm": 0.43013447523117065, + "learning_rate": 7.277462854008629e-05, + "loss": 0.0095, + "step": 38200 + }, + { + "epoch": 2.967766990291262, + "grad_norm": 0.39065325260162354, + "learning_rate": 7.275990745221904e-05, + "loss": 0.0119, + "step": 38210 + }, + { + "epoch": 2.9685436893203883, + "grad_norm": 0.26708438992500305, + "learning_rate": 7.274518387536252e-05, + "loss": 0.0075, + "step": 38220 + }, + { + "epoch": 2.9693203883495145, + "grad_norm": 0.2814711928367615, + "learning_rate": 7.273045781112689e-05, + "loss": 0.0096, + "step": 38230 + }, + { + "epoch": 2.9700970873786408, + "grad_norm": 0.22441834211349487, + "learning_rate": 7.271572926112257e-05, + "loss": 0.01, + "step": 38240 + }, + { + "epoch": 2.970873786407767, + "grad_norm": 0.19772565364837646, + "learning_rate": 7.270099822696024e-05, + "loss": 0.0071, + "step": 38250 + }, + { + "epoch": 2.9716504854368933, + "grad_norm": 0.33537009358406067, + "learning_rate": 7.26862647102509e-05, + "loss": 0.0111, + "step": 38260 + }, + { + "epoch": 2.9724271844660195, + "grad_norm": 0.4719844460487366, + "learning_rate": 7.267152871260573e-05, + "loss": 0.0109, + "step": 38270 + }, + { + "epoch": 2.9732038834951457, + "grad_norm": 0.3098474144935608, + "learning_rate": 7.265679023563629e-05, + "loss": 0.0094, + "step": 38280 + }, + { + "epoch": 2.973980582524272, + "grad_norm": 0.29456135630607605, + "learning_rate": 7.264204928095433e-05, + "loss": 0.0108, + "step": 38290 + }, + { + "epoch": 2.974757281553398, + "grad_norm": 0.4825846254825592, + "learning_rate": 7.262730585017188e-05, + "loss": 0.0078, + "step": 38300 + }, + { + "epoch": 2.975533980582524, + "grad_norm": 0.2836378216743469, + "learning_rate": 7.261255994490131e-05, + "loss": 0.0114, + "step": 38310 + }, + { + "epoch": 2.9763106796116503, + "grad_norm": 0.2710714042186737, + "learning_rate": 7.259781156675517e-05, + "loss": 0.0078, + "step": 38320 + }, + { + "epoch": 2.9770873786407765, + "grad_norm": 0.2125166803598404, + "learning_rate": 7.258306071734633e-05, + "loss": 0.0108, + "step": 38330 + }, + { + "epoch": 2.977864077669903, + "grad_norm": 0.15312612056732178, + "learning_rate": 7.256830739828792e-05, + "loss": 0.0083, + "step": 38340 + }, + { + "epoch": 2.978640776699029, + "grad_norm": 0.29221227765083313, + "learning_rate": 7.255355161119336e-05, + "loss": 0.0087, + "step": 38350 + }, + { + "epoch": 2.9794174757281553, + "grad_norm": 0.19540953636169434, + "learning_rate": 7.253879335767628e-05, + "loss": 0.009, + "step": 38360 + }, + { + "epoch": 2.9801941747572815, + "grad_norm": 0.2994844317436218, + "learning_rate": 7.252403263935067e-05, + "loss": 0.0088, + "step": 38370 + }, + { + "epoch": 2.980970873786408, + "grad_norm": 0.3590647876262665, + "learning_rate": 7.25092694578307e-05, + "loss": 0.0122, + "step": 38380 + }, + { + "epoch": 2.981747572815534, + "grad_norm": 0.1638023555278778, + "learning_rate": 7.249450381473087e-05, + "loss": 0.0086, + "step": 38390 + }, + { + "epoch": 2.9825242718446603, + "grad_norm": 0.38780051469802856, + "learning_rate": 7.247973571166593e-05, + "loss": 0.0095, + "step": 38400 + }, + { + "epoch": 2.9833009708737865, + "grad_norm": 0.22518117725849152, + "learning_rate": 7.246496515025089e-05, + "loss": 0.0071, + "step": 38410 + }, + { + "epoch": 2.984077669902913, + "grad_norm": 0.2725607752799988, + "learning_rate": 7.245019213210101e-05, + "loss": 0.0084, + "step": 38420 + }, + { + "epoch": 2.984854368932039, + "grad_norm": 0.21356084942817688, + "learning_rate": 7.24354166588319e-05, + "loss": 0.0086, + "step": 38430 + }, + { + "epoch": 2.9856310679611653, + "grad_norm": 0.208950474858284, + "learning_rate": 7.242063873205935e-05, + "loss": 0.0074, + "step": 38440 + }, + { + "epoch": 2.9864077669902915, + "grad_norm": 0.27580833435058594, + "learning_rate": 7.240585835339946e-05, + "loss": 0.0085, + "step": 38450 + }, + { + "epoch": 2.9871844660194173, + "grad_norm": 0.280872642993927, + "learning_rate": 7.239107552446857e-05, + "loss": 0.0105, + "step": 38460 + }, + { + "epoch": 2.9879611650485436, + "grad_norm": 0.365183562040329, + "learning_rate": 7.237629024688334e-05, + "loss": 0.0063, + "step": 38470 + }, + { + "epoch": 2.98873786407767, + "grad_norm": 0.18267826735973358, + "learning_rate": 7.236150252226064e-05, + "loss": 0.0102, + "step": 38480 + }, + { + "epoch": 2.989514563106796, + "grad_norm": 0.6867082118988037, + "learning_rate": 7.234671235221765e-05, + "loss": 0.0072, + "step": 38490 + }, + { + "epoch": 2.9902912621359223, + "grad_norm": 0.34802231192588806, + "learning_rate": 7.233191973837179e-05, + "loss": 0.0111, + "step": 38500 + }, + { + "epoch": 2.9910679611650486, + "grad_norm": 0.2821543514728546, + "learning_rate": 7.231712468234078e-05, + "loss": 0.0103, + "step": 38510 + }, + { + "epoch": 2.991844660194175, + "grad_norm": 0.32472917437553406, + "learning_rate": 7.230232718574254e-05, + "loss": 0.0094, + "step": 38520 + }, + { + "epoch": 2.992621359223301, + "grad_norm": 0.22129203379154205, + "learning_rate": 7.228752725019535e-05, + "loss": 0.0077, + "step": 38530 + }, + { + "epoch": 2.9933980582524273, + "grad_norm": 0.2301577627658844, + "learning_rate": 7.227272487731769e-05, + "loss": 0.0084, + "step": 38540 + }, + { + "epoch": 2.994174757281553, + "grad_norm": 0.1890740543603897, + "learning_rate": 7.225792006872831e-05, + "loss": 0.0086, + "step": 38550 + }, + { + "epoch": 2.9949514563106794, + "grad_norm": 0.27472570538520813, + "learning_rate": 7.224311282604628e-05, + "loss": 0.0098, + "step": 38560 + }, + { + "epoch": 2.9957281553398056, + "grad_norm": 0.21607963740825653, + "learning_rate": 7.222830315089085e-05, + "loss": 0.0064, + "step": 38570 + }, + { + "epoch": 2.996504854368932, + "grad_norm": 0.23397736251354218, + "learning_rate": 7.22134910448816e-05, + "loss": 0.0052, + "step": 38580 + }, + { + "epoch": 2.997281553398058, + "grad_norm": 0.4033343195915222, + "learning_rate": 7.219867650963839e-05, + "loss": 0.0096, + "step": 38590 + }, + { + "epoch": 2.9980582524271844, + "grad_norm": 0.4738840162754059, + "learning_rate": 7.21838595467813e-05, + "loss": 0.0081, + "step": 38600 + }, + { + "epoch": 2.9988349514563106, + "grad_norm": 0.263448029756546, + "learning_rate": 7.216904015793067e-05, + "loss": 0.0121, + "step": 38610 + }, + { + "epoch": 2.999611650485437, + "grad_norm": 0.19457164406776428, + "learning_rate": 7.215421834470713e-05, + "loss": 0.0068, + "step": 38620 + }, + { + "epoch": 3.000388349514563, + "grad_norm": 0.21686695516109467, + "learning_rate": 7.213939410873158e-05, + "loss": 0.0113, + "step": 38630 + }, + { + "epoch": 3.0011650485436894, + "grad_norm": 0.2531437873840332, + "learning_rate": 7.212456745162519e-05, + "loss": 0.0096, + "step": 38640 + }, + { + "epoch": 3.0019417475728156, + "grad_norm": 0.8339480757713318, + "learning_rate": 7.210973837500937e-05, + "loss": 0.0071, + "step": 38650 + }, + { + "epoch": 3.002718446601942, + "grad_norm": 0.3004761040210724, + "learning_rate": 7.209490688050578e-05, + "loss": 0.0107, + "step": 38660 + }, + { + "epoch": 3.003495145631068, + "grad_norm": 0.3506331145763397, + "learning_rate": 7.208007296973641e-05, + "loss": 0.0084, + "step": 38670 + }, + { + "epoch": 3.0042718446601944, + "grad_norm": 0.3504910171031952, + "learning_rate": 7.206523664432345e-05, + "loss": 0.0093, + "step": 38680 + }, + { + "epoch": 3.00504854368932, + "grad_norm": 0.20603357255458832, + "learning_rate": 7.205039790588939e-05, + "loss": 0.008, + "step": 38690 + }, + { + "epoch": 3.0058252427184464, + "grad_norm": 0.2104438692331314, + "learning_rate": 7.203555675605697e-05, + "loss": 0.009, + "step": 38700 + }, + { + "epoch": 3.0066019417475727, + "grad_norm": 0.25468289852142334, + "learning_rate": 7.202071319644917e-05, + "loss": 0.0091, + "step": 38710 + }, + { + "epoch": 3.007378640776699, + "grad_norm": 0.19944250583648682, + "learning_rate": 7.200586722868932e-05, + "loss": 0.0061, + "step": 38720 + }, + { + "epoch": 3.008155339805825, + "grad_norm": 0.30318886041641235, + "learning_rate": 7.199101885440088e-05, + "loss": 0.0129, + "step": 38730 + }, + { + "epoch": 3.0089320388349514, + "grad_norm": 0.3942016065120697, + "learning_rate": 7.197616807520771e-05, + "loss": 0.0098, + "step": 38740 + }, + { + "epoch": 3.0097087378640777, + "grad_norm": 0.20832039415836334, + "learning_rate": 7.196131489273381e-05, + "loss": 0.0081, + "step": 38750 + }, + { + "epoch": 3.010485436893204, + "grad_norm": 0.22005048394203186, + "learning_rate": 7.194645930860355e-05, + "loss": 0.0075, + "step": 38760 + }, + { + "epoch": 3.01126213592233, + "grad_norm": 0.37265950441360474, + "learning_rate": 7.19316013244415e-05, + "loss": 0.0087, + "step": 38770 + }, + { + "epoch": 3.0120388349514564, + "grad_norm": 0.24519719183444977, + "learning_rate": 7.191674094187248e-05, + "loss": 0.01, + "step": 38780 + }, + { + "epoch": 3.0128155339805827, + "grad_norm": 0.317708820104599, + "learning_rate": 7.190187816252165e-05, + "loss": 0.0079, + "step": 38790 + }, + { + "epoch": 3.013592233009709, + "grad_norm": 0.22497475147247314, + "learning_rate": 7.188701298801435e-05, + "loss": 0.0088, + "step": 38800 + }, + { + "epoch": 3.014368932038835, + "grad_norm": 0.3406515121459961, + "learning_rate": 7.18721454199762e-05, + "loss": 0.01, + "step": 38810 + }, + { + "epoch": 3.015145631067961, + "grad_norm": 0.268740713596344, + "learning_rate": 7.185727546003312e-05, + "loss": 0.0078, + "step": 38820 + }, + { + "epoch": 3.015922330097087, + "grad_norm": 0.14603644609451294, + "learning_rate": 7.184240310981126e-05, + "loss": 0.0059, + "step": 38830 + }, + { + "epoch": 3.0166990291262135, + "grad_norm": 0.23345835506916046, + "learning_rate": 7.182752837093704e-05, + "loss": 0.0072, + "step": 38840 + }, + { + "epoch": 3.0174757281553397, + "grad_norm": 0.2528505027294159, + "learning_rate": 7.181265124503711e-05, + "loss": 0.0078, + "step": 38850 + }, + { + "epoch": 3.018252427184466, + "grad_norm": 0.33648067712783813, + "learning_rate": 7.179777173373847e-05, + "loss": 0.0071, + "step": 38860 + }, + { + "epoch": 3.019029126213592, + "grad_norm": 0.2816639542579651, + "learning_rate": 7.178288983866826e-05, + "loss": 0.0119, + "step": 38870 + }, + { + "epoch": 3.0198058252427185, + "grad_norm": 0.17236308753490448, + "learning_rate": 7.176800556145397e-05, + "loss": 0.0076, + "step": 38880 + }, + { + "epoch": 3.0205825242718447, + "grad_norm": 0.1273905485868454, + "learning_rate": 7.175311890372334e-05, + "loss": 0.0053, + "step": 38890 + }, + { + "epoch": 3.021359223300971, + "grad_norm": 0.1045820340514183, + "learning_rate": 7.17382298671043e-05, + "loss": 0.0084, + "step": 38900 + }, + { + "epoch": 3.022135922330097, + "grad_norm": 0.6512429714202881, + "learning_rate": 7.172333845322515e-05, + "loss": 0.007, + "step": 38910 + }, + { + "epoch": 3.0229126213592235, + "grad_norm": 0.16965734958648682, + "learning_rate": 7.170844466371436e-05, + "loss": 0.0078, + "step": 38920 + }, + { + "epoch": 3.0236893203883497, + "grad_norm": 0.3258030116558075, + "learning_rate": 7.16935485002007e-05, + "loss": 0.0055, + "step": 38930 + }, + { + "epoch": 3.024466019417476, + "grad_norm": 0.30663278698921204, + "learning_rate": 7.167864996431319e-05, + "loss": 0.006, + "step": 38940 + }, + { + "epoch": 3.0252427184466018, + "grad_norm": 0.28523141145706177, + "learning_rate": 7.166374905768111e-05, + "loss": 0.0079, + "step": 38950 + }, + { + "epoch": 3.026019417475728, + "grad_norm": 0.28314733505249023, + "learning_rate": 7.164884578193401e-05, + "loss": 0.007, + "step": 38960 + }, + { + "epoch": 3.0267961165048543, + "grad_norm": 0.4283628463745117, + "learning_rate": 7.163394013870168e-05, + "loss": 0.0065, + "step": 38970 + }, + { + "epoch": 3.0275728155339805, + "grad_norm": 0.4952712953090668, + "learning_rate": 7.161903212961419e-05, + "loss": 0.0087, + "step": 38980 + }, + { + "epoch": 3.0283495145631067, + "grad_norm": 0.2722402513027191, + "learning_rate": 7.160412175630183e-05, + "loss": 0.0074, + "step": 38990 + }, + { + "epoch": 3.029126213592233, + "grad_norm": 0.37591156363487244, + "learning_rate": 7.158920902039521e-05, + "loss": 0.0093, + "step": 39000 + }, + { + "epoch": 3.0299029126213592, + "grad_norm": 0.31060266494750977, + "learning_rate": 7.157429392352514e-05, + "loss": 0.0106, + "step": 39010 + }, + { + "epoch": 3.0306796116504855, + "grad_norm": 0.29430073499679565, + "learning_rate": 7.155937646732275e-05, + "loss": 0.0094, + "step": 39020 + }, + { + "epoch": 3.0314563106796117, + "grad_norm": 0.3394031226634979, + "learning_rate": 7.154445665341933e-05, + "loss": 0.0091, + "step": 39030 + }, + { + "epoch": 3.032233009708738, + "grad_norm": 0.5887660384178162, + "learning_rate": 7.152953448344654e-05, + "loss": 0.009, + "step": 39040 + }, + { + "epoch": 3.0330097087378642, + "grad_norm": 0.17035289108753204, + "learning_rate": 7.151460995903624e-05, + "loss": 0.0089, + "step": 39050 + }, + { + "epoch": 3.0337864077669905, + "grad_norm": 0.2095930427312851, + "learning_rate": 7.149968308182052e-05, + "loss": 0.0064, + "step": 39060 + }, + { + "epoch": 3.0345631067961163, + "grad_norm": 0.1656578779220581, + "learning_rate": 7.14847538534318e-05, + "loss": 0.0059, + "step": 39070 + }, + { + "epoch": 3.0353398058252425, + "grad_norm": 0.28254273533821106, + "learning_rate": 7.14698222755027e-05, + "loss": 0.0078, + "step": 39080 + }, + { + "epoch": 3.036116504854369, + "grad_norm": 0.44596067070961, + "learning_rate": 7.14548883496661e-05, + "loss": 0.0057, + "step": 39090 + }, + { + "epoch": 3.036893203883495, + "grad_norm": 0.33227869868278503, + "learning_rate": 7.143995207755517e-05, + "loss": 0.0054, + "step": 39100 + }, + { + "epoch": 3.0376699029126213, + "grad_norm": 0.27901774644851685, + "learning_rate": 7.142501346080333e-05, + "loss": 0.0105, + "step": 39110 + }, + { + "epoch": 3.0384466019417475, + "grad_norm": 0.7113941311836243, + "learning_rate": 7.141007250104421e-05, + "loss": 0.007, + "step": 39120 + }, + { + "epoch": 3.039223300970874, + "grad_norm": 0.3065520226955414, + "learning_rate": 7.139512919991176e-05, + "loss": 0.0084, + "step": 39130 + }, + { + "epoch": 3.04, + "grad_norm": 0.24800367653369904, + "learning_rate": 7.138018355904014e-05, + "loss": 0.0092, + "step": 39140 + }, + { + "epoch": 3.0407766990291263, + "grad_norm": 0.33616572618484497, + "learning_rate": 7.13652355800638e-05, + "loss": 0.0094, + "step": 39150 + }, + { + "epoch": 3.0415533980582525, + "grad_norm": 0.29676002264022827, + "learning_rate": 7.135028526461743e-05, + "loss": 0.009, + "step": 39160 + }, + { + "epoch": 3.042330097087379, + "grad_norm": 0.3506304621696472, + "learning_rate": 7.133533261433594e-05, + "loss": 0.0097, + "step": 39170 + }, + { + "epoch": 3.043106796116505, + "grad_norm": 0.2355334311723709, + "learning_rate": 7.132037763085457e-05, + "loss": 0.0065, + "step": 39180 + }, + { + "epoch": 3.0438834951456313, + "grad_norm": 0.3022172451019287, + "learning_rate": 7.130542031580875e-05, + "loss": 0.0096, + "step": 39190 + }, + { + "epoch": 3.044660194174757, + "grad_norm": 0.4610370397567749, + "learning_rate": 7.12904606708342e-05, + "loss": 0.0098, + "step": 39200 + }, + { + "epoch": 3.0454368932038833, + "grad_norm": 0.34072762727737427, + "learning_rate": 7.127549869756687e-05, + "loss": 0.0078, + "step": 39210 + }, + { + "epoch": 3.0462135922330096, + "grad_norm": 0.16110290586948395, + "learning_rate": 7.126053439764299e-05, + "loss": 0.0092, + "step": 39220 + }, + { + "epoch": 3.046990291262136, + "grad_norm": 0.1925916224718094, + "learning_rate": 7.124556777269904e-05, + "loss": 0.0076, + "step": 39230 + }, + { + "epoch": 3.047766990291262, + "grad_norm": 0.15666994452476501, + "learning_rate": 7.123059882437174e-05, + "loss": 0.0088, + "step": 39240 + }, + { + "epoch": 3.0485436893203883, + "grad_norm": 0.21245168149471283, + "learning_rate": 7.121562755429807e-05, + "loss": 0.012, + "step": 39250 + }, + { + "epoch": 3.0493203883495146, + "grad_norm": 0.2114252895116806, + "learning_rate": 7.120065396411528e-05, + "loss": 0.0062, + "step": 39260 + }, + { + "epoch": 3.050097087378641, + "grad_norm": 0.16497382521629333, + "learning_rate": 7.118567805546084e-05, + "loss": 0.0085, + "step": 39270 + }, + { + "epoch": 3.050873786407767, + "grad_norm": 0.23402543365955353, + "learning_rate": 7.117069982997248e-05, + "loss": 0.0059, + "step": 39280 + }, + { + "epoch": 3.0516504854368933, + "grad_norm": 0.3192528188228607, + "learning_rate": 7.115571928928825e-05, + "loss": 0.0077, + "step": 39290 + }, + { + "epoch": 3.0524271844660196, + "grad_norm": 0.26549282670021057, + "learning_rate": 7.114073643504635e-05, + "loss": 0.0065, + "step": 39300 + }, + { + "epoch": 3.053203883495146, + "grad_norm": 0.2683822810649872, + "learning_rate": 7.11257512688853e-05, + "loss": 0.0069, + "step": 39310 + }, + { + "epoch": 3.0539805825242716, + "grad_norm": 0.26711708307266235, + "learning_rate": 7.111076379244384e-05, + "loss": 0.0112, + "step": 39320 + }, + { + "epoch": 3.054757281553398, + "grad_norm": 0.14662262797355652, + "learning_rate": 7.109577400736101e-05, + "loss": 0.0068, + "step": 39330 + }, + { + "epoch": 3.055533980582524, + "grad_norm": 0.3465411365032196, + "learning_rate": 7.108078191527605e-05, + "loss": 0.0124, + "step": 39340 + }, + { + "epoch": 3.0563106796116504, + "grad_norm": 0.32638588547706604, + "learning_rate": 7.106578751782847e-05, + "loss": 0.0067, + "step": 39350 + }, + { + "epoch": 3.0570873786407766, + "grad_norm": 0.41179007291793823, + "learning_rate": 7.105079081665803e-05, + "loss": 0.0083, + "step": 39360 + }, + { + "epoch": 3.057864077669903, + "grad_norm": 0.38381847739219666, + "learning_rate": 7.103579181340476e-05, + "loss": 0.0065, + "step": 39370 + }, + { + "epoch": 3.058640776699029, + "grad_norm": 0.30665045976638794, + "learning_rate": 7.102079050970893e-05, + "loss": 0.0086, + "step": 39380 + }, + { + "epoch": 3.0594174757281554, + "grad_norm": 0.2589832842350006, + "learning_rate": 7.100578690721105e-05, + "loss": 0.007, + "step": 39390 + }, + { + "epoch": 3.0601941747572816, + "grad_norm": 0.2007114589214325, + "learning_rate": 7.09907810075519e-05, + "loss": 0.0073, + "step": 39400 + }, + { + "epoch": 3.060970873786408, + "grad_norm": 0.17748790979385376, + "learning_rate": 7.097577281237249e-05, + "loss": 0.0063, + "step": 39410 + }, + { + "epoch": 3.061747572815534, + "grad_norm": 0.19210661947727203, + "learning_rate": 7.09607623233141e-05, + "loss": 0.009, + "step": 39420 + }, + { + "epoch": 3.0625242718446604, + "grad_norm": 0.2504993677139282, + "learning_rate": 7.094574954201827e-05, + "loss": 0.0086, + "step": 39430 + }, + { + "epoch": 3.0633009708737866, + "grad_norm": 0.3217412531375885, + "learning_rate": 7.093073447012675e-05, + "loss": 0.0072, + "step": 39440 + }, + { + "epoch": 3.0640776699029124, + "grad_norm": 0.14515842497348785, + "learning_rate": 7.09157171092816e-05, + "loss": 0.0069, + "step": 39450 + }, + { + "epoch": 3.0648543689320387, + "grad_norm": 0.12437552958726883, + "learning_rate": 7.090069746112504e-05, + "loss": 0.0068, + "step": 39460 + }, + { + "epoch": 3.065631067961165, + "grad_norm": 0.3075614869594574, + "learning_rate": 7.088567552729965e-05, + "loss": 0.0081, + "step": 39470 + }, + { + "epoch": 3.066407766990291, + "grad_norm": 0.1521320939064026, + "learning_rate": 7.087065130944818e-05, + "loss": 0.0073, + "step": 39480 + }, + { + "epoch": 3.0671844660194174, + "grad_norm": 0.3200215995311737, + "learning_rate": 7.085562480921366e-05, + "loss": 0.0095, + "step": 39490 + }, + { + "epoch": 3.0679611650485437, + "grad_norm": 0.16031189262866974, + "learning_rate": 7.084059602823937e-05, + "loss": 0.0095, + "step": 39500 + }, + { + "epoch": 3.06873786407767, + "grad_norm": 0.19616329669952393, + "learning_rate": 7.082556496816882e-05, + "loss": 0.0082, + "step": 39510 + }, + { + "epoch": 3.069514563106796, + "grad_norm": 0.12334226816892624, + "learning_rate": 7.081053163064582e-05, + "loss": 0.0103, + "step": 39520 + }, + { + "epoch": 3.0702912621359224, + "grad_norm": 0.26998820900917053, + "learning_rate": 7.079549601731434e-05, + "loss": 0.0102, + "step": 39530 + }, + { + "epoch": 3.0710679611650487, + "grad_norm": 0.2649594247341156, + "learning_rate": 7.07804581298187e-05, + "loss": 0.0088, + "step": 39540 + }, + { + "epoch": 3.071844660194175, + "grad_norm": 0.22864332795143127, + "learning_rate": 7.07654179698034e-05, + "loss": 0.0055, + "step": 39550 + }, + { + "epoch": 3.072621359223301, + "grad_norm": 0.1931753307580948, + "learning_rate": 7.075037553891321e-05, + "loss": 0.0072, + "step": 39560 + }, + { + "epoch": 3.073398058252427, + "grad_norm": 0.32321473956108093, + "learning_rate": 7.073533083879315e-05, + "loss": 0.0099, + "step": 39570 + }, + { + "epoch": 3.074174757281553, + "grad_norm": 0.14240187406539917, + "learning_rate": 7.072028387108849e-05, + "loss": 0.0084, + "step": 39580 + }, + { + "epoch": 3.0749514563106795, + "grad_norm": 0.32844772934913635, + "learning_rate": 7.070523463744472e-05, + "loss": 0.0078, + "step": 39590 + }, + { + "epoch": 3.0757281553398057, + "grad_norm": 0.19924066960811615, + "learning_rate": 7.069018313950763e-05, + "loss": 0.0087, + "step": 39600 + }, + { + "epoch": 3.076504854368932, + "grad_norm": 0.1480225771665573, + "learning_rate": 7.067512937892322e-05, + "loss": 0.0082, + "step": 39610 + }, + { + "epoch": 3.077281553398058, + "grad_norm": 0.10295341163873672, + "learning_rate": 7.066007335733775e-05, + "loss": 0.0086, + "step": 39620 + }, + { + "epoch": 3.0780582524271844, + "grad_norm": 0.5873264670372009, + "learning_rate": 7.064501507639772e-05, + "loss": 0.008, + "step": 39630 + }, + { + "epoch": 3.0788349514563107, + "grad_norm": 0.1929519921541214, + "learning_rate": 7.062995453774987e-05, + "loss": 0.0079, + "step": 39640 + }, + { + "epoch": 3.079611650485437, + "grad_norm": 0.20422811806201935, + "learning_rate": 7.061489174304121e-05, + "loss": 0.0077, + "step": 39650 + }, + { + "epoch": 3.080388349514563, + "grad_norm": 0.3015933632850647, + "learning_rate": 7.0599826693919e-05, + "loss": 0.0107, + "step": 39660 + }, + { + "epoch": 3.0811650485436894, + "grad_norm": 0.6217702031135559, + "learning_rate": 7.05847593920307e-05, + "loss": 0.0086, + "step": 39670 + }, + { + "epoch": 3.0819417475728157, + "grad_norm": 0.21106861531734467, + "learning_rate": 7.056968983902406e-05, + "loss": 0.0093, + "step": 39680 + }, + { + "epoch": 3.082718446601942, + "grad_norm": 0.2677818834781647, + "learning_rate": 7.055461803654706e-05, + "loss": 0.0075, + "step": 39690 + }, + { + "epoch": 3.0834951456310677, + "grad_norm": 0.18334774672985077, + "learning_rate": 7.053954398624794e-05, + "loss": 0.006, + "step": 39700 + }, + { + "epoch": 3.084271844660194, + "grad_norm": 0.22553737461566925, + "learning_rate": 7.052446768977518e-05, + "loss": 0.0071, + "step": 39710 + }, + { + "epoch": 3.0850485436893202, + "grad_norm": 0.2977901101112366, + "learning_rate": 7.050938914877748e-05, + "loss": 0.0067, + "step": 39720 + }, + { + "epoch": 3.0858252427184465, + "grad_norm": 0.37795379757881165, + "learning_rate": 7.049430836490382e-05, + "loss": 0.0099, + "step": 39730 + }, + { + "epoch": 3.0866019417475727, + "grad_norm": 0.2021433264017105, + "learning_rate": 7.047922533980342e-05, + "loss": 0.0071, + "step": 39740 + }, + { + "epoch": 3.087378640776699, + "grad_norm": 0.4557875096797943, + "learning_rate": 7.046414007512571e-05, + "loss": 0.0078, + "step": 39750 + }, + { + "epoch": 3.0881553398058252, + "grad_norm": 0.21479739248752594, + "learning_rate": 7.044905257252042e-05, + "loss": 0.0075, + "step": 39760 + }, + { + "epoch": 3.0889320388349515, + "grad_norm": 0.38110724091529846, + "learning_rate": 7.04339628336375e-05, + "loss": 0.008, + "step": 39770 + }, + { + "epoch": 3.0897087378640777, + "grad_norm": 0.5363356471061707, + "learning_rate": 7.041887086012711e-05, + "loss": 0.0094, + "step": 39780 + }, + { + "epoch": 3.090485436893204, + "grad_norm": 0.6763386130332947, + "learning_rate": 7.040377665363969e-05, + "loss": 0.0081, + "step": 39790 + }, + { + "epoch": 3.0912621359223302, + "grad_norm": 0.20996813476085663, + "learning_rate": 7.038868021582594e-05, + "loss": 0.0074, + "step": 39800 + }, + { + "epoch": 3.0920388349514565, + "grad_norm": 0.14663177728652954, + "learning_rate": 7.037358154833679e-05, + "loss": 0.0096, + "step": 39810 + }, + { + "epoch": 3.0928155339805827, + "grad_norm": 0.3250434696674347, + "learning_rate": 7.035848065282339e-05, + "loss": 0.0092, + "step": 39820 + }, + { + "epoch": 3.0935922330097085, + "grad_norm": 0.17580896615982056, + "learning_rate": 7.034337753093714e-05, + "loss": 0.0062, + "step": 39830 + }, + { + "epoch": 3.094368932038835, + "grad_norm": 0.5340560674667358, + "learning_rate": 7.032827218432972e-05, + "loss": 0.0094, + "step": 39840 + }, + { + "epoch": 3.095145631067961, + "grad_norm": 0.3000243008136749, + "learning_rate": 7.031316461465302e-05, + "loss": 0.0089, + "step": 39850 + }, + { + "epoch": 3.0959223300970873, + "grad_norm": 0.24200180172920227, + "learning_rate": 7.029805482355915e-05, + "loss": 0.0095, + "step": 39860 + }, + { + "epoch": 3.0966990291262135, + "grad_norm": 0.2009091079235077, + "learning_rate": 7.028294281270055e-05, + "loss": 0.0092, + "step": 39870 + }, + { + "epoch": 3.09747572815534, + "grad_norm": 0.2311142086982727, + "learning_rate": 7.026782858372982e-05, + "loss": 0.0075, + "step": 39880 + }, + { + "epoch": 3.098252427184466, + "grad_norm": 0.38170063495635986, + "learning_rate": 7.025271213829982e-05, + "loss": 0.0082, + "step": 39890 + }, + { + "epoch": 3.0990291262135923, + "grad_norm": 0.1775358021259308, + "learning_rate": 7.023759347806366e-05, + "loss": 0.0065, + "step": 39900 + }, + { + "epoch": 3.0998058252427185, + "grad_norm": 0.1168278157711029, + "learning_rate": 7.022247260467472e-05, + "loss": 0.0108, + "step": 39910 + }, + { + "epoch": 3.1005825242718448, + "grad_norm": 0.34953945875167847, + "learning_rate": 7.020734951978658e-05, + "loss": 0.0083, + "step": 39920 + }, + { + "epoch": 3.101359223300971, + "grad_norm": 0.3004077672958374, + "learning_rate": 7.019222422505307e-05, + "loss": 0.0066, + "step": 39930 + }, + { + "epoch": 3.1021359223300973, + "grad_norm": 0.43989819288253784, + "learning_rate": 7.017709672212827e-05, + "loss": 0.0066, + "step": 39940 + }, + { + "epoch": 3.1029126213592235, + "grad_norm": 0.23884212970733643, + "learning_rate": 7.016196701266652e-05, + "loss": 0.0062, + "step": 39950 + }, + { + "epoch": 3.1036893203883493, + "grad_norm": 0.27007928490638733, + "learning_rate": 7.014683509832238e-05, + "loss": 0.0079, + "step": 39960 + }, + { + "epoch": 3.1044660194174756, + "grad_norm": 0.22473730146884918, + "learning_rate": 7.013170098075063e-05, + "loss": 0.0062, + "step": 39970 + }, + { + "epoch": 3.105242718446602, + "grad_norm": 0.1474427729845047, + "learning_rate": 7.011656466160632e-05, + "loss": 0.0075, + "step": 39980 + }, + { + "epoch": 3.106019417475728, + "grad_norm": 0.1848635971546173, + "learning_rate": 7.010142614254475e-05, + "loss": 0.0082, + "step": 39990 + }, + { + "epoch": 3.1067961165048543, + "grad_norm": 0.4970974624156952, + "learning_rate": 7.008628542522147e-05, + "loss": 0.0091, + "step": 40000 + }, + { + "epoch": 3.1075728155339806, + "grad_norm": 0.2876264154911041, + "learning_rate": 7.007114251129218e-05, + "loss": 0.0074, + "step": 40010 + }, + { + "epoch": 3.108349514563107, + "grad_norm": 0.10632750391960144, + "learning_rate": 7.005599740241293e-05, + "loss": 0.0048, + "step": 40020 + }, + { + "epoch": 3.109126213592233, + "grad_norm": 0.1388462334871292, + "learning_rate": 7.004085010023996e-05, + "loss": 0.0094, + "step": 40030 + }, + { + "epoch": 3.1099029126213593, + "grad_norm": 0.20931553840637207, + "learning_rate": 7.002570060642976e-05, + "loss": 0.0072, + "step": 40040 + }, + { + "epoch": 3.1106796116504856, + "grad_norm": 0.2585000693798065, + "learning_rate": 7.001054892263903e-05, + "loss": 0.0076, + "step": 40050 + }, + { + "epoch": 3.111456310679612, + "grad_norm": 0.17393557727336884, + "learning_rate": 6.999539505052477e-05, + "loss": 0.0072, + "step": 40060 + }, + { + "epoch": 3.112233009708738, + "grad_norm": 0.1756247878074646, + "learning_rate": 6.998023899174418e-05, + "loss": 0.0069, + "step": 40070 + }, + { + "epoch": 3.113009708737864, + "grad_norm": 0.2132689356803894, + "learning_rate": 6.996508074795467e-05, + "loss": 0.0078, + "step": 40080 + }, + { + "epoch": 3.11378640776699, + "grad_norm": 0.19013367593288422, + "learning_rate": 6.994992032081396e-05, + "loss": 0.0068, + "step": 40090 + }, + { + "epoch": 3.1145631067961164, + "grad_norm": 0.5381582379341125, + "learning_rate": 6.993475771197995e-05, + "loss": 0.0051, + "step": 40100 + }, + { + "epoch": 3.1153398058252426, + "grad_norm": 0.09199031442403793, + "learning_rate": 6.991959292311082e-05, + "loss": 0.0092, + "step": 40110 + }, + { + "epoch": 3.116116504854369, + "grad_norm": 0.2463151216506958, + "learning_rate": 6.990442595586495e-05, + "loss": 0.0079, + "step": 40120 + }, + { + "epoch": 3.116893203883495, + "grad_norm": 0.2219184935092926, + "learning_rate": 6.988925681190098e-05, + "loss": 0.0065, + "step": 40130 + }, + { + "epoch": 3.1176699029126214, + "grad_norm": 0.2619200348854065, + "learning_rate": 6.987408549287778e-05, + "loss": 0.006, + "step": 40140 + }, + { + "epoch": 3.1184466019417476, + "grad_norm": 0.14552025496959686, + "learning_rate": 6.985891200045449e-05, + "loss": 0.0063, + "step": 40150 + }, + { + "epoch": 3.119223300970874, + "grad_norm": 0.18397872149944305, + "learning_rate": 6.984373633629045e-05, + "loss": 0.0075, + "step": 40160 + }, + { + "epoch": 3.12, + "grad_norm": 0.20857028663158417, + "learning_rate": 6.98285585020452e-05, + "loss": 0.0091, + "step": 40170 + }, + { + "epoch": 3.1207766990291264, + "grad_norm": 0.18595150113105774, + "learning_rate": 6.981337849937864e-05, + "loss": 0.0067, + "step": 40180 + }, + { + "epoch": 3.1215533980582526, + "grad_norm": 0.2275722175836563, + "learning_rate": 6.979819632995078e-05, + "loss": 0.0063, + "step": 40190 + }, + { + "epoch": 3.122330097087379, + "grad_norm": 0.23838947713375092, + "learning_rate": 6.978301199542193e-05, + "loss": 0.0063, + "step": 40200 + }, + { + "epoch": 3.1231067961165047, + "grad_norm": 0.14665010571479797, + "learning_rate": 6.976782549745263e-05, + "loss": 0.0077, + "step": 40210 + }, + { + "epoch": 3.123883495145631, + "grad_norm": 0.15860654413700104, + "learning_rate": 6.975263683770367e-05, + "loss": 0.0067, + "step": 40220 + }, + { + "epoch": 3.124660194174757, + "grad_norm": 0.26382994651794434, + "learning_rate": 6.973744601783604e-05, + "loss": 0.0054, + "step": 40230 + }, + { + "epoch": 3.1254368932038834, + "grad_norm": 0.11625919491052628, + "learning_rate": 6.972225303951097e-05, + "loss": 0.0086, + "step": 40240 + }, + { + "epoch": 3.1262135922330097, + "grad_norm": 0.1973661482334137, + "learning_rate": 6.970705790438998e-05, + "loss": 0.0092, + "step": 40250 + }, + { + "epoch": 3.126990291262136, + "grad_norm": 0.13930422067642212, + "learning_rate": 6.969186061413477e-05, + "loss": 0.0088, + "step": 40260 + }, + { + "epoch": 3.127766990291262, + "grad_norm": 0.2961063086986542, + "learning_rate": 6.967666117040727e-05, + "loss": 0.0076, + "step": 40270 + }, + { + "epoch": 3.1285436893203884, + "grad_norm": 0.3274814784526825, + "learning_rate": 6.966145957486972e-05, + "loss": 0.008, + "step": 40280 + }, + { + "epoch": 3.1293203883495146, + "grad_norm": 0.7160571217536926, + "learning_rate": 6.964625582918449e-05, + "loss": 0.0098, + "step": 40290 + }, + { + "epoch": 3.130097087378641, + "grad_norm": 0.17892171442508698, + "learning_rate": 6.963104993501425e-05, + "loss": 0.0093, + "step": 40300 + }, + { + "epoch": 3.130873786407767, + "grad_norm": 0.29603150486946106, + "learning_rate": 6.961584189402192e-05, + "loss": 0.0133, + "step": 40310 + }, + { + "epoch": 3.1316504854368934, + "grad_norm": 0.3290659487247467, + "learning_rate": 6.96006317078706e-05, + "loss": 0.0099, + "step": 40320 + }, + { + "epoch": 3.132427184466019, + "grad_norm": 0.23934414982795715, + "learning_rate": 6.958541937822367e-05, + "loss": 0.0088, + "step": 40330 + }, + { + "epoch": 3.1332038834951454, + "grad_norm": 0.28126490116119385, + "learning_rate": 6.957020490674473e-05, + "loss": 0.0103, + "step": 40340 + }, + { + "epoch": 3.1339805825242717, + "grad_norm": 0.2238032966852188, + "learning_rate": 6.95549882950976e-05, + "loss": 0.012, + "step": 40350 + }, + { + "epoch": 3.134757281553398, + "grad_norm": 0.2685145139694214, + "learning_rate": 6.953976954494635e-05, + "loss": 0.0067, + "step": 40360 + }, + { + "epoch": 3.135533980582524, + "grad_norm": 0.29116466641426086, + "learning_rate": 6.952454865795528e-05, + "loss": 0.0085, + "step": 40370 + }, + { + "epoch": 3.1363106796116504, + "grad_norm": 0.9232564568519592, + "learning_rate": 6.950932563578892e-05, + "loss": 0.0093, + "step": 40380 + }, + { + "epoch": 3.1370873786407767, + "grad_norm": 0.23339492082595825, + "learning_rate": 6.949410048011206e-05, + "loss": 0.0066, + "step": 40390 + }, + { + "epoch": 3.137864077669903, + "grad_norm": 0.6069256663322449, + "learning_rate": 6.947887319258966e-05, + "loss": 0.0058, + "step": 40400 + }, + { + "epoch": 3.138640776699029, + "grad_norm": 0.2073862999677658, + "learning_rate": 6.946364377488696e-05, + "loss": 0.0073, + "step": 40410 + }, + { + "epoch": 3.1394174757281554, + "grad_norm": 0.32664164900779724, + "learning_rate": 6.944841222866947e-05, + "loss": 0.0078, + "step": 40420 + }, + { + "epoch": 3.1401941747572817, + "grad_norm": 0.2754111588001251, + "learning_rate": 6.943317855560284e-05, + "loss": 0.0074, + "step": 40430 + }, + { + "epoch": 3.140970873786408, + "grad_norm": 0.31321412324905396, + "learning_rate": 6.941794275735302e-05, + "loss": 0.0086, + "step": 40440 + }, + { + "epoch": 3.141747572815534, + "grad_norm": 0.2685461938381195, + "learning_rate": 6.94027048355862e-05, + "loss": 0.0086, + "step": 40450 + }, + { + "epoch": 3.14252427184466, + "grad_norm": 0.8054752349853516, + "learning_rate": 6.938746479196871e-05, + "loss": 0.0062, + "step": 40460 + }, + { + "epoch": 3.1433009708737862, + "grad_norm": 0.3452872037887573, + "learning_rate": 6.937222262816724e-05, + "loss": 0.0094, + "step": 40470 + }, + { + "epoch": 3.1440776699029125, + "grad_norm": 0.22958724200725555, + "learning_rate": 6.935697834584865e-05, + "loss": 0.0087, + "step": 40480 + }, + { + "epoch": 3.1448543689320387, + "grad_norm": 0.31729117035865784, + "learning_rate": 6.934173194667998e-05, + "loss": 0.0077, + "step": 40490 + }, + { + "epoch": 3.145631067961165, + "grad_norm": 0.26045387983322144, + "learning_rate": 6.93264834323286e-05, + "loss": 0.0065, + "step": 40500 + }, + { + "epoch": 3.1464077669902912, + "grad_norm": 0.1466287225484848, + "learning_rate": 6.931123280446206e-05, + "loss": 0.0089, + "step": 40510 + }, + { + "epoch": 3.1471844660194175, + "grad_norm": 0.3011729419231415, + "learning_rate": 6.929598006474811e-05, + "loss": 0.0089, + "step": 40520 + }, + { + "epoch": 3.1479611650485437, + "grad_norm": 0.19525086879730225, + "learning_rate": 6.92807252148548e-05, + "loss": 0.0093, + "step": 40530 + }, + { + "epoch": 3.14873786407767, + "grad_norm": 0.6350200772285461, + "learning_rate": 6.926546825645037e-05, + "loss": 0.0078, + "step": 40540 + }, + { + "epoch": 3.1495145631067962, + "grad_norm": 0.41706040501594543, + "learning_rate": 6.92502091912033e-05, + "loss": 0.0099, + "step": 40550 + }, + { + "epoch": 3.1502912621359225, + "grad_norm": 0.15127654373645782, + "learning_rate": 6.923494802078229e-05, + "loss": 0.0069, + "step": 40560 + }, + { + "epoch": 3.1510679611650487, + "grad_norm": 0.2679073214530945, + "learning_rate": 6.921968474685628e-05, + "loss": 0.0099, + "step": 40570 + }, + { + "epoch": 3.1518446601941745, + "grad_norm": 0.2200445979833603, + "learning_rate": 6.920441937109445e-05, + "loss": 0.0074, + "step": 40580 + }, + { + "epoch": 3.152621359223301, + "grad_norm": 0.28783419728279114, + "learning_rate": 6.91891518951662e-05, + "loss": 0.009, + "step": 40590 + }, + { + "epoch": 3.153398058252427, + "grad_norm": 0.22526951134204865, + "learning_rate": 6.917388232074114e-05, + "loss": 0.0098, + "step": 40600 + }, + { + "epoch": 3.1541747572815533, + "grad_norm": 0.2920188009738922, + "learning_rate": 6.915861064948914e-05, + "loss": 0.0085, + "step": 40610 + }, + { + "epoch": 3.1549514563106795, + "grad_norm": 0.3309069275856018, + "learning_rate": 6.914333688308025e-05, + "loss": 0.0095, + "step": 40620 + }, + { + "epoch": 3.1557281553398058, + "grad_norm": 0.25145086646080017, + "learning_rate": 6.912806102318487e-05, + "loss": 0.0082, + "step": 40630 + }, + { + "epoch": 3.156504854368932, + "grad_norm": 0.2974226772785187, + "learning_rate": 6.911278307147347e-05, + "loss": 0.0084, + "step": 40640 + }, + { + "epoch": 3.1572815533980583, + "grad_norm": 0.324217826128006, + "learning_rate": 6.909750302961684e-05, + "loss": 0.0082, + "step": 40650 + }, + { + "epoch": 3.1580582524271845, + "grad_norm": 0.18418169021606445, + "learning_rate": 6.9082220899286e-05, + "loss": 0.0055, + "step": 40660 + }, + { + "epoch": 3.1588349514563108, + "grad_norm": 0.14966464042663574, + "learning_rate": 6.906693668215219e-05, + "loss": 0.0072, + "step": 40670 + }, + { + "epoch": 3.159611650485437, + "grad_norm": 0.20717068016529083, + "learning_rate": 6.905165037988683e-05, + "loss": 0.0082, + "step": 40680 + }, + { + "epoch": 3.1603883495145633, + "grad_norm": 0.32758116722106934, + "learning_rate": 6.903636199416164e-05, + "loss": 0.0081, + "step": 40690 + }, + { + "epoch": 3.1611650485436895, + "grad_norm": 0.16024576127529144, + "learning_rate": 6.902107152664851e-05, + "loss": 0.0076, + "step": 40700 + }, + { + "epoch": 3.1619417475728158, + "grad_norm": 0.15168437361717224, + "learning_rate": 6.90057789790196e-05, + "loss": 0.0059, + "step": 40710 + }, + { + "epoch": 3.1627184466019416, + "grad_norm": 0.25055262446403503, + "learning_rate": 6.899048435294728e-05, + "loss": 0.0106, + "step": 40720 + }, + { + "epoch": 3.163495145631068, + "grad_norm": 0.7802700996398926, + "learning_rate": 6.897518765010415e-05, + "loss": 0.0091, + "step": 40730 + }, + { + "epoch": 3.164271844660194, + "grad_norm": 0.35393497347831726, + "learning_rate": 6.895988887216302e-05, + "loss": 0.0086, + "step": 40740 + }, + { + "epoch": 3.1650485436893203, + "grad_norm": 0.3289221227169037, + "learning_rate": 6.894458802079694e-05, + "loss": 0.0107, + "step": 40750 + }, + { + "epoch": 3.1658252427184466, + "grad_norm": 0.23015441000461578, + "learning_rate": 6.892928509767922e-05, + "loss": 0.01, + "step": 40760 + }, + { + "epoch": 3.166601941747573, + "grad_norm": 0.318489134311676, + "learning_rate": 6.891398010448333e-05, + "loss": 0.0101, + "step": 40770 + }, + { + "epoch": 3.167378640776699, + "grad_norm": 0.2058803290128708, + "learning_rate": 6.8898673042883e-05, + "loss": 0.0095, + "step": 40780 + }, + { + "epoch": 3.1681553398058253, + "grad_norm": 0.2623686194419861, + "learning_rate": 6.888336391455222e-05, + "loss": 0.0087, + "step": 40790 + }, + { + "epoch": 3.1689320388349516, + "grad_norm": 0.26347780227661133, + "learning_rate": 6.886805272116513e-05, + "loss": 0.0079, + "step": 40800 + }, + { + "epoch": 3.169708737864078, + "grad_norm": 0.21082863211631775, + "learning_rate": 6.885273946439617e-05, + "loss": 0.0075, + "step": 40810 + }, + { + "epoch": 3.170485436893204, + "grad_norm": 0.2654772400856018, + "learning_rate": 6.883742414591998e-05, + "loss": 0.0064, + "step": 40820 + }, + { + "epoch": 3.1712621359223303, + "grad_norm": 0.22396408021450043, + "learning_rate": 6.88221067674114e-05, + "loss": 0.0073, + "step": 40830 + }, + { + "epoch": 3.172038834951456, + "grad_norm": 0.2642694115638733, + "learning_rate": 6.88067873305455e-05, + "loss": 0.0083, + "step": 40840 + }, + { + "epoch": 3.1728155339805824, + "grad_norm": 0.24044637382030487, + "learning_rate": 6.879146583699765e-05, + "loss": 0.0078, + "step": 40850 + }, + { + "epoch": 3.1735922330097086, + "grad_norm": 0.19623225927352905, + "learning_rate": 6.877614228844334e-05, + "loss": 0.0059, + "step": 40860 + }, + { + "epoch": 3.174368932038835, + "grad_norm": 0.20434586703777313, + "learning_rate": 6.876081668655832e-05, + "loss": 0.0088, + "step": 40870 + }, + { + "epoch": 3.175145631067961, + "grad_norm": 0.21365924179553986, + "learning_rate": 6.874548903301861e-05, + "loss": 0.0075, + "step": 40880 + }, + { + "epoch": 3.1759223300970874, + "grad_norm": 0.23745986819267273, + "learning_rate": 6.873015932950039e-05, + "loss": 0.0088, + "step": 40890 + }, + { + "epoch": 3.1766990291262136, + "grad_norm": 0.2663077116012573, + "learning_rate": 6.871482757768012e-05, + "loss": 0.0095, + "step": 40900 + }, + { + "epoch": 3.17747572815534, + "grad_norm": 0.10202276706695557, + "learning_rate": 6.869949377923442e-05, + "loss": 0.0055, + "step": 40910 + }, + { + "epoch": 3.178252427184466, + "grad_norm": 0.19709955155849457, + "learning_rate": 6.868415793584022e-05, + "loss": 0.0086, + "step": 40920 + }, + { + "epoch": 3.1790291262135923, + "grad_norm": 0.16072745621204376, + "learning_rate": 6.86688200491746e-05, + "loss": 0.011, + "step": 40930 + }, + { + "epoch": 3.1798058252427186, + "grad_norm": 0.7069708704948425, + "learning_rate": 6.865348012091488e-05, + "loss": 0.0063, + "step": 40940 + }, + { + "epoch": 3.180582524271845, + "grad_norm": 0.14563526213169098, + "learning_rate": 6.86381381527386e-05, + "loss": 0.0098, + "step": 40950 + }, + { + "epoch": 3.181359223300971, + "grad_norm": 0.27676963806152344, + "learning_rate": 6.862279414632359e-05, + "loss": 0.0079, + "step": 40960 + }, + { + "epoch": 3.182135922330097, + "grad_norm": 0.3028434216976166, + "learning_rate": 6.860744810334778e-05, + "loss": 0.007, + "step": 40970 + }, + { + "epoch": 3.182912621359223, + "grad_norm": 0.3211749494075775, + "learning_rate": 6.859210002548943e-05, + "loss": 0.0081, + "step": 40980 + }, + { + "epoch": 3.1836893203883494, + "grad_norm": 0.24662357568740845, + "learning_rate": 6.8576749914427e-05, + "loss": 0.0044, + "step": 40990 + }, + { + "epoch": 3.1844660194174756, + "grad_norm": 0.18265566229820251, + "learning_rate": 6.85613977718391e-05, + "loss": 0.0074, + "step": 41000 + }, + { + "epoch": 3.185242718446602, + "grad_norm": 0.34870895743370056, + "learning_rate": 6.854604359940464e-05, + "loss": 0.0095, + "step": 41010 + }, + { + "epoch": 3.186019417475728, + "grad_norm": 0.24679820239543915, + "learning_rate": 6.853068739880276e-05, + "loss": 0.0067, + "step": 41020 + }, + { + "epoch": 3.1867961165048544, + "grad_norm": 0.3441217243671417, + "learning_rate": 6.851532917171276e-05, + "loss": 0.0089, + "step": 41030 + }, + { + "epoch": 3.1875728155339806, + "grad_norm": 0.22360709309577942, + "learning_rate": 6.849996891981418e-05, + "loss": 0.0061, + "step": 41040 + }, + { + "epoch": 3.188349514563107, + "grad_norm": 0.32984867691993713, + "learning_rate": 6.848460664478684e-05, + "loss": 0.0077, + "step": 41050 + }, + { + "epoch": 3.189126213592233, + "grad_norm": 0.2536618709564209, + "learning_rate": 6.846924234831069e-05, + "loss": 0.0095, + "step": 41060 + }, + { + "epoch": 3.1899029126213594, + "grad_norm": 0.26645857095718384, + "learning_rate": 6.845387603206597e-05, + "loss": 0.0078, + "step": 41070 + }, + { + "epoch": 3.1906796116504856, + "grad_norm": 0.7714701294898987, + "learning_rate": 6.843850769773312e-05, + "loss": 0.008, + "step": 41080 + }, + { + "epoch": 3.1914563106796114, + "grad_norm": 0.2698996365070343, + "learning_rate": 6.842313734699278e-05, + "loss": 0.0075, + "step": 41090 + }, + { + "epoch": 3.1922330097087377, + "grad_norm": 0.45612484216690063, + "learning_rate": 6.840776498152584e-05, + "loss": 0.0076, + "step": 41100 + }, + { + "epoch": 3.193009708737864, + "grad_norm": 0.24259518086910248, + "learning_rate": 6.83923906030134e-05, + "loss": 0.0088, + "step": 41110 + }, + { + "epoch": 3.19378640776699, + "grad_norm": 0.16381385922431946, + "learning_rate": 6.837701421313677e-05, + "loss": 0.0059, + "step": 41120 + }, + { + "epoch": 3.1945631067961164, + "grad_norm": 0.20315328240394592, + "learning_rate": 6.83616358135775e-05, + "loss": 0.0074, + "step": 41130 + }, + { + "epoch": 3.1953398058252427, + "grad_norm": 0.29087719321250916, + "learning_rate": 6.834625540601734e-05, + "loss": 0.007, + "step": 41140 + }, + { + "epoch": 3.196116504854369, + "grad_norm": 0.17691676318645477, + "learning_rate": 6.833087299213829e-05, + "loss": 0.0083, + "step": 41150 + }, + { + "epoch": 3.196893203883495, + "grad_norm": 0.3395529091358185, + "learning_rate": 6.83154885736225e-05, + "loss": 0.0098, + "step": 41160 + }, + { + "epoch": 3.1976699029126214, + "grad_norm": 0.33924180269241333, + "learning_rate": 6.830010215215244e-05, + "loss": 0.0064, + "step": 41170 + }, + { + "epoch": 3.1984466019417477, + "grad_norm": 0.2587917447090149, + "learning_rate": 6.828471372941072e-05, + "loss": 0.0078, + "step": 41180 + }, + { + "epoch": 3.199223300970874, + "grad_norm": 0.18193091452121735, + "learning_rate": 6.826932330708018e-05, + "loss": 0.0075, + "step": 41190 + }, + { + "epoch": 3.2, + "grad_norm": 0.3387988209724426, + "learning_rate": 6.825393088684393e-05, + "loss": 0.0095, + "step": 41200 + }, + { + "epoch": 3.2007766990291264, + "grad_norm": 0.2460080236196518, + "learning_rate": 6.823853647038524e-05, + "loss": 0.0087, + "step": 41210 + }, + { + "epoch": 3.2015533980582522, + "grad_norm": 0.23452399671077728, + "learning_rate": 6.822314005938762e-05, + "loss": 0.0078, + "step": 41220 + }, + { + "epoch": 3.2023300970873785, + "grad_norm": 0.17616650462150574, + "learning_rate": 6.820774165553479e-05, + "loss": 0.0075, + "step": 41230 + }, + { + "epoch": 3.2031067961165047, + "grad_norm": 0.2931734025478363, + "learning_rate": 6.819234126051072e-05, + "loss": 0.0071, + "step": 41240 + }, + { + "epoch": 3.203883495145631, + "grad_norm": 0.13082027435302734, + "learning_rate": 6.817693887599956e-05, + "loss": 0.0105, + "step": 41250 + }, + { + "epoch": 3.2046601941747572, + "grad_norm": 0.44071534276008606, + "learning_rate": 6.816153450368568e-05, + "loss": 0.0069, + "step": 41260 + }, + { + "epoch": 3.2054368932038835, + "grad_norm": 0.37260884046554565, + "learning_rate": 6.814612814525372e-05, + "loss": 0.0064, + "step": 41270 + }, + { + "epoch": 3.2062135922330097, + "grad_norm": 0.3706704080104828, + "learning_rate": 6.813071980238846e-05, + "loss": 0.0068, + "step": 41280 + }, + { + "epoch": 3.206990291262136, + "grad_norm": 0.3105030953884125, + "learning_rate": 6.811530947677492e-05, + "loss": 0.0093, + "step": 41290 + }, + { + "epoch": 3.207766990291262, + "grad_norm": 0.19778640568256378, + "learning_rate": 6.809989717009839e-05, + "loss": 0.0082, + "step": 41300 + }, + { + "epoch": 3.2085436893203885, + "grad_norm": 0.27684977650642395, + "learning_rate": 6.808448288404431e-05, + "loss": 0.0078, + "step": 41310 + }, + { + "epoch": 3.2093203883495147, + "grad_norm": 0.31270432472229004, + "learning_rate": 6.806906662029838e-05, + "loss": 0.0067, + "step": 41320 + }, + { + "epoch": 3.210097087378641, + "grad_norm": 0.24404439330101013, + "learning_rate": 6.805364838054647e-05, + "loss": 0.008, + "step": 41330 + }, + { + "epoch": 3.2108737864077668, + "grad_norm": 0.10765136033296585, + "learning_rate": 6.803822816647474e-05, + "loss": 0.0059, + "step": 41340 + }, + { + "epoch": 3.211650485436893, + "grad_norm": 0.15456622838974, + "learning_rate": 6.802280597976949e-05, + "loss": 0.008, + "step": 41350 + }, + { + "epoch": 3.2124271844660193, + "grad_norm": 0.14841631054878235, + "learning_rate": 6.800738182211727e-05, + "loss": 0.0086, + "step": 41360 + }, + { + "epoch": 3.2132038834951455, + "grad_norm": 0.3713252544403076, + "learning_rate": 6.799195569520485e-05, + "loss": 0.0089, + "step": 41370 + }, + { + "epoch": 3.2139805825242718, + "grad_norm": 0.2988492548465729, + "learning_rate": 6.79765276007192e-05, + "loss": 0.0098, + "step": 41380 + }, + { + "epoch": 3.214757281553398, + "grad_norm": 0.1547280251979828, + "learning_rate": 6.796109754034753e-05, + "loss": 0.0082, + "step": 41390 + }, + { + "epoch": 3.2155339805825243, + "grad_norm": 0.18544769287109375, + "learning_rate": 6.794566551577724e-05, + "loss": 0.0073, + "step": 41400 + }, + { + "epoch": 3.2163106796116505, + "grad_norm": 0.14352528750896454, + "learning_rate": 6.793023152869594e-05, + "loss": 0.0083, + "step": 41410 + }, + { + "epoch": 3.2170873786407768, + "grad_norm": 0.1838179975748062, + "learning_rate": 6.791479558079147e-05, + "loss": 0.0099, + "step": 41420 + }, + { + "epoch": 3.217864077669903, + "grad_norm": 0.37874990701675415, + "learning_rate": 6.789935767375191e-05, + "loss": 0.0122, + "step": 41430 + }, + { + "epoch": 3.2186407766990293, + "grad_norm": 0.25070375204086304, + "learning_rate": 6.78839178092655e-05, + "loss": 0.0088, + "step": 41440 + }, + { + "epoch": 3.2194174757281555, + "grad_norm": 0.35771608352661133, + "learning_rate": 6.786847598902072e-05, + "loss": 0.0098, + "step": 41450 + }, + { + "epoch": 3.2201941747572818, + "grad_norm": 0.23418082296848297, + "learning_rate": 6.785303221470627e-05, + "loss": 0.0056, + "step": 41460 + }, + { + "epoch": 3.2209708737864076, + "grad_norm": 0.20989742875099182, + "learning_rate": 6.783758648801108e-05, + "loss": 0.0065, + "step": 41470 + }, + { + "epoch": 3.221747572815534, + "grad_norm": 0.23460513353347778, + "learning_rate": 6.782213881062422e-05, + "loss": 0.0066, + "step": 41480 + }, + { + "epoch": 3.22252427184466, + "grad_norm": 0.37431952357292175, + "learning_rate": 6.780668918423508e-05, + "loss": 0.0089, + "step": 41490 + }, + { + "epoch": 3.2233009708737863, + "grad_norm": 0.49125516414642334, + "learning_rate": 6.779123761053317e-05, + "loss": 0.0102, + "step": 41500 + }, + { + "epoch": 3.2240776699029126, + "grad_norm": 0.398306667804718, + "learning_rate": 6.777578409120826e-05, + "loss": 0.0061, + "step": 41510 + }, + { + "epoch": 3.224854368932039, + "grad_norm": 0.17786823213100433, + "learning_rate": 6.776032862795033e-05, + "loss": 0.0092, + "step": 41520 + }, + { + "epoch": 3.225631067961165, + "grad_norm": 0.27751168608665466, + "learning_rate": 6.774487122244956e-05, + "loss": 0.0104, + "step": 41530 + }, + { + "epoch": 3.2264077669902913, + "grad_norm": 1.0153850317001343, + "learning_rate": 6.772941187639637e-05, + "loss": 0.0088, + "step": 41540 + }, + { + "epoch": 3.2271844660194176, + "grad_norm": 0.3064456582069397, + "learning_rate": 6.771395059148134e-05, + "loss": 0.0113, + "step": 41550 + }, + { + "epoch": 3.227961165048544, + "grad_norm": 0.40481650829315186, + "learning_rate": 6.76984873693953e-05, + "loss": 0.009, + "step": 41560 + }, + { + "epoch": 3.22873786407767, + "grad_norm": 0.3073768615722656, + "learning_rate": 6.768302221182931e-05, + "loss": 0.0085, + "step": 41570 + }, + { + "epoch": 3.2295145631067963, + "grad_norm": 0.4186054766178131, + "learning_rate": 6.766755512047457e-05, + "loss": 0.0107, + "step": 41580 + }, + { + "epoch": 3.230291262135922, + "grad_norm": 0.39286354184150696, + "learning_rate": 6.765208609702259e-05, + "loss": 0.0075, + "step": 41590 + }, + { + "epoch": 3.2310679611650484, + "grad_norm": 0.4187982976436615, + "learning_rate": 6.763661514316499e-05, + "loss": 0.0107, + "step": 41600 + }, + { + "epoch": 3.2318446601941746, + "grad_norm": 0.2861064076423645, + "learning_rate": 6.76211422605937e-05, + "loss": 0.0083, + "step": 41610 + }, + { + "epoch": 3.232621359223301, + "grad_norm": 0.23571476340293884, + "learning_rate": 6.760566745100076e-05, + "loss": 0.0087, + "step": 41620 + }, + { + "epoch": 3.233398058252427, + "grad_norm": 0.25153955817222595, + "learning_rate": 6.75901907160785e-05, + "loss": 0.0089, + "step": 41630 + }, + { + "epoch": 3.2341747572815533, + "grad_norm": 0.21886515617370605, + "learning_rate": 6.757471205751943e-05, + "loss": 0.0109, + "step": 41640 + }, + { + "epoch": 3.2349514563106796, + "grad_norm": 0.3603833317756653, + "learning_rate": 6.75592314770163e-05, + "loss": 0.0083, + "step": 41650 + }, + { + "epoch": 3.235728155339806, + "grad_norm": 0.17435584962368011, + "learning_rate": 6.7543748976262e-05, + "loss": 0.0065, + "step": 41660 + }, + { + "epoch": 3.236504854368932, + "grad_norm": 0.22872807085514069, + "learning_rate": 6.752826455694968e-05, + "loss": 0.0058, + "step": 41670 + }, + { + "epoch": 3.2372815533980583, + "grad_norm": 0.203835129737854, + "learning_rate": 6.751277822077271e-05, + "loss": 0.0066, + "step": 41680 + }, + { + "epoch": 3.2380582524271846, + "grad_norm": 0.19361795485019684, + "learning_rate": 6.749728996942465e-05, + "loss": 0.0097, + "step": 41690 + }, + { + "epoch": 3.238834951456311, + "grad_norm": 0.15880198776721954, + "learning_rate": 6.748179980459924e-05, + "loss": 0.0067, + "step": 41700 + }, + { + "epoch": 3.239611650485437, + "grad_norm": 0.27615880966186523, + "learning_rate": 6.746630772799052e-05, + "loss": 0.0081, + "step": 41710 + }, + { + "epoch": 3.2403883495145633, + "grad_norm": 0.22910776734352112, + "learning_rate": 6.745081374129262e-05, + "loss": 0.0063, + "step": 41720 + }, + { + "epoch": 3.241165048543689, + "grad_norm": 0.19584053754806519, + "learning_rate": 6.743531784619998e-05, + "loss": 0.0082, + "step": 41730 + }, + { + "epoch": 3.2419417475728154, + "grad_norm": 0.13948598504066467, + "learning_rate": 6.741982004440719e-05, + "loss": 0.0073, + "step": 41740 + }, + { + "epoch": 3.2427184466019416, + "grad_norm": 0.34006398916244507, + "learning_rate": 6.740432033760907e-05, + "loss": 0.0059, + "step": 41750 + }, + { + "epoch": 3.243495145631068, + "grad_norm": 0.29532527923583984, + "learning_rate": 6.738881872750066e-05, + "loss": 0.0069, + "step": 41760 + }, + { + "epoch": 3.244271844660194, + "grad_norm": 0.17224527895450592, + "learning_rate": 6.737331521577715e-05, + "loss": 0.0066, + "step": 41770 + }, + { + "epoch": 3.2450485436893204, + "grad_norm": 0.36370351910591125, + "learning_rate": 6.735780980413403e-05, + "loss": 0.0072, + "step": 41780 + }, + { + "epoch": 3.2458252427184466, + "grad_norm": 0.27829229831695557, + "learning_rate": 6.734230249426692e-05, + "loss": 0.0079, + "step": 41790 + }, + { + "epoch": 3.246601941747573, + "grad_norm": 0.27612200379371643, + "learning_rate": 6.732679328787168e-05, + "loss": 0.0058, + "step": 41800 + }, + { + "epoch": 3.247378640776699, + "grad_norm": 0.4886054992675781, + "learning_rate": 6.731128218664438e-05, + "loss": 0.0098, + "step": 41810 + }, + { + "epoch": 3.2481553398058254, + "grad_norm": 0.3050101101398468, + "learning_rate": 6.729576919228129e-05, + "loss": 0.0072, + "step": 41820 + }, + { + "epoch": 3.2489320388349516, + "grad_norm": 0.2482646405696869, + "learning_rate": 6.728025430647888e-05, + "loss": 0.0089, + "step": 41830 + }, + { + "epoch": 3.2497087378640774, + "grad_norm": 0.20041295886039734, + "learning_rate": 6.726473753093383e-05, + "loss": 0.009, + "step": 41840 + }, + { + "epoch": 3.2504854368932037, + "grad_norm": 0.35417819023132324, + "learning_rate": 6.724921886734305e-05, + "loss": 0.0061, + "step": 41850 + }, + { + "epoch": 3.25126213592233, + "grad_norm": 0.4211118817329407, + "learning_rate": 6.723369831740363e-05, + "loss": 0.008, + "step": 41860 + }, + { + "epoch": 3.252038834951456, + "grad_norm": 0.30484333634376526, + "learning_rate": 6.721817588281288e-05, + "loss": 0.0108, + "step": 41870 + }, + { + "epoch": 3.2528155339805824, + "grad_norm": 0.197379007935524, + "learning_rate": 6.720265156526828e-05, + "loss": 0.0094, + "step": 41880 + }, + { + "epoch": 3.2535922330097087, + "grad_norm": 0.2667190730571747, + "learning_rate": 6.718712536646758e-05, + "loss": 0.0082, + "step": 41890 + }, + { + "epoch": 3.254368932038835, + "grad_norm": 0.1524379998445511, + "learning_rate": 6.71715972881087e-05, + "loss": 0.0101, + "step": 41900 + }, + { + "epoch": 3.255145631067961, + "grad_norm": 0.26204022765159607, + "learning_rate": 6.715606733188974e-05, + "loss": 0.006, + "step": 41910 + }, + { + "epoch": 3.2559223300970874, + "grad_norm": 0.18891659379005432, + "learning_rate": 6.714053549950909e-05, + "loss": 0.0092, + "step": 41920 + }, + { + "epoch": 3.2566990291262137, + "grad_norm": 0.12869605422019958, + "learning_rate": 6.71250017926652e-05, + "loss": 0.0081, + "step": 41930 + }, + { + "epoch": 3.25747572815534, + "grad_norm": 0.38768014311790466, + "learning_rate": 6.71094662130569e-05, + "loss": 0.0066, + "step": 41940 + }, + { + "epoch": 3.258252427184466, + "grad_norm": 0.2635297179222107, + "learning_rate": 6.709392876238307e-05, + "loss": 0.0068, + "step": 41950 + }, + { + "epoch": 3.2590291262135924, + "grad_norm": 0.2008301466703415, + "learning_rate": 6.707838944234289e-05, + "loss": 0.0074, + "step": 41960 + }, + { + "epoch": 3.2598058252427187, + "grad_norm": 0.12451300770044327, + "learning_rate": 6.706284825463574e-05, + "loss": 0.0055, + "step": 41970 + }, + { + "epoch": 3.2605825242718445, + "grad_norm": 0.22562798857688904, + "learning_rate": 6.704730520096114e-05, + "loss": 0.0088, + "step": 41980 + }, + { + "epoch": 3.2613592233009707, + "grad_norm": 0.3420323431491852, + "learning_rate": 6.703176028301888e-05, + "loss": 0.0068, + "step": 41990 + }, + { + "epoch": 3.262135922330097, + "grad_norm": 0.1565765142440796, + "learning_rate": 6.701621350250892e-05, + "loss": 0.0078, + "step": 42000 + }, + { + "epoch": 3.262912621359223, + "grad_norm": 0.43935394287109375, + "learning_rate": 6.700066486113144e-05, + "loss": 0.0086, + "step": 42010 + }, + { + "epoch": 3.2636893203883495, + "grad_norm": 0.31361067295074463, + "learning_rate": 6.69851143605868e-05, + "loss": 0.007, + "step": 42020 + }, + { + "epoch": 3.2644660194174757, + "grad_norm": 0.4164636731147766, + "learning_rate": 6.696956200257561e-05, + "loss": 0.0071, + "step": 42030 + }, + { + "epoch": 3.265242718446602, + "grad_norm": 0.27257174253463745, + "learning_rate": 6.695400778879863e-05, + "loss": 0.0079, + "step": 42040 + }, + { + "epoch": 3.266019417475728, + "grad_norm": 0.2834935188293457, + "learning_rate": 6.693845172095683e-05, + "loss": 0.0061, + "step": 42050 + }, + { + "epoch": 3.2667961165048545, + "grad_norm": 0.1848030686378479, + "learning_rate": 6.692289380075142e-05, + "loss": 0.0078, + "step": 42060 + }, + { + "epoch": 3.2675728155339807, + "grad_norm": 0.1850883513689041, + "learning_rate": 6.690733402988379e-05, + "loss": 0.0082, + "step": 42070 + }, + { + "epoch": 3.268349514563107, + "grad_norm": 0.352674275636673, + "learning_rate": 6.689177241005553e-05, + "loss": 0.0091, + "step": 42080 + }, + { + "epoch": 3.2691262135922328, + "grad_norm": 0.28496643900871277, + "learning_rate": 6.687620894296844e-05, + "loss": 0.0088, + "step": 42090 + }, + { + "epoch": 3.269902912621359, + "grad_norm": 0.2648363709449768, + "learning_rate": 6.686064363032451e-05, + "loss": 0.0083, + "step": 42100 + }, + { + "epoch": 3.2706796116504853, + "grad_norm": 0.6118050813674927, + "learning_rate": 6.684507647382596e-05, + "loss": 0.0077, + "step": 42110 + }, + { + "epoch": 3.2714563106796115, + "grad_norm": 0.2598336637020111, + "learning_rate": 6.682950747517513e-05, + "loss": 0.0091, + "step": 42120 + }, + { + "epoch": 3.2722330097087378, + "grad_norm": 0.28565922379493713, + "learning_rate": 6.68139366360747e-05, + "loss": 0.0065, + "step": 42130 + }, + { + "epoch": 3.273009708737864, + "grad_norm": 0.21777695417404175, + "learning_rate": 6.679836395822744e-05, + "loss": 0.0068, + "step": 42140 + }, + { + "epoch": 3.2737864077669903, + "grad_norm": 0.3321709930896759, + "learning_rate": 6.678278944333633e-05, + "loss": 0.0129, + "step": 42150 + }, + { + "epoch": 3.2745631067961165, + "grad_norm": 0.30193641781806946, + "learning_rate": 6.676721309310462e-05, + "loss": 0.0076, + "step": 42160 + }, + { + "epoch": 3.2753398058252428, + "grad_norm": 0.3076436221599579, + "learning_rate": 6.67516349092357e-05, + "loss": 0.0086, + "step": 42170 + }, + { + "epoch": 3.276116504854369, + "grad_norm": 0.18790361285209656, + "learning_rate": 6.673605489343315e-05, + "loss": 0.0099, + "step": 42180 + }, + { + "epoch": 3.2768932038834953, + "grad_norm": 0.3185842037200928, + "learning_rate": 6.672047304740081e-05, + "loss": 0.0086, + "step": 42190 + }, + { + "epoch": 3.2776699029126215, + "grad_norm": 0.3139133155345917, + "learning_rate": 6.670488937284268e-05, + "loss": 0.0086, + "step": 42200 + }, + { + "epoch": 3.2784466019417478, + "grad_norm": 0.13425585627555847, + "learning_rate": 6.668930387146294e-05, + "loss": 0.0067, + "step": 42210 + }, + { + "epoch": 3.279223300970874, + "grad_norm": 0.1457650065422058, + "learning_rate": 6.667371654496605e-05, + "loss": 0.0049, + "step": 42220 + }, + { + "epoch": 3.2800000000000002, + "grad_norm": 0.2791220247745514, + "learning_rate": 6.665812739505659e-05, + "loss": 0.0062, + "step": 42230 + }, + { + "epoch": 3.280776699029126, + "grad_norm": 0.15845420956611633, + "learning_rate": 6.664253642343935e-05, + "loss": 0.0103, + "step": 42240 + }, + { + "epoch": 3.2815533980582523, + "grad_norm": 0.7059871554374695, + "learning_rate": 6.662694363181935e-05, + "loss": 0.0101, + "step": 42250 + }, + { + "epoch": 3.2823300970873786, + "grad_norm": 0.2390138953924179, + "learning_rate": 6.66113490219018e-05, + "loss": 0.0088, + "step": 42260 + }, + { + "epoch": 3.283106796116505, + "grad_norm": 0.11725740879774094, + "learning_rate": 6.659575259539207e-05, + "loss": 0.0052, + "step": 42270 + }, + { + "epoch": 3.283883495145631, + "grad_norm": 0.35465264320373535, + "learning_rate": 6.65801543539958e-05, + "loss": 0.0081, + "step": 42280 + }, + { + "epoch": 3.2846601941747573, + "grad_norm": 0.13851900398731232, + "learning_rate": 6.65645542994188e-05, + "loss": 0.0086, + "step": 42290 + }, + { + "epoch": 3.2854368932038835, + "grad_norm": 0.23740831017494202, + "learning_rate": 6.654895243336702e-05, + "loss": 0.0057, + "step": 42300 + }, + { + "epoch": 3.28621359223301, + "grad_norm": 0.28865525126457214, + "learning_rate": 6.65333487575467e-05, + "loss": 0.0089, + "step": 42310 + }, + { + "epoch": 3.286990291262136, + "grad_norm": 0.20704543590545654, + "learning_rate": 6.651774327366422e-05, + "loss": 0.0061, + "step": 42320 + }, + { + "epoch": 3.2877669902912623, + "grad_norm": 0.29128262400627136, + "learning_rate": 6.650213598342613e-05, + "loss": 0.0087, + "step": 42330 + }, + { + "epoch": 3.2885436893203885, + "grad_norm": 0.20082373917102814, + "learning_rate": 6.648652688853931e-05, + "loss": 0.007, + "step": 42340 + }, + { + "epoch": 3.2893203883495143, + "grad_norm": 0.34981298446655273, + "learning_rate": 6.647091599071066e-05, + "loss": 0.005, + "step": 42350 + }, + { + "epoch": 3.2900970873786406, + "grad_norm": 0.27188703417778015, + "learning_rate": 6.645530329164743e-05, + "loss": 0.0076, + "step": 42360 + }, + { + "epoch": 3.290873786407767, + "grad_norm": 0.3017987310886383, + "learning_rate": 6.643968879305694e-05, + "loss": 0.0069, + "step": 42370 + }, + { + "epoch": 3.291650485436893, + "grad_norm": 0.283353328704834, + "learning_rate": 6.642407249664683e-05, + "loss": 0.0075, + "step": 42380 + }, + { + "epoch": 3.2924271844660193, + "grad_norm": 0.11902421712875366, + "learning_rate": 6.640845440412483e-05, + "loss": 0.0071, + "step": 42390 + }, + { + "epoch": 3.2932038834951456, + "grad_norm": 0.17327074706554413, + "learning_rate": 6.639283451719893e-05, + "loss": 0.0098, + "step": 42400 + }, + { + "epoch": 3.293980582524272, + "grad_norm": 0.2055923491716385, + "learning_rate": 6.637721283757729e-05, + "loss": 0.007, + "step": 42410 + }, + { + "epoch": 3.294757281553398, + "grad_norm": 0.1753554791212082, + "learning_rate": 6.636158936696828e-05, + "loss": 0.0074, + "step": 42420 + }, + { + "epoch": 3.2955339805825243, + "grad_norm": 0.283370703458786, + "learning_rate": 6.634596410708047e-05, + "loss": 0.0085, + "step": 42430 + }, + { + "epoch": 3.2963106796116506, + "grad_norm": 0.349971741437912, + "learning_rate": 6.633033705962257e-05, + "loss": 0.0066, + "step": 42440 + }, + { + "epoch": 3.297087378640777, + "grad_norm": 0.1882169097661972, + "learning_rate": 6.631470822630359e-05, + "loss": 0.0072, + "step": 42450 + }, + { + "epoch": 3.297864077669903, + "grad_norm": 0.2688639163970947, + "learning_rate": 6.629907760883263e-05, + "loss": 0.0086, + "step": 42460 + }, + { + "epoch": 3.2986407766990293, + "grad_norm": 0.2825281023979187, + "learning_rate": 6.628344520891905e-05, + "loss": 0.009, + "step": 42470 + }, + { + "epoch": 3.2994174757281556, + "grad_norm": 0.34693825244903564, + "learning_rate": 6.62678110282724e-05, + "loss": 0.0082, + "step": 42480 + }, + { + "epoch": 3.3001941747572814, + "grad_norm": 0.25407150387763977, + "learning_rate": 6.625217506860237e-05, + "loss": 0.0073, + "step": 42490 + }, + { + "epoch": 3.3009708737864076, + "grad_norm": 0.2508038878440857, + "learning_rate": 6.623653733161892e-05, + "loss": 0.0094, + "step": 42500 + }, + { + "epoch": 3.301747572815534, + "grad_norm": 0.19847740232944489, + "learning_rate": 6.622089781903217e-05, + "loss": 0.0084, + "step": 42510 + }, + { + "epoch": 3.30252427184466, + "grad_norm": 0.10096978396177292, + "learning_rate": 6.620525653255241e-05, + "loss": 0.0065, + "step": 42520 + }, + { + "epoch": 3.3033009708737864, + "grad_norm": 0.4413469731807709, + "learning_rate": 6.618961347389015e-05, + "loss": 0.0078, + "step": 42530 + }, + { + "epoch": 3.3040776699029126, + "grad_norm": 0.18944387137889862, + "learning_rate": 6.617396864475613e-05, + "loss": 0.0074, + "step": 42540 + }, + { + "epoch": 3.304854368932039, + "grad_norm": 0.27465999126434326, + "learning_rate": 6.61583220468612e-05, + "loss": 0.0062, + "step": 42550 + }, + { + "epoch": 3.305631067961165, + "grad_norm": 0.25946176052093506, + "learning_rate": 6.614267368191645e-05, + "loss": 0.0079, + "step": 42560 + }, + { + "epoch": 3.3064077669902914, + "grad_norm": 0.2139652818441391, + "learning_rate": 6.61270235516332e-05, + "loss": 0.0072, + "step": 42570 + }, + { + "epoch": 3.3071844660194176, + "grad_norm": 0.1138659417629242, + "learning_rate": 6.611137165772287e-05, + "loss": 0.0053, + "step": 42580 + }, + { + "epoch": 3.307961165048544, + "grad_norm": 0.26721394062042236, + "learning_rate": 6.60957180018972e-05, + "loss": 0.0071, + "step": 42590 + }, + { + "epoch": 3.3087378640776697, + "grad_norm": 0.21176253259181976, + "learning_rate": 6.608006258586797e-05, + "loss": 0.0087, + "step": 42600 + }, + { + "epoch": 3.309514563106796, + "grad_norm": 0.1492680162191391, + "learning_rate": 6.606440541134731e-05, + "loss": 0.0124, + "step": 42610 + }, + { + "epoch": 3.310291262135922, + "grad_norm": 0.3430321216583252, + "learning_rate": 6.60487464800474e-05, + "loss": 0.0134, + "step": 42620 + }, + { + "epoch": 3.3110679611650484, + "grad_norm": 0.23387813568115234, + "learning_rate": 6.603308579368071e-05, + "loss": 0.0077, + "step": 42630 + }, + { + "epoch": 3.3118446601941747, + "grad_norm": 0.15849430859088898, + "learning_rate": 6.601742335395987e-05, + "loss": 0.0074, + "step": 42640 + }, + { + "epoch": 3.312621359223301, + "grad_norm": 0.19834508001804352, + "learning_rate": 6.600175916259769e-05, + "loss": 0.008, + "step": 42650 + }, + { + "epoch": 3.313398058252427, + "grad_norm": 0.37061044573783875, + "learning_rate": 6.598609322130718e-05, + "loss": 0.0072, + "step": 42660 + }, + { + "epoch": 3.3141747572815534, + "grad_norm": 0.38462355732917786, + "learning_rate": 6.597042553180155e-05, + "loss": 0.007, + "step": 42670 + }, + { + "epoch": 3.3149514563106797, + "grad_norm": 0.29273757338523865, + "learning_rate": 6.59547560957942e-05, + "loss": 0.0075, + "step": 42680 + }, + { + "epoch": 3.315728155339806, + "grad_norm": 0.33089470863342285, + "learning_rate": 6.59390849149987e-05, + "loss": 0.0099, + "step": 42690 + }, + { + "epoch": 3.316504854368932, + "grad_norm": 0.32603394985198975, + "learning_rate": 6.592341199112886e-05, + "loss": 0.0088, + "step": 42700 + }, + { + "epoch": 3.3172815533980584, + "grad_norm": 0.9392871856689453, + "learning_rate": 6.590773732589861e-05, + "loss": 0.0065, + "step": 42710 + }, + { + "epoch": 3.3180582524271847, + "grad_norm": 0.37672159075737, + "learning_rate": 6.589206092102213e-05, + "loss": 0.01, + "step": 42720 + }, + { + "epoch": 3.318834951456311, + "grad_norm": 0.1867024004459381, + "learning_rate": 6.587638277821375e-05, + "loss": 0.0078, + "step": 42730 + }, + { + "epoch": 3.3196116504854367, + "grad_norm": 0.24476046860218048, + "learning_rate": 6.586070289918803e-05, + "loss": 0.0093, + "step": 42740 + }, + { + "epoch": 3.320388349514563, + "grad_norm": 0.2666022479534149, + "learning_rate": 6.584502128565968e-05, + "loss": 0.0078, + "step": 42750 + }, + { + "epoch": 3.321165048543689, + "grad_norm": 0.3771688938140869, + "learning_rate": 6.582933793934363e-05, + "loss": 0.0096, + "step": 42760 + }, + { + "epoch": 3.3219417475728155, + "grad_norm": 0.30158019065856934, + "learning_rate": 6.581365286195498e-05, + "loss": 0.0064, + "step": 42770 + }, + { + "epoch": 3.3227184466019417, + "grad_norm": 0.222904235124588, + "learning_rate": 6.579796605520903e-05, + "loss": 0.0088, + "step": 42780 + }, + { + "epoch": 3.323495145631068, + "grad_norm": 0.302985280752182, + "learning_rate": 6.578227752082127e-05, + "loss": 0.0095, + "step": 42790 + }, + { + "epoch": 3.324271844660194, + "grad_norm": 0.17918668687343597, + "learning_rate": 6.576658726050735e-05, + "loss": 0.0093, + "step": 42800 + }, + { + "epoch": 3.3250485436893205, + "grad_norm": 0.2702880799770355, + "learning_rate": 6.575089527598316e-05, + "loss": 0.0097, + "step": 42810 + }, + { + "epoch": 3.3258252427184467, + "grad_norm": 0.22495229542255402, + "learning_rate": 6.573520156896476e-05, + "loss": 0.0083, + "step": 42820 + }, + { + "epoch": 3.326601941747573, + "grad_norm": 0.33809787034988403, + "learning_rate": 6.571950614116835e-05, + "loss": 0.0077, + "step": 42830 + }, + { + "epoch": 3.327378640776699, + "grad_norm": 0.47722673416137695, + "learning_rate": 6.570380899431039e-05, + "loss": 0.0086, + "step": 42840 + }, + { + "epoch": 3.328155339805825, + "grad_norm": 0.4407382607460022, + "learning_rate": 6.568811013010749e-05, + "loss": 0.0081, + "step": 42850 + }, + { + "epoch": 3.3289320388349513, + "grad_norm": 0.1398766189813614, + "learning_rate": 6.567240955027645e-05, + "loss": 0.0078, + "step": 42860 + }, + { + "epoch": 3.3297087378640775, + "grad_norm": 0.11463119834661484, + "learning_rate": 6.565670725653427e-05, + "loss": 0.0072, + "step": 42870 + }, + { + "epoch": 3.3304854368932038, + "grad_norm": 0.13119903206825256, + "learning_rate": 6.564100325059812e-05, + "loss": 0.0056, + "step": 42880 + }, + { + "epoch": 3.33126213592233, + "grad_norm": 0.20163121819496155, + "learning_rate": 6.562529753418538e-05, + "loss": 0.0077, + "step": 42890 + }, + { + "epoch": 3.3320388349514563, + "grad_norm": 0.14290329813957214, + "learning_rate": 6.56095901090136e-05, + "loss": 0.0071, + "step": 42900 + }, + { + "epoch": 3.3328155339805825, + "grad_norm": 0.23009337484836578, + "learning_rate": 6.55938809768005e-05, + "loss": 0.0068, + "step": 42910 + }, + { + "epoch": 3.3335922330097087, + "grad_norm": 0.23735947906970978, + "learning_rate": 6.557817013926402e-05, + "loss": 0.0082, + "step": 42920 + }, + { + "epoch": 3.334368932038835, + "grad_norm": 0.21374066174030304, + "learning_rate": 6.556245759812227e-05, + "loss": 0.0081, + "step": 42930 + }, + { + "epoch": 3.3351456310679612, + "grad_norm": 0.2829653024673462, + "learning_rate": 6.554674335509357e-05, + "loss": 0.011, + "step": 42940 + }, + { + "epoch": 3.3359223300970875, + "grad_norm": 0.16620300710201263, + "learning_rate": 6.553102741189638e-05, + "loss": 0.0068, + "step": 42950 + }, + { + "epoch": 3.3366990291262137, + "grad_norm": 0.18725928664207458, + "learning_rate": 6.55153097702494e-05, + "loss": 0.0067, + "step": 42960 + }, + { + "epoch": 3.33747572815534, + "grad_norm": 0.3207841217517853, + "learning_rate": 6.549959043187146e-05, + "loss": 0.0059, + "step": 42970 + }, + { + "epoch": 3.3382524271844662, + "grad_norm": 0.12220098078250885, + "learning_rate": 6.54838693984816e-05, + "loss": 0.0071, + "step": 42980 + }, + { + "epoch": 3.339029126213592, + "grad_norm": 0.2146521508693695, + "learning_rate": 6.546814667179909e-05, + "loss": 0.0071, + "step": 42990 + }, + { + "epoch": 3.3398058252427183, + "grad_norm": 0.3037680387496948, + "learning_rate": 6.545242225354328e-05, + "loss": 0.0064, + "step": 43000 + }, + { + "epoch": 3.3405825242718445, + "grad_norm": 0.2333119809627533, + "learning_rate": 6.543669614543383e-05, + "loss": 0.0054, + "step": 43010 + }, + { + "epoch": 3.341359223300971, + "grad_norm": 0.4116780459880829, + "learning_rate": 6.542096834919049e-05, + "loss": 0.0056, + "step": 43020 + }, + { + "epoch": 3.342135922330097, + "grad_norm": 0.25465136766433716, + "learning_rate": 6.540523886653324e-05, + "loss": 0.0072, + "step": 43030 + }, + { + "epoch": 3.3429126213592233, + "grad_norm": 0.28208833932876587, + "learning_rate": 6.538950769918222e-05, + "loss": 0.0075, + "step": 43040 + }, + { + "epoch": 3.3436893203883495, + "grad_norm": 0.2566225528717041, + "learning_rate": 6.537377484885779e-05, + "loss": 0.0063, + "step": 43050 + }, + { + "epoch": 3.344466019417476, + "grad_norm": 0.20081260800361633, + "learning_rate": 6.535804031728045e-05, + "loss": 0.0081, + "step": 43060 + }, + { + "epoch": 3.345242718446602, + "grad_norm": 0.28622275590896606, + "learning_rate": 6.534230410617092e-05, + "loss": 0.0081, + "step": 43070 + }, + { + "epoch": 3.3460194174757283, + "grad_norm": 0.1414164900779724, + "learning_rate": 6.532656621725007e-05, + "loss": 0.0079, + "step": 43080 + }, + { + "epoch": 3.3467961165048545, + "grad_norm": 0.1992192566394806, + "learning_rate": 6.531082665223898e-05, + "loss": 0.0065, + "step": 43090 + }, + { + "epoch": 3.3475728155339803, + "grad_norm": 0.31906986236572266, + "learning_rate": 6.529508541285889e-05, + "loss": 0.006, + "step": 43100 + }, + { + "epoch": 3.3483495145631066, + "grad_norm": 0.3070645034313202, + "learning_rate": 6.527934250083128e-05, + "loss": 0.0067, + "step": 43110 + }, + { + "epoch": 3.349126213592233, + "grad_norm": 0.3202793002128601, + "learning_rate": 6.526359791787775e-05, + "loss": 0.0072, + "step": 43120 + }, + { + "epoch": 3.349902912621359, + "grad_norm": 0.1800348162651062, + "learning_rate": 6.524785166572009e-05, + "loss": 0.0076, + "step": 43130 + }, + { + "epoch": 3.3506796116504853, + "grad_norm": 0.3333243429660797, + "learning_rate": 6.523210374608029e-05, + "loss": 0.0055, + "step": 43140 + }, + { + "epoch": 3.3514563106796116, + "grad_norm": 0.34073570370674133, + "learning_rate": 6.521635416068054e-05, + "loss": 0.0084, + "step": 43150 + }, + { + "epoch": 3.352233009708738, + "grad_norm": 0.21048800647258759, + "learning_rate": 6.520060291124317e-05, + "loss": 0.0067, + "step": 43160 + }, + { + "epoch": 3.353009708737864, + "grad_norm": 0.3087782859802246, + "learning_rate": 6.518484999949072e-05, + "loss": 0.0083, + "step": 43170 + }, + { + "epoch": 3.3537864077669903, + "grad_norm": 0.20648474991321564, + "learning_rate": 6.516909542714591e-05, + "loss": 0.007, + "step": 43180 + }, + { + "epoch": 3.3545631067961166, + "grad_norm": 0.28344571590423584, + "learning_rate": 6.515333919593161e-05, + "loss": 0.0067, + "step": 43190 + }, + { + "epoch": 3.355339805825243, + "grad_norm": 0.2269468605518341, + "learning_rate": 6.513758130757094e-05, + "loss": 0.0077, + "step": 43200 + }, + { + "epoch": 3.356116504854369, + "grad_norm": 0.3021705746650696, + "learning_rate": 6.512182176378713e-05, + "loss": 0.0064, + "step": 43210 + }, + { + "epoch": 3.3568932038834953, + "grad_norm": 0.34522780776023865, + "learning_rate": 6.510606056630362e-05, + "loss": 0.0087, + "step": 43220 + }, + { + "epoch": 3.3576699029126216, + "grad_norm": 0.29671889543533325, + "learning_rate": 6.509029771684405e-05, + "loss": 0.0098, + "step": 43230 + }, + { + "epoch": 3.358446601941748, + "grad_norm": 0.27046552300453186, + "learning_rate": 6.50745332171322e-05, + "loss": 0.0073, + "step": 43240 + }, + { + "epoch": 3.3592233009708736, + "grad_norm": 0.1820564717054367, + "learning_rate": 6.505876706889207e-05, + "loss": 0.0081, + "step": 43250 + }, + { + "epoch": 3.36, + "grad_norm": 0.2832212448120117, + "learning_rate": 6.504299927384781e-05, + "loss": 0.0079, + "step": 43260 + }, + { + "epoch": 3.360776699029126, + "grad_norm": 0.2548358738422394, + "learning_rate": 6.502722983372378e-05, + "loss": 0.0078, + "step": 43270 + }, + { + "epoch": 3.3615533980582524, + "grad_norm": 0.16953256726264954, + "learning_rate": 6.501145875024446e-05, + "loss": 0.0082, + "step": 43280 + }, + { + "epoch": 3.3623300970873786, + "grad_norm": 1.1361656188964844, + "learning_rate": 6.499568602513459e-05, + "loss": 0.0084, + "step": 43290 + }, + { + "epoch": 3.363106796116505, + "grad_norm": 0.23682312667369843, + "learning_rate": 6.497991166011903e-05, + "loss": 0.0088, + "step": 43300 + }, + { + "epoch": 3.363883495145631, + "grad_norm": 0.43736153841018677, + "learning_rate": 6.49641356569229e-05, + "loss": 0.0064, + "step": 43310 + }, + { + "epoch": 3.3646601941747574, + "grad_norm": 0.4968525171279907, + "learning_rate": 6.494835801727135e-05, + "loss": 0.0085, + "step": 43320 + }, + { + "epoch": 3.3654368932038836, + "grad_norm": 0.2542344033718109, + "learning_rate": 6.493257874288986e-05, + "loss": 0.0067, + "step": 43330 + }, + { + "epoch": 3.36621359223301, + "grad_norm": 0.2945004403591156, + "learning_rate": 6.491679783550399e-05, + "loss": 0.0079, + "step": 43340 + }, + { + "epoch": 3.366990291262136, + "grad_norm": 0.2343362420797348, + "learning_rate": 6.490101529683957e-05, + "loss": 0.0071, + "step": 43350 + }, + { + "epoch": 3.367766990291262, + "grad_norm": 0.2008380889892578, + "learning_rate": 6.488523112862248e-05, + "loss": 0.0068, + "step": 43360 + }, + { + "epoch": 3.368543689320388, + "grad_norm": 0.20944635570049286, + "learning_rate": 6.486944533257891e-05, + "loss": 0.0066, + "step": 43370 + }, + { + "epoch": 3.3693203883495144, + "grad_norm": 0.2022116333246231, + "learning_rate": 6.485365791043517e-05, + "loss": 0.007, + "step": 43380 + }, + { + "epoch": 3.3700970873786407, + "grad_norm": 0.23796214163303375, + "learning_rate": 6.483786886391772e-05, + "loss": 0.0063, + "step": 43390 + }, + { + "epoch": 3.370873786407767, + "grad_norm": 0.17176410555839539, + "learning_rate": 6.482207819475323e-05, + "loss": 0.0077, + "step": 43400 + }, + { + "epoch": 3.371650485436893, + "grad_norm": 0.2868838310241699, + "learning_rate": 6.480628590466857e-05, + "loss": 0.0073, + "step": 43410 + }, + { + "epoch": 3.3724271844660194, + "grad_norm": 0.28690046072006226, + "learning_rate": 6.479049199539074e-05, + "loss": 0.0074, + "step": 43420 + }, + { + "epoch": 3.3732038834951457, + "grad_norm": 0.4579726457595825, + "learning_rate": 6.477469646864693e-05, + "loss": 0.0104, + "step": 43430 + }, + { + "epoch": 3.373980582524272, + "grad_norm": 0.25222542881965637, + "learning_rate": 6.475889932616454e-05, + "loss": 0.0074, + "step": 43440 + }, + { + "epoch": 3.374757281553398, + "grad_norm": 0.38951700925827026, + "learning_rate": 6.474310056967111e-05, + "loss": 0.0104, + "step": 43450 + }, + { + "epoch": 3.3755339805825244, + "grad_norm": 0.36322903633117676, + "learning_rate": 6.472730020089437e-05, + "loss": 0.0066, + "step": 43460 + }, + { + "epoch": 3.3763106796116507, + "grad_norm": 0.3141816258430481, + "learning_rate": 6.471149822156222e-05, + "loss": 0.0092, + "step": 43470 + }, + { + "epoch": 3.377087378640777, + "grad_norm": 0.36826837062835693, + "learning_rate": 6.469569463340275e-05, + "loss": 0.0084, + "step": 43480 + }, + { + "epoch": 3.377864077669903, + "grad_norm": 0.3715125024318695, + "learning_rate": 6.467988943814418e-05, + "loss": 0.0083, + "step": 43490 + }, + { + "epoch": 3.378640776699029, + "grad_norm": 0.245940163731575, + "learning_rate": 6.4664082637515e-05, + "loss": 0.0075, + "step": 43500 + }, + { + "epoch": 3.379417475728155, + "grad_norm": 0.27343589067459106, + "learning_rate": 6.464827423324377e-05, + "loss": 0.0079, + "step": 43510 + }, + { + "epoch": 3.3801941747572815, + "grad_norm": 0.4258023202419281, + "learning_rate": 6.463246422705931e-05, + "loss": 0.0079, + "step": 43520 + }, + { + "epoch": 3.3809708737864077, + "grad_norm": 0.3540845215320587, + "learning_rate": 6.461665262069055e-05, + "loss": 0.0074, + "step": 43530 + }, + { + "epoch": 3.381747572815534, + "grad_norm": 0.3430729806423187, + "learning_rate": 6.460083941586665e-05, + "loss": 0.0072, + "step": 43540 + }, + { + "epoch": 3.38252427184466, + "grad_norm": 0.4281177520751953, + "learning_rate": 6.45850246143169e-05, + "loss": 0.0091, + "step": 43550 + }, + { + "epoch": 3.3833009708737865, + "grad_norm": 0.5156943798065186, + "learning_rate": 6.456920821777077e-05, + "loss": 0.0102, + "step": 43560 + }, + { + "epoch": 3.3840776699029127, + "grad_norm": 0.1551656424999237, + "learning_rate": 6.455339022795796e-05, + "loss": 0.0044, + "step": 43570 + }, + { + "epoch": 3.384854368932039, + "grad_norm": 0.5093649625778198, + "learning_rate": 6.453757064660825e-05, + "loss": 0.0082, + "step": 43580 + }, + { + "epoch": 3.385631067961165, + "grad_norm": 0.13958197832107544, + "learning_rate": 6.452174947545169e-05, + "loss": 0.0066, + "step": 43590 + }, + { + "epoch": 3.3864077669902914, + "grad_norm": 0.17214848101139069, + "learning_rate": 6.450592671621842e-05, + "loss": 0.0063, + "step": 43600 + }, + { + "epoch": 3.3871844660194173, + "grad_norm": 0.38107895851135254, + "learning_rate": 6.449010237063882e-05, + "loss": 0.0085, + "step": 43610 + }, + { + "epoch": 3.3879611650485435, + "grad_norm": 0.24810074269771576, + "learning_rate": 6.447427644044343e-05, + "loss": 0.007, + "step": 43620 + }, + { + "epoch": 3.3887378640776697, + "grad_norm": 0.24742382764816284, + "learning_rate": 6.44584489273629e-05, + "loss": 0.007, + "step": 43630 + }, + { + "epoch": 3.389514563106796, + "grad_norm": 0.322051078081131, + "learning_rate": 6.444261983312817e-05, + "loss": 0.006, + "step": 43640 + }, + { + "epoch": 3.3902912621359222, + "grad_norm": 0.29147255420684814, + "learning_rate": 6.442678915947023e-05, + "loss": 0.0074, + "step": 43650 + }, + { + "epoch": 3.3910679611650485, + "grad_norm": 0.2035292088985443, + "learning_rate": 6.441095690812032e-05, + "loss": 0.0071, + "step": 43660 + }, + { + "epoch": 3.3918446601941747, + "grad_norm": 0.41788482666015625, + "learning_rate": 6.439512308080982e-05, + "loss": 0.0092, + "step": 43670 + }, + { + "epoch": 3.392621359223301, + "grad_norm": 0.23436202108860016, + "learning_rate": 6.43792876792703e-05, + "loss": 0.0079, + "step": 43680 + }, + { + "epoch": 3.3933980582524272, + "grad_norm": 0.3074994683265686, + "learning_rate": 6.43634507052335e-05, + "loss": 0.0084, + "step": 43690 + }, + { + "epoch": 3.3941747572815535, + "grad_norm": 0.16377118229866028, + "learning_rate": 6.434761216043135e-05, + "loss": 0.0088, + "step": 43700 + }, + { + "epoch": 3.3949514563106797, + "grad_norm": 0.2503325641155243, + "learning_rate": 6.433177204659589e-05, + "loss": 0.008, + "step": 43710 + }, + { + "epoch": 3.395728155339806, + "grad_norm": 0.1501058042049408, + "learning_rate": 6.431593036545938e-05, + "loss": 0.0071, + "step": 43720 + }, + { + "epoch": 3.3965048543689322, + "grad_norm": 0.304055780172348, + "learning_rate": 6.430008711875426e-05, + "loss": 0.0097, + "step": 43730 + }, + { + "epoch": 3.3972815533980585, + "grad_norm": 0.4118429124355316, + "learning_rate": 6.42842423082131e-05, + "loss": 0.0073, + "step": 43740 + }, + { + "epoch": 3.3980582524271843, + "grad_norm": 0.23404249548912048, + "learning_rate": 6.426839593556868e-05, + "loss": 0.0083, + "step": 43750 + }, + { + "epoch": 3.3988349514563105, + "grad_norm": 0.2557471692562103, + "learning_rate": 6.425254800255392e-05, + "loss": 0.0088, + "step": 43760 + }, + { + "epoch": 3.399611650485437, + "grad_norm": 0.5577152371406555, + "learning_rate": 6.423669851090196e-05, + "loss": 0.0058, + "step": 43770 + }, + { + "epoch": 3.400388349514563, + "grad_norm": 0.28074342012405396, + "learning_rate": 6.422084746234604e-05, + "loss": 0.007, + "step": 43780 + }, + { + "epoch": 3.4011650485436893, + "grad_norm": 0.17797185480594635, + "learning_rate": 6.420499485861965e-05, + "loss": 0.0057, + "step": 43790 + }, + { + "epoch": 3.4019417475728155, + "grad_norm": 0.1257181316614151, + "learning_rate": 6.418914070145636e-05, + "loss": 0.0069, + "step": 43800 + }, + { + "epoch": 3.402718446601942, + "grad_norm": 0.14664806425571442, + "learning_rate": 6.417328499258996e-05, + "loss": 0.0061, + "step": 43810 + }, + { + "epoch": 3.403495145631068, + "grad_norm": 0.2139866203069687, + "learning_rate": 6.415742773375446e-05, + "loss": 0.0082, + "step": 43820 + }, + { + "epoch": 3.4042718446601943, + "grad_norm": 0.33746859431266785, + "learning_rate": 6.414156892668393e-05, + "loss": 0.0073, + "step": 43830 + }, + { + "epoch": 3.4050485436893205, + "grad_norm": 0.1857280284166336, + "learning_rate": 6.412570857311267e-05, + "loss": 0.0064, + "step": 43840 + }, + { + "epoch": 3.4058252427184468, + "grad_norm": 0.26170051097869873, + "learning_rate": 6.410984667477518e-05, + "loss": 0.009, + "step": 43850 + }, + { + "epoch": 3.4066019417475726, + "grad_norm": 0.22497954964637756, + "learning_rate": 6.409398323340607e-05, + "loss": 0.0076, + "step": 43860 + }, + { + "epoch": 3.407378640776699, + "grad_norm": 0.4608434736728668, + "learning_rate": 6.407811825074012e-05, + "loss": 0.0111, + "step": 43870 + }, + { + "epoch": 3.408155339805825, + "grad_norm": 0.19715692102909088, + "learning_rate": 6.406225172851234e-05, + "loss": 0.0085, + "step": 43880 + }, + { + "epoch": 3.4089320388349513, + "grad_norm": 0.24343210458755493, + "learning_rate": 6.404638366845786e-05, + "loss": 0.0068, + "step": 43890 + }, + { + "epoch": 3.4097087378640776, + "grad_norm": 0.1877724528312683, + "learning_rate": 6.403051407231196e-05, + "loss": 0.0073, + "step": 43900 + }, + { + "epoch": 3.410485436893204, + "grad_norm": 0.4415053725242615, + "learning_rate": 6.401464294181016e-05, + "loss": 0.0054, + "step": 43910 + }, + { + "epoch": 3.41126213592233, + "grad_norm": 1.2428110837936401, + "learning_rate": 6.399877027868808e-05, + "loss": 0.0076, + "step": 43920 + }, + { + "epoch": 3.4120388349514563, + "grad_norm": 0.24176166951656342, + "learning_rate": 6.39828960846815e-05, + "loss": 0.007, + "step": 43930 + }, + { + "epoch": 3.4128155339805826, + "grad_norm": 0.2112484574317932, + "learning_rate": 6.396702036152644e-05, + "loss": 0.0109, + "step": 43940 + }, + { + "epoch": 3.413592233009709, + "grad_norm": 0.4638822376728058, + "learning_rate": 6.395114311095905e-05, + "loss": 0.009, + "step": 43950 + }, + { + "epoch": 3.414368932038835, + "grad_norm": 0.4210664927959442, + "learning_rate": 6.393526433471559e-05, + "loss": 0.008, + "step": 43960 + }, + { + "epoch": 3.4151456310679613, + "grad_norm": 0.2791820466518402, + "learning_rate": 6.391938403453262e-05, + "loss": 0.009, + "step": 43970 + }, + { + "epoch": 3.4159223300970876, + "grad_norm": 0.20826973021030426, + "learning_rate": 6.390350221214671e-05, + "loss": 0.0067, + "step": 43980 + }, + { + "epoch": 3.416699029126214, + "grad_norm": 0.3976549506187439, + "learning_rate": 6.38876188692947e-05, + "loss": 0.009, + "step": 43990 + }, + { + "epoch": 3.4174757281553396, + "grad_norm": 0.2906109392642975, + "learning_rate": 6.387173400771358e-05, + "loss": 0.0084, + "step": 44000 + }, + { + "epoch": 3.418252427184466, + "grad_norm": 0.23923589289188385, + "learning_rate": 6.385584762914047e-05, + "loss": 0.0061, + "step": 44010 + }, + { + "epoch": 3.419029126213592, + "grad_norm": 0.274291455745697, + "learning_rate": 6.38399597353127e-05, + "loss": 0.0072, + "step": 44020 + }, + { + "epoch": 3.4198058252427184, + "grad_norm": 0.5721265077590942, + "learning_rate": 6.382407032796775e-05, + "loss": 0.0076, + "step": 44030 + }, + { + "epoch": 3.4205825242718446, + "grad_norm": 0.27241694927215576, + "learning_rate": 6.380817940884325e-05, + "loss": 0.006, + "step": 44040 + }, + { + "epoch": 3.421359223300971, + "grad_norm": 0.23358087241649628, + "learning_rate": 6.379228697967702e-05, + "loss": 0.0073, + "step": 44050 + }, + { + "epoch": 3.422135922330097, + "grad_norm": 0.16666923463344574, + "learning_rate": 6.3776393042207e-05, + "loss": 0.0065, + "step": 44060 + }, + { + "epoch": 3.4229126213592234, + "grad_norm": 1.1512279510498047, + "learning_rate": 6.376049759817137e-05, + "loss": 0.0122, + "step": 44070 + }, + { + "epoch": 3.4236893203883496, + "grad_norm": 0.3753276765346527, + "learning_rate": 6.37446006493084e-05, + "loss": 0.0064, + "step": 44080 + }, + { + "epoch": 3.424466019417476, + "grad_norm": 0.2924320697784424, + "learning_rate": 6.372870219735656e-05, + "loss": 0.0093, + "step": 44090 + }, + { + "epoch": 3.425242718446602, + "grad_norm": 0.28812116384506226, + "learning_rate": 6.371280224405452e-05, + "loss": 0.0088, + "step": 44100 + }, + { + "epoch": 3.426019417475728, + "grad_norm": 0.25518518686294556, + "learning_rate": 6.369690079114101e-05, + "loss": 0.0077, + "step": 44110 + }, + { + "epoch": 3.426796116504854, + "grad_norm": 0.13424164056777954, + "learning_rate": 6.368099784035504e-05, + "loss": 0.0061, + "step": 44120 + }, + { + "epoch": 3.4275728155339804, + "grad_norm": 0.22844377160072327, + "learning_rate": 6.366509339343572e-05, + "loss": 0.0053, + "step": 44130 + }, + { + "epoch": 3.4283495145631067, + "grad_norm": 0.20656102895736694, + "learning_rate": 6.364918745212232e-05, + "loss": 0.0095, + "step": 44140 + }, + { + "epoch": 3.429126213592233, + "grad_norm": 0.24624402821063995, + "learning_rate": 6.363328001815431e-05, + "loss": 0.0077, + "step": 44150 + }, + { + "epoch": 3.429902912621359, + "grad_norm": 0.24911287426948547, + "learning_rate": 6.361737109327128e-05, + "loss": 0.007, + "step": 44160 + }, + { + "epoch": 3.4306796116504854, + "grad_norm": 0.23704712092876434, + "learning_rate": 6.360146067921301e-05, + "loss": 0.009, + "step": 44170 + }, + { + "epoch": 3.4314563106796117, + "grad_norm": 0.17893561720848083, + "learning_rate": 6.35855487777195e-05, + "loss": 0.0079, + "step": 44180 + }, + { + "epoch": 3.432233009708738, + "grad_norm": 0.23432482779026031, + "learning_rate": 6.356963539053076e-05, + "loss": 0.0056, + "step": 44190 + }, + { + "epoch": 3.433009708737864, + "grad_norm": 0.14090770483016968, + "learning_rate": 6.355372051938711e-05, + "loss": 0.0042, + "step": 44200 + }, + { + "epoch": 3.4337864077669904, + "grad_norm": 0.1144217699766159, + "learning_rate": 6.353780416602894e-05, + "loss": 0.0043, + "step": 44210 + }, + { + "epoch": 3.4345631067961166, + "grad_norm": 0.2246111035346985, + "learning_rate": 6.352188633219689e-05, + "loss": 0.0077, + "step": 44220 + }, + { + "epoch": 3.435339805825243, + "grad_norm": 0.29449108242988586, + "learning_rate": 6.350596701963166e-05, + "loss": 0.0064, + "step": 44230 + }, + { + "epoch": 3.436116504854369, + "grad_norm": 1.1395963430404663, + "learning_rate": 6.349004623007419e-05, + "loss": 0.0076, + "step": 44240 + }, + { + "epoch": 3.436893203883495, + "grad_norm": 0.24663646519184113, + "learning_rate": 6.347412396526555e-05, + "loss": 0.0072, + "step": 44250 + }, + { + "epoch": 3.437669902912621, + "grad_norm": 0.2557012438774109, + "learning_rate": 6.345820022694696e-05, + "loss": 0.0086, + "step": 44260 + }, + { + "epoch": 3.4384466019417474, + "grad_norm": 0.3082277178764343, + "learning_rate": 6.344227501685984e-05, + "loss": 0.0063, + "step": 44270 + }, + { + "epoch": 3.4392233009708737, + "grad_norm": 0.6652493476867676, + "learning_rate": 6.342634833674572e-05, + "loss": 0.0111, + "step": 44280 + }, + { + "epoch": 3.44, + "grad_norm": 0.41706180572509766, + "learning_rate": 6.341042018834635e-05, + "loss": 0.006, + "step": 44290 + }, + { + "epoch": 3.440776699029126, + "grad_norm": 0.20312264561653137, + "learning_rate": 6.339449057340359e-05, + "loss": 0.0074, + "step": 44300 + }, + { + "epoch": 3.4415533980582524, + "grad_norm": 0.32892152667045593, + "learning_rate": 6.337855949365945e-05, + "loss": 0.0094, + "step": 44310 + }, + { + "epoch": 3.4423300970873787, + "grad_norm": 0.17963780462741852, + "learning_rate": 6.336262695085619e-05, + "loss": 0.0065, + "step": 44320 + }, + { + "epoch": 3.443106796116505, + "grad_norm": 0.34325990080833435, + "learning_rate": 6.334669294673612e-05, + "loss": 0.009, + "step": 44330 + }, + { + "epoch": 3.443883495145631, + "grad_norm": 0.5100052952766418, + "learning_rate": 6.333075748304179e-05, + "loss": 0.0101, + "step": 44340 + }, + { + "epoch": 3.4446601941747574, + "grad_norm": 0.17510002851486206, + "learning_rate": 6.331482056151585e-05, + "loss": 0.0055, + "step": 44350 + }, + { + "epoch": 3.4454368932038837, + "grad_norm": 0.22003690898418427, + "learning_rate": 6.329888218390117e-05, + "loss": 0.0084, + "step": 44360 + }, + { + "epoch": 3.4462135922330095, + "grad_norm": 0.31101855635643005, + "learning_rate": 6.328294235194072e-05, + "loss": 0.0079, + "step": 44370 + }, + { + "epoch": 3.4469902912621357, + "grad_norm": 0.4108719229698181, + "learning_rate": 6.326700106737765e-05, + "loss": 0.0086, + "step": 44380 + }, + { + "epoch": 3.447766990291262, + "grad_norm": 0.24899683892726898, + "learning_rate": 6.325105833195531e-05, + "loss": 0.0092, + "step": 44390 + }, + { + "epoch": 3.4485436893203882, + "grad_norm": 0.44718295335769653, + "learning_rate": 6.323511414741715e-05, + "loss": 0.0061, + "step": 44400 + }, + { + "epoch": 3.4493203883495145, + "grad_norm": 0.924491286277771, + "learning_rate": 6.321916851550678e-05, + "loss": 0.0075, + "step": 44410 + }, + { + "epoch": 3.4500970873786407, + "grad_norm": 0.44026076793670654, + "learning_rate": 6.320322143796806e-05, + "loss": 0.0095, + "step": 44420 + }, + { + "epoch": 3.450873786407767, + "grad_norm": 0.2288869172334671, + "learning_rate": 6.318727291654488e-05, + "loss": 0.0091, + "step": 44430 + }, + { + "epoch": 3.4516504854368932, + "grad_norm": 0.43813562393188477, + "learning_rate": 6.317132295298134e-05, + "loss": 0.0052, + "step": 44440 + }, + { + "epoch": 3.4524271844660195, + "grad_norm": 0.1171521320939064, + "learning_rate": 6.315537154902173e-05, + "loss": 0.0054, + "step": 44450 + }, + { + "epoch": 3.4532038834951457, + "grad_norm": 0.22245396673679352, + "learning_rate": 6.313941870641048e-05, + "loss": 0.0075, + "step": 44460 + }, + { + "epoch": 3.453980582524272, + "grad_norm": 0.19935283064842224, + "learning_rate": 6.312346442689214e-05, + "loss": 0.0076, + "step": 44470 + }, + { + "epoch": 3.4547572815533982, + "grad_norm": 0.6443875432014465, + "learning_rate": 6.310750871221147e-05, + "loss": 0.0065, + "step": 44480 + }, + { + "epoch": 3.4555339805825245, + "grad_norm": 0.30841782689094543, + "learning_rate": 6.309155156411335e-05, + "loss": 0.0065, + "step": 44490 + }, + { + "epoch": 3.4563106796116507, + "grad_norm": 0.2589596211910248, + "learning_rate": 6.307559298434284e-05, + "loss": 0.0068, + "step": 44500 + }, + { + "epoch": 3.4570873786407765, + "grad_norm": 0.2301298826932907, + "learning_rate": 6.305963297464512e-05, + "loss": 0.0054, + "step": 44510 + }, + { + "epoch": 3.457864077669903, + "grad_norm": 0.5212901830673218, + "learning_rate": 6.304367153676561e-05, + "loss": 0.0067, + "step": 44520 + }, + { + "epoch": 3.458640776699029, + "grad_norm": 0.284749299287796, + "learning_rate": 6.302770867244978e-05, + "loss": 0.0083, + "step": 44530 + }, + { + "epoch": 3.4594174757281553, + "grad_norm": 0.16085569560527802, + "learning_rate": 6.301174438344328e-05, + "loss": 0.0048, + "step": 44540 + }, + { + "epoch": 3.4601941747572815, + "grad_norm": 0.2547130584716797, + "learning_rate": 6.299577867149202e-05, + "loss": 0.0079, + "step": 44550 + }, + { + "epoch": 3.4609708737864078, + "grad_norm": 0.17349888384342194, + "learning_rate": 6.297981153834192e-05, + "loss": 0.0065, + "step": 44560 + }, + { + "epoch": 3.461747572815534, + "grad_norm": 0.16758953034877777, + "learning_rate": 6.296384298573916e-05, + "loss": 0.0062, + "step": 44570 + }, + { + "epoch": 3.4625242718446603, + "grad_norm": 0.13781151175498962, + "learning_rate": 6.294787301543001e-05, + "loss": 0.0072, + "step": 44580 + }, + { + "epoch": 3.4633009708737865, + "grad_norm": 0.3197147250175476, + "learning_rate": 6.293190162916095e-05, + "loss": 0.013, + "step": 44590 + }, + { + "epoch": 3.4640776699029128, + "grad_norm": 0.6852118968963623, + "learning_rate": 6.291592882867855e-05, + "loss": 0.0082, + "step": 44600 + }, + { + "epoch": 3.464854368932039, + "grad_norm": 0.25162380933761597, + "learning_rate": 6.28999546157296e-05, + "loss": 0.0063, + "step": 44610 + }, + { + "epoch": 3.465631067961165, + "grad_norm": 0.13680513203144073, + "learning_rate": 6.288397899206102e-05, + "loss": 0.0085, + "step": 44620 + }, + { + "epoch": 3.466407766990291, + "grad_norm": 0.19252152740955353, + "learning_rate": 6.286800195941984e-05, + "loss": 0.0065, + "step": 44630 + }, + { + "epoch": 3.4671844660194173, + "grad_norm": 0.31445059180259705, + "learning_rate": 6.285202351955334e-05, + "loss": 0.008, + "step": 44640 + }, + { + "epoch": 3.4679611650485436, + "grad_norm": 0.17462658882141113, + "learning_rate": 6.283604367420887e-05, + "loss": 0.0069, + "step": 44650 + }, + { + "epoch": 3.46873786407767, + "grad_norm": 0.33470165729522705, + "learning_rate": 6.282006242513394e-05, + "loss": 0.0098, + "step": 44660 + }, + { + "epoch": 3.469514563106796, + "grad_norm": 0.18157321214675903, + "learning_rate": 6.280407977407628e-05, + "loss": 0.0107, + "step": 44670 + }, + { + "epoch": 3.4702912621359223, + "grad_norm": 0.44378790259361267, + "learning_rate": 6.27880957227837e-05, + "loss": 0.0095, + "step": 44680 + }, + { + "epoch": 3.4710679611650486, + "grad_norm": 0.35654139518737793, + "learning_rate": 6.27721102730042e-05, + "loss": 0.0109, + "step": 44690 + }, + { + "epoch": 3.471844660194175, + "grad_norm": 0.23684623837471008, + "learning_rate": 6.275612342648592e-05, + "loss": 0.0094, + "step": 44700 + }, + { + "epoch": 3.472621359223301, + "grad_norm": 0.11863391101360321, + "learning_rate": 6.274013518497716e-05, + "loss": 0.0071, + "step": 44710 + }, + { + "epoch": 3.4733980582524273, + "grad_norm": 0.3603357970714569, + "learning_rate": 6.272414555022636e-05, + "loss": 0.0071, + "step": 44720 + }, + { + "epoch": 3.4741747572815536, + "grad_norm": 0.28642281889915466, + "learning_rate": 6.270815452398215e-05, + "loss": 0.0073, + "step": 44730 + }, + { + "epoch": 3.47495145631068, + "grad_norm": 0.16811241209506989, + "learning_rate": 6.269216210799326e-05, + "loss": 0.0069, + "step": 44740 + }, + { + "epoch": 3.475728155339806, + "grad_norm": 0.3611561357975006, + "learning_rate": 6.26761683040086e-05, + "loss": 0.0074, + "step": 44750 + }, + { + "epoch": 3.476504854368932, + "grad_norm": 0.4094214141368866, + "learning_rate": 6.266017311377723e-05, + "loss": 0.0064, + "step": 44760 + }, + { + "epoch": 3.477281553398058, + "grad_norm": 0.4016675055027008, + "learning_rate": 6.264417653904839e-05, + "loss": 0.007, + "step": 44770 + }, + { + "epoch": 3.4780582524271844, + "grad_norm": 0.21231426298618317, + "learning_rate": 6.262817858157139e-05, + "loss": 0.0089, + "step": 44780 + }, + { + "epoch": 3.4788349514563106, + "grad_norm": 0.24151821434497833, + "learning_rate": 6.261217924309576e-05, + "loss": 0.0076, + "step": 44790 + }, + { + "epoch": 3.479611650485437, + "grad_norm": 0.2332872599363327, + "learning_rate": 6.259617852537118e-05, + "loss": 0.0108, + "step": 44800 + }, + { + "epoch": 3.480388349514563, + "grad_norm": 0.28382202982902527, + "learning_rate": 6.258017643014747e-05, + "loss": 0.01, + "step": 44810 + }, + { + "epoch": 3.4811650485436894, + "grad_norm": 0.1630265861749649, + "learning_rate": 6.256417295917456e-05, + "loss": 0.0087, + "step": 44820 + }, + { + "epoch": 3.4819417475728156, + "grad_norm": 0.2505657374858856, + "learning_rate": 6.254816811420258e-05, + "loss": 0.0103, + "step": 44830 + }, + { + "epoch": 3.482718446601942, + "grad_norm": 0.2966558635234833, + "learning_rate": 6.253216189698183e-05, + "loss": 0.0082, + "step": 44840 + }, + { + "epoch": 3.483495145631068, + "grad_norm": 0.3139156997203827, + "learning_rate": 6.251615430926267e-05, + "loss": 0.0096, + "step": 44850 + }, + { + "epoch": 3.4842718446601944, + "grad_norm": 0.23847444355487823, + "learning_rate": 6.25001453527957e-05, + "loss": 0.0065, + "step": 44860 + }, + { + "epoch": 3.48504854368932, + "grad_norm": 0.39596959948539734, + "learning_rate": 6.248413502933164e-05, + "loss": 0.0086, + "step": 44870 + }, + { + "epoch": 3.4858252427184464, + "grad_norm": 0.21842029690742493, + "learning_rate": 6.246812334062133e-05, + "loss": 0.0081, + "step": 44880 + }, + { + "epoch": 3.4866019417475727, + "grad_norm": 0.17769357562065125, + "learning_rate": 6.245211028841579e-05, + "loss": 0.0091, + "step": 44890 + }, + { + "epoch": 3.487378640776699, + "grad_norm": 0.23019452393054962, + "learning_rate": 6.24360958744662e-05, + "loss": 0.0076, + "step": 44900 + }, + { + "epoch": 3.488155339805825, + "grad_norm": 0.14968550205230713, + "learning_rate": 6.242008010052387e-05, + "loss": 0.0146, + "step": 44910 + }, + { + "epoch": 3.4889320388349514, + "grad_norm": 0.916192889213562, + "learning_rate": 6.240406296834024e-05, + "loss": 0.0098, + "step": 44920 + }, + { + "epoch": 3.4897087378640776, + "grad_norm": 0.23978295922279358, + "learning_rate": 6.238804447966694e-05, + "loss": 0.0077, + "step": 44930 + }, + { + "epoch": 3.490485436893204, + "grad_norm": 0.24115005135536194, + "learning_rate": 6.237202463625573e-05, + "loss": 0.0077, + "step": 44940 + }, + { + "epoch": 3.49126213592233, + "grad_norm": 0.3244486153125763, + "learning_rate": 6.235600343985848e-05, + "loss": 0.0084, + "step": 44950 + }, + { + "epoch": 3.4920388349514564, + "grad_norm": 0.17826752364635468, + "learning_rate": 6.233998089222729e-05, + "loss": 0.0073, + "step": 44960 + }, + { + "epoch": 3.4928155339805826, + "grad_norm": 0.23594258725643158, + "learning_rate": 6.232395699511433e-05, + "loss": 0.0079, + "step": 44970 + }, + { + "epoch": 3.493592233009709, + "grad_norm": 0.17932051420211792, + "learning_rate": 6.230793175027199e-05, + "loss": 0.0078, + "step": 44980 + }, + { + "epoch": 3.494368932038835, + "grad_norm": 0.477548211812973, + "learning_rate": 6.22919051594527e-05, + "loss": 0.0088, + "step": 44990 + }, + { + "epoch": 3.4951456310679614, + "grad_norm": 0.15568050742149353, + "learning_rate": 6.227587722440917e-05, + "loss": 0.0092, + "step": 45000 + }, + { + "epoch": 3.495922330097087, + "grad_norm": 0.26811739802360535, + "learning_rate": 6.225984794689414e-05, + "loss": 0.008, + "step": 45010 + }, + { + "epoch": 3.4966990291262134, + "grad_norm": 0.2556384205818176, + "learning_rate": 6.22438173286606e-05, + "loss": 0.0085, + "step": 45020 + }, + { + "epoch": 3.4974757281553397, + "grad_norm": 0.36281818151474, + "learning_rate": 6.222778537146159e-05, + "loss": 0.008, + "step": 45030 + }, + { + "epoch": 3.498252427184466, + "grad_norm": 0.28732770681381226, + "learning_rate": 6.221175207705037e-05, + "loss": 0.0089, + "step": 45040 + }, + { + "epoch": 3.499029126213592, + "grad_norm": 0.22243045270442963, + "learning_rate": 6.219571744718027e-05, + "loss": 0.0076, + "step": 45050 + }, + { + "epoch": 3.4998058252427184, + "grad_norm": 0.23083335161209106, + "learning_rate": 6.217968148360487e-05, + "loss": 0.0079, + "step": 45060 + }, + { + "epoch": 3.5005825242718447, + "grad_norm": 0.2016742080450058, + "learning_rate": 6.216364418807782e-05, + "loss": 0.0095, + "step": 45070 + }, + { + "epoch": 3.501359223300971, + "grad_norm": 0.14170271158218384, + "learning_rate": 6.214760556235292e-05, + "loss": 0.0063, + "step": 45080 + }, + { + "epoch": 3.502135922330097, + "grad_norm": 0.8855358362197876, + "learning_rate": 6.213156560818414e-05, + "loss": 0.009, + "step": 45090 + }, + { + "epoch": 3.5029126213592234, + "grad_norm": 0.1886293888092041, + "learning_rate": 6.211552432732559e-05, + "loss": 0.0057, + "step": 45100 + }, + { + "epoch": 3.5036893203883492, + "grad_norm": 0.9286242127418518, + "learning_rate": 6.20994817215315e-05, + "loss": 0.0073, + "step": 45110 + }, + { + "epoch": 3.5044660194174755, + "grad_norm": 0.464047372341156, + "learning_rate": 6.208343779255628e-05, + "loss": 0.0063, + "step": 45120 + }, + { + "epoch": 3.5052427184466017, + "grad_norm": 0.38643208146095276, + "learning_rate": 6.206739254215449e-05, + "loss": 0.0102, + "step": 45130 + }, + { + "epoch": 3.506019417475728, + "grad_norm": 0.2994089424610138, + "learning_rate": 6.205134597208077e-05, + "loss": 0.0086, + "step": 45140 + }, + { + "epoch": 3.5067961165048542, + "grad_norm": 0.23428507149219513, + "learning_rate": 6.203529808409e-05, + "loss": 0.006, + "step": 45150 + }, + { + "epoch": 3.5075728155339805, + "grad_norm": 0.19198381900787354, + "learning_rate": 6.201924887993712e-05, + "loss": 0.0092, + "step": 45160 + }, + { + "epoch": 3.5083495145631067, + "grad_norm": 0.20199519395828247, + "learning_rate": 6.200319836137725e-05, + "loss": 0.0075, + "step": 45170 + }, + { + "epoch": 3.509126213592233, + "grad_norm": 0.2040439248085022, + "learning_rate": 6.198714653016565e-05, + "loss": 0.0088, + "step": 45180 + }, + { + "epoch": 3.5099029126213592, + "grad_norm": 0.3174402415752411, + "learning_rate": 6.197109338805774e-05, + "loss": 0.0084, + "step": 45190 + }, + { + "epoch": 3.5106796116504855, + "grad_norm": 0.20328426361083984, + "learning_rate": 6.195503893680903e-05, + "loss": 0.0061, + "step": 45200 + }, + { + "epoch": 3.5114563106796117, + "grad_norm": 0.21891650557518005, + "learning_rate": 6.193898317817524e-05, + "loss": 0.0103, + "step": 45210 + }, + { + "epoch": 3.512233009708738, + "grad_norm": 0.17950738966464996, + "learning_rate": 6.192292611391221e-05, + "loss": 0.0079, + "step": 45220 + }, + { + "epoch": 3.5130097087378642, + "grad_norm": 0.26746582984924316, + "learning_rate": 6.190686774577591e-05, + "loss": 0.0088, + "step": 45230 + }, + { + "epoch": 3.5137864077669905, + "grad_norm": 0.2186070680618286, + "learning_rate": 6.189080807552245e-05, + "loss": 0.008, + "step": 45240 + }, + { + "epoch": 3.5145631067961167, + "grad_norm": 0.4497990310192108, + "learning_rate": 6.187474710490809e-05, + "loss": 0.0089, + "step": 45250 + }, + { + "epoch": 3.515339805825243, + "grad_norm": 0.14913465082645416, + "learning_rate": 6.185868483568926e-05, + "loss": 0.0067, + "step": 45260 + }, + { + "epoch": 3.5161165048543688, + "grad_norm": 0.3035331964492798, + "learning_rate": 6.184262126962245e-05, + "loss": 0.0089, + "step": 45270 + }, + { + "epoch": 3.516893203883495, + "grad_norm": 0.18529663980007172, + "learning_rate": 6.182655640846442e-05, + "loss": 0.0074, + "step": 45280 + }, + { + "epoch": 3.5176699029126213, + "grad_norm": 0.5491353273391724, + "learning_rate": 6.181049025397196e-05, + "loss": 0.009, + "step": 45290 + }, + { + "epoch": 3.5184466019417475, + "grad_norm": 0.205623596906662, + "learning_rate": 6.179442280790202e-05, + "loss": 0.0091, + "step": 45300 + }, + { + "epoch": 3.5192233009708738, + "grad_norm": 0.15263672173023224, + "learning_rate": 6.177835407201174e-05, + "loss": 0.0081, + "step": 45310 + }, + { + "epoch": 3.52, + "grad_norm": 0.3444114029407501, + "learning_rate": 6.176228404805839e-05, + "loss": 0.0066, + "step": 45320 + }, + { + "epoch": 3.5207766990291263, + "grad_norm": 0.2493935227394104, + "learning_rate": 6.174621273779932e-05, + "loss": 0.0078, + "step": 45330 + }, + { + "epoch": 3.5215533980582525, + "grad_norm": 0.7044627666473389, + "learning_rate": 6.17301401429921e-05, + "loss": 0.0068, + "step": 45340 + }, + { + "epoch": 3.5223300970873788, + "grad_norm": 0.18648678064346313, + "learning_rate": 6.171406626539436e-05, + "loss": 0.0051, + "step": 45350 + }, + { + "epoch": 3.523106796116505, + "grad_norm": 0.20677898824214935, + "learning_rate": 6.169799110676398e-05, + "loss": 0.007, + "step": 45360 + }, + { + "epoch": 3.523883495145631, + "grad_norm": 0.2611708641052246, + "learning_rate": 6.168191466885885e-05, + "loss": 0.0086, + "step": 45370 + }, + { + "epoch": 3.524660194174757, + "grad_norm": 0.43787139654159546, + "learning_rate": 6.166583695343711e-05, + "loss": 0.0117, + "step": 45380 + }, + { + "epoch": 3.5254368932038833, + "grad_norm": 0.2545941174030304, + "learning_rate": 6.164975796225698e-05, + "loss": 0.0072, + "step": 45390 + }, + { + "epoch": 3.5262135922330096, + "grad_norm": 0.2508987486362457, + "learning_rate": 6.163367769707683e-05, + "loss": 0.0075, + "step": 45400 + }, + { + "epoch": 3.526990291262136, + "grad_norm": 0.27970758080482483, + "learning_rate": 6.161759615965519e-05, + "loss": 0.0111, + "step": 45410 + }, + { + "epoch": 3.527766990291262, + "grad_norm": 0.28882914781570435, + "learning_rate": 6.16015133517507e-05, + "loss": 0.0075, + "step": 45420 + }, + { + "epoch": 3.5285436893203883, + "grad_norm": 0.30646783113479614, + "learning_rate": 6.158542927512214e-05, + "loss": 0.006, + "step": 45430 + }, + { + "epoch": 3.5293203883495146, + "grad_norm": 0.3399932086467743, + "learning_rate": 6.156934393152846e-05, + "loss": 0.007, + "step": 45440 + }, + { + "epoch": 3.530097087378641, + "grad_norm": 0.38972562551498413, + "learning_rate": 6.15532573227287e-05, + "loss": 0.0104, + "step": 45450 + }, + { + "epoch": 3.530873786407767, + "grad_norm": 0.17042411863803864, + "learning_rate": 6.153716945048212e-05, + "loss": 0.0106, + "step": 45460 + }, + { + "epoch": 3.5316504854368933, + "grad_norm": 0.24776355922222137, + "learning_rate": 6.152108031654802e-05, + "loss": 0.0093, + "step": 45470 + }, + { + "epoch": 3.5324271844660196, + "grad_norm": 0.43687689304351807, + "learning_rate": 6.15049899226859e-05, + "loss": 0.0068, + "step": 45480 + }, + { + "epoch": 3.533203883495146, + "grad_norm": 0.23411789536476135, + "learning_rate": 6.148889827065537e-05, + "loss": 0.0094, + "step": 45490 + }, + { + "epoch": 3.533980582524272, + "grad_norm": 0.4651491045951843, + "learning_rate": 6.147280536221622e-05, + "loss": 0.0091, + "step": 45500 + }, + { + "epoch": 3.5347572815533983, + "grad_norm": 0.21837441623210907, + "learning_rate": 6.145671119912832e-05, + "loss": 0.0068, + "step": 45510 + }, + { + "epoch": 3.5355339805825245, + "grad_norm": 0.2687418460845947, + "learning_rate": 6.144061578315169e-05, + "loss": 0.0062, + "step": 45520 + }, + { + "epoch": 3.5363106796116504, + "grad_norm": 0.15488874912261963, + "learning_rate": 6.142451911604654e-05, + "loss": 0.0062, + "step": 45530 + }, + { + "epoch": 3.5370873786407766, + "grad_norm": 0.22866834700107574, + "learning_rate": 6.140842119957315e-05, + "loss": 0.0064, + "step": 45540 + }, + { + "epoch": 3.537864077669903, + "grad_norm": 0.3463602364063263, + "learning_rate": 6.139232203549197e-05, + "loss": 0.0098, + "step": 45550 + }, + { + "epoch": 3.538640776699029, + "grad_norm": 0.22317342460155487, + "learning_rate": 6.137622162556357e-05, + "loss": 0.0081, + "step": 45560 + }, + { + "epoch": 3.5394174757281553, + "grad_norm": 0.1762794554233551, + "learning_rate": 6.13601199715487e-05, + "loss": 0.0079, + "step": 45570 + }, + { + "epoch": 3.5401941747572816, + "grad_norm": 0.18510319292545319, + "learning_rate": 6.134401707520816e-05, + "loss": 0.0065, + "step": 45580 + }, + { + "epoch": 3.540970873786408, + "grad_norm": 0.2445257604122162, + "learning_rate": 6.132791293830298e-05, + "loss": 0.0061, + "step": 45590 + }, + { + "epoch": 3.541747572815534, + "grad_norm": 0.33117157220840454, + "learning_rate": 6.131180756259428e-05, + "loss": 0.0061, + "step": 45600 + }, + { + "epoch": 3.5425242718446603, + "grad_norm": 0.34728047251701355, + "learning_rate": 6.129570094984331e-05, + "loss": 0.0064, + "step": 45610 + }, + { + "epoch": 3.543300970873786, + "grad_norm": 0.2804288864135742, + "learning_rate": 6.127959310181145e-05, + "loss": 0.0086, + "step": 45620 + }, + { + "epoch": 3.5440776699029124, + "grad_norm": 0.3937860131263733, + "learning_rate": 6.126348402026026e-05, + "loss": 0.009, + "step": 45630 + }, + { + "epoch": 3.5448543689320386, + "grad_norm": 0.1535976529121399, + "learning_rate": 6.12473737069514e-05, + "loss": 0.0067, + "step": 45640 + }, + { + "epoch": 3.545631067961165, + "grad_norm": 0.2772797644138336, + "learning_rate": 6.123126216364665e-05, + "loss": 0.0062, + "step": 45650 + }, + { + "epoch": 3.546407766990291, + "grad_norm": 0.1594782918691635, + "learning_rate": 6.121514939210797e-05, + "loss": 0.006, + "step": 45660 + }, + { + "epoch": 3.5471844660194174, + "grad_norm": 0.4014279246330261, + "learning_rate": 6.119903539409741e-05, + "loss": 0.0067, + "step": 45670 + }, + { + "epoch": 3.5479611650485436, + "grad_norm": 0.15599659085273743, + "learning_rate": 6.118292017137716e-05, + "loss": 0.0068, + "step": 45680 + }, + { + "epoch": 3.54873786407767, + "grad_norm": 0.22885562479496002, + "learning_rate": 6.116680372570959e-05, + "loss": 0.0085, + "step": 45690 + }, + { + "epoch": 3.549514563106796, + "grad_norm": 0.30605730414390564, + "learning_rate": 6.115068605885713e-05, + "loss": 0.0064, + "step": 45700 + }, + { + "epoch": 3.5502912621359224, + "grad_norm": 0.27197518944740295, + "learning_rate": 6.113456717258243e-05, + "loss": 0.0093, + "step": 45710 + }, + { + "epoch": 3.5510679611650486, + "grad_norm": 0.3224762976169586, + "learning_rate": 6.11184470686482e-05, + "loss": 0.0053, + "step": 45720 + }, + { + "epoch": 3.551844660194175, + "grad_norm": 0.296051025390625, + "learning_rate": 6.110232574881731e-05, + "loss": 0.0079, + "step": 45730 + }, + { + "epoch": 3.552621359223301, + "grad_norm": 0.14915384352207184, + "learning_rate": 6.108620321485277e-05, + "loss": 0.0056, + "step": 45740 + }, + { + "epoch": 3.5533980582524274, + "grad_norm": 0.32997068762779236, + "learning_rate": 6.107007946851773e-05, + "loss": 0.009, + "step": 45750 + }, + { + "epoch": 3.5541747572815536, + "grad_norm": 0.1825762242078781, + "learning_rate": 6.105395451157542e-05, + "loss": 0.0068, + "step": 45760 + }, + { + "epoch": 3.55495145631068, + "grad_norm": 0.1966678947210312, + "learning_rate": 6.103782834578928e-05, + "loss": 0.0078, + "step": 45770 + }, + { + "epoch": 3.5557281553398057, + "grad_norm": 0.2136584371328354, + "learning_rate": 6.102170097292281e-05, + "loss": 0.0069, + "step": 45780 + }, + { + "epoch": 3.556504854368932, + "grad_norm": 0.1620594710111618, + "learning_rate": 6.1005572394739716e-05, + "loss": 0.0091, + "step": 45790 + }, + { + "epoch": 3.557281553398058, + "grad_norm": 0.1480574607849121, + "learning_rate": 6.0989442613003765e-05, + "loss": 0.0092, + "step": 45800 + }, + { + "epoch": 3.5580582524271844, + "grad_norm": 0.20576679706573486, + "learning_rate": 6.097331162947888e-05, + "loss": 0.0088, + "step": 45810 + }, + { + "epoch": 3.5588349514563107, + "grad_norm": 0.4206257164478302, + "learning_rate": 6.095717944592914e-05, + "loss": 0.0068, + "step": 45820 + }, + { + "epoch": 3.559611650485437, + "grad_norm": 0.25504666566848755, + "learning_rate": 6.094104606411873e-05, + "loss": 0.0086, + "step": 45830 + }, + { + "epoch": 3.560388349514563, + "grad_norm": 0.3977965712547302, + "learning_rate": 6.0924911485811966e-05, + "loss": 0.0099, + "step": 45840 + }, + { + "epoch": 3.5611650485436894, + "grad_norm": 0.2832046151161194, + "learning_rate": 6.090877571277331e-05, + "loss": 0.0085, + "step": 45850 + }, + { + "epoch": 3.5619417475728157, + "grad_norm": 0.16791340708732605, + "learning_rate": 6.089263874676736e-05, + "loss": 0.0086, + "step": 45860 + }, + { + "epoch": 3.5627184466019415, + "grad_norm": 0.4830203950405121, + "learning_rate": 6.0876500589558796e-05, + "loss": 0.0073, + "step": 45870 + }, + { + "epoch": 3.5634951456310677, + "grad_norm": 0.23018132150173187, + "learning_rate": 6.086036124291248e-05, + "loss": 0.0058, + "step": 45880 + }, + { + "epoch": 3.564271844660194, + "grad_norm": 0.2749389111995697, + "learning_rate": 6.084422070859339e-05, + "loss": 0.009, + "step": 45890 + }, + { + "epoch": 3.5650485436893202, + "grad_norm": 0.6014438271522522, + "learning_rate": 6.082807898836663e-05, + "loss": 0.0073, + "step": 45900 + }, + { + "epoch": 3.5658252427184465, + "grad_norm": 0.40012019872665405, + "learning_rate": 6.081193608399742e-05, + "loss": 0.008, + "step": 45910 + }, + { + "epoch": 3.5666019417475727, + "grad_norm": 0.5847072601318359, + "learning_rate": 6.0795791997251164e-05, + "loss": 0.0075, + "step": 45920 + }, + { + "epoch": 3.567378640776699, + "grad_norm": 0.1784086972475052, + "learning_rate": 6.0779646729893294e-05, + "loss": 0.0062, + "step": 45930 + }, + { + "epoch": 3.568155339805825, + "grad_norm": 0.19705405831336975, + "learning_rate": 6.0763500283689476e-05, + "loss": 0.0067, + "step": 45940 + }, + { + "epoch": 3.5689320388349515, + "grad_norm": 0.8909388184547424, + "learning_rate": 6.0747352660405455e-05, + "loss": 0.0096, + "step": 45950 + }, + { + "epoch": 3.5697087378640777, + "grad_norm": 0.26308369636535645, + "learning_rate": 6.073120386180709e-05, + "loss": 0.0091, + "step": 45960 + }, + { + "epoch": 3.570485436893204, + "grad_norm": 0.19961291551589966, + "learning_rate": 6.0715053889660425e-05, + "loss": 0.0068, + "step": 45970 + }, + { + "epoch": 3.57126213592233, + "grad_norm": 0.2957819998264313, + "learning_rate": 6.069890274573157e-05, + "loss": 0.0068, + "step": 45980 + }, + { + "epoch": 3.5720388349514565, + "grad_norm": 0.22706863284111023, + "learning_rate": 6.068275043178679e-05, + "loss": 0.0077, + "step": 45990 + }, + { + "epoch": 3.5728155339805827, + "grad_norm": 0.18712922930717468, + "learning_rate": 6.066659694959248e-05, + "loss": 0.0061, + "step": 46000 + }, + { + "epoch": 3.573592233009709, + "grad_norm": 0.21056419610977173, + "learning_rate": 6.0650442300915176e-05, + "loss": 0.0056, + "step": 46010 + }, + { + "epoch": 3.574368932038835, + "grad_norm": 0.2735230326652527, + "learning_rate": 6.063428648752152e-05, + "loss": 0.0071, + "step": 46020 + }, + { + "epoch": 3.575145631067961, + "grad_norm": 0.1944059431552887, + "learning_rate": 6.0618129511178266e-05, + "loss": 0.0092, + "step": 46030 + }, + { + "epoch": 3.5759223300970873, + "grad_norm": 0.2147952914237976, + "learning_rate": 6.060197137365234e-05, + "loss": 0.0071, + "step": 46040 + }, + { + "epoch": 3.5766990291262135, + "grad_norm": 0.20216785371303558, + "learning_rate": 6.058581207671077e-05, + "loss": 0.0069, + "step": 46050 + }, + { + "epoch": 3.5774757281553398, + "grad_norm": 0.33574390411376953, + "learning_rate": 6.056965162212072e-05, + "loss": 0.0092, + "step": 46060 + }, + { + "epoch": 3.578252427184466, + "grad_norm": 0.1688910275697708, + "learning_rate": 6.055349001164943e-05, + "loss": 0.0068, + "step": 46070 + }, + { + "epoch": 3.5790291262135923, + "grad_norm": 0.21689939498901367, + "learning_rate": 6.0537327247064347e-05, + "loss": 0.0101, + "step": 46080 + }, + { + "epoch": 3.5798058252427185, + "grad_norm": 0.17193208634853363, + "learning_rate": 6.0521163330133014e-05, + "loss": 0.0056, + "step": 46090 + }, + { + "epoch": 3.5805825242718448, + "grad_norm": 0.3462528586387634, + "learning_rate": 6.050499826262306e-05, + "loss": 0.0088, + "step": 46100 + }, + { + "epoch": 3.581359223300971, + "grad_norm": 0.32876625657081604, + "learning_rate": 6.0488832046302294e-05, + "loss": 0.0084, + "step": 46110 + }, + { + "epoch": 3.582135922330097, + "grad_norm": 0.24176858365535736, + "learning_rate": 6.0472664682938626e-05, + "loss": 0.0061, + "step": 46120 + }, + { + "epoch": 3.582912621359223, + "grad_norm": 0.15590780973434448, + "learning_rate": 6.045649617430009e-05, + "loss": 0.0069, + "step": 46130 + }, + { + "epoch": 3.5836893203883493, + "grad_norm": 0.23931312561035156, + "learning_rate": 6.0440326522154866e-05, + "loss": 0.0069, + "step": 46140 + }, + { + "epoch": 3.5844660194174756, + "grad_norm": 0.2591603398323059, + "learning_rate": 6.0424155728271224e-05, + "loss": 0.0048, + "step": 46150 + }, + { + "epoch": 3.585242718446602, + "grad_norm": 0.1949755996465683, + "learning_rate": 6.040798379441758e-05, + "loss": 0.0079, + "step": 46160 + }, + { + "epoch": 3.586019417475728, + "grad_norm": 0.4029542803764343, + "learning_rate": 6.0391810722362485e-05, + "loss": 0.0079, + "step": 46170 + }, + { + "epoch": 3.5867961165048543, + "grad_norm": 0.21656924486160278, + "learning_rate": 6.037563651387458e-05, + "loss": 0.0068, + "step": 46180 + }, + { + "epoch": 3.5875728155339806, + "grad_norm": 0.3498605489730835, + "learning_rate": 6.0359461170722666e-05, + "loss": 0.0057, + "step": 46190 + }, + { + "epoch": 3.588349514563107, + "grad_norm": 0.31822219491004944, + "learning_rate": 6.034328469467566e-05, + "loss": 0.0092, + "step": 46200 + }, + { + "epoch": 3.589126213592233, + "grad_norm": 0.3214229941368103, + "learning_rate": 6.0327107087502596e-05, + "loss": 0.009, + "step": 46210 + }, + { + "epoch": 3.5899029126213593, + "grad_norm": 0.13838478922843933, + "learning_rate": 6.031092835097262e-05, + "loss": 0.0063, + "step": 46220 + }, + { + "epoch": 3.5906796116504855, + "grad_norm": 0.24785906076431274, + "learning_rate": 6.0294748486855024e-05, + "loss": 0.0082, + "step": 46230 + }, + { + "epoch": 3.591456310679612, + "grad_norm": 0.4202131927013397, + "learning_rate": 6.0278567496919216e-05, + "loss": 0.0071, + "step": 46240 + }, + { + "epoch": 3.592233009708738, + "grad_norm": 0.264703631401062, + "learning_rate": 6.026238538293472e-05, + "loss": 0.0083, + "step": 46250 + }, + { + "epoch": 3.5930097087378643, + "grad_norm": 0.224604994058609, + "learning_rate": 6.024620214667118e-05, + "loss": 0.0052, + "step": 46260 + }, + { + "epoch": 3.5937864077669905, + "grad_norm": 0.6040720343589783, + "learning_rate": 6.0230017789898384e-05, + "loss": 0.0072, + "step": 46270 + }, + { + "epoch": 3.5945631067961163, + "grad_norm": 0.2720598578453064, + "learning_rate": 6.021383231438622e-05, + "loss": 0.007, + "step": 46280 + }, + { + "epoch": 3.5953398058252426, + "grad_norm": 0.12544241547584534, + "learning_rate": 6.0197645721904704e-05, + "loss": 0.0085, + "step": 46290 + }, + { + "epoch": 3.596116504854369, + "grad_norm": 0.16644544899463654, + "learning_rate": 6.0181458014224e-05, + "loss": 0.0084, + "step": 46300 + }, + { + "epoch": 3.596893203883495, + "grad_norm": 0.17133429646492004, + "learning_rate": 6.016526919311435e-05, + "loss": 0.0082, + "step": 46310 + }, + { + "epoch": 3.5976699029126213, + "grad_norm": 0.22683030366897583, + "learning_rate": 6.014907926034613e-05, + "loss": 0.006, + "step": 46320 + }, + { + "epoch": 3.5984466019417476, + "grad_norm": 0.3485395908355713, + "learning_rate": 6.0132888217689875e-05, + "loss": 0.0081, + "step": 46330 + }, + { + "epoch": 3.599223300970874, + "grad_norm": 0.31998077034950256, + "learning_rate": 6.01166960669162e-05, + "loss": 0.0076, + "step": 46340 + }, + { + "epoch": 3.6, + "grad_norm": 0.2001885324716568, + "learning_rate": 6.0100502809795845e-05, + "loss": 0.0069, + "step": 46350 + }, + { + "epoch": 3.6007766990291263, + "grad_norm": 0.47302713990211487, + "learning_rate": 6.00843084480997e-05, + "loss": 0.0076, + "step": 46360 + }, + { + "epoch": 3.6015533980582526, + "grad_norm": 0.25209006667137146, + "learning_rate": 6.0068112983598736e-05, + "loss": 0.0065, + "step": 46370 + }, + { + "epoch": 3.6023300970873784, + "grad_norm": 0.21761766076087952, + "learning_rate": 6.005191641806407e-05, + "loss": 0.0075, + "step": 46380 + }, + { + "epoch": 3.6031067961165046, + "grad_norm": 0.14595849812030792, + "learning_rate": 6.003571875326694e-05, + "loss": 0.0076, + "step": 46390 + }, + { + "epoch": 3.603883495145631, + "grad_norm": 0.19280663132667542, + "learning_rate": 6.00195199909787e-05, + "loss": 0.0057, + "step": 46400 + }, + { + "epoch": 3.604660194174757, + "grad_norm": 0.22251251339912415, + "learning_rate": 6.000332013297082e-05, + "loss": 0.008, + "step": 46410 + }, + { + "epoch": 3.6054368932038834, + "grad_norm": 0.22919602692127228, + "learning_rate": 5.998711918101487e-05, + "loss": 0.0058, + "step": 46420 + }, + { + "epoch": 3.6062135922330096, + "grad_norm": 0.24640575051307678, + "learning_rate": 5.997091713688261e-05, + "loss": 0.006, + "step": 46430 + }, + { + "epoch": 3.606990291262136, + "grad_norm": 0.2111426591873169, + "learning_rate": 5.9954714002345836e-05, + "loss": 0.0089, + "step": 46440 + }, + { + "epoch": 3.607766990291262, + "grad_norm": 0.2122148871421814, + "learning_rate": 5.993850977917649e-05, + "loss": 0.0062, + "step": 46450 + }, + { + "epoch": 3.6085436893203884, + "grad_norm": 0.31056126952171326, + "learning_rate": 5.992230446914667e-05, + "loss": 0.0066, + "step": 46460 + }, + { + "epoch": 3.6093203883495146, + "grad_norm": 0.16177742183208466, + "learning_rate": 5.990609807402855e-05, + "loss": 0.0058, + "step": 46470 + }, + { + "epoch": 3.610097087378641, + "grad_norm": 0.23395615816116333, + "learning_rate": 5.988989059559443e-05, + "loss": 0.0073, + "step": 46480 + }, + { + "epoch": 3.610873786407767, + "grad_norm": 0.17486828565597534, + "learning_rate": 5.987368203561675e-05, + "loss": 0.0072, + "step": 46490 + }, + { + "epoch": 3.6116504854368934, + "grad_norm": 0.22156627476215363, + "learning_rate": 5.9857472395868055e-05, + "loss": 0.0065, + "step": 46500 + }, + { + "epoch": 3.6124271844660196, + "grad_norm": 0.19689249992370605, + "learning_rate": 5.9841261678120983e-05, + "loss": 0.0078, + "step": 46510 + }, + { + "epoch": 3.613203883495146, + "grad_norm": 0.36201322078704834, + "learning_rate": 5.982504988414834e-05, + "loss": 0.0058, + "step": 46520 + }, + { + "epoch": 3.6139805825242717, + "grad_norm": 0.64539635181427, + "learning_rate": 5.9808837015723015e-05, + "loss": 0.0061, + "step": 46530 + }, + { + "epoch": 3.614757281553398, + "grad_norm": 0.16416919231414795, + "learning_rate": 5.9792623074618016e-05, + "loss": 0.0086, + "step": 46540 + }, + { + "epoch": 3.615533980582524, + "grad_norm": 0.22258807718753815, + "learning_rate": 5.977640806260648e-05, + "loss": 0.0098, + "step": 46550 + }, + { + "epoch": 3.6163106796116504, + "grad_norm": 0.30478787422180176, + "learning_rate": 5.976019198146166e-05, + "loss": 0.0135, + "step": 46560 + }, + { + "epoch": 3.6170873786407767, + "grad_norm": 0.1962708681821823, + "learning_rate": 5.974397483295692e-05, + "loss": 0.0087, + "step": 46570 + }, + { + "epoch": 3.617864077669903, + "grad_norm": 0.28479957580566406, + "learning_rate": 5.972775661886575e-05, + "loss": 0.0056, + "step": 46580 + }, + { + "epoch": 3.618640776699029, + "grad_norm": 0.21386393904685974, + "learning_rate": 5.971153734096173e-05, + "loss": 0.0095, + "step": 46590 + }, + { + "epoch": 3.6194174757281554, + "grad_norm": 0.1656128615140915, + "learning_rate": 5.969531700101859e-05, + "loss": 0.0067, + "step": 46600 + }, + { + "epoch": 3.6201941747572817, + "grad_norm": 0.15980468690395355, + "learning_rate": 5.9679095600810155e-05, + "loss": 0.0081, + "step": 46610 + }, + { + "epoch": 3.620970873786408, + "grad_norm": 0.2585274279117584, + "learning_rate": 5.9662873142110384e-05, + "loss": 0.008, + "step": 46620 + }, + { + "epoch": 3.6217475728155337, + "grad_norm": 0.12328478693962097, + "learning_rate": 5.964664962669333e-05, + "loss": 0.0089, + "step": 46630 + }, + { + "epoch": 3.62252427184466, + "grad_norm": 0.23210176825523376, + "learning_rate": 5.9630425056333186e-05, + "loss": 0.0084, + "step": 46640 + }, + { + "epoch": 3.623300970873786, + "grad_norm": 0.2716667652130127, + "learning_rate": 5.961419943280422e-05, + "loss": 0.0076, + "step": 46650 + }, + { + "epoch": 3.6240776699029125, + "grad_norm": 0.21346798539161682, + "learning_rate": 5.959797275788087e-05, + "loss": 0.0066, + "step": 46660 + }, + { + "epoch": 3.6248543689320387, + "grad_norm": 0.23851914703845978, + "learning_rate": 5.958174503333765e-05, + "loss": 0.0099, + "step": 46670 + }, + { + "epoch": 3.625631067961165, + "grad_norm": 0.23617035150527954, + "learning_rate": 5.9565516260949195e-05, + "loss": 0.0059, + "step": 46680 + }, + { + "epoch": 3.626407766990291, + "grad_norm": 0.18581244349479675, + "learning_rate": 5.9549286442490273e-05, + "loss": 0.0053, + "step": 46690 + }, + { + "epoch": 3.6271844660194175, + "grad_norm": 0.2617039084434509, + "learning_rate": 5.953305557973572e-05, + "loss": 0.0066, + "step": 46700 + }, + { + "epoch": 3.6279611650485437, + "grad_norm": 0.32729968428611755, + "learning_rate": 5.9516823674460564e-05, + "loss": 0.0077, + "step": 46710 + }, + { + "epoch": 3.62873786407767, + "grad_norm": 0.17946413159370422, + "learning_rate": 5.9500590728439875e-05, + "loss": 0.0073, + "step": 46720 + }, + { + "epoch": 3.629514563106796, + "grad_norm": 0.23663318157196045, + "learning_rate": 5.948435674344886e-05, + "loss": 0.0093, + "step": 46730 + }, + { + "epoch": 3.6302912621359225, + "grad_norm": 0.23787377774715424, + "learning_rate": 5.946812172126285e-05, + "loss": 0.008, + "step": 46740 + }, + { + "epoch": 3.6310679611650487, + "grad_norm": 0.2753640413284302, + "learning_rate": 5.945188566365729e-05, + "loss": 0.0077, + "step": 46750 + }, + { + "epoch": 3.631844660194175, + "grad_norm": 0.21209339797496796, + "learning_rate": 5.943564857240773e-05, + "loss": 0.007, + "step": 46760 + }, + { + "epoch": 3.632621359223301, + "grad_norm": 0.36544761061668396, + "learning_rate": 5.941941044928981e-05, + "loss": 0.0063, + "step": 46770 + }, + { + "epoch": 3.6333980582524275, + "grad_norm": 0.27631938457489014, + "learning_rate": 5.940317129607935e-05, + "loss": 0.0085, + "step": 46780 + }, + { + "epoch": 3.6341747572815533, + "grad_norm": 0.13784928619861603, + "learning_rate": 5.9386931114552204e-05, + "loss": 0.0065, + "step": 46790 + }, + { + "epoch": 3.6349514563106795, + "grad_norm": 0.21536359190940857, + "learning_rate": 5.937068990648438e-05, + "loss": 0.0067, + "step": 46800 + }, + { + "epoch": 3.6357281553398058, + "grad_norm": 0.2542590796947479, + "learning_rate": 5.935444767365199e-05, + "loss": 0.0055, + "step": 46810 + }, + { + "epoch": 3.636504854368932, + "grad_norm": 0.1569870561361313, + "learning_rate": 5.933820441783129e-05, + "loss": 0.0055, + "step": 46820 + }, + { + "epoch": 3.6372815533980583, + "grad_norm": 0.25650647282600403, + "learning_rate": 5.932196014079857e-05, + "loss": 0.0097, + "step": 46830 + }, + { + "epoch": 3.6380582524271845, + "grad_norm": 0.18620575964450836, + "learning_rate": 5.930571484433032e-05, + "loss": 0.0074, + "step": 46840 + }, + { + "epoch": 3.6388349514563108, + "grad_norm": 0.37006571888923645, + "learning_rate": 5.928946853020309e-05, + "loss": 0.0081, + "step": 46850 + }, + { + "epoch": 3.639611650485437, + "grad_norm": 0.5171764492988586, + "learning_rate": 5.927322120019352e-05, + "loss": 0.0068, + "step": 46860 + }, + { + "epoch": 3.6403883495145632, + "grad_norm": 0.19146640598773956, + "learning_rate": 5.925697285607844e-05, + "loss": 0.0088, + "step": 46870 + }, + { + "epoch": 3.641165048543689, + "grad_norm": 0.2661668360233307, + "learning_rate": 5.924072349963472e-05, + "loss": 0.0082, + "step": 46880 + }, + { + "epoch": 3.6419417475728153, + "grad_norm": 0.17404814064502716, + "learning_rate": 5.922447313263938e-05, + "loss": 0.0067, + "step": 46890 + }, + { + "epoch": 3.6427184466019416, + "grad_norm": 0.322822242975235, + "learning_rate": 5.920822175686952e-05, + "loss": 0.0087, + "step": 46900 + }, + { + "epoch": 3.643495145631068, + "grad_norm": 0.1609807014465332, + "learning_rate": 5.919196937410239e-05, + "loss": 0.0069, + "step": 46910 + }, + { + "epoch": 3.644271844660194, + "grad_norm": 0.29889193177223206, + "learning_rate": 5.917571598611529e-05, + "loss": 0.012, + "step": 46920 + }, + { + "epoch": 3.6450485436893203, + "grad_norm": 0.24490249156951904, + "learning_rate": 5.9159461594685686e-05, + "loss": 0.0092, + "step": 46930 + }, + { + "epoch": 3.6458252427184465, + "grad_norm": 0.16913756728172302, + "learning_rate": 5.914320620159114e-05, + "loss": 0.0111, + "step": 46940 + }, + { + "epoch": 3.646601941747573, + "grad_norm": 0.4396164119243622, + "learning_rate": 5.912694980860931e-05, + "loss": 0.0085, + "step": 46950 + }, + { + "epoch": 3.647378640776699, + "grad_norm": 0.2784363031387329, + "learning_rate": 5.9110692417517964e-05, + "loss": 0.0069, + "step": 46960 + }, + { + "epoch": 3.6481553398058253, + "grad_norm": 0.16328570246696472, + "learning_rate": 5.9094434030095e-05, + "loss": 0.0074, + "step": 46970 + }, + { + "epoch": 3.6489320388349515, + "grad_norm": 0.19356995820999146, + "learning_rate": 5.907817464811841e-05, + "loss": 0.0044, + "step": 46980 + }, + { + "epoch": 3.649708737864078, + "grad_norm": 0.1841251254081726, + "learning_rate": 5.906191427336627e-05, + "loss": 0.0072, + "step": 46990 + }, + { + "epoch": 3.650485436893204, + "grad_norm": 0.20613382756710052, + "learning_rate": 5.9045652907616835e-05, + "loss": 0.0072, + "step": 47000 + }, + { + "epoch": 3.6512621359223303, + "grad_norm": 0.2908659875392914, + "learning_rate": 5.902939055264838e-05, + "loss": 0.0063, + "step": 47010 + }, + { + "epoch": 3.6520388349514565, + "grad_norm": 0.13221542537212372, + "learning_rate": 5.901312721023935e-05, + "loss": 0.0088, + "step": 47020 + }, + { + "epoch": 3.652815533980583, + "grad_norm": 0.20787790417671204, + "learning_rate": 5.8996862882168294e-05, + "loss": 0.0081, + "step": 47030 + }, + { + "epoch": 3.6535922330097086, + "grad_norm": 0.2480660229921341, + "learning_rate": 5.8980597570213826e-05, + "loss": 0.0076, + "step": 47040 + }, + { + "epoch": 3.654368932038835, + "grad_norm": 0.39180871844291687, + "learning_rate": 5.896433127615471e-05, + "loss": 0.0098, + "step": 47050 + }, + { + "epoch": 3.655145631067961, + "grad_norm": 0.21424436569213867, + "learning_rate": 5.894806400176981e-05, + "loss": 0.0064, + "step": 47060 + }, + { + "epoch": 3.6559223300970873, + "grad_norm": 0.559912383556366, + "learning_rate": 5.893179574883808e-05, + "loss": 0.0077, + "step": 47070 + }, + { + "epoch": 3.6566990291262136, + "grad_norm": 0.21105509996414185, + "learning_rate": 5.8915526519138585e-05, + "loss": 0.0083, + "step": 47080 + }, + { + "epoch": 3.65747572815534, + "grad_norm": 0.30955228209495544, + "learning_rate": 5.889925631445053e-05, + "loss": 0.0093, + "step": 47090 + }, + { + "epoch": 3.658252427184466, + "grad_norm": 0.14977295696735382, + "learning_rate": 5.888298513655318e-05, + "loss": 0.0084, + "step": 47100 + }, + { + "epoch": 3.6590291262135923, + "grad_norm": 0.3253262937068939, + "learning_rate": 5.886671298722592e-05, + "loss": 0.0061, + "step": 47110 + }, + { + "epoch": 3.6598058252427186, + "grad_norm": 0.20330961048603058, + "learning_rate": 5.885043986824828e-05, + "loss": 0.0093, + "step": 47120 + }, + { + "epoch": 3.6605825242718444, + "grad_norm": 0.3650040030479431, + "learning_rate": 5.883416578139982e-05, + "loss": 0.0081, + "step": 47130 + }, + { + "epoch": 3.6613592233009706, + "grad_norm": 0.20129986107349396, + "learning_rate": 5.881789072846029e-05, + "loss": 0.0061, + "step": 47140 + }, + { + "epoch": 3.662135922330097, + "grad_norm": 0.19689662754535675, + "learning_rate": 5.8801614711209474e-05, + "loss": 0.0074, + "step": 47150 + }, + { + "epoch": 3.662912621359223, + "grad_norm": 0.31883004307746887, + "learning_rate": 5.878533773142732e-05, + "loss": 0.0089, + "step": 47160 + }, + { + "epoch": 3.6636893203883494, + "grad_norm": 0.26131686568260193, + "learning_rate": 5.8769059790893834e-05, + "loss": 0.007, + "step": 47170 + }, + { + "epoch": 3.6644660194174756, + "grad_norm": 0.6873645186424255, + "learning_rate": 5.8752780891389145e-05, + "loss": 0.0056, + "step": 47180 + }, + { + "epoch": 3.665242718446602, + "grad_norm": 0.20114333927631378, + "learning_rate": 5.873650103469351e-05, + "loss": 0.0073, + "step": 47190 + }, + { + "epoch": 3.666019417475728, + "grad_norm": 0.21958027780056, + "learning_rate": 5.872022022258725e-05, + "loss": 0.0063, + "step": 47200 + }, + { + "epoch": 3.6667961165048544, + "grad_norm": 0.2498294562101364, + "learning_rate": 5.8703938456850814e-05, + "loss": 0.0064, + "step": 47210 + }, + { + "epoch": 3.6675728155339806, + "grad_norm": 0.21631267666816711, + "learning_rate": 5.868765573926477e-05, + "loss": 0.0047, + "step": 47220 + }, + { + "epoch": 3.668349514563107, + "grad_norm": 0.3452627956867218, + "learning_rate": 5.867137207160974e-05, + "loss": 0.006, + "step": 47230 + }, + { + "epoch": 3.669126213592233, + "grad_norm": 0.20323732495307922, + "learning_rate": 5.86550874556665e-05, + "loss": 0.0074, + "step": 47240 + }, + { + "epoch": 3.6699029126213594, + "grad_norm": 0.20743532478809357, + "learning_rate": 5.863880189321592e-05, + "loss": 0.0063, + "step": 47250 + }, + { + "epoch": 3.6706796116504856, + "grad_norm": 1.0374664068222046, + "learning_rate": 5.862251538603894e-05, + "loss": 0.0061, + "step": 47260 + }, + { + "epoch": 3.671456310679612, + "grad_norm": 0.4492397606372833, + "learning_rate": 5.8606227935916656e-05, + "loss": 0.0048, + "step": 47270 + }, + { + "epoch": 3.672233009708738, + "grad_norm": 0.1612284928560257, + "learning_rate": 5.858993954463021e-05, + "loss": 0.0066, + "step": 47280 + }, + { + "epoch": 3.673009708737864, + "grad_norm": 0.2287362515926361, + "learning_rate": 5.8573650213960896e-05, + "loss": 0.0062, + "step": 47290 + }, + { + "epoch": 3.67378640776699, + "grad_norm": 0.4950065612792969, + "learning_rate": 5.8557359945690084e-05, + "loss": 0.0053, + "step": 47300 + }, + { + "epoch": 3.6745631067961164, + "grad_norm": 0.19381853938102722, + "learning_rate": 5.8541068741599246e-05, + "loss": 0.0069, + "step": 47310 + }, + { + "epoch": 3.6753398058252427, + "grad_norm": 0.2467993199825287, + "learning_rate": 5.8524776603469985e-05, + "loss": 0.0066, + "step": 47320 + }, + { + "epoch": 3.676116504854369, + "grad_norm": 0.23805497586727142, + "learning_rate": 5.850848353308397e-05, + "loss": 0.0075, + "step": 47330 + }, + { + "epoch": 3.676893203883495, + "grad_norm": 0.4566667973995209, + "learning_rate": 5.849218953222297e-05, + "loss": 0.0082, + "step": 47340 + }, + { + "epoch": 3.6776699029126214, + "grad_norm": 0.18590308725833893, + "learning_rate": 5.847589460266891e-05, + "loss": 0.0071, + "step": 47350 + }, + { + "epoch": 3.6784466019417477, + "grad_norm": 0.4866655766963959, + "learning_rate": 5.845959874620375e-05, + "loss": 0.0052, + "step": 47360 + }, + { + "epoch": 3.679223300970874, + "grad_norm": 0.22986312210559845, + "learning_rate": 5.84433019646096e-05, + "loss": 0.0086, + "step": 47370 + }, + { + "epoch": 3.68, + "grad_norm": 0.14696596562862396, + "learning_rate": 5.842700425966863e-05, + "loss": 0.006, + "step": 47380 + }, + { + "epoch": 3.680776699029126, + "grad_norm": 0.2587880790233612, + "learning_rate": 5.841070563316315e-05, + "loss": 0.0076, + "step": 47390 + }, + { + "epoch": 3.681553398058252, + "grad_norm": 0.9422637820243835, + "learning_rate": 5.839440608687554e-05, + "loss": 0.0087, + "step": 47400 + }, + { + "epoch": 3.6823300970873785, + "grad_norm": 0.20279183983802795, + "learning_rate": 5.837810562258831e-05, + "loss": 0.0053, + "step": 47410 + }, + { + "epoch": 3.6831067961165047, + "grad_norm": 0.2392706573009491, + "learning_rate": 5.836180424208405e-05, + "loss": 0.0096, + "step": 47420 + }, + { + "epoch": 3.683883495145631, + "grad_norm": 0.391342431306839, + "learning_rate": 5.834550194714543e-05, + "loss": 0.01, + "step": 47430 + }, + { + "epoch": 3.684660194174757, + "grad_norm": 0.2466515600681305, + "learning_rate": 5.832919873955528e-05, + "loss": 0.0072, + "step": 47440 + }, + { + "epoch": 3.6854368932038835, + "grad_norm": 0.31856414675712585, + "learning_rate": 5.831289462109648e-05, + "loss": 0.0051, + "step": 47450 + }, + { + "epoch": 3.6862135922330097, + "grad_norm": 0.3312437832355499, + "learning_rate": 5.829658959355202e-05, + "loss": 0.0078, + "step": 47460 + }, + { + "epoch": 3.686990291262136, + "grad_norm": 0.1925630420446396, + "learning_rate": 5.8280283658704995e-05, + "loss": 0.0068, + "step": 47470 + }, + { + "epoch": 3.687766990291262, + "grad_norm": 0.305706262588501, + "learning_rate": 5.826397681833859e-05, + "loss": 0.0105, + "step": 47480 + }, + { + "epoch": 3.6885436893203885, + "grad_norm": 0.15943686664104462, + "learning_rate": 5.824766907423612e-05, + "loss": 0.0096, + "step": 47490 + }, + { + "epoch": 3.6893203883495147, + "grad_norm": 0.24397794902324677, + "learning_rate": 5.8231360428180945e-05, + "loss": 0.0071, + "step": 47500 + }, + { + "epoch": 3.690097087378641, + "grad_norm": 0.2782078981399536, + "learning_rate": 5.821505088195658e-05, + "loss": 0.0085, + "step": 47510 + }, + { + "epoch": 3.690873786407767, + "grad_norm": 0.3071383237838745, + "learning_rate": 5.819874043734661e-05, + "loss": 0.0083, + "step": 47520 + }, + { + "epoch": 3.6916504854368934, + "grad_norm": 0.3377082347869873, + "learning_rate": 5.8182429096134695e-05, + "loss": 0.0067, + "step": 47530 + }, + { + "epoch": 3.6924271844660193, + "grad_norm": 0.2458437979221344, + "learning_rate": 5.816611686010465e-05, + "loss": 0.0065, + "step": 47540 + }, + { + "epoch": 3.6932038834951455, + "grad_norm": 0.20448404550552368, + "learning_rate": 5.814980373104033e-05, + "loss": 0.0076, + "step": 47550 + }, + { + "epoch": 3.6939805825242717, + "grad_norm": 0.21910730004310608, + "learning_rate": 5.813348971072572e-05, + "loss": 0.008, + "step": 47560 + }, + { + "epoch": 3.694757281553398, + "grad_norm": 0.1450657844543457, + "learning_rate": 5.811717480094492e-05, + "loss": 0.005, + "step": 47570 + }, + { + "epoch": 3.6955339805825242, + "grad_norm": 0.3354778289794922, + "learning_rate": 5.810085900348209e-05, + "loss": 0.0067, + "step": 47580 + }, + { + "epoch": 3.6963106796116505, + "grad_norm": 0.26344361901283264, + "learning_rate": 5.8084542320121483e-05, + "loss": 0.0052, + "step": 47590 + }, + { + "epoch": 3.6970873786407767, + "grad_norm": 0.1935458481311798, + "learning_rate": 5.8068224752647497e-05, + "loss": 0.0062, + "step": 47600 + }, + { + "epoch": 3.697864077669903, + "grad_norm": 0.196321502327919, + "learning_rate": 5.805190630284457e-05, + "loss": 0.0065, + "step": 47610 + }, + { + "epoch": 3.6986407766990292, + "grad_norm": 0.16246066987514496, + "learning_rate": 5.803558697249729e-05, + "loss": 0.0089, + "step": 47620 + }, + { + "epoch": 3.6994174757281555, + "grad_norm": 0.23031838238239288, + "learning_rate": 5.8019266763390276e-05, + "loss": 0.007, + "step": 47630 + }, + { + "epoch": 3.7001941747572813, + "grad_norm": 0.2630064785480499, + "learning_rate": 5.800294567730833e-05, + "loss": 0.005, + "step": 47640 + }, + { + "epoch": 3.7009708737864075, + "grad_norm": 0.1734534204006195, + "learning_rate": 5.798662371603628e-05, + "loss": 0.0073, + "step": 47650 + }, + { + "epoch": 3.701747572815534, + "grad_norm": 0.28227853775024414, + "learning_rate": 5.797030088135904e-05, + "loss": 0.0067, + "step": 47660 + }, + { + "epoch": 3.70252427184466, + "grad_norm": 0.5459275841712952, + "learning_rate": 5.7953977175061705e-05, + "loss": 0.0046, + "step": 47670 + }, + { + "epoch": 3.7033009708737863, + "grad_norm": 0.2366085797548294, + "learning_rate": 5.793765259892938e-05, + "loss": 0.0071, + "step": 47680 + }, + { + "epoch": 3.7040776699029125, + "grad_norm": 0.16811759769916534, + "learning_rate": 5.792132715474729e-05, + "loss": 0.0051, + "step": 47690 + }, + { + "epoch": 3.704854368932039, + "grad_norm": 0.27906733751296997, + "learning_rate": 5.790500084430078e-05, + "loss": 0.0074, + "step": 47700 + }, + { + "epoch": 3.705631067961165, + "grad_norm": 0.25187402963638306, + "learning_rate": 5.7888673669375274e-05, + "loss": 0.0081, + "step": 47710 + }, + { + "epoch": 3.7064077669902913, + "grad_norm": 0.2866177260875702, + "learning_rate": 5.787234563175625e-05, + "loss": 0.0075, + "step": 47720 + }, + { + "epoch": 3.7071844660194175, + "grad_norm": 0.20391353964805603, + "learning_rate": 5.7856016733229355e-05, + "loss": 0.0065, + "step": 47730 + }, + { + "epoch": 3.707961165048544, + "grad_norm": 0.20745790004730225, + "learning_rate": 5.7839686975580297e-05, + "loss": 0.008, + "step": 47740 + }, + { + "epoch": 3.70873786407767, + "grad_norm": 0.8148251175880432, + "learning_rate": 5.782335636059484e-05, + "loss": 0.0084, + "step": 47750 + }, + { + "epoch": 3.7095145631067963, + "grad_norm": 0.20346413552761078, + "learning_rate": 5.780702489005889e-05, + "loss": 0.0065, + "step": 47760 + }, + { + "epoch": 3.7102912621359225, + "grad_norm": 0.5280622243881226, + "learning_rate": 5.779069256575845e-05, + "loss": 0.0069, + "step": 47770 + }, + { + "epoch": 3.711067961165049, + "grad_norm": 0.13746114075183868, + "learning_rate": 5.7774359389479574e-05, + "loss": 0.0053, + "step": 47780 + }, + { + "epoch": 3.711844660194175, + "grad_norm": 0.33336004614830017, + "learning_rate": 5.775802536300845e-05, + "loss": 0.0064, + "step": 47790 + }, + { + "epoch": 3.712621359223301, + "grad_norm": 0.12755674123764038, + "learning_rate": 5.774169048813134e-05, + "loss": 0.0055, + "step": 47800 + }, + { + "epoch": 3.713398058252427, + "grad_norm": 0.10839279741048813, + "learning_rate": 5.77253547666346e-05, + "loss": 0.0045, + "step": 47810 + }, + { + "epoch": 3.7141747572815533, + "grad_norm": 0.1880037635564804, + "learning_rate": 5.770901820030465e-05, + "loss": 0.0063, + "step": 47820 + }, + { + "epoch": 3.7149514563106796, + "grad_norm": 0.28801336884498596, + "learning_rate": 5.769268079092809e-05, + "loss": 0.0059, + "step": 47830 + }, + { + "epoch": 3.715728155339806, + "grad_norm": 0.168148010969162, + "learning_rate": 5.767634254029151e-05, + "loss": 0.0064, + "step": 47840 + }, + { + "epoch": 3.716504854368932, + "grad_norm": 0.261512815952301, + "learning_rate": 5.7660003450181655e-05, + "loss": 0.0091, + "step": 47850 + }, + { + "epoch": 3.7172815533980583, + "grad_norm": 0.1806025356054306, + "learning_rate": 5.764366352238534e-05, + "loss": 0.0089, + "step": 47860 + }, + { + "epoch": 3.7180582524271846, + "grad_norm": 0.1801312267780304, + "learning_rate": 5.7627322758689474e-05, + "loss": 0.0063, + "step": 47870 + }, + { + "epoch": 3.718834951456311, + "grad_norm": 0.3165931701660156, + "learning_rate": 5.761098116088105e-05, + "loss": 0.0065, + "step": 47880 + }, + { + "epoch": 3.7196116504854366, + "grad_norm": 0.6790018081665039, + "learning_rate": 5.759463873074717e-05, + "loss": 0.0086, + "step": 47890 + }, + { + "epoch": 3.720388349514563, + "grad_norm": 0.2021014243364334, + "learning_rate": 5.757829547007504e-05, + "loss": 0.0075, + "step": 47900 + }, + { + "epoch": 3.721165048543689, + "grad_norm": 0.2200383096933365, + "learning_rate": 5.756195138065189e-05, + "loss": 0.0064, + "step": 47910 + }, + { + "epoch": 3.7219417475728154, + "grad_norm": 0.24278956651687622, + "learning_rate": 5.754560646426511e-05, + "loss": 0.0092, + "step": 47920 + }, + { + "epoch": 3.7227184466019416, + "grad_norm": 0.3027634024620056, + "learning_rate": 5.752926072270216e-05, + "loss": 0.0093, + "step": 47930 + }, + { + "epoch": 3.723495145631068, + "grad_norm": 0.1961820274591446, + "learning_rate": 5.7512914157750563e-05, + "loss": 0.0062, + "step": 47940 + }, + { + "epoch": 3.724271844660194, + "grad_norm": 0.2413041591644287, + "learning_rate": 5.749656677119798e-05, + "loss": 0.007, + "step": 47950 + }, + { + "epoch": 3.7250485436893204, + "grad_norm": 0.2657265365123749, + "learning_rate": 5.7480218564832125e-05, + "loss": 0.0092, + "step": 47960 + }, + { + "epoch": 3.7258252427184466, + "grad_norm": 0.234809011220932, + "learning_rate": 5.746386954044082e-05, + "loss": 0.0082, + "step": 47970 + }, + { + "epoch": 3.726601941747573, + "grad_norm": 0.34316274523735046, + "learning_rate": 5.744751969981195e-05, + "loss": 0.0076, + "step": 47980 + }, + { + "epoch": 3.727378640776699, + "grad_norm": 0.36675721406936646, + "learning_rate": 5.7431169044733526e-05, + "loss": 0.0075, + "step": 47990 + }, + { + "epoch": 3.7281553398058254, + "grad_norm": 0.24852584302425385, + "learning_rate": 5.741481757699364e-05, + "loss": 0.0081, + "step": 48000 + }, + { + "epoch": 3.7289320388349516, + "grad_norm": 0.27813786268234253, + "learning_rate": 5.7398465298380434e-05, + "loss": 0.0071, + "step": 48010 + }, + { + "epoch": 3.729708737864078, + "grad_norm": 0.17238304018974304, + "learning_rate": 5.7382112210682193e-05, + "loss": 0.0084, + "step": 48020 + }, + { + "epoch": 3.730485436893204, + "grad_norm": 0.2702910006046295, + "learning_rate": 5.7365758315687266e-05, + "loss": 0.0072, + "step": 48030 + }, + { + "epoch": 3.7312621359223304, + "grad_norm": 0.13951760530471802, + "learning_rate": 5.734940361518407e-05, + "loss": 0.0055, + "step": 48040 + }, + { + "epoch": 3.732038834951456, + "grad_norm": 0.2338186800479889, + "learning_rate": 5.733304811096116e-05, + "loss": 0.0062, + "step": 48050 + }, + { + "epoch": 3.7328155339805824, + "grad_norm": 0.26826366782188416, + "learning_rate": 5.731669180480713e-05, + "loss": 0.0073, + "step": 48060 + }, + { + "epoch": 3.7335922330097087, + "grad_norm": 0.20958024263381958, + "learning_rate": 5.730033469851067e-05, + "loss": 0.0072, + "step": 48070 + }, + { + "epoch": 3.734368932038835, + "grad_norm": 0.2015407234430313, + "learning_rate": 5.7283976793860607e-05, + "loss": 0.0061, + "step": 48080 + }, + { + "epoch": 3.735145631067961, + "grad_norm": 0.3140290677547455, + "learning_rate": 5.726761809264577e-05, + "loss": 0.0092, + "step": 48090 + }, + { + "epoch": 3.7359223300970874, + "grad_norm": 0.15137270092964172, + "learning_rate": 5.7251258596655155e-05, + "loss": 0.0065, + "step": 48100 + }, + { + "epoch": 3.7366990291262137, + "grad_norm": 0.20715971291065216, + "learning_rate": 5.72348983076778e-05, + "loss": 0.0062, + "step": 48110 + }, + { + "epoch": 3.73747572815534, + "grad_norm": 0.19264240562915802, + "learning_rate": 5.7218537227502854e-05, + "loss": 0.0087, + "step": 48120 + }, + { + "epoch": 3.738252427184466, + "grad_norm": 0.18250741064548492, + "learning_rate": 5.720217535791951e-05, + "loss": 0.0085, + "step": 48130 + }, + { + "epoch": 3.739029126213592, + "grad_norm": 0.18094518780708313, + "learning_rate": 5.718581270071711e-05, + "loss": 0.0079, + "step": 48140 + }, + { + "epoch": 3.739805825242718, + "grad_norm": 0.20885390043258667, + "learning_rate": 5.716944925768505e-05, + "loss": 0.0079, + "step": 48150 + }, + { + "epoch": 3.7405825242718445, + "grad_norm": 0.2257995903491974, + "learning_rate": 5.7153085030612786e-05, + "loss": 0.0064, + "step": 48160 + }, + { + "epoch": 3.7413592233009707, + "grad_norm": 0.2035859376192093, + "learning_rate": 5.713672002128989e-05, + "loss": 0.0068, + "step": 48170 + }, + { + "epoch": 3.742135922330097, + "grad_norm": 0.4847932755947113, + "learning_rate": 5.712035423150602e-05, + "loss": 0.0052, + "step": 48180 + }, + { + "epoch": 3.742912621359223, + "grad_norm": 0.3098622262477875, + "learning_rate": 5.710398766305094e-05, + "loss": 0.0054, + "step": 48190 + }, + { + "epoch": 3.7436893203883495, + "grad_norm": 0.27300745248794556, + "learning_rate": 5.708762031771442e-05, + "loss": 0.0081, + "step": 48200 + }, + { + "epoch": 3.7444660194174757, + "grad_norm": 0.15791746973991394, + "learning_rate": 5.707125219728642e-05, + "loss": 0.0075, + "step": 48210 + }, + { + "epoch": 3.745242718446602, + "grad_norm": 0.2500877380371094, + "learning_rate": 5.7054883303556905e-05, + "loss": 0.0086, + "step": 48220 + }, + { + "epoch": 3.746019417475728, + "grad_norm": 0.5279837250709534, + "learning_rate": 5.703851363831595e-05, + "loss": 0.0069, + "step": 48230 + }, + { + "epoch": 3.7467961165048544, + "grad_norm": 0.34831738471984863, + "learning_rate": 5.702214320335374e-05, + "loss": 0.0078, + "step": 48240 + }, + { + "epoch": 3.7475728155339807, + "grad_norm": 0.34465456008911133, + "learning_rate": 5.7005772000460514e-05, + "loss": 0.0049, + "step": 48250 + }, + { + "epoch": 3.748349514563107, + "grad_norm": 0.13585253059864044, + "learning_rate": 5.698940003142656e-05, + "loss": 0.0058, + "step": 48260 + }, + { + "epoch": 3.749126213592233, + "grad_norm": 0.198966383934021, + "learning_rate": 5.697302729804236e-05, + "loss": 0.0057, + "step": 48270 + }, + { + "epoch": 3.7499029126213594, + "grad_norm": 0.13482260704040527, + "learning_rate": 5.695665380209837e-05, + "loss": 0.0066, + "step": 48280 + }, + { + "epoch": 3.7506796116504857, + "grad_norm": 0.24388079345226288, + "learning_rate": 5.6940279545385165e-05, + "loss": 0.0067, + "step": 48290 + }, + { + "epoch": 3.7514563106796115, + "grad_norm": 0.20697487890720367, + "learning_rate": 5.692390452969344e-05, + "loss": 0.0058, + "step": 48300 + }, + { + "epoch": 3.7522330097087377, + "grad_norm": 0.15771672129631042, + "learning_rate": 5.690752875681392e-05, + "loss": 0.0105, + "step": 48310 + }, + { + "epoch": 3.753009708737864, + "grad_norm": 0.30309852957725525, + "learning_rate": 5.6891152228537435e-05, + "loss": 0.0077, + "step": 48320 + }, + { + "epoch": 3.7537864077669902, + "grad_norm": 0.2383124679327011, + "learning_rate": 5.687477494665492e-05, + "loss": 0.0066, + "step": 48330 + }, + { + "epoch": 3.7545631067961165, + "grad_norm": 0.31647783517837524, + "learning_rate": 5.685839691295733e-05, + "loss": 0.0074, + "step": 48340 + }, + { + "epoch": 3.7553398058252427, + "grad_norm": 0.22093069553375244, + "learning_rate": 5.6842018129235786e-05, + "loss": 0.0069, + "step": 48350 + }, + { + "epoch": 3.756116504854369, + "grad_norm": 0.24893692135810852, + "learning_rate": 5.6825638597281404e-05, + "loss": 0.0074, + "step": 48360 + }, + { + "epoch": 3.7568932038834952, + "grad_norm": 0.20708389580249786, + "learning_rate": 5.680925831888546e-05, + "loss": 0.0056, + "step": 48370 + }, + { + "epoch": 3.7576699029126215, + "grad_norm": 0.14210912585258484, + "learning_rate": 5.6792877295839274e-05, + "loss": 0.0065, + "step": 48380 + }, + { + "epoch": 3.7584466019417477, + "grad_norm": 0.27141284942626953, + "learning_rate": 5.6776495529934224e-05, + "loss": 0.009, + "step": 48390 + }, + { + "epoch": 3.7592233009708735, + "grad_norm": 0.11253691464662552, + "learning_rate": 5.6760113022961824e-05, + "loss": 0.0093, + "step": 48400 + }, + { + "epoch": 3.76, + "grad_norm": 0.2768924832344055, + "learning_rate": 5.6743729776713617e-05, + "loss": 0.0081, + "step": 48410 + }, + { + "epoch": 3.760776699029126, + "grad_norm": 0.3424949049949646, + "learning_rate": 5.672734579298126e-05, + "loss": 0.0061, + "step": 48420 + }, + { + "epoch": 3.7615533980582523, + "grad_norm": 0.315232515335083, + "learning_rate": 5.671096107355649e-05, + "loss": 0.0066, + "step": 48430 + }, + { + "epoch": 3.7623300970873785, + "grad_norm": 0.27367907762527466, + "learning_rate": 5.66945756202311e-05, + "loss": 0.0066, + "step": 48440 + }, + { + "epoch": 3.763106796116505, + "grad_norm": 0.3663608133792877, + "learning_rate": 5.667818943479699e-05, + "loss": 0.0065, + "step": 48450 + }, + { + "epoch": 3.763883495145631, + "grad_norm": 0.2279641330242157, + "learning_rate": 5.666180251904612e-05, + "loss": 0.006, + "step": 48460 + }, + { + "epoch": 3.7646601941747573, + "grad_norm": 0.41060879826545715, + "learning_rate": 5.6645414874770555e-05, + "loss": 0.0071, + "step": 48470 + }, + { + "epoch": 3.7654368932038835, + "grad_norm": 0.3045080900192261, + "learning_rate": 5.66290265037624e-05, + "loss": 0.0074, + "step": 48480 + }, + { + "epoch": 3.7662135922330098, + "grad_norm": 0.7105335593223572, + "learning_rate": 5.661263740781386e-05, + "loss": 0.0082, + "step": 48490 + }, + { + "epoch": 3.766990291262136, + "grad_norm": 0.13932380080223083, + "learning_rate": 5.6596247588717254e-05, + "loss": 0.0069, + "step": 48500 + }, + { + "epoch": 3.7677669902912623, + "grad_norm": 0.13005639612674713, + "learning_rate": 5.6579857048264926e-05, + "loss": 0.005, + "step": 48510 + }, + { + "epoch": 3.7685436893203885, + "grad_norm": 0.1986207365989685, + "learning_rate": 5.6563465788249314e-05, + "loss": 0.0052, + "step": 48520 + }, + { + "epoch": 3.7693203883495148, + "grad_norm": 0.25740477442741394, + "learning_rate": 5.6547073810462956e-05, + "loss": 0.0088, + "step": 48530 + }, + { + "epoch": 3.770097087378641, + "grad_norm": 0.15731823444366455, + "learning_rate": 5.653068111669846e-05, + "loss": 0.0065, + "step": 48540 + }, + { + "epoch": 3.770873786407767, + "grad_norm": 0.1856263279914856, + "learning_rate": 5.651428770874848e-05, + "loss": 0.0075, + "step": 48550 + }, + { + "epoch": 3.771650485436893, + "grad_norm": 0.13581599295139313, + "learning_rate": 5.64978935884058e-05, + "loss": 0.0057, + "step": 48560 + }, + { + "epoch": 3.7724271844660193, + "grad_norm": 0.1819159835577011, + "learning_rate": 5.6481498757463244e-05, + "loss": 0.0055, + "step": 48570 + }, + { + "epoch": 3.7732038834951456, + "grad_norm": 0.23946994543075562, + "learning_rate": 5.646510321771373e-05, + "loss": 0.0087, + "step": 48580 + }, + { + "epoch": 3.773980582524272, + "grad_norm": 0.18215760588645935, + "learning_rate": 5.644870697095024e-05, + "loss": 0.0091, + "step": 48590 + }, + { + "epoch": 3.774757281553398, + "grad_norm": 0.18264661729335785, + "learning_rate": 5.643231001896586e-05, + "loss": 0.0071, + "step": 48600 + }, + { + "epoch": 3.7755339805825243, + "grad_norm": 0.2658619284629822, + "learning_rate": 5.6415912363553726e-05, + "loss": 0.0118, + "step": 48610 + }, + { + "epoch": 3.7763106796116506, + "grad_norm": 0.22597597539424896, + "learning_rate": 5.639951400650706e-05, + "loss": 0.0057, + "step": 48620 + }, + { + "epoch": 3.777087378640777, + "grad_norm": 0.2703701853752136, + "learning_rate": 5.6383114949619165e-05, + "loss": 0.0084, + "step": 48630 + }, + { + "epoch": 3.777864077669903, + "grad_norm": 0.33904173970222473, + "learning_rate": 5.636671519468342e-05, + "loss": 0.0058, + "step": 48640 + }, + { + "epoch": 3.778640776699029, + "grad_norm": 0.22517311573028564, + "learning_rate": 5.635031474349327e-05, + "loss": 0.0059, + "step": 48650 + }, + { + "epoch": 3.779417475728155, + "grad_norm": 0.14952991902828217, + "learning_rate": 5.6333913597842246e-05, + "loss": 0.0063, + "step": 48660 + }, + { + "epoch": 3.7801941747572814, + "grad_norm": 0.24935562908649445, + "learning_rate": 5.6317511759523955e-05, + "loss": 0.0082, + "step": 48670 + }, + { + "epoch": 3.7809708737864076, + "grad_norm": 0.3496425747871399, + "learning_rate": 5.630110923033207e-05, + "loss": 0.0068, + "step": 48680 + }, + { + "epoch": 3.781747572815534, + "grad_norm": 0.2921091914176941, + "learning_rate": 5.628470601206036e-05, + "loss": 0.0108, + "step": 48690 + }, + { + "epoch": 3.78252427184466, + "grad_norm": 0.5996983051300049, + "learning_rate": 5.626830210650266e-05, + "loss": 0.0077, + "step": 48700 + }, + { + "epoch": 3.7833009708737864, + "grad_norm": 0.3917669355869293, + "learning_rate": 5.625189751545285e-05, + "loss": 0.0076, + "step": 48710 + }, + { + "epoch": 3.7840776699029126, + "grad_norm": 0.16406767070293427, + "learning_rate": 5.6235492240704936e-05, + "loss": 0.0064, + "step": 48720 + }, + { + "epoch": 3.784854368932039, + "grad_norm": 0.18732130527496338, + "learning_rate": 5.621908628405296e-05, + "loss": 0.0063, + "step": 48730 + }, + { + "epoch": 3.785631067961165, + "grad_norm": 0.3837549388408661, + "learning_rate": 5.620267964729106e-05, + "loss": 0.0049, + "step": 48740 + }, + { + "epoch": 3.7864077669902914, + "grad_norm": 0.14442314207553864, + "learning_rate": 5.618627233221344e-05, + "loss": 0.0067, + "step": 48750 + }, + { + "epoch": 3.7871844660194176, + "grad_norm": 0.14054210484027863, + "learning_rate": 5.616986434061438e-05, + "loss": 0.0058, + "step": 48760 + }, + { + "epoch": 3.787961165048544, + "grad_norm": 0.20277513563632965, + "learning_rate": 5.615345567428822e-05, + "loss": 0.0068, + "step": 48770 + }, + { + "epoch": 3.78873786407767, + "grad_norm": 0.35154902935028076, + "learning_rate": 5.613704633502941e-05, + "loss": 0.0083, + "step": 48780 + }, + { + "epoch": 3.7895145631067964, + "grad_norm": 0.31359830498695374, + "learning_rate": 5.612063632463245e-05, + "loss": 0.0085, + "step": 48790 + }, + { + "epoch": 3.7902912621359226, + "grad_norm": 0.21296000480651855, + "learning_rate": 5.610422564489188e-05, + "loss": 0.0102, + "step": 48800 + }, + { + "epoch": 3.7910679611650484, + "grad_norm": 0.33817052841186523, + "learning_rate": 5.608781429760239e-05, + "loss": 0.013, + "step": 48810 + }, + { + "epoch": 3.7918446601941747, + "grad_norm": 0.2034517228603363, + "learning_rate": 5.607140228455866e-05, + "loss": 0.007, + "step": 48820 + }, + { + "epoch": 3.792621359223301, + "grad_norm": 0.16956833004951477, + "learning_rate": 5.605498960755553e-05, + "loss": 0.0069, + "step": 48830 + }, + { + "epoch": 3.793398058252427, + "grad_norm": 0.3073500692844391, + "learning_rate": 5.603857626838782e-05, + "loss": 0.007, + "step": 48840 + }, + { + "epoch": 3.7941747572815534, + "grad_norm": 0.37184467911720276, + "learning_rate": 5.60221622688505e-05, + "loss": 0.0087, + "step": 48850 + }, + { + "epoch": 3.7949514563106796, + "grad_norm": 0.43419918417930603, + "learning_rate": 5.6005747610738565e-05, + "loss": 0.009, + "step": 48860 + }, + { + "epoch": 3.795728155339806, + "grad_norm": 0.4549820125102997, + "learning_rate": 5.59893322958471e-05, + "loss": 0.0064, + "step": 48870 + }, + { + "epoch": 3.796504854368932, + "grad_norm": 0.1465606689453125, + "learning_rate": 5.5972916325971256e-05, + "loss": 0.0064, + "step": 48880 + }, + { + "epoch": 3.7972815533980584, + "grad_norm": 0.31521618366241455, + "learning_rate": 5.595649970290628e-05, + "loss": 0.0075, + "step": 48890 + }, + { + "epoch": 3.798058252427184, + "grad_norm": 0.18561521172523499, + "learning_rate": 5.5940082428447426e-05, + "loss": 0.0066, + "step": 48900 + }, + { + "epoch": 3.7988349514563104, + "grad_norm": 0.6209407448768616, + "learning_rate": 5.592366450439012e-05, + "loss": 0.0053, + "step": 48910 + }, + { + "epoch": 3.7996116504854367, + "grad_norm": 0.241334930062294, + "learning_rate": 5.590724593252975e-05, + "loss": 0.0046, + "step": 48920 + }, + { + "epoch": 3.800388349514563, + "grad_norm": 0.2792927920818329, + "learning_rate": 5.589082671466184e-05, + "loss": 0.0071, + "step": 48930 + }, + { + "epoch": 3.801165048543689, + "grad_norm": 0.570650041103363, + "learning_rate": 5.587440685258199e-05, + "loss": 0.0079, + "step": 48940 + }, + { + "epoch": 3.8019417475728154, + "grad_norm": 0.17527607083320618, + "learning_rate": 5.585798634808583e-05, + "loss": 0.0067, + "step": 48950 + }, + { + "epoch": 3.8027184466019417, + "grad_norm": 0.13537733256816864, + "learning_rate": 5.584156520296909e-05, + "loss": 0.007, + "step": 48960 + }, + { + "epoch": 3.803495145631068, + "grad_norm": 0.29843398928642273, + "learning_rate": 5.582514341902757e-05, + "loss": 0.0059, + "step": 48970 + }, + { + "epoch": 3.804271844660194, + "grad_norm": 0.19910679757595062, + "learning_rate": 5.580872099805713e-05, + "loss": 0.0075, + "step": 48980 + }, + { + "epoch": 3.8050485436893204, + "grad_norm": 0.18348343670368195, + "learning_rate": 5.5792297941853674e-05, + "loss": 0.0069, + "step": 48990 + }, + { + "epoch": 3.8058252427184467, + "grad_norm": 0.323581725358963, + "learning_rate": 5.5775874252213247e-05, + "loss": 0.0069, + "step": 49000 + }, + { + "epoch": 3.806601941747573, + "grad_norm": 0.3519204258918762, + "learning_rate": 5.575944993093189e-05, + "loss": 0.0064, + "step": 49010 + }, + { + "epoch": 3.807378640776699, + "grad_norm": 0.23295605182647705, + "learning_rate": 5.574302497980574e-05, + "loss": 0.0085, + "step": 49020 + }, + { + "epoch": 3.8081553398058254, + "grad_norm": 0.37943559885025024, + "learning_rate": 5.5726599400631005e-05, + "loss": 0.0087, + "step": 49030 + }, + { + "epoch": 3.8089320388349517, + "grad_norm": 0.11768294125795364, + "learning_rate": 5.571017319520399e-05, + "loss": 0.008, + "step": 49040 + }, + { + "epoch": 3.809708737864078, + "grad_norm": 0.2816692292690277, + "learning_rate": 5.5693746365320984e-05, + "loss": 0.007, + "step": 49050 + }, + { + "epoch": 3.8104854368932037, + "grad_norm": 0.22830170392990112, + "learning_rate": 5.567731891277848e-05, + "loss": 0.0067, + "step": 49060 + }, + { + "epoch": 3.81126213592233, + "grad_norm": 0.17623370885849, + "learning_rate": 5.566089083937288e-05, + "loss": 0.0057, + "step": 49070 + }, + { + "epoch": 3.8120388349514562, + "grad_norm": 0.3296937048435211, + "learning_rate": 5.564446214690079e-05, + "loss": 0.0054, + "step": 49080 + }, + { + "epoch": 3.8128155339805825, + "grad_norm": 0.20897342264652252, + "learning_rate": 5.56280328371588e-05, + "loss": 0.0074, + "step": 49090 + }, + { + "epoch": 3.8135922330097087, + "grad_norm": 0.23187874257564545, + "learning_rate": 5.5611602911943596e-05, + "loss": 0.0075, + "step": 49100 + }, + { + "epoch": 3.814368932038835, + "grad_norm": 0.1273510903120041, + "learning_rate": 5.559517237305194e-05, + "loss": 0.0061, + "step": 49110 + }, + { + "epoch": 3.8151456310679612, + "grad_norm": 0.292046457529068, + "learning_rate": 5.557874122228064e-05, + "loss": 0.0079, + "step": 49120 + }, + { + "epoch": 3.8159223300970875, + "grad_norm": 0.1801685094833374, + "learning_rate": 5.55623094614266e-05, + "loss": 0.006, + "step": 49130 + }, + { + "epoch": 3.8166990291262137, + "grad_norm": 0.35628223419189453, + "learning_rate": 5.5545877092286743e-05, + "loss": 0.0068, + "step": 49140 + }, + { + "epoch": 3.8174757281553395, + "grad_norm": 0.2360442578792572, + "learning_rate": 5.552944411665811e-05, + "loss": 0.0098, + "step": 49150 + }, + { + "epoch": 3.818252427184466, + "grad_norm": 0.260624498128891, + "learning_rate": 5.5513010536337795e-05, + "loss": 0.0056, + "step": 49160 + }, + { + "epoch": 3.819029126213592, + "grad_norm": 0.2692478597164154, + "learning_rate": 5.5496576353122934e-05, + "loss": 0.0052, + "step": 49170 + }, + { + "epoch": 3.8198058252427183, + "grad_norm": 0.52022385597229, + "learning_rate": 5.548014156881074e-05, + "loss": 0.0066, + "step": 49180 + }, + { + "epoch": 3.8205825242718445, + "grad_norm": 0.19669926166534424, + "learning_rate": 5.5463706185198494e-05, + "loss": 0.0064, + "step": 49190 + }, + { + "epoch": 3.8213592233009708, + "grad_norm": 0.20925045013427734, + "learning_rate": 5.544727020408358e-05, + "loss": 0.0068, + "step": 49200 + }, + { + "epoch": 3.822135922330097, + "grad_norm": 0.26894429326057434, + "learning_rate": 5.5430833627263366e-05, + "loss": 0.0084, + "step": 49210 + }, + { + "epoch": 3.8229126213592233, + "grad_norm": 0.2019014060497284, + "learning_rate": 5.5414396456535354e-05, + "loss": 0.0077, + "step": 49220 + }, + { + "epoch": 3.8236893203883495, + "grad_norm": 0.5314840078353882, + "learning_rate": 5.539795869369709e-05, + "loss": 0.0078, + "step": 49230 + }, + { + "epoch": 3.8244660194174758, + "grad_norm": 0.8562100529670715, + "learning_rate": 5.538152034054618e-05, + "loss": 0.0097, + "step": 49240 + }, + { + "epoch": 3.825242718446602, + "grad_norm": 0.23729901015758514, + "learning_rate": 5.536508139888028e-05, + "loss": 0.008, + "step": 49250 + }, + { + "epoch": 3.8260194174757283, + "grad_norm": 0.21673744916915894, + "learning_rate": 5.534864187049716e-05, + "loss": 0.0053, + "step": 49260 + }, + { + "epoch": 3.8267961165048545, + "grad_norm": 0.35166218876838684, + "learning_rate": 5.5332201757194615e-05, + "loss": 0.0067, + "step": 49270 + }, + { + "epoch": 3.8275728155339808, + "grad_norm": 0.17853666841983795, + "learning_rate": 5.531576106077049e-05, + "loss": 0.0068, + "step": 49280 + }, + { + "epoch": 3.828349514563107, + "grad_norm": 0.2888318598270416, + "learning_rate": 5.529931978302272e-05, + "loss": 0.0075, + "step": 49290 + }, + { + "epoch": 3.8291262135922333, + "grad_norm": 0.3088608980178833, + "learning_rate": 5.528287792574932e-05, + "loss": 0.007, + "step": 49300 + }, + { + "epoch": 3.829902912621359, + "grad_norm": 0.22075684368610382, + "learning_rate": 5.5266435490748324e-05, + "loss": 0.0047, + "step": 49310 + }, + { + "epoch": 3.8306796116504853, + "grad_norm": 0.22890514135360718, + "learning_rate": 5.524999247981787e-05, + "loss": 0.0063, + "step": 49320 + }, + { + "epoch": 3.8314563106796116, + "grad_norm": 0.19454891979694366, + "learning_rate": 5.523354889475613e-05, + "loss": 0.0049, + "step": 49330 + }, + { + "epoch": 3.832233009708738, + "grad_norm": 0.242399200797081, + "learning_rate": 5.521710473736134e-05, + "loss": 0.0058, + "step": 49340 + }, + { + "epoch": 3.833009708737864, + "grad_norm": 0.4337763488292694, + "learning_rate": 5.5200660009431826e-05, + "loss": 0.0081, + "step": 49350 + }, + { + "epoch": 3.8337864077669903, + "grad_norm": 0.27211856842041016, + "learning_rate": 5.518421471276596e-05, + "loss": 0.0062, + "step": 49360 + }, + { + "epoch": 3.8345631067961166, + "grad_norm": 0.3626125156879425, + "learning_rate": 5.5167768849162174e-05, + "loss": 0.0069, + "step": 49370 + }, + { + "epoch": 3.835339805825243, + "grad_norm": 0.26773831248283386, + "learning_rate": 5.515132242041893e-05, + "loss": 0.0074, + "step": 49380 + }, + { + "epoch": 3.836116504854369, + "grad_norm": 0.4153629541397095, + "learning_rate": 5.513487542833483e-05, + "loss": 0.0069, + "step": 49390 + }, + { + "epoch": 3.8368932038834953, + "grad_norm": 0.2509966492652893, + "learning_rate": 5.5118427874708466e-05, + "loss": 0.0082, + "step": 49400 + }, + { + "epoch": 3.837669902912621, + "grad_norm": 0.22513926029205322, + "learning_rate": 5.510197976133853e-05, + "loss": 0.009, + "step": 49410 + }, + { + "epoch": 3.8384466019417474, + "grad_norm": 0.342658132314682, + "learning_rate": 5.508553109002376e-05, + "loss": 0.0076, + "step": 49420 + }, + { + "epoch": 3.8392233009708736, + "grad_norm": 0.12995734810829163, + "learning_rate": 5.5069081862562957e-05, + "loss": 0.0083, + "step": 49430 + }, + { + "epoch": 3.84, + "grad_norm": 0.3290637135505676, + "learning_rate": 5.5052632080754965e-05, + "loss": 0.0049, + "step": 49440 + }, + { + "epoch": 3.840776699029126, + "grad_norm": 0.3866083323955536, + "learning_rate": 5.503618174639874e-05, + "loss": 0.0068, + "step": 49450 + }, + { + "epoch": 3.8415533980582524, + "grad_norm": 0.508634626865387, + "learning_rate": 5.501973086129325e-05, + "loss": 0.0072, + "step": 49460 + }, + { + "epoch": 3.8423300970873786, + "grad_norm": 0.3269915282726288, + "learning_rate": 5.500327942723753e-05, + "loss": 0.0057, + "step": 49470 + }, + { + "epoch": 3.843106796116505, + "grad_norm": 0.15914912521839142, + "learning_rate": 5.498682744603071e-05, + "loss": 0.0066, + "step": 49480 + }, + { + "epoch": 3.843883495145631, + "grad_norm": 0.17620952427387238, + "learning_rate": 5.497037491947194e-05, + "loss": 0.0061, + "step": 49490 + }, + { + "epoch": 3.8446601941747574, + "grad_norm": 0.13335978984832764, + "learning_rate": 5.4953921849360424e-05, + "loss": 0.0075, + "step": 49500 + }, + { + "epoch": 3.8454368932038836, + "grad_norm": 0.2038533240556717, + "learning_rate": 5.493746823749547e-05, + "loss": 0.0069, + "step": 49510 + }, + { + "epoch": 3.84621359223301, + "grad_norm": 0.20728366076946259, + "learning_rate": 5.49210140856764e-05, + "loss": 0.0079, + "step": 49520 + }, + { + "epoch": 3.846990291262136, + "grad_norm": 0.17790542542934418, + "learning_rate": 5.4904559395702647e-05, + "loss": 0.006, + "step": 49530 + }, + { + "epoch": 3.8477669902912623, + "grad_norm": 0.34953615069389343, + "learning_rate": 5.488810416937364e-05, + "loss": 0.0087, + "step": 49540 + }, + { + "epoch": 3.8485436893203886, + "grad_norm": 0.2456725686788559, + "learning_rate": 5.4871648408488926e-05, + "loss": 0.0062, + "step": 49550 + }, + { + "epoch": 3.8493203883495144, + "grad_norm": 0.12915319204330444, + "learning_rate": 5.485519211484807e-05, + "loss": 0.0074, + "step": 49560 + }, + { + "epoch": 3.8500970873786406, + "grad_norm": 0.1719759851694107, + "learning_rate": 5.4838735290250676e-05, + "loss": 0.0063, + "step": 49570 + }, + { + "epoch": 3.850873786407767, + "grad_norm": 0.1700073629617691, + "learning_rate": 5.482227793649649e-05, + "loss": 0.007, + "step": 49580 + }, + { + "epoch": 3.851650485436893, + "grad_norm": 0.27381622791290283, + "learning_rate": 5.480582005538524e-05, + "loss": 0.0083, + "step": 49590 + }, + { + "epoch": 3.8524271844660194, + "grad_norm": 0.23473021388053894, + "learning_rate": 5.478936164871671e-05, + "loss": 0.008, + "step": 49600 + }, + { + "epoch": 3.8532038834951456, + "grad_norm": 0.2337971329689026, + "learning_rate": 5.477290271829081e-05, + "loss": 0.0067, + "step": 49610 + }, + { + "epoch": 3.853980582524272, + "grad_norm": 0.5259634256362915, + "learning_rate": 5.475644326590744e-05, + "loss": 0.0075, + "step": 49620 + }, + { + "epoch": 3.854757281553398, + "grad_norm": 0.18795952200889587, + "learning_rate": 5.473998329336658e-05, + "loss": 0.0075, + "step": 49630 + }, + { + "epoch": 3.8555339805825244, + "grad_norm": 0.3462913930416107, + "learning_rate": 5.4723522802468286e-05, + "loss": 0.0077, + "step": 49640 + }, + { + "epoch": 3.8563106796116506, + "grad_norm": 0.2635004222393036, + "learning_rate": 5.470706179501264e-05, + "loss": 0.007, + "step": 49650 + }, + { + "epoch": 3.8570873786407764, + "grad_norm": 0.26264891028404236, + "learning_rate": 5.4690600272799776e-05, + "loss": 0.0082, + "step": 49660 + }, + { + "epoch": 3.8578640776699027, + "grad_norm": 0.14320887625217438, + "learning_rate": 5.467413823762993e-05, + "loss": 0.0073, + "step": 49670 + }, + { + "epoch": 3.858640776699029, + "grad_norm": 0.31417116522789, + "learning_rate": 5.465767569130335e-05, + "loss": 0.0074, + "step": 49680 + }, + { + "epoch": 3.859417475728155, + "grad_norm": 0.14322030544281006, + "learning_rate": 5.464121263562036e-05, + "loss": 0.0069, + "step": 49690 + }, + { + "epoch": 3.8601941747572814, + "grad_norm": 0.1911761313676834, + "learning_rate": 5.4624749072381343e-05, + "loss": 0.0074, + "step": 49700 + }, + { + "epoch": 3.8609708737864077, + "grad_norm": 0.1694534868001938, + "learning_rate": 5.460828500338672e-05, + "loss": 0.0071, + "step": 49710 + }, + { + "epoch": 3.861747572815534, + "grad_norm": 0.26300865411758423, + "learning_rate": 5.459182043043698e-05, + "loss": 0.0051, + "step": 49720 + }, + { + "epoch": 3.86252427184466, + "grad_norm": 0.20258614420890808, + "learning_rate": 5.457535535533265e-05, + "loss": 0.0085, + "step": 49730 + }, + { + "epoch": 3.8633009708737864, + "grad_norm": 0.21733994781970978, + "learning_rate": 5.4558889779874334e-05, + "loss": 0.0075, + "step": 49740 + }, + { + "epoch": 3.8640776699029127, + "grad_norm": 0.20611870288848877, + "learning_rate": 5.454242370586269e-05, + "loss": 0.0072, + "step": 49750 + }, + { + "epoch": 3.864854368932039, + "grad_norm": 0.24729903042316437, + "learning_rate": 5.452595713509843e-05, + "loss": 0.008, + "step": 49760 + }, + { + "epoch": 3.865631067961165, + "grad_norm": 0.23087885975837708, + "learning_rate": 5.4509490069382274e-05, + "loss": 0.0073, + "step": 49770 + }, + { + "epoch": 3.8664077669902914, + "grad_norm": 0.3584502637386322, + "learning_rate": 5.449302251051509e-05, + "loss": 0.0065, + "step": 49780 + }, + { + "epoch": 3.8671844660194177, + "grad_norm": 0.34142714738845825, + "learning_rate": 5.44765544602977e-05, + "loss": 0.0064, + "step": 49790 + }, + { + "epoch": 3.867961165048544, + "grad_norm": 0.1432046741247177, + "learning_rate": 5.446008592053107e-05, + "loss": 0.0085, + "step": 49800 + }, + { + "epoch": 3.86873786407767, + "grad_norm": 0.23146896064281464, + "learning_rate": 5.4443616893016135e-05, + "loss": 0.0067, + "step": 49810 + }, + { + "epoch": 3.869514563106796, + "grad_norm": 0.5157952308654785, + "learning_rate": 5.4427147379553934e-05, + "loss": 0.0082, + "step": 49820 + }, + { + "epoch": 3.8702912621359222, + "grad_norm": 0.18623462319374084, + "learning_rate": 5.441067738194556e-05, + "loss": 0.0098, + "step": 49830 + }, + { + "epoch": 3.8710679611650485, + "grad_norm": 0.15688146650791168, + "learning_rate": 5.439420690199214e-05, + "loss": 0.0071, + "step": 49840 + }, + { + "epoch": 3.8718446601941747, + "grad_norm": 0.2778684198856354, + "learning_rate": 5.437773594149484e-05, + "loss": 0.0064, + "step": 49850 + }, + { + "epoch": 3.872621359223301, + "grad_norm": 0.19820579886436462, + "learning_rate": 5.436126450225495e-05, + "loss": 0.0074, + "step": 49860 + }, + { + "epoch": 3.8733980582524272, + "grad_norm": 0.4491075277328491, + "learning_rate": 5.434479258607371e-05, + "loss": 0.0073, + "step": 49870 + }, + { + "epoch": 3.8741747572815535, + "grad_norm": 0.21650920808315277, + "learning_rate": 5.43283201947525e-05, + "loss": 0.0083, + "step": 49880 + }, + { + "epoch": 3.8749514563106797, + "grad_norm": 0.2670389711856842, + "learning_rate": 5.431184733009268e-05, + "loss": 0.0061, + "step": 49890 + }, + { + "epoch": 3.875728155339806, + "grad_norm": 0.16747553646564484, + "learning_rate": 5.4295373993895736e-05, + "loss": 0.0062, + "step": 49900 + }, + { + "epoch": 3.8765048543689318, + "grad_norm": 0.19268934428691864, + "learning_rate": 5.4278900187963157e-05, + "loss": 0.0084, + "step": 49910 + }, + { + "epoch": 3.877281553398058, + "grad_norm": 0.27938565611839294, + "learning_rate": 5.426242591409646e-05, + "loss": 0.0073, + "step": 49920 + }, + { + "epoch": 3.8780582524271843, + "grad_norm": 0.3761257529258728, + "learning_rate": 5.42459511740973e-05, + "loss": 0.0067, + "step": 49930 + }, + { + "epoch": 3.8788349514563105, + "grad_norm": 0.21452972292900085, + "learning_rate": 5.422947596976729e-05, + "loss": 0.0067, + "step": 49940 + }, + { + "epoch": 3.8796116504854368, + "grad_norm": 0.2596321105957031, + "learning_rate": 5.4213000302908134e-05, + "loss": 0.0071, + "step": 49950 + }, + { + "epoch": 3.880388349514563, + "grad_norm": 0.19410263001918793, + "learning_rate": 5.419652417532162e-05, + "loss": 0.0057, + "step": 49960 + }, + { + "epoch": 3.8811650485436893, + "grad_norm": 0.33154159784317017, + "learning_rate": 5.4180047588809534e-05, + "loss": 0.0066, + "step": 49970 + }, + { + "epoch": 3.8819417475728155, + "grad_norm": 0.6064377427101135, + "learning_rate": 5.4163570545173704e-05, + "loss": 0.0062, + "step": 49980 + }, + { + "epoch": 3.8827184466019418, + "grad_norm": 0.3323858380317688, + "learning_rate": 5.414709304621608e-05, + "loss": 0.0093, + "step": 49990 + }, + { + "epoch": 3.883495145631068, + "grad_norm": 0.7221822738647461, + "learning_rate": 5.413061509373858e-05, + "loss": 0.0066, + "step": 50000 + }, + { + "epoch": 3.8842718446601943, + "grad_norm": 0.14902102947235107, + "learning_rate": 5.411413668954324e-05, + "loss": 0.0076, + "step": 50010 + }, + { + "epoch": 3.8850485436893205, + "grad_norm": 0.23579464852809906, + "learning_rate": 5.4097657835432083e-05, + "loss": 0.0067, + "step": 50020 + }, + { + "epoch": 3.8858252427184468, + "grad_norm": 0.24576203525066376, + "learning_rate": 5.408117853320723e-05, + "loss": 0.0068, + "step": 50030 + }, + { + "epoch": 3.886601941747573, + "grad_norm": 0.39033395051956177, + "learning_rate": 5.406469878467084e-05, + "loss": 0.008, + "step": 50040 + }, + { + "epoch": 3.8873786407766993, + "grad_norm": 0.2213078737258911, + "learning_rate": 5.404821859162509e-05, + "loss": 0.006, + "step": 50050 + }, + { + "epoch": 3.8881553398058255, + "grad_norm": 0.16802066564559937, + "learning_rate": 5.403173795587225e-05, + "loss": 0.0077, + "step": 50060 + }, + { + "epoch": 3.8889320388349513, + "grad_norm": 0.3609966039657593, + "learning_rate": 5.4015256879214606e-05, + "loss": 0.0067, + "step": 50070 + }, + { + "epoch": 3.8897087378640776, + "grad_norm": 0.18383269011974335, + "learning_rate": 5.39987753634545e-05, + "loss": 0.0066, + "step": 50080 + }, + { + "epoch": 3.890485436893204, + "grad_norm": 0.19248121976852417, + "learning_rate": 5.3982293410394335e-05, + "loss": 0.0083, + "step": 50090 + }, + { + "epoch": 3.89126213592233, + "grad_norm": 0.2715592384338379, + "learning_rate": 5.396581102183655e-05, + "loss": 0.0082, + "step": 50100 + }, + { + "epoch": 3.8920388349514563, + "grad_norm": 0.15401580929756165, + "learning_rate": 5.3949328199583634e-05, + "loss": 0.0071, + "step": 50110 + }, + { + "epoch": 3.8928155339805826, + "grad_norm": 0.22612538933753967, + "learning_rate": 5.393284494543813e-05, + "loss": 0.0058, + "step": 50120 + }, + { + "epoch": 3.893592233009709, + "grad_norm": 0.2113766223192215, + "learning_rate": 5.391636126120262e-05, + "loss": 0.0092, + "step": 50130 + }, + { + "epoch": 3.894368932038835, + "grad_norm": 0.17959174513816833, + "learning_rate": 5.389987714867971e-05, + "loss": 0.0091, + "step": 50140 + }, + { + "epoch": 3.8951456310679613, + "grad_norm": 0.1630610078573227, + "learning_rate": 5.3883392609672114e-05, + "loss": 0.0065, + "step": 50150 + }, + { + "epoch": 3.895922330097087, + "grad_norm": 0.24772703647613525, + "learning_rate": 5.3866907645982545e-05, + "loss": 0.006, + "step": 50160 + }, + { + "epoch": 3.8966990291262134, + "grad_norm": 0.14541426301002502, + "learning_rate": 5.385042225941375e-05, + "loss": 0.0059, + "step": 50170 + }, + { + "epoch": 3.8974757281553396, + "grad_norm": 0.22750838100910187, + "learning_rate": 5.3833936451768585e-05, + "loss": 0.0057, + "step": 50180 + }, + { + "epoch": 3.898252427184466, + "grad_norm": 0.6037907600402832, + "learning_rate": 5.3817450224849884e-05, + "loss": 0.0084, + "step": 50190 + }, + { + "epoch": 3.899029126213592, + "grad_norm": 0.482422798871994, + "learning_rate": 5.380096358046056e-05, + "loss": 0.0062, + "step": 50200 + }, + { + "epoch": 3.8998058252427183, + "grad_norm": 0.26661041378974915, + "learning_rate": 5.378447652040359e-05, + "loss": 0.0064, + "step": 50210 + }, + { + "epoch": 3.9005825242718446, + "grad_norm": 0.37884238362312317, + "learning_rate": 5.376798904648195e-05, + "loss": 0.0081, + "step": 50220 + }, + { + "epoch": 3.901359223300971, + "grad_norm": 0.5044795870780945, + "learning_rate": 5.375150116049869e-05, + "loss": 0.0066, + "step": 50230 + }, + { + "epoch": 3.902135922330097, + "grad_norm": 0.1837521493434906, + "learning_rate": 5.373501286425691e-05, + "loss": 0.0059, + "step": 50240 + }, + { + "epoch": 3.9029126213592233, + "grad_norm": 0.9819009900093079, + "learning_rate": 5.3718524159559726e-05, + "loss": 0.0083, + "step": 50250 + }, + { + "epoch": 3.9036893203883496, + "grad_norm": 0.6754916906356812, + "learning_rate": 5.370203504821034e-05, + "loss": 0.0076, + "step": 50260 + }, + { + "epoch": 3.904466019417476, + "grad_norm": 0.3228212296962738, + "learning_rate": 5.368554553201196e-05, + "loss": 0.0086, + "step": 50270 + }, + { + "epoch": 3.905242718446602, + "grad_norm": 0.23694360256195068, + "learning_rate": 5.3669055612767874e-05, + "loss": 0.0063, + "step": 50280 + }, + { + "epoch": 3.9060194174757283, + "grad_norm": 0.4691196382045746, + "learning_rate": 5.3652565292281376e-05, + "loss": 0.0094, + "step": 50290 + }, + { + "epoch": 3.9067961165048546, + "grad_norm": 0.21321263909339905, + "learning_rate": 5.3636074572355824e-05, + "loss": 0.0055, + "step": 50300 + }, + { + "epoch": 3.907572815533981, + "grad_norm": 0.1271190643310547, + "learning_rate": 5.361958345479463e-05, + "loss": 0.0055, + "step": 50310 + }, + { + "epoch": 3.9083495145631066, + "grad_norm": 0.2108016163110733, + "learning_rate": 5.360309194140123e-05, + "loss": 0.0079, + "step": 50320 + }, + { + "epoch": 3.909126213592233, + "grad_norm": 0.31498295068740845, + "learning_rate": 5.3586600033979106e-05, + "loss": 0.0046, + "step": 50330 + }, + { + "epoch": 3.909902912621359, + "grad_norm": 0.17970915138721466, + "learning_rate": 5.357010773433181e-05, + "loss": 0.0072, + "step": 50340 + }, + { + "epoch": 3.9106796116504854, + "grad_norm": 0.5402306914329529, + "learning_rate": 5.3553615044262895e-05, + "loss": 0.0071, + "step": 50350 + }, + { + "epoch": 3.9114563106796116, + "grad_norm": 0.21540400385856628, + "learning_rate": 5.3537121965575964e-05, + "loss": 0.0081, + "step": 50360 + }, + { + "epoch": 3.912233009708738, + "grad_norm": 0.2105969786643982, + "learning_rate": 5.352062850007471e-05, + "loss": 0.0071, + "step": 50370 + }, + { + "epoch": 3.913009708737864, + "grad_norm": 0.3516288995742798, + "learning_rate": 5.350413464956282e-05, + "loss": 0.0069, + "step": 50380 + }, + { + "epoch": 3.9137864077669904, + "grad_norm": 0.22278396785259247, + "learning_rate": 5.348764041584403e-05, + "loss": 0.0068, + "step": 50390 + }, + { + "epoch": 3.9145631067961166, + "grad_norm": 0.32385605573654175, + "learning_rate": 5.347114580072212e-05, + "loss": 0.006, + "step": 50400 + }, + { + "epoch": 3.915339805825243, + "grad_norm": 0.13940221071243286, + "learning_rate": 5.345465080600093e-05, + "loss": 0.0075, + "step": 50410 + }, + { + "epoch": 3.9161165048543687, + "grad_norm": 0.2360846847295761, + "learning_rate": 5.3438155433484326e-05, + "loss": 0.0085, + "step": 50420 + }, + { + "epoch": 3.916893203883495, + "grad_norm": 0.17041471600532532, + "learning_rate": 5.3421659684976197e-05, + "loss": 0.0074, + "step": 50430 + }, + { + "epoch": 3.917669902912621, + "grad_norm": 0.2053675651550293, + "learning_rate": 5.340516356228052e-05, + "loss": 0.0093, + "step": 50440 + }, + { + "epoch": 3.9184466019417474, + "grad_norm": 0.20536957681179047, + "learning_rate": 5.338866706720128e-05, + "loss": 0.0045, + "step": 50450 + }, + { + "epoch": 3.9192233009708737, + "grad_norm": 0.14869126677513123, + "learning_rate": 5.337217020154249e-05, + "loss": 0.0083, + "step": 50460 + }, + { + "epoch": 3.92, + "grad_norm": 0.4963908791542053, + "learning_rate": 5.335567296710825e-05, + "loss": 0.0079, + "step": 50470 + }, + { + "epoch": 3.920776699029126, + "grad_norm": 0.21201243996620178, + "learning_rate": 5.333917536570265e-05, + "loss": 0.0075, + "step": 50480 + }, + { + "epoch": 3.9215533980582524, + "grad_norm": 0.30969128012657166, + "learning_rate": 5.332267739912986e-05, + "loss": 0.0069, + "step": 50490 + }, + { + "epoch": 3.9223300970873787, + "grad_norm": 0.31086739897727966, + "learning_rate": 5.330617906919405e-05, + "loss": 0.0089, + "step": 50500 + }, + { + "epoch": 3.923106796116505, + "grad_norm": 0.21746523678302765, + "learning_rate": 5.328968037769949e-05, + "loss": 0.0081, + "step": 50510 + }, + { + "epoch": 3.923883495145631, + "grad_norm": 0.24895143508911133, + "learning_rate": 5.327318132645042e-05, + "loss": 0.0076, + "step": 50520 + }, + { + "epoch": 3.9246601941747574, + "grad_norm": 0.30395856499671936, + "learning_rate": 5.325668191725116e-05, + "loss": 0.0095, + "step": 50530 + }, + { + "epoch": 3.9254368932038837, + "grad_norm": 0.36849677562713623, + "learning_rate": 5.324018215190606e-05, + "loss": 0.0087, + "step": 50540 + }, + { + "epoch": 3.92621359223301, + "grad_norm": 0.1854267567396164, + "learning_rate": 5.3223682032219515e-05, + "loss": 0.0066, + "step": 50550 + }, + { + "epoch": 3.926990291262136, + "grad_norm": 0.5114126801490784, + "learning_rate": 5.320718155999595e-05, + "loss": 0.008, + "step": 50560 + }, + { + "epoch": 3.927766990291262, + "grad_norm": 0.18009471893310547, + "learning_rate": 5.3190680737039835e-05, + "loss": 0.0067, + "step": 50570 + }, + { + "epoch": 3.928543689320388, + "grad_norm": 0.16272956132888794, + "learning_rate": 5.317417956515567e-05, + "loss": 0.0069, + "step": 50580 + }, + { + "epoch": 3.9293203883495145, + "grad_norm": 0.32422110438346863, + "learning_rate": 5.315767804614798e-05, + "loss": 0.0073, + "step": 50590 + }, + { + "epoch": 3.9300970873786407, + "grad_norm": 0.4112941324710846, + "learning_rate": 5.3141176181821395e-05, + "loss": 0.0077, + "step": 50600 + }, + { + "epoch": 3.930873786407767, + "grad_norm": 0.17245927453041077, + "learning_rate": 5.31246739739805e-05, + "loss": 0.0064, + "step": 50610 + }, + { + "epoch": 3.931650485436893, + "grad_norm": 0.22594456374645233, + "learning_rate": 5.310817142442994e-05, + "loss": 0.0052, + "step": 50620 + }, + { + "epoch": 3.9324271844660195, + "grad_norm": 1.6676913499832153, + "learning_rate": 5.309166853497445e-05, + "loss": 0.0081, + "step": 50630 + }, + { + "epoch": 3.9332038834951457, + "grad_norm": 0.2101881206035614, + "learning_rate": 5.307516530741873e-05, + "loss": 0.0078, + "step": 50640 + }, + { + "epoch": 3.933980582524272, + "grad_norm": 0.43968310952186584, + "learning_rate": 5.305866174356754e-05, + "loss": 0.0068, + "step": 50650 + }, + { + "epoch": 3.934757281553398, + "grad_norm": 0.20703813433647156, + "learning_rate": 5.304215784522571e-05, + "loss": 0.0065, + "step": 50660 + }, + { + "epoch": 3.935533980582524, + "grad_norm": 0.18152618408203125, + "learning_rate": 5.302565361419808e-05, + "loss": 0.006, + "step": 50670 + }, + { + "epoch": 3.9363106796116503, + "grad_norm": 0.21790315210819244, + "learning_rate": 5.3009149052289507e-05, + "loss": 0.0057, + "step": 50680 + }, + { + "epoch": 3.9370873786407765, + "grad_norm": 0.4055776000022888, + "learning_rate": 5.299264416130493e-05, + "loss": 0.0098, + "step": 50690 + }, + { + "epoch": 3.9378640776699028, + "grad_norm": 0.42873615026474, + "learning_rate": 5.297613894304928e-05, + "loss": 0.008, + "step": 50700 + }, + { + "epoch": 3.938640776699029, + "grad_norm": 0.30394110083580017, + "learning_rate": 5.2959633399327534e-05, + "loss": 0.0069, + "step": 50710 + }, + { + "epoch": 3.9394174757281553, + "grad_norm": 0.2084234207868576, + "learning_rate": 5.294312753194476e-05, + "loss": 0.006, + "step": 50720 + }, + { + "epoch": 3.9401941747572815, + "grad_norm": 0.38279587030410767, + "learning_rate": 5.292662134270596e-05, + "loss": 0.0064, + "step": 50730 + }, + { + "epoch": 3.9409708737864078, + "grad_norm": 0.28177231550216675, + "learning_rate": 5.291011483341626e-05, + "loss": 0.0082, + "step": 50740 + }, + { + "epoch": 3.941747572815534, + "grad_norm": 0.28949740529060364, + "learning_rate": 5.2893608005880767e-05, + "loss": 0.0072, + "step": 50750 + }, + { + "epoch": 3.9425242718446603, + "grad_norm": 0.2173728048801422, + "learning_rate": 5.287710086190467e-05, + "loss": 0.0054, + "step": 50760 + }, + { + "epoch": 3.9433009708737865, + "grad_norm": 0.0713832899928093, + "learning_rate": 5.286059340329314e-05, + "loss": 0.0068, + "step": 50770 + }, + { + "epoch": 3.9440776699029128, + "grad_norm": 0.15718494355678558, + "learning_rate": 5.284408563185141e-05, + "loss": 0.0065, + "step": 50780 + }, + { + "epoch": 3.944854368932039, + "grad_norm": 0.7821317911148071, + "learning_rate": 5.282757754938476e-05, + "loss": 0.0065, + "step": 50790 + }, + { + "epoch": 3.9456310679611653, + "grad_norm": 0.25043633580207825, + "learning_rate": 5.281106915769849e-05, + "loss": 0.0065, + "step": 50800 + }, + { + "epoch": 3.9464077669902915, + "grad_norm": 0.7534846067428589, + "learning_rate": 5.2794560458597897e-05, + "loss": 0.0051, + "step": 50810 + }, + { + "epoch": 3.9471844660194177, + "grad_norm": 0.23177969455718994, + "learning_rate": 5.27780514538884e-05, + "loss": 0.0061, + "step": 50820 + }, + { + "epoch": 3.9479611650485436, + "grad_norm": 0.1571986824274063, + "learning_rate": 5.2761542145375365e-05, + "loss": 0.0053, + "step": 50830 + }, + { + "epoch": 3.94873786407767, + "grad_norm": 0.36626842617988586, + "learning_rate": 5.274503253486421e-05, + "loss": 0.0094, + "step": 50840 + }, + { + "epoch": 3.949514563106796, + "grad_norm": 0.3548094928264618, + "learning_rate": 5.272852262416046e-05, + "loss": 0.007, + "step": 50850 + }, + { + "epoch": 3.9502912621359223, + "grad_norm": 0.281990110874176, + "learning_rate": 5.2712012415069555e-05, + "loss": 0.0079, + "step": 50860 + }, + { + "epoch": 3.9510679611650485, + "grad_norm": 0.1756623238325119, + "learning_rate": 5.269550190939705e-05, + "loss": 0.0053, + "step": 50870 + }, + { + "epoch": 3.951844660194175, + "grad_norm": 0.19924761354923248, + "learning_rate": 5.267899110894852e-05, + "loss": 0.0079, + "step": 50880 + }, + { + "epoch": 3.952621359223301, + "grad_norm": 0.1796964406967163, + "learning_rate": 5.266248001552955e-05, + "loss": 0.0083, + "step": 50890 + }, + { + "epoch": 3.9533980582524273, + "grad_norm": 0.6215372085571289, + "learning_rate": 5.264596863094575e-05, + "loss": 0.0067, + "step": 50900 + }, + { + "epoch": 3.9541747572815535, + "grad_norm": 0.3289305567741394, + "learning_rate": 5.262945695700282e-05, + "loss": 0.0103, + "step": 50910 + }, + { + "epoch": 3.9549514563106793, + "grad_norm": 0.42126592993736267, + "learning_rate": 5.261294499550643e-05, + "loss": 0.0069, + "step": 50920 + }, + { + "epoch": 3.9557281553398056, + "grad_norm": 0.47283998131752014, + "learning_rate": 5.25964327482623e-05, + "loss": 0.0089, + "step": 50930 + }, + { + "epoch": 3.956504854368932, + "grad_norm": 0.3073953092098236, + "learning_rate": 5.257992021707617e-05, + "loss": 0.0058, + "step": 50940 + }, + { + "epoch": 3.957281553398058, + "grad_norm": 0.297558456659317, + "learning_rate": 5.256340740375387e-05, + "loss": 0.0075, + "step": 50950 + }, + { + "epoch": 3.9580582524271843, + "grad_norm": 0.2156520038843155, + "learning_rate": 5.254689431010117e-05, + "loss": 0.0091, + "step": 50960 + }, + { + "epoch": 3.9588349514563106, + "grad_norm": 0.30611157417297363, + "learning_rate": 5.253038093792395e-05, + "loss": 0.0076, + "step": 50970 + }, + { + "epoch": 3.959611650485437, + "grad_norm": 0.23673832416534424, + "learning_rate": 5.251386728902806e-05, + "loss": 0.0081, + "step": 50980 + }, + { + "epoch": 3.960388349514563, + "grad_norm": 0.22622962296009064, + "learning_rate": 5.2497353365219446e-05, + "loss": 0.0069, + "step": 50990 + }, + { + "epoch": 3.9611650485436893, + "grad_norm": 0.29322779178619385, + "learning_rate": 5.2480839168304e-05, + "loss": 0.0078, + "step": 51000 + }, + { + "epoch": 3.9619417475728156, + "grad_norm": 0.43089836835861206, + "learning_rate": 5.246432470008773e-05, + "loss": 0.0073, + "step": 51010 + }, + { + "epoch": 3.962718446601942, + "grad_norm": 0.4897266626358032, + "learning_rate": 5.2447809962376616e-05, + "loss": 0.008, + "step": 51020 + }, + { + "epoch": 3.963495145631068, + "grad_norm": 0.19995178282260895, + "learning_rate": 5.243129495697666e-05, + "loss": 0.0069, + "step": 51030 + }, + { + "epoch": 3.9642718446601943, + "grad_norm": 0.1587418168783188, + "learning_rate": 5.241477968569397e-05, + "loss": 0.0096, + "step": 51040 + }, + { + "epoch": 3.9650485436893206, + "grad_norm": 0.13197468221187592, + "learning_rate": 5.23982641503346e-05, + "loss": 0.0084, + "step": 51050 + }, + { + "epoch": 3.965825242718447, + "grad_norm": 0.2092394232749939, + "learning_rate": 5.238174835270465e-05, + "loss": 0.0069, + "step": 51060 + }, + { + "epoch": 3.966601941747573, + "grad_norm": 0.2142460197210312, + "learning_rate": 5.23652322946103e-05, + "loss": 0.0059, + "step": 51070 + }, + { + "epoch": 3.967378640776699, + "grad_norm": 0.2512320578098297, + "learning_rate": 5.234871597785771e-05, + "loss": 0.0067, + "step": 51080 + }, + { + "epoch": 3.968155339805825, + "grad_norm": 0.23119854927062988, + "learning_rate": 5.233219940425306e-05, + "loss": 0.009, + "step": 51090 + }, + { + "epoch": 3.9689320388349514, + "grad_norm": 0.23160488903522491, + "learning_rate": 5.2315682575602576e-05, + "loss": 0.0048, + "step": 51100 + }, + { + "epoch": 3.9697087378640776, + "grad_norm": 0.21241603791713715, + "learning_rate": 5.229916549371254e-05, + "loss": 0.0072, + "step": 51110 + }, + { + "epoch": 3.970485436893204, + "grad_norm": 0.2242727279663086, + "learning_rate": 5.228264816038924e-05, + "loss": 0.0062, + "step": 51120 + }, + { + "epoch": 3.97126213592233, + "grad_norm": 0.2589258849620819, + "learning_rate": 5.2266130577438935e-05, + "loss": 0.0069, + "step": 51130 + }, + { + "epoch": 3.9720388349514564, + "grad_norm": 0.2623409330844879, + "learning_rate": 5.224961274666802e-05, + "loss": 0.0054, + "step": 51140 + }, + { + "epoch": 3.9728155339805826, + "grad_norm": 0.20891517400741577, + "learning_rate": 5.223309466988283e-05, + "loss": 0.0078, + "step": 51150 + }, + { + "epoch": 3.973592233009709, + "grad_norm": 0.2629842758178711, + "learning_rate": 5.221657634888976e-05, + "loss": 0.0063, + "step": 51160 + }, + { + "epoch": 3.9743689320388347, + "grad_norm": 0.2670094966888428, + "learning_rate": 5.2200057785495235e-05, + "loss": 0.0115, + "step": 51170 + }, + { + "epoch": 3.975145631067961, + "grad_norm": 0.15196719765663147, + "learning_rate": 5.2183538981505706e-05, + "loss": 0.0052, + "step": 51180 + }, + { + "epoch": 3.975922330097087, + "grad_norm": 0.15680575370788574, + "learning_rate": 5.216701993872762e-05, + "loss": 0.0078, + "step": 51190 + }, + { + "epoch": 3.9766990291262134, + "grad_norm": 0.2971748411655426, + "learning_rate": 5.215050065896752e-05, + "loss": 0.0068, + "step": 51200 + }, + { + "epoch": 3.9774757281553397, + "grad_norm": 0.3739662766456604, + "learning_rate": 5.2133981144031875e-05, + "loss": 0.0066, + "step": 51210 + }, + { + "epoch": 3.978252427184466, + "grad_norm": 0.3860378861427307, + "learning_rate": 5.211746139572727e-05, + "loss": 0.0084, + "step": 51220 + }, + { + "epoch": 3.979029126213592, + "grad_norm": 0.30471763014793396, + "learning_rate": 5.210094141586026e-05, + "loss": 0.0063, + "step": 51230 + }, + { + "epoch": 3.9798058252427184, + "grad_norm": 0.2525365948677063, + "learning_rate": 5.2084421206237476e-05, + "loss": 0.0062, + "step": 51240 + }, + { + "epoch": 3.9805825242718447, + "grad_norm": 0.1977781057357788, + "learning_rate": 5.20679007686655e-05, + "loss": 0.0057, + "step": 51250 + }, + { + "epoch": 3.981359223300971, + "grad_norm": 0.4182509183883667, + "learning_rate": 5.205138010495102e-05, + "loss": 0.0075, + "step": 51260 + }, + { + "epoch": 3.982135922330097, + "grad_norm": 0.4104330241680145, + "learning_rate": 5.20348592169007e-05, + "loss": 0.0066, + "step": 51270 + }, + { + "epoch": 3.9829126213592234, + "grad_norm": 0.09855539351701736, + "learning_rate": 5.201833810632123e-05, + "loss": 0.0069, + "step": 51280 + }, + { + "epoch": 3.9836893203883497, + "grad_norm": 0.3098177909851074, + "learning_rate": 5.200181677501933e-05, + "loss": 0.0057, + "step": 51290 + }, + { + "epoch": 3.984466019417476, + "grad_norm": 0.24624435603618622, + "learning_rate": 5.1985295224801777e-05, + "loss": 0.0069, + "step": 51300 + }, + { + "epoch": 3.985242718446602, + "grad_norm": 0.14847750961780548, + "learning_rate": 5.196877345747531e-05, + "loss": 0.0042, + "step": 51310 + }, + { + "epoch": 3.9860194174757284, + "grad_norm": 0.30574831366539, + "learning_rate": 5.195225147484674e-05, + "loss": 0.0072, + "step": 51320 + }, + { + "epoch": 3.986796116504854, + "grad_norm": 0.4374869465827942, + "learning_rate": 5.193572927872288e-05, + "loss": 0.0064, + "step": 51330 + }, + { + "epoch": 3.9875728155339805, + "grad_norm": 0.7183743119239807, + "learning_rate": 5.191920687091059e-05, + "loss": 0.0067, + "step": 51340 + }, + { + "epoch": 3.9883495145631067, + "grad_norm": 0.17475540935993195, + "learning_rate": 5.1902684253216695e-05, + "loss": 0.0058, + "step": 51350 + }, + { + "epoch": 3.989126213592233, + "grad_norm": 0.3621765971183777, + "learning_rate": 5.1886161427448134e-05, + "loss": 0.0057, + "step": 51360 + }, + { + "epoch": 3.989902912621359, + "grad_norm": 0.19449147582054138, + "learning_rate": 5.186963839541179e-05, + "loss": 0.0067, + "step": 51370 + }, + { + "epoch": 3.9906796116504855, + "grad_norm": 0.2032705694437027, + "learning_rate": 5.185311515891459e-05, + "loss": 0.0053, + "step": 51380 + }, + { + "epoch": 3.9914563106796117, + "grad_norm": 0.33946970105171204, + "learning_rate": 5.1836591719763516e-05, + "loss": 0.0073, + "step": 51390 + }, + { + "epoch": 3.992233009708738, + "grad_norm": 0.24926184117794037, + "learning_rate": 5.182006807976554e-05, + "loss": 0.0057, + "step": 51400 + }, + { + "epoch": 3.993009708737864, + "grad_norm": 0.20870248973369598, + "learning_rate": 5.180354424072763e-05, + "loss": 0.006, + "step": 51410 + }, + { + "epoch": 3.9937864077669905, + "grad_norm": 0.20268505811691284, + "learning_rate": 5.178702020445686e-05, + "loss": 0.0074, + "step": 51420 + }, + { + "epoch": 3.9945631067961163, + "grad_norm": 0.15768469870090485, + "learning_rate": 5.177049597276024e-05, + "loss": 0.0056, + "step": 51430 + }, + { + "epoch": 3.9953398058252425, + "grad_norm": 0.3613198697566986, + "learning_rate": 5.175397154744482e-05, + "loss": 0.0075, + "step": 51440 + }, + { + "epoch": 3.9961165048543688, + "grad_norm": 0.72941654920578, + "learning_rate": 5.1737446930317736e-05, + "loss": 0.0051, + "step": 51450 + }, + { + "epoch": 3.996893203883495, + "grad_norm": 0.9422253966331482, + "learning_rate": 5.172092212318604e-05, + "loss": 0.0069, + "step": 51460 + }, + { + "epoch": 3.9976699029126213, + "grad_norm": 0.19209614396095276, + "learning_rate": 5.170439712785692e-05, + "loss": 0.0063, + "step": 51470 + }, + { + "epoch": 3.9984466019417475, + "grad_norm": 0.28401583433151245, + "learning_rate": 5.1687871946137467e-05, + "loss": 0.0089, + "step": 51480 + }, + { + "epoch": 3.9992233009708738, + "grad_norm": 0.44882288575172424, + "learning_rate": 5.16713465798349e-05, + "loss": 0.0056, + "step": 51490 + }, + { + "epoch": 4.0, + "grad_norm": 0.1764368712902069, + "learning_rate": 5.165482103075637e-05, + "loss": 0.0043, + "step": 51500 + }, + { + "epoch": 4.000776699029126, + "grad_norm": 0.18910230696201324, + "learning_rate": 5.163829530070909e-05, + "loss": 0.0065, + "step": 51510 + }, + { + "epoch": 4.0015533980582525, + "grad_norm": 0.19300992786884308, + "learning_rate": 5.162176939150032e-05, + "loss": 0.0055, + "step": 51520 + }, + { + "epoch": 4.002330097087379, + "grad_norm": 0.2408275306224823, + "learning_rate": 5.160524330493728e-05, + "loss": 0.006, + "step": 51530 + }, + { + "epoch": 4.003106796116505, + "grad_norm": 0.2790663242340088, + "learning_rate": 5.158871704282724e-05, + "loss": 0.0069, + "step": 51540 + }, + { + "epoch": 4.003883495145631, + "grad_norm": 0.14068034291267395, + "learning_rate": 5.15721906069775e-05, + "loss": 0.0051, + "step": 51550 + }, + { + "epoch": 4.0046601941747575, + "grad_norm": 0.22917687892913818, + "learning_rate": 5.155566399919536e-05, + "loss": 0.0074, + "step": 51560 + }, + { + "epoch": 4.005436893203884, + "grad_norm": 0.28923964500427246, + "learning_rate": 5.153913722128813e-05, + "loss": 0.0062, + "step": 51570 + }, + { + "epoch": 4.00621359223301, + "grad_norm": 0.2249757945537567, + "learning_rate": 5.1522610275063176e-05, + "loss": 0.0107, + "step": 51580 + }, + { + "epoch": 4.006990291262136, + "grad_norm": 0.1589442491531372, + "learning_rate": 5.150608316232786e-05, + "loss": 0.0065, + "step": 51590 + }, + { + "epoch": 4.0077669902912625, + "grad_norm": 0.8414888978004456, + "learning_rate": 5.148955588488955e-05, + "loss": 0.0071, + "step": 51600 + }, + { + "epoch": 4.008543689320389, + "grad_norm": 0.3198179006576538, + "learning_rate": 5.147302844455564e-05, + "loss": 0.0074, + "step": 51610 + }, + { + "epoch": 4.009320388349515, + "grad_norm": 0.1101248636841774, + "learning_rate": 5.145650084313356e-05, + "loss": 0.0041, + "step": 51620 + }, + { + "epoch": 4.01009708737864, + "grad_norm": 0.16587434709072113, + "learning_rate": 5.1439973082430734e-05, + "loss": 0.0057, + "step": 51630 + }, + { + "epoch": 4.010873786407767, + "grad_norm": 0.3036366105079651, + "learning_rate": 5.142344516425461e-05, + "loss": 0.0057, + "step": 51640 + }, + { + "epoch": 4.011650485436893, + "grad_norm": 0.356746643781662, + "learning_rate": 5.1406917090412673e-05, + "loss": 0.0068, + "step": 51650 + }, + { + "epoch": 4.012427184466019, + "grad_norm": 0.13062043488025665, + "learning_rate": 5.1390388862712404e-05, + "loss": 0.0071, + "step": 51660 + }, + { + "epoch": 4.013203883495145, + "grad_norm": 0.2199086993932724, + "learning_rate": 5.137386048296129e-05, + "loss": 0.0064, + "step": 51670 + }, + { + "epoch": 4.013980582524272, + "grad_norm": 0.16608333587646484, + "learning_rate": 5.135733195296686e-05, + "loss": 0.0074, + "step": 51680 + }, + { + "epoch": 4.014757281553398, + "grad_norm": 0.2833988070487976, + "learning_rate": 5.134080327453664e-05, + "loss": 0.0087, + "step": 51690 + }, + { + "epoch": 4.015533980582524, + "grad_norm": 0.5827057957649231, + "learning_rate": 5.1324274449478204e-05, + "loss": 0.0057, + "step": 51700 + }, + { + "epoch": 4.01631067961165, + "grad_norm": 0.2894308567047119, + "learning_rate": 5.130774547959909e-05, + "loss": 0.0063, + "step": 51710 + }, + { + "epoch": 4.017087378640777, + "grad_norm": 0.29086437821388245, + "learning_rate": 5.129121636670692e-05, + "loss": 0.0095, + "step": 51720 + }, + { + "epoch": 4.017864077669903, + "grad_norm": 0.3254387378692627, + "learning_rate": 5.1274687112609246e-05, + "loss": 0.0072, + "step": 51730 + }, + { + "epoch": 4.018640776699029, + "grad_norm": 0.37849730253219604, + "learning_rate": 5.125815771911372e-05, + "loss": 0.0071, + "step": 51740 + }, + { + "epoch": 4.019417475728155, + "grad_norm": 0.4599495232105255, + "learning_rate": 5.1241628188027956e-05, + "loss": 0.0062, + "step": 51750 + }, + { + "epoch": 4.020194174757282, + "grad_norm": 0.18651582300662994, + "learning_rate": 5.1225098521159585e-05, + "loss": 0.0055, + "step": 51760 + }, + { + "epoch": 4.020970873786408, + "grad_norm": 0.09045688807964325, + "learning_rate": 5.12085687203163e-05, + "loss": 0.0046, + "step": 51770 + }, + { + "epoch": 4.021747572815534, + "grad_norm": 0.26649197936058044, + "learning_rate": 5.119203878730575e-05, + "loss": 0.0084, + "step": 51780 + }, + { + "epoch": 4.02252427184466, + "grad_norm": 0.36953404545783997, + "learning_rate": 5.117550872393564e-05, + "loss": 0.0062, + "step": 51790 + }, + { + "epoch": 4.023300970873787, + "grad_norm": 0.21759364008903503, + "learning_rate": 5.1158978532013645e-05, + "loss": 0.0065, + "step": 51800 + }, + { + "epoch": 4.024077669902913, + "grad_norm": 0.48889830708503723, + "learning_rate": 5.114244821334751e-05, + "loss": 0.0054, + "step": 51810 + }, + { + "epoch": 4.024854368932039, + "grad_norm": 0.17980769276618958, + "learning_rate": 5.112591776974497e-05, + "loss": 0.0057, + "step": 51820 + }, + { + "epoch": 4.025631067961165, + "grad_norm": 0.18640422821044922, + "learning_rate": 5.110938720301374e-05, + "loss": 0.0059, + "step": 51830 + }, + { + "epoch": 4.026407766990292, + "grad_norm": 0.16664762794971466, + "learning_rate": 5.1092856514961595e-05, + "loss": 0.0094, + "step": 51840 + }, + { + "epoch": 4.027184466019418, + "grad_norm": 0.19657361507415771, + "learning_rate": 5.107632570739631e-05, + "loss": 0.0066, + "step": 51850 + }, + { + "epoch": 4.027961165048544, + "grad_norm": 0.20856493711471558, + "learning_rate": 5.1059794782125645e-05, + "loss": 0.0054, + "step": 51860 + }, + { + "epoch": 4.02873786407767, + "grad_norm": 0.6892892122268677, + "learning_rate": 5.104326374095744e-05, + "loss": 0.0085, + "step": 51870 + }, + { + "epoch": 4.029514563106796, + "grad_norm": 0.11219687014818192, + "learning_rate": 5.102673258569947e-05, + "loss": 0.0051, + "step": 51880 + }, + { + "epoch": 4.030291262135922, + "grad_norm": 0.3550974726676941, + "learning_rate": 5.101020131815956e-05, + "loss": 0.0087, + "step": 51890 + }, + { + "epoch": 4.031067961165048, + "grad_norm": 0.25691163539886475, + "learning_rate": 5.099366994014556e-05, + "loss": 0.009, + "step": 51900 + }, + { + "epoch": 4.031844660194174, + "grad_norm": 0.20682433247566223, + "learning_rate": 5.097713845346532e-05, + "loss": 0.0065, + "step": 51910 + }, + { + "epoch": 4.032621359223301, + "grad_norm": 0.22263449430465698, + "learning_rate": 5.096060685992668e-05, + "loss": 0.0065, + "step": 51920 + }, + { + "epoch": 4.033398058252427, + "grad_norm": 0.28593379259109497, + "learning_rate": 5.0944075161337514e-05, + "loss": 0.0095, + "step": 51930 + }, + { + "epoch": 4.034174757281553, + "grad_norm": 0.20152737200260162, + "learning_rate": 5.09275433595057e-05, + "loss": 0.0082, + "step": 51940 + }, + { + "epoch": 4.034951456310679, + "grad_norm": 0.560692548751831, + "learning_rate": 5.0911011456239157e-05, + "loss": 0.0118, + "step": 51950 + }, + { + "epoch": 4.035728155339806, + "grad_norm": 0.47986364364624023, + "learning_rate": 5.089447945334574e-05, + "loss": 0.0068, + "step": 51960 + }, + { + "epoch": 4.036504854368932, + "grad_norm": 0.17236094176769257, + "learning_rate": 5.087794735263343e-05, + "loss": 0.0075, + "step": 51970 + }, + { + "epoch": 4.037281553398058, + "grad_norm": 0.3281225562095642, + "learning_rate": 5.086141515591011e-05, + "loss": 0.0079, + "step": 51980 + }, + { + "epoch": 4.038058252427184, + "grad_norm": 0.21374793350696564, + "learning_rate": 5.084488286498371e-05, + "loss": 0.0069, + "step": 51990 + }, + { + "epoch": 4.038834951456311, + "grad_norm": 0.281960666179657, + "learning_rate": 5.08283504816622e-05, + "loss": 0.0063, + "step": 52000 + }, + { + "epoch": 4.039611650485437, + "grad_norm": 0.36417847871780396, + "learning_rate": 5.081181800775353e-05, + "loss": 0.0075, + "step": 52010 + }, + { + "epoch": 4.040388349514563, + "grad_norm": 0.24642500281333923, + "learning_rate": 5.079528544506565e-05, + "loss": 0.0082, + "step": 52020 + }, + { + "epoch": 4.041165048543689, + "grad_norm": 0.4102639853954315, + "learning_rate": 5.077875279540657e-05, + "loss": 0.0065, + "step": 52030 + }, + { + "epoch": 4.041941747572816, + "grad_norm": 0.29025983810424805, + "learning_rate": 5.076222006058424e-05, + "loss": 0.009, + "step": 52040 + }, + { + "epoch": 4.042718446601942, + "grad_norm": 0.25326278805732727, + "learning_rate": 5.074568724240667e-05, + "loss": 0.0082, + "step": 52050 + }, + { + "epoch": 4.043495145631068, + "grad_norm": 0.14763490855693817, + "learning_rate": 5.072915434268187e-05, + "loss": 0.0066, + "step": 52060 + }, + { + "epoch": 4.044271844660194, + "grad_norm": 0.4311290383338928, + "learning_rate": 5.071262136321785e-05, + "loss": 0.0081, + "step": 52070 + }, + { + "epoch": 4.045048543689321, + "grad_norm": 0.2612124979496002, + "learning_rate": 5.069608830582262e-05, + "loss": 0.0047, + "step": 52080 + }, + { + "epoch": 4.045825242718447, + "grad_norm": 0.12974992394447327, + "learning_rate": 5.067955517230423e-05, + "loss": 0.0085, + "step": 52090 + }, + { + "epoch": 4.046601941747573, + "grad_norm": 0.15684224665164948, + "learning_rate": 5.06630219644707e-05, + "loss": 0.0069, + "step": 52100 + }, + { + "epoch": 4.047378640776699, + "grad_norm": 0.2751147150993347, + "learning_rate": 5.064648868413008e-05, + "loss": 0.0083, + "step": 52110 + }, + { + "epoch": 4.048155339805826, + "grad_norm": 0.2672174572944641, + "learning_rate": 5.062995533309044e-05, + "loss": 0.0075, + "step": 52120 + }, + { + "epoch": 4.048932038834952, + "grad_norm": 0.34670117497444153, + "learning_rate": 5.061342191315983e-05, + "loss": 0.0059, + "step": 52130 + }, + { + "epoch": 4.049708737864077, + "grad_norm": 0.13028478622436523, + "learning_rate": 5.0596888426146325e-05, + "loss": 0.0061, + "step": 52140 + }, + { + "epoch": 4.0504854368932035, + "grad_norm": 0.46364423632621765, + "learning_rate": 5.058035487385798e-05, + "loss": 0.0069, + "step": 52150 + }, + { + "epoch": 4.05126213592233, + "grad_norm": 0.3487398624420166, + "learning_rate": 5.0563821258102915e-05, + "loss": 0.0062, + "step": 52160 + }, + { + "epoch": 4.052038834951456, + "grad_norm": 0.21982160210609436, + "learning_rate": 5.05472875806892e-05, + "loss": 0.005, + "step": 52170 + }, + { + "epoch": 4.052815533980582, + "grad_norm": 0.17897570133209229, + "learning_rate": 5.053075384342494e-05, + "loss": 0.0052, + "step": 52180 + }, + { + "epoch": 4.0535922330097085, + "grad_norm": 0.22101712226867676, + "learning_rate": 5.0514220048118234e-05, + "loss": 0.005, + "step": 52190 + }, + { + "epoch": 4.054368932038835, + "grad_norm": 0.1353878378868103, + "learning_rate": 5.049768619657721e-05, + "loss": 0.0082, + "step": 52200 + }, + { + "epoch": 4.055145631067961, + "grad_norm": 0.3229408860206604, + "learning_rate": 5.048115229060996e-05, + "loss": 0.0076, + "step": 52210 + }, + { + "epoch": 4.055922330097087, + "grad_norm": 0.14692850410938263, + "learning_rate": 5.046461833202462e-05, + "loss": 0.0059, + "step": 52220 + }, + { + "epoch": 4.0566990291262135, + "grad_norm": 0.16961538791656494, + "learning_rate": 5.044808432262932e-05, + "loss": 0.0072, + "step": 52230 + }, + { + "epoch": 4.05747572815534, + "grad_norm": 0.24236968159675598, + "learning_rate": 5.043155026423219e-05, + "loss": 0.0065, + "step": 52240 + }, + { + "epoch": 4.058252427184466, + "grad_norm": 0.14914950728416443, + "learning_rate": 5.041501615864137e-05, + "loss": 0.008, + "step": 52250 + }, + { + "epoch": 4.059029126213592, + "grad_norm": 0.23581267893314362, + "learning_rate": 5.039848200766502e-05, + "loss": 0.006, + "step": 52260 + }, + { + "epoch": 4.0598058252427185, + "grad_norm": 0.2128678858280182, + "learning_rate": 5.038194781311125e-05, + "loss": 0.0075, + "step": 52270 + }, + { + "epoch": 4.060582524271845, + "grad_norm": 0.23310166597366333, + "learning_rate": 5.0365413576788264e-05, + "loss": 0.0063, + "step": 52280 + }, + { + "epoch": 4.061359223300971, + "grad_norm": 0.3660750389099121, + "learning_rate": 5.0348879300504194e-05, + "loss": 0.0096, + "step": 52290 + }, + { + "epoch": 4.062135922330097, + "grad_norm": 0.24282321333885193, + "learning_rate": 5.033234498606721e-05, + "loss": 0.0078, + "step": 52300 + }, + { + "epoch": 4.0629126213592235, + "grad_norm": 0.48727139830589294, + "learning_rate": 5.0315810635285476e-05, + "loss": 0.0086, + "step": 52310 + }, + { + "epoch": 4.06368932038835, + "grad_norm": 0.15082019567489624, + "learning_rate": 5.0299276249967164e-05, + "loss": 0.0059, + "step": 52320 + }, + { + "epoch": 4.064466019417476, + "grad_norm": 0.27909424901008606, + "learning_rate": 5.0282741831920454e-05, + "loss": 0.0079, + "step": 52330 + }, + { + "epoch": 4.065242718446602, + "grad_norm": 0.28786230087280273, + "learning_rate": 5.026620738295351e-05, + "loss": 0.0062, + "step": 52340 + }, + { + "epoch": 4.0660194174757285, + "grad_norm": 0.1506688892841339, + "learning_rate": 5.024967290487455e-05, + "loss": 0.0059, + "step": 52350 + }, + { + "epoch": 4.066796116504855, + "grad_norm": 0.14493593573570251, + "learning_rate": 5.023313839949172e-05, + "loss": 0.0046, + "step": 52360 + }, + { + "epoch": 4.067572815533981, + "grad_norm": 0.13993264734745026, + "learning_rate": 5.0216603868613224e-05, + "loss": 0.0061, + "step": 52370 + }, + { + "epoch": 4.068349514563107, + "grad_norm": 0.6302933096885681, + "learning_rate": 5.020006931404726e-05, + "loss": 0.0085, + "step": 52380 + }, + { + "epoch": 4.069126213592233, + "grad_norm": 0.11718573421239853, + "learning_rate": 5.0183534737602014e-05, + "loss": 0.0072, + "step": 52390 + }, + { + "epoch": 4.069902912621359, + "grad_norm": 0.4316020607948303, + "learning_rate": 5.0167000141085686e-05, + "loss": 0.0073, + "step": 52400 + }, + { + "epoch": 4.070679611650485, + "grad_norm": 0.1392776072025299, + "learning_rate": 5.015046552630647e-05, + "loss": 0.0088, + "step": 52410 + }, + { + "epoch": 4.071456310679611, + "grad_norm": 0.17991310358047485, + "learning_rate": 5.013393089507258e-05, + "loss": 0.0057, + "step": 52420 + }, + { + "epoch": 4.072233009708738, + "grad_norm": 0.20488987863063812, + "learning_rate": 5.011739624919219e-05, + "loss": 0.0052, + "step": 52430 + }, + { + "epoch": 4.073009708737864, + "grad_norm": 0.19889773428440094, + "learning_rate": 5.0100861590473545e-05, + "loss": 0.0072, + "step": 52440 + }, + { + "epoch": 4.07378640776699, + "grad_norm": 0.34438052773475647, + "learning_rate": 5.008432692072482e-05, + "loss": 0.0072, + "step": 52450 + }, + { + "epoch": 4.074563106796116, + "grad_norm": 0.1406421661376953, + "learning_rate": 5.006779224175423e-05, + "loss": 0.0096, + "step": 52460 + }, + { + "epoch": 4.075339805825243, + "grad_norm": 0.29037386178970337, + "learning_rate": 5.005125755536998e-05, + "loss": 0.0069, + "step": 52470 + }, + { + "epoch": 4.076116504854369, + "grad_norm": 0.2224888950586319, + "learning_rate": 5.003472286338029e-05, + "loss": 0.0051, + "step": 52480 + }, + { + "epoch": 4.076893203883495, + "grad_norm": 0.23126892745494843, + "learning_rate": 5.001818816759335e-05, + "loss": 0.0064, + "step": 52490 + }, + { + "epoch": 4.077669902912621, + "grad_norm": 0.28130143880844116, + "learning_rate": 5.000165346981738e-05, + "loss": 0.0061, + "step": 52500 + }, + { + "epoch": 4.078446601941748, + "grad_norm": 0.18472129106521606, + "learning_rate": 4.9985118771860595e-05, + "loss": 0.0138, + "step": 52510 + }, + { + "epoch": 4.079223300970874, + "grad_norm": 0.15747079253196716, + "learning_rate": 4.996858407553119e-05, + "loss": 0.0073, + "step": 52520 + }, + { + "epoch": 4.08, + "grad_norm": 0.1306104212999344, + "learning_rate": 4.99520493826374e-05, + "loss": 0.0054, + "step": 52530 + }, + { + "epoch": 4.080776699029126, + "grad_norm": 0.23252834379673004, + "learning_rate": 4.9935514694987384e-05, + "loss": 0.0082, + "step": 52540 + }, + { + "epoch": 4.081553398058253, + "grad_norm": 0.16733604669570923, + "learning_rate": 4.991898001438941e-05, + "loss": 0.0072, + "step": 52550 + }, + { + "epoch": 4.082330097087379, + "grad_norm": 0.2867908179759979, + "learning_rate": 4.990244534265166e-05, + "loss": 0.0077, + "step": 52560 + }, + { + "epoch": 4.083106796116505, + "grad_norm": 0.2708650529384613, + "learning_rate": 4.988591068158234e-05, + "loss": 0.0061, + "step": 52570 + }, + { + "epoch": 4.083883495145631, + "grad_norm": 0.2209704965353012, + "learning_rate": 4.986937603298965e-05, + "loss": 0.0058, + "step": 52580 + }, + { + "epoch": 4.084660194174758, + "grad_norm": 0.3222880959510803, + "learning_rate": 4.985284139868181e-05, + "loss": 0.0062, + "step": 52590 + }, + { + "epoch": 4.085436893203884, + "grad_norm": 0.5106932520866394, + "learning_rate": 4.9836306780467e-05, + "loss": 0.0038, + "step": 52600 + }, + { + "epoch": 4.08621359223301, + "grad_norm": 0.40749257802963257, + "learning_rate": 4.981977218015343e-05, + "loss": 0.0115, + "step": 52610 + }, + { + "epoch": 4.086990291262136, + "grad_norm": 0.307389497756958, + "learning_rate": 4.980323759954932e-05, + "loss": 0.0086, + "step": 52620 + }, + { + "epoch": 4.087766990291263, + "grad_norm": 0.31348589062690735, + "learning_rate": 4.978670304046284e-05, + "loss": 0.0066, + "step": 52630 + }, + { + "epoch": 4.088543689320388, + "grad_norm": 0.24682873487472534, + "learning_rate": 4.9770168504702195e-05, + "loss": 0.0083, + "step": 52640 + }, + { + "epoch": 4.089320388349514, + "grad_norm": 0.333442360162735, + "learning_rate": 4.975363399407559e-05, + "loss": 0.0096, + "step": 52650 + }, + { + "epoch": 4.09009708737864, + "grad_norm": 0.10396328568458557, + "learning_rate": 4.9737099510391194e-05, + "loss": 0.0054, + "step": 52660 + }, + { + "epoch": 4.090873786407767, + "grad_norm": 0.24753199517726898, + "learning_rate": 4.972056505545719e-05, + "loss": 0.0109, + "step": 52670 + }, + { + "epoch": 4.091650485436893, + "grad_norm": 0.3427453935146332, + "learning_rate": 4.9704030631081775e-05, + "loss": 0.0061, + "step": 52680 + }, + { + "epoch": 4.092427184466019, + "grad_norm": 0.15282011032104492, + "learning_rate": 4.968749623907313e-05, + "loss": 0.0066, + "step": 52690 + }, + { + "epoch": 4.093203883495145, + "grad_norm": 0.27978503704071045, + "learning_rate": 4.967096188123943e-05, + "loss": 0.0069, + "step": 52700 + }, + { + "epoch": 4.093980582524272, + "grad_norm": 0.2723485231399536, + "learning_rate": 4.965442755938884e-05, + "loss": 0.0066, + "step": 52710 + }, + { + "epoch": 4.094757281553398, + "grad_norm": 0.15650731325149536, + "learning_rate": 4.963789327532954e-05, + "loss": 0.0034, + "step": 52720 + }, + { + "epoch": 4.095533980582524, + "grad_norm": 0.2081364095211029, + "learning_rate": 4.962135903086968e-05, + "loss": 0.0053, + "step": 52730 + }, + { + "epoch": 4.09631067961165, + "grad_norm": 0.30168697237968445, + "learning_rate": 4.960482482781741e-05, + "loss": 0.0053, + "step": 52740 + }, + { + "epoch": 4.097087378640777, + "grad_norm": 0.1096717044711113, + "learning_rate": 4.958829066798092e-05, + "loss": 0.0049, + "step": 52750 + }, + { + "epoch": 4.097864077669903, + "grad_norm": 0.3060143291950226, + "learning_rate": 4.957175655316835e-05, + "loss": 0.0066, + "step": 52760 + }, + { + "epoch": 4.098640776699029, + "grad_norm": 0.2619533836841583, + "learning_rate": 4.9555222485187834e-05, + "loss": 0.0077, + "step": 52770 + }, + { + "epoch": 4.099417475728155, + "grad_norm": 0.41808566451072693, + "learning_rate": 4.953868846584752e-05, + "loss": 0.0098, + "step": 52780 + }, + { + "epoch": 4.100194174757282, + "grad_norm": 0.2853069305419922, + "learning_rate": 4.9522154496955544e-05, + "loss": 0.006, + "step": 52790 + }, + { + "epoch": 4.100970873786408, + "grad_norm": 0.2166573405265808, + "learning_rate": 4.950562058032001e-05, + "loss": 0.01, + "step": 52800 + }, + { + "epoch": 4.101747572815534, + "grad_norm": 0.21768076717853546, + "learning_rate": 4.948908671774909e-05, + "loss": 0.0078, + "step": 52810 + }, + { + "epoch": 4.10252427184466, + "grad_norm": 0.16559424996376038, + "learning_rate": 4.947255291105087e-05, + "loss": 0.006, + "step": 52820 + }, + { + "epoch": 4.103300970873787, + "grad_norm": 0.3408847153186798, + "learning_rate": 4.945601916203348e-05, + "loss": 0.0066, + "step": 52830 + }, + { + "epoch": 4.104077669902913, + "grad_norm": 0.36592113971710205, + "learning_rate": 4.9439485472505016e-05, + "loss": 0.009, + "step": 52840 + }, + { + "epoch": 4.104854368932039, + "grad_norm": 0.17330266535282135, + "learning_rate": 4.942295184427356e-05, + "loss": 0.0059, + "step": 52850 + }, + { + "epoch": 4.105631067961165, + "grad_norm": 0.2397024780511856, + "learning_rate": 4.9406418279147234e-05, + "loss": 0.0068, + "step": 52860 + }, + { + "epoch": 4.106407766990292, + "grad_norm": 0.22912944853305817, + "learning_rate": 4.9389884778934116e-05, + "loss": 0.0085, + "step": 52870 + }, + { + "epoch": 4.107184466019418, + "grad_norm": 0.23295019567012787, + "learning_rate": 4.937335134544228e-05, + "loss": 0.0059, + "step": 52880 + }, + { + "epoch": 4.107961165048543, + "grad_norm": 0.14712728559970856, + "learning_rate": 4.935681798047979e-05, + "loss": 0.0055, + "step": 52890 + }, + { + "epoch": 4.1087378640776695, + "grad_norm": 0.3070560395717621, + "learning_rate": 4.934028468585473e-05, + "loss": 0.0036, + "step": 52900 + }, + { + "epoch": 4.109514563106796, + "grad_norm": 0.3136637806892395, + "learning_rate": 4.932375146337514e-05, + "loss": 0.006, + "step": 52910 + }, + { + "epoch": 4.110291262135922, + "grad_norm": 0.6125689744949341, + "learning_rate": 4.930721831484908e-05, + "loss": 0.007, + "step": 52920 + }, + { + "epoch": 4.111067961165048, + "grad_norm": 0.3128233253955841, + "learning_rate": 4.929068524208456e-05, + "loss": 0.0054, + "step": 52930 + }, + { + "epoch": 4.1118446601941745, + "grad_norm": 0.3638380765914917, + "learning_rate": 4.927415224688965e-05, + "loss": 0.005, + "step": 52940 + }, + { + "epoch": 4.112621359223301, + "grad_norm": 0.13011007010936737, + "learning_rate": 4.9257619331072375e-05, + "loss": 0.0058, + "step": 52950 + }, + { + "epoch": 4.113398058252427, + "grad_norm": 0.339031845331192, + "learning_rate": 4.924108649644072e-05, + "loss": 0.0059, + "step": 52960 + }, + { + "epoch": 4.114174757281553, + "grad_norm": 0.14965400099754333, + "learning_rate": 4.922455374480271e-05, + "loss": 0.0047, + "step": 52970 + }, + { + "epoch": 4.1149514563106795, + "grad_norm": 0.30186790227890015, + "learning_rate": 4.920802107796634e-05, + "loss": 0.0062, + "step": 52980 + }, + { + "epoch": 4.115728155339806, + "grad_norm": 0.15367989242076874, + "learning_rate": 4.9191488497739564e-05, + "loss": 0.0054, + "step": 52990 + }, + { + "epoch": 4.116504854368932, + "grad_norm": 0.21302515268325806, + "learning_rate": 4.917495600593042e-05, + "loss": 0.0071, + "step": 53000 + }, + { + "epoch": 4.117281553398058, + "grad_norm": 0.34230712056159973, + "learning_rate": 4.915842360434684e-05, + "loss": 0.0096, + "step": 53010 + }, + { + "epoch": 4.1180582524271845, + "grad_norm": 0.14102071523666382, + "learning_rate": 4.914189129479679e-05, + "loss": 0.0069, + "step": 53020 + }, + { + "epoch": 4.118834951456311, + "grad_norm": 0.30511805415153503, + "learning_rate": 4.912535907908822e-05, + "loss": 0.0062, + "step": 53030 + }, + { + "epoch": 4.119611650485437, + "grad_norm": 0.39794138073921204, + "learning_rate": 4.910882695902907e-05, + "loss": 0.0042, + "step": 53040 + }, + { + "epoch": 4.120388349514563, + "grad_norm": 0.18775200843811035, + "learning_rate": 4.9092294936427234e-05, + "loss": 0.0102, + "step": 53050 + }, + { + "epoch": 4.1211650485436895, + "grad_norm": 0.5078274607658386, + "learning_rate": 4.9075763013090685e-05, + "loss": 0.0076, + "step": 53060 + }, + { + "epoch": 4.121941747572816, + "grad_norm": 0.22983452677726746, + "learning_rate": 4.9059231190827305e-05, + "loss": 0.0047, + "step": 53070 + }, + { + "epoch": 4.122718446601942, + "grad_norm": 0.3358045220375061, + "learning_rate": 4.904269947144499e-05, + "loss": 0.007, + "step": 53080 + }, + { + "epoch": 4.123495145631068, + "grad_norm": 0.2744038701057434, + "learning_rate": 4.902616785675161e-05, + "loss": 0.0064, + "step": 53090 + }, + { + "epoch": 4.1242718446601945, + "grad_norm": 0.3232221305370331, + "learning_rate": 4.900963634855505e-05, + "loss": 0.0055, + "step": 53100 + }, + { + "epoch": 4.125048543689321, + "grad_norm": 0.16333739459514618, + "learning_rate": 4.8993104948663173e-05, + "loss": 0.006, + "step": 53110 + }, + { + "epoch": 4.125825242718447, + "grad_norm": 0.3370192348957062, + "learning_rate": 4.8976573658883823e-05, + "loss": 0.0068, + "step": 53120 + }, + { + "epoch": 4.126601941747573, + "grad_norm": 0.5748260021209717, + "learning_rate": 4.896004248102484e-05, + "loss": 0.0102, + "step": 53130 + }, + { + "epoch": 4.1273786407766995, + "grad_norm": 0.19057923555374146, + "learning_rate": 4.8943511416894045e-05, + "loss": 0.0062, + "step": 53140 + }, + { + "epoch": 4.128155339805825, + "grad_norm": 0.20984387397766113, + "learning_rate": 4.892698046829926e-05, + "loss": 0.008, + "step": 53150 + }, + { + "epoch": 4.128932038834951, + "grad_norm": 0.4474039673805237, + "learning_rate": 4.8910449637048275e-05, + "loss": 0.0069, + "step": 53160 + }, + { + "epoch": 4.129708737864077, + "grad_norm": 0.19829358160495758, + "learning_rate": 4.8893918924948884e-05, + "loss": 0.0086, + "step": 53170 + }, + { + "epoch": 4.130485436893204, + "grad_norm": 0.27490657567977905, + "learning_rate": 4.8877388333808835e-05, + "loss": 0.0047, + "step": 53180 + }, + { + "epoch": 4.13126213592233, + "grad_norm": 0.26539507508277893, + "learning_rate": 4.8860857865435925e-05, + "loss": 0.0093, + "step": 53190 + }, + { + "epoch": 4.132038834951456, + "grad_norm": 0.17778605222702026, + "learning_rate": 4.884432752163789e-05, + "loss": 0.0072, + "step": 53200 + }, + { + "epoch": 4.132815533980582, + "grad_norm": 0.1997738778591156, + "learning_rate": 4.8827797304222466e-05, + "loss": 0.0067, + "step": 53210 + }, + { + "epoch": 4.133592233009709, + "grad_norm": 0.5301939845085144, + "learning_rate": 4.8811267214997366e-05, + "loss": 0.007, + "step": 53220 + }, + { + "epoch": 4.134368932038835, + "grad_norm": 0.21357516944408417, + "learning_rate": 4.879473725577029e-05, + "loss": 0.0053, + "step": 53230 + }, + { + "epoch": 4.135145631067961, + "grad_norm": 0.19869232177734375, + "learning_rate": 4.877820742834891e-05, + "loss": 0.0072, + "step": 53240 + }, + { + "epoch": 4.135922330097087, + "grad_norm": 1.8744791746139526, + "learning_rate": 4.8761677734540956e-05, + "loss": 0.0056, + "step": 53250 + }, + { + "epoch": 4.136699029126214, + "grad_norm": 0.3635384142398834, + "learning_rate": 4.874514817615407e-05, + "loss": 0.0077, + "step": 53260 + }, + { + "epoch": 4.13747572815534, + "grad_norm": 0.307417631149292, + "learning_rate": 4.872861875499588e-05, + "loss": 0.0073, + "step": 53270 + }, + { + "epoch": 4.138252427184466, + "grad_norm": 0.5971054434776306, + "learning_rate": 4.871208947287404e-05, + "loss": 0.0075, + "step": 53280 + }, + { + "epoch": 4.139029126213592, + "grad_norm": 0.2341117113828659, + "learning_rate": 4.869556033159615e-05, + "loss": 0.0052, + "step": 53290 + }, + { + "epoch": 4.139805825242719, + "grad_norm": 0.22029055655002594, + "learning_rate": 4.8679031332969835e-05, + "loss": 0.0053, + "step": 53300 + }, + { + "epoch": 4.140582524271845, + "grad_norm": 0.19834931194782257, + "learning_rate": 4.8662502478802635e-05, + "loss": 0.008, + "step": 53310 + }, + { + "epoch": 4.141359223300971, + "grad_norm": 0.1639605164527893, + "learning_rate": 4.8645973770902176e-05, + "loss": 0.0056, + "step": 53320 + }, + { + "epoch": 4.142135922330097, + "grad_norm": 0.21830186247825623, + "learning_rate": 4.862944521107599e-05, + "loss": 0.0085, + "step": 53330 + }, + { + "epoch": 4.142912621359224, + "grad_norm": 0.14639191329479218, + "learning_rate": 4.861291680113162e-05, + "loss": 0.0081, + "step": 53340 + }, + { + "epoch": 4.14368932038835, + "grad_norm": 0.29296156764030457, + "learning_rate": 4.8596388542876563e-05, + "loss": 0.0051, + "step": 53350 + }, + { + "epoch": 4.144466019417476, + "grad_norm": 0.2991078495979309, + "learning_rate": 4.857986043811836e-05, + "loss": 0.0073, + "step": 53360 + }, + { + "epoch": 4.145242718446602, + "grad_norm": 0.18861599266529083, + "learning_rate": 4.856333248866448e-05, + "loss": 0.0048, + "step": 53370 + }, + { + "epoch": 4.1460194174757286, + "grad_norm": 0.3135896921157837, + "learning_rate": 4.85468046963224e-05, + "loss": 0.0057, + "step": 53380 + }, + { + "epoch": 4.146796116504854, + "grad_norm": 0.2780599594116211, + "learning_rate": 4.853027706289957e-05, + "loss": 0.0058, + "step": 53390 + }, + { + "epoch": 4.14757281553398, + "grad_norm": 0.33346104621887207, + "learning_rate": 4.851374959020344e-05, + "loss": 0.0077, + "step": 53400 + }, + { + "epoch": 4.148349514563106, + "grad_norm": 0.21865028142929077, + "learning_rate": 4.849722228004141e-05, + "loss": 0.0071, + "step": 53410 + }, + { + "epoch": 4.149126213592233, + "grad_norm": 0.2905230224132538, + "learning_rate": 4.848069513422091e-05, + "loss": 0.0087, + "step": 53420 + }, + { + "epoch": 4.149902912621359, + "grad_norm": 0.6441414952278137, + "learning_rate": 4.8464168154549294e-05, + "loss": 0.0054, + "step": 53430 + }, + { + "epoch": 4.150679611650485, + "grad_norm": 0.5035539865493774, + "learning_rate": 4.844764134283392e-05, + "loss": 0.0065, + "step": 53440 + }, + { + "epoch": 4.151456310679611, + "grad_norm": 0.33090782165527344, + "learning_rate": 4.843111470088218e-05, + "loss": 0.0099, + "step": 53450 + }, + { + "epoch": 4.152233009708738, + "grad_norm": 0.5349048376083374, + "learning_rate": 4.841458823050137e-05, + "loss": 0.0067, + "step": 53460 + }, + { + "epoch": 4.153009708737864, + "grad_norm": 0.26332888007164, + "learning_rate": 4.8398061933498816e-05, + "loss": 0.0066, + "step": 53470 + }, + { + "epoch": 4.15378640776699, + "grad_norm": 0.2894834280014038, + "learning_rate": 4.83815358116818e-05, + "loss": 0.0082, + "step": 53480 + }, + { + "epoch": 4.154563106796116, + "grad_norm": 0.36908894777297974, + "learning_rate": 4.8365009866857584e-05, + "loss": 0.007, + "step": 53490 + }, + { + "epoch": 4.155339805825243, + "grad_norm": 0.15954159200191498, + "learning_rate": 4.834848410083342e-05, + "loss": 0.0048, + "step": 53500 + }, + { + "epoch": 4.156116504854369, + "grad_norm": 0.23133358359336853, + "learning_rate": 4.833195851541657e-05, + "loss": 0.009, + "step": 53510 + }, + { + "epoch": 4.156893203883495, + "grad_norm": 0.3734413683414459, + "learning_rate": 4.831543311241423e-05, + "loss": 0.0061, + "step": 53520 + }, + { + "epoch": 4.157669902912621, + "grad_norm": 0.27077361941337585, + "learning_rate": 4.829890789363359e-05, + "loss": 0.0102, + "step": 53530 + }, + { + "epoch": 4.158446601941748, + "grad_norm": 0.16813205182552338, + "learning_rate": 4.828238286088182e-05, + "loss": 0.006, + "step": 53540 + }, + { + "epoch": 4.159223300970874, + "grad_norm": 0.173055037856102, + "learning_rate": 4.826585801596609e-05, + "loss": 0.0069, + "step": 53550 + }, + { + "epoch": 4.16, + "grad_norm": 0.1409546434879303, + "learning_rate": 4.8249333360693486e-05, + "loss": 0.0066, + "step": 53560 + }, + { + "epoch": 4.160776699029126, + "grad_norm": 0.31630778312683105, + "learning_rate": 4.8232808896871185e-05, + "loss": 0.0103, + "step": 53570 + }, + { + "epoch": 4.161553398058253, + "grad_norm": 0.22387149930000305, + "learning_rate": 4.821628462630624e-05, + "loss": 0.0069, + "step": 53580 + }, + { + "epoch": 4.162330097087379, + "grad_norm": 0.3790440559387207, + "learning_rate": 4.8199760550805735e-05, + "loss": 0.006, + "step": 53590 + }, + { + "epoch": 4.163106796116505, + "grad_norm": 0.2758257985115051, + "learning_rate": 4.818323667217669e-05, + "loss": 0.0063, + "step": 53600 + }, + { + "epoch": 4.163883495145631, + "grad_norm": 0.4222116470336914, + "learning_rate": 4.8166712992226165e-05, + "loss": 0.0056, + "step": 53610 + }, + { + "epoch": 4.164660194174758, + "grad_norm": 0.19425402581691742, + "learning_rate": 4.8150189512761154e-05, + "loss": 0.0056, + "step": 53620 + }, + { + "epoch": 4.165436893203884, + "grad_norm": 0.4653010666370392, + "learning_rate": 4.813366623558862e-05, + "loss": 0.0064, + "step": 53630 + }, + { + "epoch": 4.16621359223301, + "grad_norm": 0.2818908095359802, + "learning_rate": 4.811714316251554e-05, + "loss": 0.0051, + "step": 53640 + }, + { + "epoch": 4.1669902912621355, + "grad_norm": 0.31993579864501953, + "learning_rate": 4.810062029534886e-05, + "loss": 0.0057, + "step": 53650 + }, + { + "epoch": 4.167766990291262, + "grad_norm": 0.29353857040405273, + "learning_rate": 4.8084097635895505e-05, + "loss": 0.0083, + "step": 53660 + }, + { + "epoch": 4.168543689320388, + "grad_norm": 0.25723469257354736, + "learning_rate": 4.8067575185962335e-05, + "loss": 0.0079, + "step": 53670 + }, + { + "epoch": 4.169320388349514, + "grad_norm": 0.23745299875736237, + "learning_rate": 4.805105294735625e-05, + "loss": 0.007, + "step": 53680 + }, + { + "epoch": 4.1700970873786405, + "grad_norm": 0.2972932755947113, + "learning_rate": 4.803453092188406e-05, + "loss": 0.0072, + "step": 53690 + }, + { + "epoch": 4.170873786407767, + "grad_norm": 0.24634979665279388, + "learning_rate": 4.801800911135263e-05, + "loss": 0.0092, + "step": 53700 + }, + { + "epoch": 4.171650485436893, + "grad_norm": 0.2198278307914734, + "learning_rate": 4.8001487517568744e-05, + "loss": 0.0065, + "step": 53710 + }, + { + "epoch": 4.172427184466019, + "grad_norm": 0.24874183535575867, + "learning_rate": 4.798496614233918e-05, + "loss": 0.007, + "step": 53720 + }, + { + "epoch": 4.1732038834951455, + "grad_norm": 0.3479019105434418, + "learning_rate": 4.796844498747069e-05, + "loss": 0.0054, + "step": 53730 + }, + { + "epoch": 4.173980582524272, + "grad_norm": 0.1466895192861557, + "learning_rate": 4.795192405477e-05, + "loss": 0.007, + "step": 53740 + }, + { + "epoch": 4.174757281553398, + "grad_norm": 0.20834548771381378, + "learning_rate": 4.79354033460438e-05, + "loss": 0.0054, + "step": 53750 + }, + { + "epoch": 4.175533980582524, + "grad_norm": 0.35977521538734436, + "learning_rate": 4.79188828630988e-05, + "loss": 0.006, + "step": 53760 + }, + { + "epoch": 4.1763106796116505, + "grad_norm": 0.15311485528945923, + "learning_rate": 4.790236260774165e-05, + "loss": 0.0073, + "step": 53770 + }, + { + "epoch": 4.177087378640777, + "grad_norm": 0.2069607526063919, + "learning_rate": 4.788584258177896e-05, + "loss": 0.0046, + "step": 53780 + }, + { + "epoch": 4.177864077669903, + "grad_norm": 0.23023393750190735, + "learning_rate": 4.7869322787017356e-05, + "loss": 0.0056, + "step": 53790 + }, + { + "epoch": 4.178640776699029, + "grad_norm": 0.10772368311882019, + "learning_rate": 4.78528032252634e-05, + "loss": 0.0058, + "step": 53800 + }, + { + "epoch": 4.1794174757281555, + "grad_norm": 0.3023962378501892, + "learning_rate": 4.783628389832366e-05, + "loss": 0.0079, + "step": 53810 + }, + { + "epoch": 4.180194174757282, + "grad_norm": 0.25337690114974976, + "learning_rate": 4.7819764808004644e-05, + "loss": 0.0067, + "step": 53820 + }, + { + "epoch": 4.180970873786408, + "grad_norm": 0.4336925148963928, + "learning_rate": 4.780324595611289e-05, + "loss": 0.0069, + "step": 53830 + }, + { + "epoch": 4.181747572815534, + "grad_norm": 0.39681553840637207, + "learning_rate": 4.7786727344454854e-05, + "loss": 0.0062, + "step": 53840 + }, + { + "epoch": 4.1825242718446605, + "grad_norm": 0.1717703640460968, + "learning_rate": 4.777020897483697e-05, + "loss": 0.0073, + "step": 53850 + }, + { + "epoch": 4.183300970873787, + "grad_norm": 0.2029854953289032, + "learning_rate": 4.775369084906569e-05, + "loss": 0.0078, + "step": 53860 + }, + { + "epoch": 4.184077669902913, + "grad_norm": 0.10865217447280884, + "learning_rate": 4.773717296894742e-05, + "loss": 0.0062, + "step": 53870 + }, + { + "epoch": 4.184854368932039, + "grad_norm": 0.19468343257904053, + "learning_rate": 4.772065533628848e-05, + "loss": 0.0055, + "step": 53880 + }, + { + "epoch": 4.1856310679611655, + "grad_norm": 0.39755767583847046, + "learning_rate": 4.7704137952895265e-05, + "loss": 0.0062, + "step": 53890 + }, + { + "epoch": 4.186407766990291, + "grad_norm": 0.2309858649969101, + "learning_rate": 4.768762082057407e-05, + "loss": 0.0052, + "step": 53900 + }, + { + "epoch": 4.187184466019417, + "grad_norm": 0.4305833876132965, + "learning_rate": 4.7671103941131186e-05, + "loss": 0.0058, + "step": 53910 + }, + { + "epoch": 4.187961165048543, + "grad_norm": 0.14314605295658112, + "learning_rate": 4.765458731637287e-05, + "loss": 0.007, + "step": 53920 + }, + { + "epoch": 4.18873786407767, + "grad_norm": 0.18457163870334625, + "learning_rate": 4.763807094810537e-05, + "loss": 0.0058, + "step": 53930 + }, + { + "epoch": 4.189514563106796, + "grad_norm": 0.4498855769634247, + "learning_rate": 4.762155483813487e-05, + "loss": 0.0065, + "step": 53940 + }, + { + "epoch": 4.190291262135922, + "grad_norm": 0.19704768061637878, + "learning_rate": 4.760503898826754e-05, + "loss": 0.0061, + "step": 53950 + }, + { + "epoch": 4.191067961165048, + "grad_norm": 0.2116072177886963, + "learning_rate": 4.758852340030957e-05, + "loss": 0.0077, + "step": 53960 + }, + { + "epoch": 4.191844660194175, + "grad_norm": 0.9856889247894287, + "learning_rate": 4.757200807606705e-05, + "loss": 0.0054, + "step": 53970 + }, + { + "epoch": 4.192621359223301, + "grad_norm": 0.3361445665359497, + "learning_rate": 4.755549301734607e-05, + "loss": 0.0081, + "step": 53980 + }, + { + "epoch": 4.193398058252427, + "grad_norm": 0.7295672297477722, + "learning_rate": 4.753897822595271e-05, + "loss": 0.0066, + "step": 53990 + }, + { + "epoch": 4.194174757281553, + "grad_norm": 0.2090049386024475, + "learning_rate": 4.7522463703692994e-05, + "loss": 0.0082, + "step": 54000 + }, + { + "epoch": 4.19495145631068, + "grad_norm": 0.2722243368625641, + "learning_rate": 4.75059494523729e-05, + "loss": 0.0073, + "step": 54010 + }, + { + "epoch": 4.195728155339806, + "grad_norm": 0.21226568520069122, + "learning_rate": 4.748943547379843e-05, + "loss": 0.0056, + "step": 54020 + }, + { + "epoch": 4.196504854368932, + "grad_norm": 0.30020493268966675, + "learning_rate": 4.7472921769775535e-05, + "loss": 0.008, + "step": 54030 + }, + { + "epoch": 4.197281553398058, + "grad_norm": 0.2118697464466095, + "learning_rate": 4.745640834211011e-05, + "loss": 0.0076, + "step": 54040 + }, + { + "epoch": 4.1980582524271846, + "grad_norm": 0.42896896600723267, + "learning_rate": 4.7439895192608054e-05, + "loss": 0.0057, + "step": 54050 + }, + { + "epoch": 4.198834951456311, + "grad_norm": 0.10379669815301895, + "learning_rate": 4.7423382323075186e-05, + "loss": 0.008, + "step": 54060 + }, + { + "epoch": 4.199611650485437, + "grad_norm": 0.2506769001483917, + "learning_rate": 4.740686973531737e-05, + "loss": 0.0065, + "step": 54070 + }, + { + "epoch": 4.200388349514563, + "grad_norm": 0.5188058614730835, + "learning_rate": 4.7390357431140375e-05, + "loss": 0.009, + "step": 54080 + }, + { + "epoch": 4.2011650485436896, + "grad_norm": 0.4181920289993286, + "learning_rate": 4.737384541234998e-05, + "loss": 0.0071, + "step": 54090 + }, + { + "epoch": 4.201941747572816, + "grad_norm": 0.12242881208658218, + "learning_rate": 4.735733368075189e-05, + "loss": 0.0067, + "step": 54100 + }, + { + "epoch": 4.202718446601942, + "grad_norm": 0.2832646071910858, + "learning_rate": 4.734082223815182e-05, + "loss": 0.0072, + "step": 54110 + }, + { + "epoch": 4.203495145631068, + "grad_norm": 0.3747376799583435, + "learning_rate": 4.7324311086355436e-05, + "loss": 0.006, + "step": 54120 + }, + { + "epoch": 4.2042718446601945, + "grad_norm": 0.3345031440258026, + "learning_rate": 4.730780022716836e-05, + "loss": 0.0077, + "step": 54130 + }, + { + "epoch": 4.205048543689321, + "grad_norm": 0.18706589937210083, + "learning_rate": 4.729128966239619e-05, + "loss": 0.0073, + "step": 54140 + }, + { + "epoch": 4.205825242718447, + "grad_norm": 0.28961074352264404, + "learning_rate": 4.7274779393844536e-05, + "loss": 0.0079, + "step": 54150 + }, + { + "epoch": 4.206601941747572, + "grad_norm": 0.2941744029521942, + "learning_rate": 4.725826942331891e-05, + "loss": 0.007, + "step": 54160 + }, + { + "epoch": 4.207378640776699, + "grad_norm": 0.18656441569328308, + "learning_rate": 4.724175975262481e-05, + "loss": 0.0065, + "step": 54170 + }, + { + "epoch": 4.208155339805825, + "grad_norm": 0.3971724212169647, + "learning_rate": 4.722525038356772e-05, + "loss": 0.0073, + "step": 54180 + }, + { + "epoch": 4.208932038834951, + "grad_norm": 0.15355761349201202, + "learning_rate": 4.7208741317953075e-05, + "loss": 0.0063, + "step": 54190 + }, + { + "epoch": 4.209708737864077, + "grad_norm": 0.12755809724330902, + "learning_rate": 4.7192232557586264e-05, + "loss": 0.0065, + "step": 54200 + }, + { + "epoch": 4.210485436893204, + "grad_norm": 0.11712638288736343, + "learning_rate": 4.717572410427269e-05, + "loss": 0.0057, + "step": 54210 + }, + { + "epoch": 4.21126213592233, + "grad_norm": 0.5760302543640137, + "learning_rate": 4.71592159598177e-05, + "loss": 0.008, + "step": 54220 + }, + { + "epoch": 4.212038834951456, + "grad_norm": 0.34644025564193726, + "learning_rate": 4.714270812602657e-05, + "loss": 0.0059, + "step": 54230 + }, + { + "epoch": 4.212815533980582, + "grad_norm": 0.18131959438323975, + "learning_rate": 4.712620060470458e-05, + "loss": 0.0046, + "step": 54240 + }, + { + "epoch": 4.213592233009709, + "grad_norm": 0.22478409111499786, + "learning_rate": 4.7109693397656986e-05, + "loss": 0.0088, + "step": 54250 + }, + { + "epoch": 4.214368932038835, + "grad_norm": 0.22764183580875397, + "learning_rate": 4.709318650668894e-05, + "loss": 0.0063, + "step": 54260 + }, + { + "epoch": 4.215145631067961, + "grad_norm": 0.1782185137271881, + "learning_rate": 4.7076679933605676e-05, + "loss": 0.0059, + "step": 54270 + }, + { + "epoch": 4.215922330097087, + "grad_norm": 0.39056089520454407, + "learning_rate": 4.70601736802123e-05, + "loss": 0.0056, + "step": 54280 + }, + { + "epoch": 4.216699029126214, + "grad_norm": 0.13463199138641357, + "learning_rate": 4.704366774831391e-05, + "loss": 0.005, + "step": 54290 + }, + { + "epoch": 4.21747572815534, + "grad_norm": 0.15268872678279877, + "learning_rate": 4.702716213971557e-05, + "loss": 0.0071, + "step": 54300 + }, + { + "epoch": 4.218252427184466, + "grad_norm": 0.21678593754768372, + "learning_rate": 4.70106568562223e-05, + "loss": 0.0056, + "step": 54310 + }, + { + "epoch": 4.219029126213592, + "grad_norm": 0.23266011476516724, + "learning_rate": 4.6994151899639105e-05, + "loss": 0.0058, + "step": 54320 + }, + { + "epoch": 4.219805825242719, + "grad_norm": 0.30731379985809326, + "learning_rate": 4.697764727177093e-05, + "loss": 0.0096, + "step": 54330 + }, + { + "epoch": 4.220582524271845, + "grad_norm": 0.13368864357471466, + "learning_rate": 4.696114297442272e-05, + "loss": 0.0072, + "step": 54340 + }, + { + "epoch": 4.221359223300971, + "grad_norm": 0.27145591378211975, + "learning_rate": 4.694463900939933e-05, + "loss": 0.0077, + "step": 54350 + }, + { + "epoch": 4.222135922330097, + "grad_norm": 0.33673644065856934, + "learning_rate": 4.692813537850564e-05, + "loss": 0.0067, + "step": 54360 + }, + { + "epoch": 4.222912621359224, + "grad_norm": 0.2464027851819992, + "learning_rate": 4.6911632083546445e-05, + "loss": 0.0058, + "step": 54370 + }, + { + "epoch": 4.22368932038835, + "grad_norm": 0.32688793540000916, + "learning_rate": 4.6895129126326525e-05, + "loss": 0.007, + "step": 54380 + }, + { + "epoch": 4.224466019417476, + "grad_norm": 0.23523510992527008, + "learning_rate": 4.687862650865059e-05, + "loss": 0.0086, + "step": 54390 + }, + { + "epoch": 4.2252427184466015, + "grad_norm": 0.2532069683074951, + "learning_rate": 4.686212423232339e-05, + "loss": 0.006, + "step": 54400 + }, + { + "epoch": 4.226019417475728, + "grad_norm": 0.33690741658210754, + "learning_rate": 4.6845622299149565e-05, + "loss": 0.0058, + "step": 54410 + }, + { + "epoch": 4.226796116504854, + "grad_norm": 0.22628077864646912, + "learning_rate": 4.682912071093374e-05, + "loss": 0.0089, + "step": 54420 + }, + { + "epoch": 4.22757281553398, + "grad_norm": 0.39493122696876526, + "learning_rate": 4.681261946948052e-05, + "loss": 0.0058, + "step": 54430 + }, + { + "epoch": 4.2283495145631065, + "grad_norm": 0.10579612851142883, + "learning_rate": 4.679611857659443e-05, + "loss": 0.006, + "step": 54440 + }, + { + "epoch": 4.229126213592233, + "grad_norm": 0.21702401340007782, + "learning_rate": 4.6779618034079975e-05, + "loss": 0.0064, + "step": 54450 + }, + { + "epoch": 4.229902912621359, + "grad_norm": 0.18178564310073853, + "learning_rate": 4.6763117843741675e-05, + "loss": 0.0056, + "step": 54460 + }, + { + "epoch": 4.230679611650485, + "grad_norm": 0.2153184711933136, + "learning_rate": 4.674661800738394e-05, + "loss": 0.0057, + "step": 54470 + }, + { + "epoch": 4.2314563106796115, + "grad_norm": 0.21321839094161987, + "learning_rate": 4.673011852681117e-05, + "loss": 0.0053, + "step": 54480 + }, + { + "epoch": 4.232233009708738, + "grad_norm": 0.28540974855422974, + "learning_rate": 4.671361940382772e-05, + "loss": 0.0098, + "step": 54490 + }, + { + "epoch": 4.233009708737864, + "grad_norm": 0.31053605675697327, + "learning_rate": 4.669712064023791e-05, + "loss": 0.0091, + "step": 54500 + }, + { + "epoch": 4.23378640776699, + "grad_norm": 0.21246513724327087, + "learning_rate": 4.668062223784602e-05, + "loss": 0.0069, + "step": 54510 + }, + { + "epoch": 4.2345631067961165, + "grad_norm": 0.21615061163902283, + "learning_rate": 4.6664124198456274e-05, + "loss": 0.0068, + "step": 54520 + }, + { + "epoch": 4.235339805825243, + "grad_norm": 0.277260959148407, + "learning_rate": 4.6647626523872905e-05, + "loss": 0.0064, + "step": 54530 + }, + { + "epoch": 4.236116504854369, + "grad_norm": 0.0917477086186409, + "learning_rate": 4.6631129215900066e-05, + "loss": 0.0045, + "step": 54540 + }, + { + "epoch": 4.236893203883495, + "grad_norm": 0.2974051237106323, + "learning_rate": 4.661463227634187e-05, + "loss": 0.0057, + "step": 54550 + }, + { + "epoch": 4.2376699029126215, + "grad_norm": 0.30811136960983276, + "learning_rate": 4.659813570700238e-05, + "loss": 0.0072, + "step": 54560 + }, + { + "epoch": 4.238446601941748, + "grad_norm": 0.5211226940155029, + "learning_rate": 4.658163950968568e-05, + "loss": 0.0085, + "step": 54570 + }, + { + "epoch": 4.239223300970874, + "grad_norm": 0.34255439043045044, + "learning_rate": 4.6565143686195726e-05, + "loss": 0.0082, + "step": 54580 + }, + { + "epoch": 4.24, + "grad_norm": 0.17571040987968445, + "learning_rate": 4.654864823833649e-05, + "loss": 0.0073, + "step": 54590 + }, + { + "epoch": 4.2407766990291265, + "grad_norm": 0.48144739866256714, + "learning_rate": 4.65321531679119e-05, + "loss": 0.0062, + "step": 54600 + }, + { + "epoch": 4.241553398058253, + "grad_norm": 0.3673208951950073, + "learning_rate": 4.6515658476725834e-05, + "loss": 0.0097, + "step": 54610 + }, + { + "epoch": 4.242330097087379, + "grad_norm": 0.21077583730220795, + "learning_rate": 4.649916416658212e-05, + "loss": 0.0082, + "step": 54620 + }, + { + "epoch": 4.243106796116505, + "grad_norm": 0.3467579483985901, + "learning_rate": 4.648267023928455e-05, + "loss": 0.0059, + "step": 54630 + }, + { + "epoch": 4.2438834951456315, + "grad_norm": 0.26484259963035583, + "learning_rate": 4.646617669663687e-05, + "loss": 0.0063, + "step": 54640 + }, + { + "epoch": 4.244660194174758, + "grad_norm": 0.34423547983169556, + "learning_rate": 4.644968354044277e-05, + "loss": 0.0065, + "step": 54650 + }, + { + "epoch": 4.245436893203883, + "grad_norm": 0.2111615538597107, + "learning_rate": 4.643319077250596e-05, + "loss": 0.0054, + "step": 54660 + }, + { + "epoch": 4.246213592233009, + "grad_norm": 0.18562641739845276, + "learning_rate": 4.6416698394630055e-05, + "loss": 0.0059, + "step": 54670 + }, + { + "epoch": 4.246990291262136, + "grad_norm": 0.21935690939426422, + "learning_rate": 4.640020640861862e-05, + "loss": 0.0068, + "step": 54680 + }, + { + "epoch": 4.247766990291262, + "grad_norm": 0.17330418527126312, + "learning_rate": 4.6383714816275195e-05, + "loss": 0.0041, + "step": 54690 + }, + { + "epoch": 4.248543689320388, + "grad_norm": 0.23280468583106995, + "learning_rate": 4.636722361940329e-05, + "loss": 0.0053, + "step": 54700 + }, + { + "epoch": 4.249320388349514, + "grad_norm": 0.19260674715042114, + "learning_rate": 4.635073281980632e-05, + "loss": 0.0056, + "step": 54710 + }, + { + "epoch": 4.250097087378641, + "grad_norm": 0.7659481763839722, + "learning_rate": 4.633424241928775e-05, + "loss": 0.0054, + "step": 54720 + }, + { + "epoch": 4.250873786407767, + "grad_norm": 0.38865822553634644, + "learning_rate": 4.6317752419650914e-05, + "loss": 0.0063, + "step": 54730 + }, + { + "epoch": 4.251650485436893, + "grad_norm": 0.17861615121364594, + "learning_rate": 4.630126282269914e-05, + "loss": 0.0055, + "step": 54740 + }, + { + "epoch": 4.252427184466019, + "grad_norm": 0.3704400956630707, + "learning_rate": 4.628477363023569e-05, + "loss": 0.0067, + "step": 54750 + }, + { + "epoch": 4.2532038834951456, + "grad_norm": 0.19553111493587494, + "learning_rate": 4.626828484406383e-05, + "loss": 0.0057, + "step": 54760 + }, + { + "epoch": 4.253980582524272, + "grad_norm": 0.2771744430065155, + "learning_rate": 4.625179646598669e-05, + "loss": 0.0061, + "step": 54770 + }, + { + "epoch": 4.254757281553398, + "grad_norm": 0.300632119178772, + "learning_rate": 4.6235308497807487e-05, + "loss": 0.0036, + "step": 54780 + }, + { + "epoch": 4.255533980582524, + "grad_norm": 0.25271832942962646, + "learning_rate": 4.6218820941329276e-05, + "loss": 0.0064, + "step": 54790 + }, + { + "epoch": 4.2563106796116505, + "grad_norm": 0.17805171012878418, + "learning_rate": 4.620233379835513e-05, + "loss": 0.0067, + "step": 54800 + }, + { + "epoch": 4.257087378640777, + "grad_norm": 0.40601646900177, + "learning_rate": 4.618584707068804e-05, + "loss": 0.0068, + "step": 54810 + }, + { + "epoch": 4.257864077669903, + "grad_norm": 0.18402250111103058, + "learning_rate": 4.616936076013099e-05, + "loss": 0.0049, + "step": 54820 + }, + { + "epoch": 4.258640776699029, + "grad_norm": 0.1296404004096985, + "learning_rate": 4.615287486848688e-05, + "loss": 0.0059, + "step": 54830 + }, + { + "epoch": 4.2594174757281555, + "grad_norm": 0.1420426070690155, + "learning_rate": 4.6136389397558584e-05, + "loss": 0.0053, + "step": 54840 + }, + { + "epoch": 4.260194174757282, + "grad_norm": 0.24393951892852783, + "learning_rate": 4.6119904349148935e-05, + "loss": 0.004, + "step": 54850 + }, + { + "epoch": 4.260970873786408, + "grad_norm": 0.29937484860420227, + "learning_rate": 4.6103419725060726e-05, + "loss": 0.006, + "step": 54860 + }, + { + "epoch": 4.261747572815534, + "grad_norm": 0.2140926569700241, + "learning_rate": 4.608693552709667e-05, + "loss": 0.0069, + "step": 54870 + }, + { + "epoch": 4.2625242718446605, + "grad_norm": 0.23982302844524384, + "learning_rate": 4.607045175705947e-05, + "loss": 0.007, + "step": 54880 + }, + { + "epoch": 4.263300970873787, + "grad_norm": 0.18787258863449097, + "learning_rate": 4.605396841675176e-05, + "loss": 0.0062, + "step": 54890 + }, + { + "epoch": 4.264077669902913, + "grad_norm": 0.27192768454551697, + "learning_rate": 4.6037485507976094e-05, + "loss": 0.0083, + "step": 54900 + }, + { + "epoch": 4.264854368932038, + "grad_norm": 0.16601021587848663, + "learning_rate": 4.602100303253509e-05, + "loss": 0.0056, + "step": 54910 + }, + { + "epoch": 4.265631067961165, + "grad_norm": 0.31152841448783875, + "learning_rate": 4.600452099223122e-05, + "loss": 0.0077, + "step": 54920 + }, + { + "epoch": 4.266407766990291, + "grad_norm": 0.18366357684135437, + "learning_rate": 4.598803938886691e-05, + "loss": 0.006, + "step": 54930 + }, + { + "epoch": 4.267184466019417, + "grad_norm": 0.31498458981513977, + "learning_rate": 4.597155822424459e-05, + "loss": 0.0051, + "step": 54940 + }, + { + "epoch": 4.267961165048543, + "grad_norm": 0.2737002670764923, + "learning_rate": 4.59550775001666e-05, + "loss": 0.0071, + "step": 54950 + }, + { + "epoch": 4.26873786407767, + "grad_norm": 0.15759222209453583, + "learning_rate": 4.5938597218435234e-05, + "loss": 0.0072, + "step": 54960 + }, + { + "epoch": 4.269514563106796, + "grad_norm": 0.558259904384613, + "learning_rate": 4.592211738085279e-05, + "loss": 0.0069, + "step": 54970 + }, + { + "epoch": 4.270291262135922, + "grad_norm": 0.09897219389677048, + "learning_rate": 4.590563798922146e-05, + "loss": 0.0063, + "step": 54980 + }, + { + "epoch": 4.271067961165048, + "grad_norm": 0.3882485628128052, + "learning_rate": 4.5889159045343404e-05, + "loss": 0.0069, + "step": 54990 + }, + { + "epoch": 4.271844660194175, + "grad_norm": 0.15203195810317993, + "learning_rate": 4.5872680551020734e-05, + "loss": 0.0056, + "step": 55000 + }, + { + "epoch": 4.272621359223301, + "grad_norm": 0.468840628862381, + "learning_rate": 4.585620250805552e-05, + "loss": 0.0073, + "step": 55010 + }, + { + "epoch": 4.273398058252427, + "grad_norm": 0.20126616954803467, + "learning_rate": 4.583972491824975e-05, + "loss": 0.0064, + "step": 55020 + }, + { + "epoch": 4.274174757281553, + "grad_norm": 0.29851511120796204, + "learning_rate": 4.582324778340541e-05, + "loss": 0.0069, + "step": 55030 + }, + { + "epoch": 4.27495145631068, + "grad_norm": 0.3554369807243347, + "learning_rate": 4.5806771105324424e-05, + "loss": 0.0065, + "step": 55040 + }, + { + "epoch": 4.275728155339806, + "grad_norm": 0.14222851395606995, + "learning_rate": 4.579029488580865e-05, + "loss": 0.0075, + "step": 55050 + }, + { + "epoch": 4.276504854368932, + "grad_norm": 0.21572886407375336, + "learning_rate": 4.5773819126659876e-05, + "loss": 0.0061, + "step": 55060 + }, + { + "epoch": 4.277281553398058, + "grad_norm": 0.19210603833198547, + "learning_rate": 4.575734382967991e-05, + "loss": 0.0063, + "step": 55070 + }, + { + "epoch": 4.278058252427185, + "grad_norm": 0.15424349904060364, + "learning_rate": 4.574086899667044e-05, + "loss": 0.0058, + "step": 55080 + }, + { + "epoch": 4.278834951456311, + "grad_norm": 0.17668382823467255, + "learning_rate": 4.5724394629433106e-05, + "loss": 0.0061, + "step": 55090 + }, + { + "epoch": 4.279611650485437, + "grad_norm": 0.26570889353752136, + "learning_rate": 4.570792072976958e-05, + "loss": 0.0051, + "step": 55100 + }, + { + "epoch": 4.280388349514563, + "grad_norm": 0.3682228624820709, + "learning_rate": 4.569144729948138e-05, + "loss": 0.0053, + "step": 55110 + }, + { + "epoch": 4.28116504854369, + "grad_norm": 0.28762659430503845, + "learning_rate": 4.5674974340370033e-05, + "loss": 0.008, + "step": 55120 + }, + { + "epoch": 4.281941747572816, + "grad_norm": 0.5610809326171875, + "learning_rate": 4.565850185423699e-05, + "loss": 0.0065, + "step": 55130 + }, + { + "epoch": 4.282718446601942, + "grad_norm": 0.27041929960250854, + "learning_rate": 4.564202984288365e-05, + "loss": 0.0077, + "step": 55140 + }, + { + "epoch": 4.283495145631068, + "grad_norm": 0.19478994607925415, + "learning_rate": 4.5625558308111356e-05, + "loss": 0.0055, + "step": 55150 + }, + { + "epoch": 4.284271844660195, + "grad_norm": 0.36482474207878113, + "learning_rate": 4.560908725172145e-05, + "loss": 0.0061, + "step": 55160 + }, + { + "epoch": 4.28504854368932, + "grad_norm": 0.34980276226997375, + "learning_rate": 4.559261667551516e-05, + "loss": 0.0052, + "step": 55170 + }, + { + "epoch": 4.285825242718446, + "grad_norm": 0.15316997468471527, + "learning_rate": 4.5576146581293685e-05, + "loss": 0.0084, + "step": 55180 + }, + { + "epoch": 4.2866019417475725, + "grad_norm": 0.2500148117542267, + "learning_rate": 4.5559676970858164e-05, + "loss": 0.0088, + "step": 55190 + }, + { + "epoch": 4.287378640776699, + "grad_norm": 0.17539449036121368, + "learning_rate": 4.5543207846009705e-05, + "loss": 0.0069, + "step": 55200 + }, + { + "epoch": 4.288155339805825, + "grad_norm": 0.21112684905529022, + "learning_rate": 4.552673920854933e-05, + "loss": 0.0073, + "step": 55210 + }, + { + "epoch": 4.288932038834951, + "grad_norm": 0.14471428096294403, + "learning_rate": 4.5510271060278006e-05, + "loss": 0.0068, + "step": 55220 + }, + { + "epoch": 4.2897087378640775, + "grad_norm": 0.4344847500324249, + "learning_rate": 4.5493803402996704e-05, + "loss": 0.0072, + "step": 55230 + }, + { + "epoch": 4.290485436893204, + "grad_norm": 0.1501057893037796, + "learning_rate": 4.5477336238506295e-05, + "loss": 0.0043, + "step": 55240 + }, + { + "epoch": 4.29126213592233, + "grad_norm": 0.3246234655380249, + "learning_rate": 4.546086956860759e-05, + "loss": 0.0057, + "step": 55250 + }, + { + "epoch": 4.292038834951456, + "grad_norm": 0.1563158482313156, + "learning_rate": 4.5444403395101374e-05, + "loss": 0.0057, + "step": 55260 + }, + { + "epoch": 4.2928155339805825, + "grad_norm": 0.1877681463956833, + "learning_rate": 4.542793771978834e-05, + "loss": 0.0085, + "step": 55270 + }, + { + "epoch": 4.293592233009709, + "grad_norm": 0.4729570746421814, + "learning_rate": 4.541147254446916e-05, + "loss": 0.0063, + "step": 55280 + }, + { + "epoch": 4.294368932038835, + "grad_norm": 0.31897053122520447, + "learning_rate": 4.539500787094445e-05, + "loss": 0.0066, + "step": 55290 + }, + { + "epoch": 4.295145631067961, + "grad_norm": 0.14494551718235016, + "learning_rate": 4.537854370101475e-05, + "loss": 0.005, + "step": 55300 + }, + { + "epoch": 4.2959223300970875, + "grad_norm": 0.1580418199300766, + "learning_rate": 4.5362080036480566e-05, + "loss": 0.0049, + "step": 55310 + }, + { + "epoch": 4.296699029126214, + "grad_norm": 0.12486911565065384, + "learning_rate": 4.534561687914235e-05, + "loss": 0.0056, + "step": 55320 + }, + { + "epoch": 4.29747572815534, + "grad_norm": 0.220463827252388, + "learning_rate": 4.5329154230800466e-05, + "loss": 0.0054, + "step": 55330 + }, + { + "epoch": 4.298252427184466, + "grad_norm": 0.24744056165218353, + "learning_rate": 4.531269209325525e-05, + "loss": 0.0081, + "step": 55340 + }, + { + "epoch": 4.2990291262135925, + "grad_norm": 0.2303517609834671, + "learning_rate": 4.529623046830696e-05, + "loss": 0.0048, + "step": 55350 + }, + { + "epoch": 4.299805825242719, + "grad_norm": 0.19278109073638916, + "learning_rate": 4.5279769357755866e-05, + "loss": 0.008, + "step": 55360 + }, + { + "epoch": 4.300582524271845, + "grad_norm": 0.4314629137516022, + "learning_rate": 4.5263308763402084e-05, + "loss": 0.0083, + "step": 55370 + }, + { + "epoch": 4.301359223300971, + "grad_norm": 0.2186250537633896, + "learning_rate": 4.524684868704574e-05, + "loss": 0.0065, + "step": 55380 + }, + { + "epoch": 4.3021359223300975, + "grad_norm": 0.21765033900737762, + "learning_rate": 4.523038913048688e-05, + "loss": 0.0067, + "step": 55390 + }, + { + "epoch": 4.302912621359224, + "grad_norm": 0.14153996109962463, + "learning_rate": 4.521393009552549e-05, + "loss": 0.0058, + "step": 55400 + }, + { + "epoch": 4.303689320388349, + "grad_norm": 0.3147026598453522, + "learning_rate": 4.5197471583961484e-05, + "loss": 0.0054, + "step": 55410 + }, + { + "epoch": 4.304466019417475, + "grad_norm": 0.6651368141174316, + "learning_rate": 4.51810135975948e-05, + "loss": 0.0076, + "step": 55420 + }, + { + "epoch": 4.305242718446602, + "grad_norm": 0.22794882953166962, + "learning_rate": 4.516455613822521e-05, + "loss": 0.0063, + "step": 55430 + }, + { + "epoch": 4.306019417475728, + "grad_norm": 0.34098902344703674, + "learning_rate": 4.514809920765249e-05, + "loss": 0.0068, + "step": 55440 + }, + { + "epoch": 4.306796116504854, + "grad_norm": 0.6010086536407471, + "learning_rate": 4.5131642807676356e-05, + "loss": 0.006, + "step": 55450 + }, + { + "epoch": 4.30757281553398, + "grad_norm": 0.12729595601558685, + "learning_rate": 4.511518694009644e-05, + "loss": 0.0067, + "step": 55460 + }, + { + "epoch": 4.3083495145631066, + "grad_norm": 0.14491580426692963, + "learning_rate": 4.509873160671231e-05, + "loss": 0.006, + "step": 55470 + }, + { + "epoch": 4.309126213592233, + "grad_norm": 0.26807716488838196, + "learning_rate": 4.508227680932356e-05, + "loss": 0.0054, + "step": 55480 + }, + { + "epoch": 4.309902912621359, + "grad_norm": 0.20057833194732666, + "learning_rate": 4.5065822549729605e-05, + "loss": 0.0056, + "step": 55490 + }, + { + "epoch": 4.310679611650485, + "grad_norm": 0.655867874622345, + "learning_rate": 4.504936882972989e-05, + "loss": 0.0073, + "step": 55500 + }, + { + "epoch": 4.3114563106796115, + "grad_norm": 0.18378837406635284, + "learning_rate": 4.503291565112374e-05, + "loss": 0.0064, + "step": 55510 + }, + { + "epoch": 4.312233009708738, + "grad_norm": 0.25507014989852905, + "learning_rate": 4.501646301571048e-05, + "loss": 0.0072, + "step": 55520 + }, + { + "epoch": 4.313009708737864, + "grad_norm": 0.38866499066352844, + "learning_rate": 4.500001092528933e-05, + "loss": 0.0038, + "step": 55530 + }, + { + "epoch": 4.31378640776699, + "grad_norm": 0.16435855627059937, + "learning_rate": 4.4983559381659455e-05, + "loss": 0.0049, + "step": 55540 + }, + { + "epoch": 4.3145631067961165, + "grad_norm": 0.2495691031217575, + "learning_rate": 4.496710838661999e-05, + "loss": 0.0095, + "step": 55550 + }, + { + "epoch": 4.315339805825243, + "grad_norm": 0.2537137269973755, + "learning_rate": 4.495065794196999e-05, + "loss": 0.0061, + "step": 55560 + }, + { + "epoch": 4.316116504854369, + "grad_norm": 0.17673520743846893, + "learning_rate": 4.4934208049508455e-05, + "loss": 0.0065, + "step": 55570 + }, + { + "epoch": 4.316893203883495, + "grad_norm": 0.37874695658683777, + "learning_rate": 4.4917758711034303e-05, + "loss": 0.0075, + "step": 55580 + }, + { + "epoch": 4.3176699029126215, + "grad_norm": 0.19086889922618866, + "learning_rate": 4.490130992834642e-05, + "loss": 0.0071, + "step": 55590 + }, + { + "epoch": 4.318446601941748, + "grad_norm": 0.4827934801578522, + "learning_rate": 4.488486170324359e-05, + "loss": 0.0059, + "step": 55600 + }, + { + "epoch": 4.319223300970874, + "grad_norm": 0.22554665803909302, + "learning_rate": 4.486841403752463e-05, + "loss": 0.0049, + "step": 55610 + }, + { + "epoch": 4.32, + "grad_norm": 0.18419204652309418, + "learning_rate": 4.4851966932988174e-05, + "loss": 0.008, + "step": 55620 + }, + { + "epoch": 4.3207766990291265, + "grad_norm": 0.19486196339130402, + "learning_rate": 4.4835520391432894e-05, + "loss": 0.0056, + "step": 55630 + }, + { + "epoch": 4.321553398058253, + "grad_norm": 0.1880510151386261, + "learning_rate": 4.481907441465733e-05, + "loss": 0.0059, + "step": 55640 + }, + { + "epoch": 4.322330097087379, + "grad_norm": 0.2095976024866104, + "learning_rate": 4.480262900446e-05, + "loss": 0.0067, + "step": 55650 + }, + { + "epoch": 4.323106796116505, + "grad_norm": 0.19264821708202362, + "learning_rate": 4.478618416263933e-05, + "loss": 0.0072, + "step": 55660 + }, + { + "epoch": 4.3238834951456315, + "grad_norm": 0.6407974362373352, + "learning_rate": 4.476973989099373e-05, + "loss": 0.0071, + "step": 55670 + }, + { + "epoch": 4.324660194174757, + "grad_norm": 0.29657766222953796, + "learning_rate": 4.475329619132153e-05, + "loss": 0.0054, + "step": 55680 + }, + { + "epoch": 4.325436893203883, + "grad_norm": 0.14184464514255524, + "learning_rate": 4.4736853065420956e-05, + "loss": 0.0072, + "step": 55690 + }, + { + "epoch": 4.326213592233009, + "grad_norm": 0.10137801617383957, + "learning_rate": 4.472041051509023e-05, + "loss": 0.0047, + "step": 55700 + }, + { + "epoch": 4.326990291262136, + "grad_norm": 0.1680448055267334, + "learning_rate": 4.470396854212746e-05, + "loss": 0.0043, + "step": 55710 + }, + { + "epoch": 4.327766990291262, + "grad_norm": 0.23527076840400696, + "learning_rate": 4.468752714833072e-05, + "loss": 0.0057, + "step": 55720 + }, + { + "epoch": 4.328543689320388, + "grad_norm": 0.20885327458381653, + "learning_rate": 4.467108633549803e-05, + "loss": 0.005, + "step": 55730 + }, + { + "epoch": 4.329320388349514, + "grad_norm": 0.1910015344619751, + "learning_rate": 4.4654646105427335e-05, + "loss": 0.0069, + "step": 55740 + }, + { + "epoch": 4.330097087378641, + "grad_norm": 0.1353210061788559, + "learning_rate": 4.463820645991651e-05, + "loss": 0.0053, + "step": 55750 + }, + { + "epoch": 4.330873786407767, + "grad_norm": 0.15911418199539185, + "learning_rate": 4.462176740076334e-05, + "loss": 0.0073, + "step": 55760 + }, + { + "epoch": 4.331650485436893, + "grad_norm": 0.14134013652801514, + "learning_rate": 4.4605328929765624e-05, + "loss": 0.0057, + "step": 55770 + }, + { + "epoch": 4.332427184466019, + "grad_norm": 0.19400708377361298, + "learning_rate": 4.458889104872102e-05, + "loss": 0.0093, + "step": 55780 + }, + { + "epoch": 4.333203883495146, + "grad_norm": 0.2040686309337616, + "learning_rate": 4.457245375942715e-05, + "loss": 0.006, + "step": 55790 + }, + { + "epoch": 4.333980582524272, + "grad_norm": 0.2292453795671463, + "learning_rate": 4.455601706368157e-05, + "loss": 0.0077, + "step": 55800 + }, + { + "epoch": 4.334757281553398, + "grad_norm": 0.1729278713464737, + "learning_rate": 4.453958096328179e-05, + "loss": 0.0055, + "step": 55810 + }, + { + "epoch": 4.335533980582524, + "grad_norm": 0.2364771068096161, + "learning_rate": 4.452314546002523e-05, + "loss": 0.0115, + "step": 55820 + }, + { + "epoch": 4.336310679611651, + "grad_norm": 0.20889225602149963, + "learning_rate": 4.4506710555709246e-05, + "loss": 0.0071, + "step": 55830 + }, + { + "epoch": 4.337087378640777, + "grad_norm": 1.1118463277816772, + "learning_rate": 4.449027625213114e-05, + "loss": 0.0066, + "step": 55840 + }, + { + "epoch": 4.337864077669903, + "grad_norm": 0.43911251425743103, + "learning_rate": 4.4473842551088135e-05, + "loss": 0.0067, + "step": 55850 + }, + { + "epoch": 4.338640776699029, + "grad_norm": 0.22938427329063416, + "learning_rate": 4.445740945437738e-05, + "loss": 0.006, + "step": 55860 + }, + { + "epoch": 4.339417475728156, + "grad_norm": 0.19651752710342407, + "learning_rate": 4.444097696379602e-05, + "loss": 0.0044, + "step": 55870 + }, + { + "epoch": 4.340194174757282, + "grad_norm": 0.32084113359451294, + "learning_rate": 4.442454508114106e-05, + "loss": 0.0088, + "step": 55880 + }, + { + "epoch": 4.340970873786408, + "grad_norm": 0.17407311499118805, + "learning_rate": 4.4408113808209465e-05, + "loss": 0.0085, + "step": 55890 + }, + { + "epoch": 4.341747572815534, + "grad_norm": 0.19333021342754364, + "learning_rate": 4.439168314679813e-05, + "loss": 0.005, + "step": 55900 + }, + { + "epoch": 4.342524271844661, + "grad_norm": 0.1771906167268753, + "learning_rate": 4.43752530987039e-05, + "loss": 0.0061, + "step": 55910 + }, + { + "epoch": 4.343300970873786, + "grad_norm": 0.3094550669193268, + "learning_rate": 4.4358823665723515e-05, + "loss": 0.0062, + "step": 55920 + }, + { + "epoch": 4.344077669902912, + "grad_norm": 0.4575853645801544, + "learning_rate": 4.434239484965371e-05, + "loss": 0.0052, + "step": 55930 + }, + { + "epoch": 4.3448543689320385, + "grad_norm": 0.18532565236091614, + "learning_rate": 4.4325966652291103e-05, + "loss": 0.0065, + "step": 55940 + }, + { + "epoch": 4.345631067961165, + "grad_norm": 0.28839632868766785, + "learning_rate": 4.430953907543225e-05, + "loss": 0.0068, + "step": 55950 + }, + { + "epoch": 4.346407766990291, + "grad_norm": 0.21384628117084503, + "learning_rate": 4.4293112120873645e-05, + "loss": 0.0045, + "step": 55960 + }, + { + "epoch": 4.347184466019417, + "grad_norm": 0.11438651382923126, + "learning_rate": 4.427668579041172e-05, + "loss": 0.0057, + "step": 55970 + }, + { + "epoch": 4.3479611650485435, + "grad_norm": 0.8949816226959229, + "learning_rate": 4.4260260085842816e-05, + "loss": 0.0112, + "step": 55980 + }, + { + "epoch": 4.34873786407767, + "grad_norm": 1.364675760269165, + "learning_rate": 4.4243835008963256e-05, + "loss": 0.0082, + "step": 55990 + }, + { + "epoch": 4.349514563106796, + "grad_norm": 0.1915826052427292, + "learning_rate": 4.422741056156925e-05, + "loss": 0.0062, + "step": 56000 + }, + { + "epoch": 4.350291262135922, + "grad_norm": 0.3443909287452698, + "learning_rate": 4.421098674545693e-05, + "loss": 0.0074, + "step": 56010 + }, + { + "epoch": 4.3510679611650485, + "grad_norm": 0.5472469925880432, + "learning_rate": 4.419456356242241e-05, + "loss": 0.0072, + "step": 56020 + }, + { + "epoch": 4.351844660194175, + "grad_norm": 0.17083534598350525, + "learning_rate": 4.4178141014261685e-05, + "loss": 0.0072, + "step": 56030 + }, + { + "epoch": 4.352621359223301, + "grad_norm": 0.31360748410224915, + "learning_rate": 4.416171910277071e-05, + "loss": 0.0071, + "step": 56040 + }, + { + "epoch": 4.353398058252427, + "grad_norm": 0.6638441681861877, + "learning_rate": 4.414529782974533e-05, + "loss": 0.0061, + "step": 56050 + }, + { + "epoch": 4.3541747572815535, + "grad_norm": 0.16768838465213776, + "learning_rate": 4.41288771969814e-05, + "loss": 0.0071, + "step": 56060 + }, + { + "epoch": 4.35495145631068, + "grad_norm": 0.38120248913764954, + "learning_rate": 4.411245720627464e-05, + "loss": 0.008, + "step": 56070 + }, + { + "epoch": 4.355728155339806, + "grad_norm": 0.26464417576789856, + "learning_rate": 4.40960378594207e-05, + "loss": 0.0075, + "step": 56080 + }, + { + "epoch": 4.356504854368932, + "grad_norm": 0.32235637307167053, + "learning_rate": 4.407961915821519e-05, + "loss": 0.0073, + "step": 56090 + }, + { + "epoch": 4.3572815533980584, + "grad_norm": 0.2352989912033081, + "learning_rate": 4.406320110445362e-05, + "loss": 0.0071, + "step": 56100 + }, + { + "epoch": 4.358058252427185, + "grad_norm": 0.3527681231498718, + "learning_rate": 4.404678369993144e-05, + "loss": 0.009, + "step": 56110 + }, + { + "epoch": 4.358834951456311, + "grad_norm": 0.18039777874946594, + "learning_rate": 4.403036694644406e-05, + "loss": 0.0082, + "step": 56120 + }, + { + "epoch": 4.359611650485437, + "grad_norm": 0.27013251185417175, + "learning_rate": 4.4013950845786764e-05, + "loss": 0.0067, + "step": 56130 + }, + { + "epoch": 4.3603883495145634, + "grad_norm": 0.4632730782032013, + "learning_rate": 4.399753539975482e-05, + "loss": 0.0062, + "step": 56140 + }, + { + "epoch": 4.36116504854369, + "grad_norm": 0.23838093876838684, + "learning_rate": 4.398112061014337e-05, + "loss": 0.0085, + "step": 56150 + }, + { + "epoch": 4.361941747572816, + "grad_norm": 0.23422104120254517, + "learning_rate": 4.396470647874753e-05, + "loss": 0.0047, + "step": 56160 + }, + { + "epoch": 4.362718446601942, + "grad_norm": 0.1860094517469406, + "learning_rate": 4.394829300736229e-05, + "loss": 0.0056, + "step": 56170 + }, + { + "epoch": 4.3634951456310676, + "grad_norm": 0.29192376136779785, + "learning_rate": 4.393188019778265e-05, + "loss": 0.0067, + "step": 56180 + }, + { + "epoch": 4.364271844660194, + "grad_norm": 0.2007220983505249, + "learning_rate": 4.391546805180347e-05, + "loss": 0.0062, + "step": 56190 + }, + { + "epoch": 4.36504854368932, + "grad_norm": 0.18820922076702118, + "learning_rate": 4.389905657121955e-05, + "loss": 0.006, + "step": 56200 + }, + { + "epoch": 4.365825242718446, + "grad_norm": 0.222894087433815, + "learning_rate": 4.3882645757825635e-05, + "loss": 0.0075, + "step": 56210 + }, + { + "epoch": 4.3666019417475725, + "grad_norm": 0.5767233371734619, + "learning_rate": 4.386623561341637e-05, + "loss": 0.0078, + "step": 56220 + }, + { + "epoch": 4.367378640776699, + "grad_norm": 0.2186586707830429, + "learning_rate": 4.384982613978637e-05, + "loss": 0.006, + "step": 56230 + }, + { + "epoch": 4.368155339805825, + "grad_norm": 0.3297567069530487, + "learning_rate": 4.383341733873012e-05, + "loss": 0.0062, + "step": 56240 + }, + { + "epoch": 4.368932038834951, + "grad_norm": 0.18445806205272675, + "learning_rate": 4.381700921204209e-05, + "loss": 0.0072, + "step": 56250 + }, + { + "epoch": 4.3697087378640775, + "grad_norm": 0.13529376685619354, + "learning_rate": 4.380060176151663e-05, + "loss": 0.0064, + "step": 56260 + }, + { + "epoch": 4.370485436893204, + "grad_norm": 0.2832140028476715, + "learning_rate": 4.378419498894805e-05, + "loss": 0.0075, + "step": 56270 + }, + { + "epoch": 4.37126213592233, + "grad_norm": 0.21083267033100128, + "learning_rate": 4.376778889613056e-05, + "loss": 0.0057, + "step": 56280 + }, + { + "epoch": 4.372038834951456, + "grad_norm": 0.20162278413772583, + "learning_rate": 4.375138348485831e-05, + "loss": 0.0053, + "step": 56290 + }, + { + "epoch": 4.3728155339805825, + "grad_norm": 0.1723327785730362, + "learning_rate": 4.3734978756925346e-05, + "loss": 0.0058, + "step": 56300 + }, + { + "epoch": 4.373592233009709, + "grad_norm": 0.3009200096130371, + "learning_rate": 4.3718574714125706e-05, + "loss": 0.0066, + "step": 56310 + }, + { + "epoch": 4.374368932038835, + "grad_norm": 0.14397203922271729, + "learning_rate": 4.370217135825329e-05, + "loss": 0.0052, + "step": 56320 + }, + { + "epoch": 4.375145631067961, + "grad_norm": 0.36341074109077454, + "learning_rate": 4.368576869110194e-05, + "loss": 0.0086, + "step": 56330 + }, + { + "epoch": 4.3759223300970875, + "grad_norm": 0.3949793875217438, + "learning_rate": 4.366936671446544e-05, + "loss": 0.0069, + "step": 56340 + }, + { + "epoch": 4.376699029126214, + "grad_norm": 0.35563015937805176, + "learning_rate": 4.3652965430137474e-05, + "loss": 0.0095, + "step": 56350 + }, + { + "epoch": 4.37747572815534, + "grad_norm": 0.20494234561920166, + "learning_rate": 4.3636564839911646e-05, + "loss": 0.0078, + "step": 56360 + }, + { + "epoch": 4.378252427184466, + "grad_norm": 0.530731201171875, + "learning_rate": 4.3620164945581545e-05, + "loss": 0.0055, + "step": 56370 + }, + { + "epoch": 4.3790291262135925, + "grad_norm": 0.40070706605911255, + "learning_rate": 4.360376574894061e-05, + "loss": 0.0079, + "step": 56380 + }, + { + "epoch": 4.379805825242719, + "grad_norm": 0.25266313552856445, + "learning_rate": 4.358736725178224e-05, + "loss": 0.0065, + "step": 56390 + }, + { + "epoch": 4.380582524271845, + "grad_norm": 0.31104961037635803, + "learning_rate": 4.357096945589974e-05, + "loss": 0.0047, + "step": 56400 + }, + { + "epoch": 4.381359223300971, + "grad_norm": 0.18776702880859375, + "learning_rate": 4.355457236308636e-05, + "loss": 0.0067, + "step": 56410 + }, + { + "epoch": 4.382135922330097, + "grad_norm": 0.1542421132326126, + "learning_rate": 4.353817597513526e-05, + "loss": 0.0062, + "step": 56420 + }, + { + "epoch": 4.382912621359223, + "grad_norm": 0.24997663497924805, + "learning_rate": 4.352178029383948e-05, + "loss": 0.0073, + "step": 56430 + }, + { + "epoch": 4.383689320388349, + "grad_norm": 0.30030059814453125, + "learning_rate": 4.3505385320992105e-05, + "loss": 0.0069, + "step": 56440 + }, + { + "epoch": 4.384466019417475, + "grad_norm": 1.5155644416809082, + "learning_rate": 4.348899105838602e-05, + "loss": 0.0056, + "step": 56450 + }, + { + "epoch": 4.385242718446602, + "grad_norm": 0.6627486944198608, + "learning_rate": 4.3472597507814087e-05, + "loss": 0.0068, + "step": 56460 + }, + { + "epoch": 4.386019417475728, + "grad_norm": 0.18822254240512848, + "learning_rate": 4.3456204671069066e-05, + "loss": 0.0065, + "step": 56470 + }, + { + "epoch": 4.386796116504854, + "grad_norm": 0.1738298386335373, + "learning_rate": 4.343981254994367e-05, + "loss": 0.0092, + "step": 56480 + }, + { + "epoch": 4.38757281553398, + "grad_norm": 0.11080721020698547, + "learning_rate": 4.34234211462305e-05, + "loss": 0.0056, + "step": 56490 + }, + { + "epoch": 4.388349514563107, + "grad_norm": 0.45056167244911194, + "learning_rate": 4.340703046172213e-05, + "loss": 0.0061, + "step": 56500 + }, + { + "epoch": 4.389126213592233, + "grad_norm": 0.4168173670768738, + "learning_rate": 4.339064049821097e-05, + "loss": 0.0055, + "step": 56510 + }, + { + "epoch": 4.389902912621359, + "grad_norm": 0.30022358894348145, + "learning_rate": 4.3374251257489446e-05, + "loss": 0.0073, + "step": 56520 + }, + { + "epoch": 4.390679611650485, + "grad_norm": 0.24877257645130157, + "learning_rate": 4.335786274134984e-05, + "loss": 0.0074, + "step": 56530 + }, + { + "epoch": 4.391456310679612, + "grad_norm": 0.2507590055465698, + "learning_rate": 4.334147495158438e-05, + "loss": 0.0055, + "step": 56540 + }, + { + "epoch": 4.392233009708738, + "grad_norm": 0.15452474355697632, + "learning_rate": 4.33250878899852e-05, + "loss": 0.005, + "step": 56550 + }, + { + "epoch": 4.393009708737864, + "grad_norm": 0.508985161781311, + "learning_rate": 4.330870155834437e-05, + "loss": 0.006, + "step": 56560 + }, + { + "epoch": 4.39378640776699, + "grad_norm": 0.2597752809524536, + "learning_rate": 4.3292315958453884e-05, + "loss": 0.0075, + "step": 56570 + }, + { + "epoch": 4.394563106796117, + "grad_norm": 0.21969088912010193, + "learning_rate": 4.3275931092105646e-05, + "loss": 0.0056, + "step": 56580 + }, + { + "epoch": 4.395339805825243, + "grad_norm": 0.49330881237983704, + "learning_rate": 4.325954696109147e-05, + "loss": 0.0081, + "step": 56590 + }, + { + "epoch": 4.396116504854369, + "grad_norm": 0.16872358322143555, + "learning_rate": 4.32431635672031e-05, + "loss": 0.0053, + "step": 56600 + }, + { + "epoch": 4.396893203883495, + "grad_norm": 0.21431919932365417, + "learning_rate": 4.3226780912232214e-05, + "loss": 0.007, + "step": 56610 + }, + { + "epoch": 4.397669902912622, + "grad_norm": 0.18002554774284363, + "learning_rate": 4.321039899797036e-05, + "loss": 0.0067, + "step": 56620 + }, + { + "epoch": 4.398446601941748, + "grad_norm": 0.35029369592666626, + "learning_rate": 4.319401782620908e-05, + "loss": 0.0062, + "step": 56630 + }, + { + "epoch": 4.399223300970874, + "grad_norm": 0.14169254899024963, + "learning_rate": 4.317763739873978e-05, + "loss": 0.0056, + "step": 56640 + }, + { + "epoch": 4.4, + "grad_norm": 0.25636449456214905, + "learning_rate": 4.31612577173538e-05, + "loss": 0.0062, + "step": 56650 + }, + { + "epoch": 4.400776699029127, + "grad_norm": 0.1508275270462036, + "learning_rate": 4.314487878384239e-05, + "loss": 0.0057, + "step": 56660 + }, + { + "epoch": 4.401553398058253, + "grad_norm": 0.17030590772628784, + "learning_rate": 4.3128500599996736e-05, + "loss": 0.0076, + "step": 56670 + }, + { + "epoch": 4.402330097087379, + "grad_norm": 0.15575815737247467, + "learning_rate": 4.31121231676079e-05, + "loss": 0.0051, + "step": 56680 + }, + { + "epoch": 4.4031067961165045, + "grad_norm": 0.4398282766342163, + "learning_rate": 4.309574648846694e-05, + "loss": 0.0053, + "step": 56690 + }, + { + "epoch": 4.403883495145631, + "grad_norm": 0.26735571026802063, + "learning_rate": 4.3079370564364755e-05, + "loss": 0.0054, + "step": 56700 + }, + { + "epoch": 4.404660194174757, + "grad_norm": 0.24450120329856873, + "learning_rate": 4.306299539709221e-05, + "loss": 0.0044, + "step": 56710 + }, + { + "epoch": 4.405436893203883, + "grad_norm": 0.6019183397293091, + "learning_rate": 4.3046620988440046e-05, + "loss": 0.0056, + "step": 56720 + }, + { + "epoch": 4.4062135922330095, + "grad_norm": 0.23720677196979523, + "learning_rate": 4.303024734019897e-05, + "loss": 0.0056, + "step": 56730 + }, + { + "epoch": 4.406990291262136, + "grad_norm": 0.21810400485992432, + "learning_rate": 4.3013874454159565e-05, + "loss": 0.005, + "step": 56740 + }, + { + "epoch": 4.407766990291262, + "grad_norm": 0.21713708341121674, + "learning_rate": 4.299750233211233e-05, + "loss": 0.0069, + "step": 56750 + }, + { + "epoch": 4.408543689320388, + "grad_norm": 0.2551230788230896, + "learning_rate": 4.2981130975847715e-05, + "loss": 0.0073, + "step": 56760 + }, + { + "epoch": 4.4093203883495145, + "grad_norm": 0.22622250020503998, + "learning_rate": 4.2964760387156076e-05, + "loss": 0.0075, + "step": 56770 + }, + { + "epoch": 4.410097087378641, + "grad_norm": 0.398552268743515, + "learning_rate": 4.294839056782767e-05, + "loss": 0.0065, + "step": 56780 + }, + { + "epoch": 4.410873786407767, + "grad_norm": 0.2283678501844406, + "learning_rate": 4.293202151965266e-05, + "loss": 0.0069, + "step": 56790 + }, + { + "epoch": 4.411650485436893, + "grad_norm": 0.38496121764183044, + "learning_rate": 4.291565324442115e-05, + "loss": 0.005, + "step": 56800 + }, + { + "epoch": 4.4124271844660194, + "grad_norm": 0.2637447714805603, + "learning_rate": 4.2899285743923135e-05, + "loss": 0.0051, + "step": 56810 + }, + { + "epoch": 4.413203883495146, + "grad_norm": 0.44857779145240784, + "learning_rate": 4.288291901994857e-05, + "loss": 0.0052, + "step": 56820 + }, + { + "epoch": 4.413980582524272, + "grad_norm": 0.3467863202095032, + "learning_rate": 4.286655307428729e-05, + "loss": 0.0081, + "step": 56830 + }, + { + "epoch": 4.414757281553398, + "grad_norm": 0.1637067049741745, + "learning_rate": 4.285018790872903e-05, + "loss": 0.0049, + "step": 56840 + }, + { + "epoch": 4.415533980582524, + "grad_norm": 0.19180449843406677, + "learning_rate": 4.283382352506347e-05, + "loss": 0.0049, + "step": 56850 + }, + { + "epoch": 4.416310679611651, + "grad_norm": 0.7338514924049377, + "learning_rate": 4.2817459925080205e-05, + "loss": 0.0044, + "step": 56860 + }, + { + "epoch": 4.417087378640777, + "grad_norm": 0.1974378079175949, + "learning_rate": 4.28010971105687e-05, + "loss": 0.0055, + "step": 56870 + }, + { + "epoch": 4.417864077669903, + "grad_norm": 0.17571291327476501, + "learning_rate": 4.278473508331841e-05, + "loss": 0.006, + "step": 56880 + }, + { + "epoch": 4.418640776699029, + "grad_norm": 0.1250140368938446, + "learning_rate": 4.276837384511864e-05, + "loss": 0.006, + "step": 56890 + }, + { + "epoch": 4.419417475728156, + "grad_norm": 0.14358565211296082, + "learning_rate": 4.275201339775864e-05, + "loss": 0.0038, + "step": 56900 + }, + { + "epoch": 4.420194174757282, + "grad_norm": 0.22052644193172455, + "learning_rate": 4.273565374302756e-05, + "loss": 0.0045, + "step": 56910 + }, + { + "epoch": 4.420970873786407, + "grad_norm": 0.23525188863277435, + "learning_rate": 4.2719294882714454e-05, + "loss": 0.005, + "step": 56920 + }, + { + "epoch": 4.4217475728155335, + "grad_norm": 0.5539165735244751, + "learning_rate": 4.270293681860831e-05, + "loss": 0.0061, + "step": 56930 + }, + { + "epoch": 4.42252427184466, + "grad_norm": 0.18944409489631653, + "learning_rate": 4.268657955249802e-05, + "loss": 0.0053, + "step": 56940 + }, + { + "epoch": 4.423300970873786, + "grad_norm": 0.15607503056526184, + "learning_rate": 4.267022308617241e-05, + "loss": 0.0063, + "step": 56950 + }, + { + "epoch": 4.424077669902912, + "grad_norm": 0.34420913457870483, + "learning_rate": 4.265386742142018e-05, + "loss": 0.0055, + "step": 56960 + }, + { + "epoch": 4.4248543689320385, + "grad_norm": 0.3519086539745331, + "learning_rate": 4.263751256002996e-05, + "loss": 0.0058, + "step": 56970 + }, + { + "epoch": 4.425631067961165, + "grad_norm": 0.2519380748271942, + "learning_rate": 4.26211585037903e-05, + "loss": 0.0056, + "step": 56980 + }, + { + "epoch": 4.426407766990291, + "grad_norm": 0.1716245859861374, + "learning_rate": 4.2604805254489656e-05, + "loss": 0.0068, + "step": 56990 + }, + { + "epoch": 4.427184466019417, + "grad_norm": 0.18608419597148895, + "learning_rate": 4.258845281391638e-05, + "loss": 0.008, + "step": 57000 + }, + { + "epoch": 4.4279611650485435, + "grad_norm": 0.16546359658241272, + "learning_rate": 4.257210118385876e-05, + "loss": 0.0047, + "step": 57010 + }, + { + "epoch": 4.42873786407767, + "grad_norm": 0.4153691232204437, + "learning_rate": 4.2555750366104994e-05, + "loss": 0.007, + "step": 57020 + }, + { + "epoch": 4.429514563106796, + "grad_norm": 0.21034123003482819, + "learning_rate": 4.2539400362443184e-05, + "loss": 0.0045, + "step": 57030 + }, + { + "epoch": 4.430291262135922, + "grad_norm": 0.14534598588943481, + "learning_rate": 4.252305117466133e-05, + "loss": 0.0054, + "step": 57040 + }, + { + "epoch": 4.4310679611650485, + "grad_norm": 0.24966847896575928, + "learning_rate": 4.2506702804547356e-05, + "loss": 0.0057, + "step": 57050 + }, + { + "epoch": 4.431844660194175, + "grad_norm": 0.2355194091796875, + "learning_rate": 4.249035525388907e-05, + "loss": 0.0065, + "step": 57060 + }, + { + "epoch": 4.432621359223301, + "grad_norm": 0.27662554383277893, + "learning_rate": 4.247400852447427e-05, + "loss": 0.0037, + "step": 57070 + }, + { + "epoch": 4.433398058252427, + "grad_norm": 0.1747913956642151, + "learning_rate": 4.245766261809059e-05, + "loss": 0.0065, + "step": 57080 + }, + { + "epoch": 4.4341747572815535, + "grad_norm": 0.22775839269161224, + "learning_rate": 4.244131753652558e-05, + "loss": 0.0072, + "step": 57090 + }, + { + "epoch": 4.43495145631068, + "grad_norm": 0.17128413915634155, + "learning_rate": 4.242497328156672e-05, + "loss": 0.0038, + "step": 57100 + }, + { + "epoch": 4.435728155339806, + "grad_norm": 0.2149246782064438, + "learning_rate": 4.240862985500139e-05, + "loss": 0.0065, + "step": 57110 + }, + { + "epoch": 4.436504854368932, + "grad_norm": 0.3229142427444458, + "learning_rate": 4.2392287258616894e-05, + "loss": 0.0061, + "step": 57120 + }, + { + "epoch": 4.4372815533980585, + "grad_norm": 0.20988139510154724, + "learning_rate": 4.2375945494200396e-05, + "loss": 0.0076, + "step": 57130 + }, + { + "epoch": 4.438058252427185, + "grad_norm": 0.15369541943073273, + "learning_rate": 4.2359604563539054e-05, + "loss": 0.0045, + "step": 57140 + }, + { + "epoch": 4.438834951456311, + "grad_norm": 0.3376014232635498, + "learning_rate": 4.2343264468419885e-05, + "loss": 0.0075, + "step": 57150 + }, + { + "epoch": 4.439611650485437, + "grad_norm": 0.19748882949352264, + "learning_rate": 4.232692521062979e-05, + "loss": 0.0064, + "step": 57160 + }, + { + "epoch": 4.4403883495145635, + "grad_norm": 0.1861790269613266, + "learning_rate": 4.231058679195562e-05, + "loss": 0.0039, + "step": 57170 + }, + { + "epoch": 4.44116504854369, + "grad_norm": 0.2379831075668335, + "learning_rate": 4.229424921418411e-05, + "loss": 0.006, + "step": 57180 + }, + { + "epoch": 4.441941747572815, + "grad_norm": 0.1616472452878952, + "learning_rate": 4.2277912479101916e-05, + "loss": 0.0042, + "step": 57190 + }, + { + "epoch": 4.442718446601941, + "grad_norm": 0.10961368680000305, + "learning_rate": 4.226157658849562e-05, + "loss": 0.0069, + "step": 57200 + }, + { + "epoch": 4.443495145631068, + "grad_norm": 0.28250426054000854, + "learning_rate": 4.2245241544151664e-05, + "loss": 0.0066, + "step": 57210 + }, + { + "epoch": 4.444271844660194, + "grad_norm": 0.3200227916240692, + "learning_rate": 4.2228907347856425e-05, + "loss": 0.0057, + "step": 57220 + }, + { + "epoch": 4.44504854368932, + "grad_norm": 0.17049573361873627, + "learning_rate": 4.2212574001396214e-05, + "loss": 0.0046, + "step": 57230 + }, + { + "epoch": 4.445825242718446, + "grad_norm": 0.3199881315231323, + "learning_rate": 4.21962415065572e-05, + "loss": 0.0051, + "step": 57240 + }, + { + "epoch": 4.446601941747573, + "grad_norm": 0.22952720522880554, + "learning_rate": 4.217990986512548e-05, + "loss": 0.0046, + "step": 57250 + }, + { + "epoch": 4.447378640776699, + "grad_norm": 0.2055302858352661, + "learning_rate": 4.216357907888704e-05, + "loss": 0.0061, + "step": 57260 + }, + { + "epoch": 4.448155339805825, + "grad_norm": 0.23850595951080322, + "learning_rate": 4.2147249149627824e-05, + "loss": 0.0052, + "step": 57270 + }, + { + "epoch": 4.448932038834951, + "grad_norm": 0.1573915034532547, + "learning_rate": 4.213092007913364e-05, + "loss": 0.0043, + "step": 57280 + }, + { + "epoch": 4.449708737864078, + "grad_norm": 0.2598455846309662, + "learning_rate": 4.2114591869190205e-05, + "loss": 0.0045, + "step": 57290 + }, + { + "epoch": 4.450485436893204, + "grad_norm": 0.36238107085227966, + "learning_rate": 4.209826452158315e-05, + "loss": 0.0056, + "step": 57300 + }, + { + "epoch": 4.45126213592233, + "grad_norm": 0.21746252477169037, + "learning_rate": 4.2081938038098e-05, + "loss": 0.0057, + "step": 57310 + }, + { + "epoch": 4.452038834951456, + "grad_norm": 0.2977639436721802, + "learning_rate": 4.2065612420520185e-05, + "loss": 0.0052, + "step": 57320 + }, + { + "epoch": 4.452815533980583, + "grad_norm": 0.9570261836051941, + "learning_rate": 4.204928767063509e-05, + "loss": 0.0063, + "step": 57330 + }, + { + "epoch": 4.453592233009709, + "grad_norm": 0.18052202463150024, + "learning_rate": 4.2032963790227944e-05, + "loss": 0.005, + "step": 57340 + }, + { + "epoch": 4.454368932038835, + "grad_norm": 0.30776551365852356, + "learning_rate": 4.2016640781083887e-05, + "loss": 0.0057, + "step": 57350 + }, + { + "epoch": 4.455145631067961, + "grad_norm": 0.26273277401924133, + "learning_rate": 4.2000318644988004e-05, + "loss": 0.0056, + "step": 57360 + }, + { + "epoch": 4.455922330097088, + "grad_norm": 0.24323634803295135, + "learning_rate": 4.198399738372524e-05, + "loss": 0.0062, + "step": 57370 + }, + { + "epoch": 4.456699029126214, + "grad_norm": 0.503639280796051, + "learning_rate": 4.196767699908045e-05, + "loss": 0.0089, + "step": 57380 + }, + { + "epoch": 4.45747572815534, + "grad_norm": 0.34101513028144836, + "learning_rate": 4.195135749283845e-05, + "loss": 0.0069, + "step": 57390 + }, + { + "epoch": 4.458252427184466, + "grad_norm": 0.174057736992836, + "learning_rate": 4.1935038866783894e-05, + "loss": 0.0057, + "step": 57400 + }, + { + "epoch": 4.459029126213593, + "grad_norm": 0.35878103971481323, + "learning_rate": 4.1918721122701355e-05, + "loss": 0.0071, + "step": 57410 + }, + { + "epoch": 4.459805825242719, + "grad_norm": 0.9739441871643066, + "learning_rate": 4.190240426237533e-05, + "loss": 0.0081, + "step": 57420 + }, + { + "epoch": 4.460582524271844, + "grad_norm": 0.1975630670785904, + "learning_rate": 4.188608828759019e-05, + "loss": 0.005, + "step": 57430 + }, + { + "epoch": 4.4613592233009705, + "grad_norm": 0.2914479672908783, + "learning_rate": 4.1869773200130246e-05, + "loss": 0.0079, + "step": 57440 + }, + { + "epoch": 4.462135922330097, + "grad_norm": 0.23960793018341064, + "learning_rate": 4.185345900177967e-05, + "loss": 0.0063, + "step": 57450 + }, + { + "epoch": 4.462912621359223, + "grad_norm": 0.3828617036342621, + "learning_rate": 4.183714569432258e-05, + "loss": 0.0065, + "step": 57460 + }, + { + "epoch": 4.463689320388349, + "grad_norm": 0.2643864154815674, + "learning_rate": 4.182083327954295e-05, + "loss": 0.007, + "step": 57470 + }, + { + "epoch": 4.4644660194174755, + "grad_norm": 0.1749376803636551, + "learning_rate": 4.1804521759224714e-05, + "loss": 0.0086, + "step": 57480 + }, + { + "epoch": 4.465242718446602, + "grad_norm": 0.35034430027008057, + "learning_rate": 4.178821113515165e-05, + "loss": 0.0064, + "step": 57490 + }, + { + "epoch": 4.466019417475728, + "grad_norm": 0.20973439514636993, + "learning_rate": 4.1771901409107476e-05, + "loss": 0.0069, + "step": 57500 + }, + { + "epoch": 4.466796116504854, + "grad_norm": 0.1750682145357132, + "learning_rate": 4.175559258287577e-05, + "loss": 0.0087, + "step": 57510 + }, + { + "epoch": 4.4675728155339804, + "grad_norm": 0.13268396258354187, + "learning_rate": 4.173928465824008e-05, + "loss": 0.0063, + "step": 57520 + }, + { + "epoch": 4.468349514563107, + "grad_norm": 0.1897769570350647, + "learning_rate": 4.17229776369838e-05, + "loss": 0.005, + "step": 57530 + }, + { + "epoch": 4.469126213592233, + "grad_norm": 0.19095730781555176, + "learning_rate": 4.170667152089025e-05, + "loss": 0.005, + "step": 57540 + }, + { + "epoch": 4.469902912621359, + "grad_norm": 0.3474772572517395, + "learning_rate": 4.1690366311742626e-05, + "loss": 0.0058, + "step": 57550 + }, + { + "epoch": 4.470679611650485, + "grad_norm": 0.2828117907047272, + "learning_rate": 4.1674062011324055e-05, + "loss": 0.0059, + "step": 57560 + }, + { + "epoch": 4.471456310679612, + "grad_norm": 0.2652389705181122, + "learning_rate": 4.165775862141752e-05, + "loss": 0.0061, + "step": 57570 + }, + { + "epoch": 4.472233009708738, + "grad_norm": 0.4060574769973755, + "learning_rate": 4.164145614380598e-05, + "loss": 0.0067, + "step": 57580 + }, + { + "epoch": 4.473009708737864, + "grad_norm": 0.3065275549888611, + "learning_rate": 4.162515458027223e-05, + "loss": 0.0065, + "step": 57590 + }, + { + "epoch": 4.47378640776699, + "grad_norm": 0.3144548833370209, + "learning_rate": 4.160885393259899e-05, + "loss": 0.0061, + "step": 57600 + }, + { + "epoch": 4.474563106796117, + "grad_norm": 0.3563598394393921, + "learning_rate": 4.159255420256886e-05, + "loss": 0.0075, + "step": 57610 + }, + { + "epoch": 4.475339805825243, + "grad_norm": 0.30957478284835815, + "learning_rate": 4.157625539196436e-05, + "loss": 0.0073, + "step": 57620 + }, + { + "epoch": 4.476116504854369, + "grad_norm": 0.32225558161735535, + "learning_rate": 4.155995750256791e-05, + "loss": 0.0088, + "step": 57630 + }, + { + "epoch": 4.476893203883495, + "grad_norm": 0.2370673567056656, + "learning_rate": 4.15436605361618e-05, + "loss": 0.0069, + "step": 57640 + }, + { + "epoch": 4.477669902912622, + "grad_norm": 0.32739007472991943, + "learning_rate": 4.152736449452827e-05, + "loss": 0.0057, + "step": 57650 + }, + { + "epoch": 4.478446601941748, + "grad_norm": 0.30130013823509216, + "learning_rate": 4.151106937944943e-05, + "loss": 0.0058, + "step": 57660 + }, + { + "epoch": 4.479223300970874, + "grad_norm": 0.31720781326293945, + "learning_rate": 4.149477519270727e-05, + "loss": 0.0063, + "step": 57670 + }, + { + "epoch": 4.48, + "grad_norm": 0.23767773807048798, + "learning_rate": 4.147848193608371e-05, + "loss": 0.0049, + "step": 57680 + }, + { + "epoch": 4.480776699029127, + "grad_norm": 0.20140764117240906, + "learning_rate": 4.146218961136056e-05, + "loss": 0.0072, + "step": 57690 + }, + { + "epoch": 4.481553398058252, + "grad_norm": 0.35308152437210083, + "learning_rate": 4.14458982203195e-05, + "loss": 0.0059, + "step": 57700 + }, + { + "epoch": 4.482330097087378, + "grad_norm": 0.20124202966690063, + "learning_rate": 4.1429607764742165e-05, + "loss": 0.0054, + "step": 57710 + }, + { + "epoch": 4.4831067961165045, + "grad_norm": 0.13285759091377258, + "learning_rate": 4.141331824641003e-05, + "loss": 0.0044, + "step": 57720 + }, + { + "epoch": 4.483883495145631, + "grad_norm": 0.16495457291603088, + "learning_rate": 4.139702966710452e-05, + "loss": 0.0066, + "step": 57730 + }, + { + "epoch": 4.484660194174757, + "grad_norm": 0.260623961687088, + "learning_rate": 4.13807420286069e-05, + "loss": 0.0109, + "step": 57740 + }, + { + "epoch": 4.485436893203883, + "grad_norm": 0.3900700509548187, + "learning_rate": 4.136445533269837e-05, + "loss": 0.0052, + "step": 57750 + }, + { + "epoch": 4.4862135922330095, + "grad_norm": 0.2652446925640106, + "learning_rate": 4.134816958116003e-05, + "loss": 0.0085, + "step": 57760 + }, + { + "epoch": 4.486990291262136, + "grad_norm": 0.2883809506893158, + "learning_rate": 4.133188477577283e-05, + "loss": 0.0051, + "step": 57770 + }, + { + "epoch": 4.487766990291262, + "grad_norm": 0.1672740876674652, + "learning_rate": 4.131560091831771e-05, + "loss": 0.0064, + "step": 57780 + }, + { + "epoch": 4.488543689320388, + "grad_norm": 0.3276269733905792, + "learning_rate": 4.1299318010575405e-05, + "loss": 0.0065, + "step": 57790 + }, + { + "epoch": 4.4893203883495145, + "grad_norm": 0.14152498543262482, + "learning_rate": 4.128303605432661e-05, + "loss": 0.0086, + "step": 57800 + }, + { + "epoch": 4.490097087378641, + "grad_norm": 0.15362311899662018, + "learning_rate": 4.126675505135188e-05, + "loss": 0.0066, + "step": 57810 + }, + { + "epoch": 4.490873786407767, + "grad_norm": 0.8724566102027893, + "learning_rate": 4.125047500343169e-05, + "loss": 0.008, + "step": 57820 + }, + { + "epoch": 4.491650485436893, + "grad_norm": 0.1432860791683197, + "learning_rate": 4.1234195912346376e-05, + "loss": 0.0042, + "step": 57830 + }, + { + "epoch": 4.4924271844660195, + "grad_norm": 0.1718229502439499, + "learning_rate": 4.1217917779876235e-05, + "loss": 0.0068, + "step": 57840 + }, + { + "epoch": 4.493203883495146, + "grad_norm": 0.28984472155570984, + "learning_rate": 4.120164060780141e-05, + "loss": 0.0093, + "step": 57850 + }, + { + "epoch": 4.493980582524272, + "grad_norm": 0.28813040256500244, + "learning_rate": 4.1185364397901935e-05, + "loss": 0.008, + "step": 57860 + }, + { + "epoch": 4.494757281553398, + "grad_norm": 0.2431521713733673, + "learning_rate": 4.116908915195776e-05, + "loss": 0.0059, + "step": 57870 + }, + { + "epoch": 4.4955339805825245, + "grad_norm": 0.20379063487052917, + "learning_rate": 4.115281487174872e-05, + "loss": 0.0089, + "step": 57880 + }, + { + "epoch": 4.496310679611651, + "grad_norm": 0.27010729908943176, + "learning_rate": 4.113654155905452e-05, + "loss": 0.0075, + "step": 57890 + }, + { + "epoch": 4.497087378640777, + "grad_norm": 0.4121648073196411, + "learning_rate": 4.112026921565483e-05, + "loss": 0.0058, + "step": 57900 + }, + { + "epoch": 4.497864077669903, + "grad_norm": 0.14647147059440613, + "learning_rate": 4.1103997843329147e-05, + "loss": 0.0045, + "step": 57910 + }, + { + "epoch": 4.4986407766990295, + "grad_norm": 0.8167838454246521, + "learning_rate": 4.108772744385689e-05, + "loss": 0.0081, + "step": 57920 + }, + { + "epoch": 4.499417475728155, + "grad_norm": 0.18605470657348633, + "learning_rate": 4.1071458019017346e-05, + "loss": 0.0058, + "step": 57930 + }, + { + "epoch": 4.500194174757281, + "grad_norm": 0.12088749557733536, + "learning_rate": 4.1055189570589734e-05, + "loss": 0.0053, + "step": 57940 + }, + { + "epoch": 4.500970873786407, + "grad_norm": 0.4418761730194092, + "learning_rate": 4.103892210035315e-05, + "loss": 0.0057, + "step": 57950 + }, + { + "epoch": 4.501747572815534, + "grad_norm": 0.35475435853004456, + "learning_rate": 4.102265561008656e-05, + "loss": 0.0076, + "step": 57960 + }, + { + "epoch": 4.50252427184466, + "grad_norm": 0.6625497937202454, + "learning_rate": 4.100639010156885e-05, + "loss": 0.0067, + "step": 57970 + }, + { + "epoch": 4.503300970873786, + "grad_norm": 0.1987023949623108, + "learning_rate": 4.099012557657882e-05, + "loss": 0.0062, + "step": 57980 + }, + { + "epoch": 4.504077669902912, + "grad_norm": 0.239919051527977, + "learning_rate": 4.09738620368951e-05, + "loss": 0.0046, + "step": 57990 + }, + { + "epoch": 4.504854368932039, + "grad_norm": 0.22530116140842438, + "learning_rate": 4.095759948429626e-05, + "loss": 0.0058, + "step": 58000 + }, + { + "epoch": 4.505631067961165, + "grad_norm": 0.13524580001831055, + "learning_rate": 4.094133792056075e-05, + "loss": 0.0074, + "step": 58010 + }, + { + "epoch": 4.506407766990291, + "grad_norm": 0.2652228772640228, + "learning_rate": 4.092507734746688e-05, + "loss": 0.007, + "step": 58020 + }, + { + "epoch": 4.507184466019417, + "grad_norm": 0.15558098256587982, + "learning_rate": 4.090881776679293e-05, + "loss": 0.0064, + "step": 58030 + }, + { + "epoch": 4.507961165048544, + "grad_norm": 0.22079958021640778, + "learning_rate": 4.0892559180316994e-05, + "loss": 0.005, + "step": 58040 + }, + { + "epoch": 4.50873786407767, + "grad_norm": 0.22835558652877808, + "learning_rate": 4.08763015898171e-05, + "loss": 0.0064, + "step": 58050 + }, + { + "epoch": 4.509514563106796, + "grad_norm": 0.24377115070819855, + "learning_rate": 4.0860044997071145e-05, + "loss": 0.0045, + "step": 58060 + }, + { + "epoch": 4.510291262135922, + "grad_norm": 0.23220744729042053, + "learning_rate": 4.084378940385693e-05, + "loss": 0.0055, + "step": 58070 + }, + { + "epoch": 4.511067961165049, + "grad_norm": 0.1655992865562439, + "learning_rate": 4.082753481195211e-05, + "loss": 0.0071, + "step": 58080 + }, + { + "epoch": 4.511844660194175, + "grad_norm": 0.4031551778316498, + "learning_rate": 4.0811281223134324e-05, + "loss": 0.0074, + "step": 58090 + }, + { + "epoch": 4.512621359223301, + "grad_norm": 0.1740717738866806, + "learning_rate": 4.0795028639181e-05, + "loss": 0.0062, + "step": 58100 + }, + { + "epoch": 4.513398058252427, + "grad_norm": 0.2657204568386078, + "learning_rate": 4.077877706186951e-05, + "loss": 0.0059, + "step": 58110 + }, + { + "epoch": 4.514174757281554, + "grad_norm": 0.1995401531457901, + "learning_rate": 4.076252649297709e-05, + "loss": 0.0087, + "step": 58120 + }, + { + "epoch": 4.51495145631068, + "grad_norm": 0.22008608281612396, + "learning_rate": 4.07462769342809e-05, + "loss": 0.0049, + "step": 58130 + }, + { + "epoch": 4.515728155339806, + "grad_norm": 0.22137568891048431, + "learning_rate": 4.073002838755793e-05, + "loss": 0.0063, + "step": 58140 + }, + { + "epoch": 4.516504854368932, + "grad_norm": 0.1285216212272644, + "learning_rate": 4.0713780854585115e-05, + "loss": 0.0038, + "step": 58150 + }, + { + "epoch": 4.517281553398059, + "grad_norm": 0.4972366392612457, + "learning_rate": 4.069753433713927e-05, + "loss": 0.0093, + "step": 58160 + }, + { + "epoch": 4.518058252427185, + "grad_norm": 0.3568299114704132, + "learning_rate": 4.068128883699709e-05, + "loss": 0.0053, + "step": 58170 + }, + { + "epoch": 4.518834951456311, + "grad_norm": 0.20131134986877441, + "learning_rate": 4.0665044355935155e-05, + "loss": 0.0058, + "step": 58180 + }, + { + "epoch": 4.519611650485437, + "grad_norm": 0.21118871867656708, + "learning_rate": 4.064880089572993e-05, + "loss": 0.0089, + "step": 58190 + }, + { + "epoch": 4.520388349514564, + "grad_norm": 0.1869697868824005, + "learning_rate": 4.063255845815779e-05, + "loss": 0.0059, + "step": 58200 + }, + { + "epoch": 4.521165048543689, + "grad_norm": 0.26487454771995544, + "learning_rate": 4.0616317044994945e-05, + "loss": 0.0058, + "step": 58210 + }, + { + "epoch": 4.521941747572815, + "grad_norm": 0.3019484579563141, + "learning_rate": 4.0600076658017585e-05, + "loss": 0.0061, + "step": 58220 + }, + { + "epoch": 4.5227184466019414, + "grad_norm": 0.10175526887178421, + "learning_rate": 4.0583837299001706e-05, + "loss": 0.0056, + "step": 58230 + }, + { + "epoch": 4.523495145631068, + "grad_norm": 0.24166274070739746, + "learning_rate": 4.056759896972322e-05, + "loss": 0.0062, + "step": 58240 + }, + { + "epoch": 4.524271844660194, + "grad_norm": 0.26805946230888367, + "learning_rate": 4.055136167195793e-05, + "loss": 0.0048, + "step": 58250 + }, + { + "epoch": 4.52504854368932, + "grad_norm": 0.10279053449630737, + "learning_rate": 4.053512540748152e-05, + "loss": 0.004, + "step": 58260 + }, + { + "epoch": 4.525825242718446, + "grad_norm": 0.24620608985424042, + "learning_rate": 4.051889017806955e-05, + "loss": 0.0054, + "step": 58270 + }, + { + "epoch": 4.526601941747573, + "grad_norm": 0.15860474109649658, + "learning_rate": 4.0502655985497516e-05, + "loss": 0.0071, + "step": 58280 + }, + { + "epoch": 4.527378640776699, + "grad_norm": 0.14878790080547333, + "learning_rate": 4.048642283154074e-05, + "loss": 0.0102, + "step": 58290 + }, + { + "epoch": 4.528155339805825, + "grad_norm": 0.2520667314529419, + "learning_rate": 4.0470190717974464e-05, + "loss": 0.0076, + "step": 58300 + }, + { + "epoch": 4.528932038834951, + "grad_norm": 0.19142146408557892, + "learning_rate": 4.045395964657381e-05, + "loss": 0.0058, + "step": 58310 + }, + { + "epoch": 4.529708737864078, + "grad_norm": 0.22828362882137299, + "learning_rate": 4.043772961911376e-05, + "loss": 0.0059, + "step": 58320 + }, + { + "epoch": 4.530485436893204, + "grad_norm": 0.30126476287841797, + "learning_rate": 4.042150063736924e-05, + "loss": 0.0048, + "step": 58330 + }, + { + "epoch": 4.53126213592233, + "grad_norm": 0.23872973024845123, + "learning_rate": 4.0405272703114976e-05, + "loss": 0.0045, + "step": 58340 + }, + { + "epoch": 4.532038834951456, + "grad_norm": 0.1794537901878357, + "learning_rate": 4.038904581812569e-05, + "loss": 0.006, + "step": 58350 + }, + { + "epoch": 4.532815533980583, + "grad_norm": 0.1419917196035385, + "learning_rate": 4.03728199841759e-05, + "loss": 0.0065, + "step": 58360 + }, + { + "epoch": 4.533592233009709, + "grad_norm": 0.26130956411361694, + "learning_rate": 4.0356595203040044e-05, + "loss": 0.0067, + "step": 58370 + }, + { + "epoch": 4.534368932038835, + "grad_norm": 0.22441866993904114, + "learning_rate": 4.0340371476492426e-05, + "loss": 0.0065, + "step": 58380 + }, + { + "epoch": 4.535145631067961, + "grad_norm": 0.17894013226032257, + "learning_rate": 4.032414880630727e-05, + "loss": 0.0037, + "step": 58390 + }, + { + "epoch": 4.535922330097088, + "grad_norm": 0.2149256020784378, + "learning_rate": 4.0307927194258646e-05, + "loss": 0.0062, + "step": 58400 + }, + { + "epoch": 4.536699029126214, + "grad_norm": 0.16217206418514252, + "learning_rate": 4.029170664212054e-05, + "loss": 0.0044, + "step": 58410 + }, + { + "epoch": 4.53747572815534, + "grad_norm": 0.15753592550754547, + "learning_rate": 4.0275487151666794e-05, + "loss": 0.0103, + "step": 58420 + }, + { + "epoch": 4.5382524271844655, + "grad_norm": 0.24664553999900818, + "learning_rate": 4.025926872467116e-05, + "loss": 0.0085, + "step": 58430 + }, + { + "epoch": 4.539029126213592, + "grad_norm": 0.28134214878082275, + "learning_rate": 4.024305136290725e-05, + "loss": 0.0063, + "step": 58440 + }, + { + "epoch": 4.539805825242718, + "grad_norm": 0.3713642656803131, + "learning_rate": 4.022683506814858e-05, + "loss": 0.0074, + "step": 58450 + }, + { + "epoch": 4.540582524271844, + "grad_norm": 0.14145775139331818, + "learning_rate": 4.021061984216853e-05, + "loss": 0.0059, + "step": 58460 + }, + { + "epoch": 4.5413592233009705, + "grad_norm": 0.25376471877098083, + "learning_rate": 4.019440568674035e-05, + "loss": 0.0066, + "step": 58470 + }, + { + "epoch": 4.542135922330097, + "grad_norm": 0.10829932242631912, + "learning_rate": 4.017819260363724e-05, + "loss": 0.0068, + "step": 58480 + }, + { + "epoch": 4.542912621359223, + "grad_norm": 0.2647787630558014, + "learning_rate": 4.016198059463223e-05, + "loss": 0.0038, + "step": 58490 + }, + { + "epoch": 4.543689320388349, + "grad_norm": 0.3494603931903839, + "learning_rate": 4.0145769661498233e-05, + "loss": 0.0065, + "step": 58500 + }, + { + "epoch": 4.5444660194174755, + "grad_norm": 0.11686909198760986, + "learning_rate": 4.012955980600804e-05, + "loss": 0.0069, + "step": 58510 + }, + { + "epoch": 4.545242718446602, + "grad_norm": 0.20103377103805542, + "learning_rate": 4.0113351029934355e-05, + "loss": 0.0067, + "step": 58520 + }, + { + "epoch": 4.546019417475728, + "grad_norm": 0.22737088799476624, + "learning_rate": 4.0097143335049714e-05, + "loss": 0.0062, + "step": 58530 + }, + { + "epoch": 4.546796116504854, + "grad_norm": 0.30966565012931824, + "learning_rate": 4.0080936723126606e-05, + "loss": 0.0088, + "step": 58540 + }, + { + "epoch": 4.5475728155339805, + "grad_norm": 0.20346908271312714, + "learning_rate": 4.006473119593735e-05, + "loss": 0.005, + "step": 58550 + }, + { + "epoch": 4.548349514563107, + "grad_norm": 0.17306353151798248, + "learning_rate": 4.004852675525415e-05, + "loss": 0.0072, + "step": 58560 + }, + { + "epoch": 4.549126213592233, + "grad_norm": 0.1455027461051941, + "learning_rate": 4.00323234028491e-05, + "loss": 0.0064, + "step": 58570 + }, + { + "epoch": 4.549902912621359, + "grad_norm": 0.24200710654258728, + "learning_rate": 4.0016121140494186e-05, + "loss": 0.0054, + "step": 58580 + }, + { + "epoch": 4.5506796116504855, + "grad_norm": 0.10240890085697174, + "learning_rate": 3.999991996996122e-05, + "loss": 0.0058, + "step": 58590 + }, + { + "epoch": 4.551456310679612, + "grad_norm": 0.3324241042137146, + "learning_rate": 3.9983719893022e-05, + "loss": 0.0066, + "step": 58600 + }, + { + "epoch": 4.552233009708738, + "grad_norm": 0.13931946456432343, + "learning_rate": 3.996752091144812e-05, + "loss": 0.0052, + "step": 58610 + }, + { + "epoch": 4.553009708737864, + "grad_norm": 0.3518921136856079, + "learning_rate": 3.9951323027011055e-05, + "loss": 0.0045, + "step": 58620 + }, + { + "epoch": 4.5537864077669905, + "grad_norm": 0.2630673944950104, + "learning_rate": 3.993512624148219e-05, + "loss": 0.0058, + "step": 58630 + }, + { + "epoch": 4.554563106796117, + "grad_norm": 0.2743470370769501, + "learning_rate": 3.99189305566328e-05, + "loss": 0.0052, + "step": 58640 + }, + { + "epoch": 4.555339805825243, + "grad_norm": 0.2093828320503235, + "learning_rate": 3.990273597423401e-05, + "loss": 0.0052, + "step": 58650 + }, + { + "epoch": 4.556116504854369, + "grad_norm": 0.13906212151050568, + "learning_rate": 3.988654249605681e-05, + "loss": 0.0054, + "step": 58660 + }, + { + "epoch": 4.5568932038834955, + "grad_norm": 0.24383559823036194, + "learning_rate": 3.987035012387213e-05, + "loss": 0.0072, + "step": 58670 + }, + { + "epoch": 4.557669902912622, + "grad_norm": 0.18891651928424835, + "learning_rate": 3.9854158859450735e-05, + "loss": 0.0075, + "step": 58680 + }, + { + "epoch": 4.558446601941748, + "grad_norm": 0.3149276375770569, + "learning_rate": 3.983796870456328e-05, + "loss": 0.0071, + "step": 58690 + }, + { + "epoch": 4.559223300970874, + "grad_norm": 0.20660129189491272, + "learning_rate": 3.982177966098028e-05, + "loss": 0.0054, + "step": 58700 + }, + { + "epoch": 4.5600000000000005, + "grad_norm": 0.2004009485244751, + "learning_rate": 3.980559173047216e-05, + "loss": 0.0068, + "step": 58710 + }, + { + "epoch": 4.560776699029126, + "grad_norm": 0.1540137678384781, + "learning_rate": 3.9789404914809176e-05, + "loss": 0.0043, + "step": 58720 + }, + { + "epoch": 4.561553398058252, + "grad_norm": 0.6970967650413513, + "learning_rate": 3.9773219215761544e-05, + "loss": 0.006, + "step": 58730 + }, + { + "epoch": 4.562330097087378, + "grad_norm": 0.6563428640365601, + "learning_rate": 3.975703463509928e-05, + "loss": 0.0078, + "step": 58740 + }, + { + "epoch": 4.563106796116505, + "grad_norm": 0.23602250218391418, + "learning_rate": 3.974085117459231e-05, + "loss": 0.0055, + "step": 58750 + }, + { + "epoch": 4.563883495145631, + "grad_norm": 0.279764860868454, + "learning_rate": 3.972466883601043e-05, + "loss": 0.007, + "step": 58760 + }, + { + "epoch": 4.564660194174757, + "grad_norm": 0.16804452240467072, + "learning_rate": 3.9708487621123316e-05, + "loss": 0.0043, + "step": 58770 + }, + { + "epoch": 4.565436893203883, + "grad_norm": 0.3960012197494507, + "learning_rate": 3.969230753170051e-05, + "loss": 0.0075, + "step": 58780 + }, + { + "epoch": 4.56621359223301, + "grad_norm": 0.2815686762332916, + "learning_rate": 3.967612856951146e-05, + "loss": 0.0069, + "step": 58790 + }, + { + "epoch": 4.566990291262136, + "grad_norm": 0.20018604397773743, + "learning_rate": 3.9659950736325476e-05, + "loss": 0.0058, + "step": 58800 + }, + { + "epoch": 4.567766990291262, + "grad_norm": 0.18882764875888824, + "learning_rate": 3.964377403391174e-05, + "loss": 0.0078, + "step": 58810 + }, + { + "epoch": 4.568543689320388, + "grad_norm": 0.12149135768413544, + "learning_rate": 3.9627598464039303e-05, + "loss": 0.0075, + "step": 58820 + }, + { + "epoch": 4.569320388349515, + "grad_norm": 0.2473757565021515, + "learning_rate": 3.9611424028477096e-05, + "loss": 0.0054, + "step": 58830 + }, + { + "epoch": 4.570097087378641, + "grad_norm": 0.2692251205444336, + "learning_rate": 3.959525072899394e-05, + "loss": 0.0063, + "step": 58840 + }, + { + "epoch": 4.570873786407767, + "grad_norm": 0.17500804364681244, + "learning_rate": 3.9579078567358506e-05, + "loss": 0.006, + "step": 58850 + }, + { + "epoch": 4.571650485436893, + "grad_norm": 0.2385723739862442, + "learning_rate": 3.956290754533939e-05, + "loss": 0.009, + "step": 58860 + }, + { + "epoch": 4.57242718446602, + "grad_norm": 0.38992902636528015, + "learning_rate": 3.954673766470501e-05, + "loss": 0.0061, + "step": 58870 + }, + { + "epoch": 4.573203883495146, + "grad_norm": 0.2597666382789612, + "learning_rate": 3.9530568927223674e-05, + "loss": 0.0047, + "step": 58880 + }, + { + "epoch": 4.573980582524272, + "grad_norm": 0.5032052397727966, + "learning_rate": 3.95144013346636e-05, + "loss": 0.0058, + "step": 58890 + }, + { + "epoch": 4.574757281553398, + "grad_norm": 0.38171252608299255, + "learning_rate": 3.949823488879282e-05, + "loss": 0.0048, + "step": 58900 + }, + { + "epoch": 4.575533980582525, + "grad_norm": 0.35881373286247253, + "learning_rate": 3.9482069591379276e-05, + "loss": 0.0053, + "step": 58910 + }, + { + "epoch": 4.576310679611651, + "grad_norm": 0.3912959694862366, + "learning_rate": 3.9465905444190784e-05, + "loss": 0.0102, + "step": 58920 + }, + { + "epoch": 4.577087378640777, + "grad_norm": 0.35728198289871216, + "learning_rate": 3.944974244899505e-05, + "loss": 0.0093, + "step": 58930 + }, + { + "epoch": 4.577864077669902, + "grad_norm": 0.3220526874065399, + "learning_rate": 3.9433580607559616e-05, + "loss": 0.009, + "step": 58940 + }, + { + "epoch": 4.578640776699029, + "grad_norm": 0.4397866427898407, + "learning_rate": 3.941741992165193e-05, + "loss": 0.0066, + "step": 58950 + }, + { + "epoch": 4.579417475728155, + "grad_norm": 0.19767643511295319, + "learning_rate": 3.940126039303928e-05, + "loss": 0.0074, + "step": 58960 + }, + { + "epoch": 4.580194174757281, + "grad_norm": 0.9300804734230042, + "learning_rate": 3.938510202348886e-05, + "loss": 0.0062, + "step": 58970 + }, + { + "epoch": 4.580970873786407, + "grad_norm": 0.6092187166213989, + "learning_rate": 3.93689448147677e-05, + "loss": 0.0044, + "step": 58980 + }, + { + "epoch": 4.581747572815534, + "grad_norm": 0.24194808304309845, + "learning_rate": 3.935278876864278e-05, + "loss": 0.0062, + "step": 58990 + }, + { + "epoch": 4.58252427184466, + "grad_norm": 0.32786959409713745, + "learning_rate": 3.933663388688087e-05, + "loss": 0.006, + "step": 59000 + }, + { + "epoch": 4.583300970873786, + "grad_norm": 0.20560501515865326, + "learning_rate": 3.9320480171248645e-05, + "loss": 0.006, + "step": 59010 + }, + { + "epoch": 4.584077669902912, + "grad_norm": 0.13221698999404907, + "learning_rate": 3.930432762351265e-05, + "loss": 0.0061, + "step": 59020 + }, + { + "epoch": 4.584854368932039, + "grad_norm": 0.3345015347003937, + "learning_rate": 3.928817624543931e-05, + "loss": 0.0072, + "step": 59030 + }, + { + "epoch": 4.585631067961165, + "grad_norm": 0.23021043837070465, + "learning_rate": 3.9272026038794885e-05, + "loss": 0.0045, + "step": 59040 + }, + { + "epoch": 4.586407766990291, + "grad_norm": 0.262589693069458, + "learning_rate": 3.9255877005345586e-05, + "loss": 0.0049, + "step": 59050 + }, + { + "epoch": 4.587184466019417, + "grad_norm": 0.20217271149158478, + "learning_rate": 3.923972914685743e-05, + "loss": 0.0053, + "step": 59060 + }, + { + "epoch": 4.587961165048544, + "grad_norm": 0.15004423260688782, + "learning_rate": 3.922358246509631e-05, + "loss": 0.0072, + "step": 59070 + }, + { + "epoch": 4.58873786407767, + "grad_norm": 0.17387887835502625, + "learning_rate": 3.920743696182801e-05, + "loss": 0.0057, + "step": 59080 + }, + { + "epoch": 4.589514563106796, + "grad_norm": 0.2679385840892792, + "learning_rate": 3.9191292638818164e-05, + "loss": 0.0052, + "step": 59090 + }, + { + "epoch": 4.590291262135922, + "grad_norm": 0.24334485828876495, + "learning_rate": 3.9175149497832316e-05, + "loss": 0.0044, + "step": 59100 + }, + { + "epoch": 4.591067961165049, + "grad_norm": 0.29180067777633667, + "learning_rate": 3.915900754063584e-05, + "loss": 0.0077, + "step": 59110 + }, + { + "epoch": 4.591844660194175, + "grad_norm": 0.2324363738298416, + "learning_rate": 3.9142866768994014e-05, + "loss": 0.0049, + "step": 59120 + }, + { + "epoch": 4.592621359223301, + "grad_norm": 0.2639780640602112, + "learning_rate": 3.912672718467193e-05, + "loss": 0.0073, + "step": 59130 + }, + { + "epoch": 4.593398058252427, + "grad_norm": 0.17349942028522491, + "learning_rate": 3.911058878943463e-05, + "loss": 0.0052, + "step": 59140 + }, + { + "epoch": 4.594174757281554, + "grad_norm": 0.14861127734184265, + "learning_rate": 3.9094451585046974e-05, + "loss": 0.0084, + "step": 59150 + }, + { + "epoch": 4.59495145631068, + "grad_norm": 0.2583911120891571, + "learning_rate": 3.90783155732737e-05, + "loss": 0.0059, + "step": 59160 + }, + { + "epoch": 4.595728155339806, + "grad_norm": 0.10557083785533905, + "learning_rate": 3.9062180755879373e-05, + "loss": 0.006, + "step": 59170 + }, + { + "epoch": 4.596504854368932, + "grad_norm": 0.08190827071666718, + "learning_rate": 3.904604713462855e-05, + "loss": 0.0038, + "step": 59180 + }, + { + "epoch": 4.597281553398059, + "grad_norm": 0.4165887236595154, + "learning_rate": 3.902991471128554e-05, + "loss": 0.0055, + "step": 59190 + }, + { + "epoch": 4.598058252427185, + "grad_norm": 0.2536377012729645, + "learning_rate": 3.901378348761456e-05, + "loss": 0.0058, + "step": 59200 + }, + { + "epoch": 4.598834951456311, + "grad_norm": 0.1719643771648407, + "learning_rate": 3.899765346537971e-05, + "loss": 0.0063, + "step": 59210 + }, + { + "epoch": 4.5996116504854365, + "grad_norm": 0.15234802663326263, + "learning_rate": 3.898152464634493e-05, + "loss": 0.0046, + "step": 59220 + }, + { + "epoch": 4.600388349514563, + "grad_norm": 0.4512311518192291, + "learning_rate": 3.896539703227402e-05, + "loss": 0.0063, + "step": 59230 + }, + { + "epoch": 4.601165048543689, + "grad_norm": 0.42665576934814453, + "learning_rate": 3.894927062493073e-05, + "loss": 0.0054, + "step": 59240 + }, + { + "epoch": 4.601941747572815, + "grad_norm": 0.23258273303508759, + "learning_rate": 3.893314542607859e-05, + "loss": 0.0058, + "step": 59250 + }, + { + "epoch": 4.6027184466019415, + "grad_norm": 0.28912270069122314, + "learning_rate": 3.891702143748103e-05, + "loss": 0.0089, + "step": 59260 + }, + { + "epoch": 4.603495145631068, + "grad_norm": 0.42684251070022583, + "learning_rate": 3.890089866090133e-05, + "loss": 0.0055, + "step": 59270 + }, + { + "epoch": 4.604271844660194, + "grad_norm": 0.409095823764801, + "learning_rate": 3.888477709810267e-05, + "loss": 0.0103, + "step": 59280 + }, + { + "epoch": 4.60504854368932, + "grad_norm": 0.0969230905175209, + "learning_rate": 3.886865675084806e-05, + "loss": 0.0046, + "step": 59290 + }, + { + "epoch": 4.6058252427184465, + "grad_norm": 0.6136560440063477, + "learning_rate": 3.885253762090043e-05, + "loss": 0.0058, + "step": 59300 + }, + { + "epoch": 4.606601941747573, + "grad_norm": 0.20771080255508423, + "learning_rate": 3.8836419710022534e-05, + "loss": 0.0059, + "step": 59310 + }, + { + "epoch": 4.607378640776699, + "grad_norm": 0.25881227850914, + "learning_rate": 3.882030301997698e-05, + "loss": 0.0061, + "step": 59320 + }, + { + "epoch": 4.608155339805825, + "grad_norm": 0.24635577201843262, + "learning_rate": 3.880418755252629e-05, + "loss": 0.0044, + "step": 59330 + }, + { + "epoch": 4.6089320388349515, + "grad_norm": 0.21875964105129242, + "learning_rate": 3.87880733094328e-05, + "loss": 0.0058, + "step": 59340 + }, + { + "epoch": 4.609708737864078, + "grad_norm": 0.18619541823863983, + "learning_rate": 3.877196029245877e-05, + "loss": 0.0055, + "step": 59350 + }, + { + "epoch": 4.610485436893204, + "grad_norm": 0.2047184854745865, + "learning_rate": 3.875584850336627e-05, + "loss": 0.0048, + "step": 59360 + }, + { + "epoch": 4.61126213592233, + "grad_norm": 0.1654309183359146, + "learning_rate": 3.873973794391728e-05, + "loss": 0.0084, + "step": 59370 + }, + { + "epoch": 4.6120388349514565, + "grad_norm": 0.15472398698329926, + "learning_rate": 3.872362861587361e-05, + "loss": 0.0043, + "step": 59380 + }, + { + "epoch": 4.612815533980583, + "grad_norm": 0.4333518445491791, + "learning_rate": 3.870752052099698e-05, + "loss": 0.0081, + "step": 59390 + }, + { + "epoch": 4.613592233009709, + "grad_norm": 0.39664769172668457, + "learning_rate": 3.8691413661048916e-05, + "loss": 0.008, + "step": 59400 + }, + { + "epoch": 4.614368932038835, + "grad_norm": 0.20783556997776031, + "learning_rate": 3.867530803779085e-05, + "loss": 0.0058, + "step": 59410 + }, + { + "epoch": 4.6151456310679615, + "grad_norm": 0.2184128761291504, + "learning_rate": 3.8659203652984054e-05, + "loss": 0.0064, + "step": 59420 + }, + { + "epoch": 4.615922330097088, + "grad_norm": 0.2129194140434265, + "learning_rate": 3.864310050838972e-05, + "loss": 0.005, + "step": 59430 + }, + { + "epoch": 4.616699029126213, + "grad_norm": 0.26679977774620056, + "learning_rate": 3.8626998605768835e-05, + "loss": 0.0064, + "step": 59440 + }, + { + "epoch": 4.617475728155339, + "grad_norm": 0.25902339816093445, + "learning_rate": 3.8610897946882294e-05, + "loss": 0.0041, + "step": 59450 + }, + { + "epoch": 4.618252427184466, + "grad_norm": 0.14860177040100098, + "learning_rate": 3.859479853349082e-05, + "loss": 0.0065, + "step": 59460 + }, + { + "epoch": 4.619029126213592, + "grad_norm": 0.25451260805130005, + "learning_rate": 3.857870036735503e-05, + "loss": 0.0095, + "step": 59470 + }, + { + "epoch": 4.619805825242718, + "grad_norm": 0.29733914136886597, + "learning_rate": 3.856260345023538e-05, + "loss": 0.0062, + "step": 59480 + }, + { + "epoch": 4.620582524271844, + "grad_norm": 0.17092381417751312, + "learning_rate": 3.854650778389224e-05, + "loss": 0.0055, + "step": 59490 + }, + { + "epoch": 4.621359223300971, + "grad_norm": 0.33220261335372925, + "learning_rate": 3.8530413370085786e-05, + "loss": 0.0052, + "step": 59500 + }, + { + "epoch": 4.622135922330097, + "grad_norm": 0.12969183921813965, + "learning_rate": 3.851432021057608e-05, + "loss": 0.0058, + "step": 59510 + }, + { + "epoch": 4.622912621359223, + "grad_norm": 0.2021344155073166, + "learning_rate": 3.849822830712306e-05, + "loss": 0.0051, + "step": 59520 + }, + { + "epoch": 4.623689320388349, + "grad_norm": 0.2554301917552948, + "learning_rate": 3.848213766148649e-05, + "loss": 0.0046, + "step": 59530 + }, + { + "epoch": 4.624466019417476, + "grad_norm": 0.3170345723628998, + "learning_rate": 3.846604827542602e-05, + "loss": 0.0066, + "step": 59540 + }, + { + "epoch": 4.625242718446602, + "grad_norm": 0.21563397347927094, + "learning_rate": 3.844996015070117e-05, + "loss": 0.0071, + "step": 59550 + }, + { + "epoch": 4.626019417475728, + "grad_norm": 0.3696484863758087, + "learning_rate": 3.843387328907132e-05, + "loss": 0.005, + "step": 59560 + }, + { + "epoch": 4.626796116504854, + "grad_norm": 0.5516195893287659, + "learning_rate": 3.8417787692295696e-05, + "loss": 0.0056, + "step": 59570 + }, + { + "epoch": 4.627572815533981, + "grad_norm": 0.06664679199457169, + "learning_rate": 3.84017033621334e-05, + "loss": 0.0044, + "step": 59580 + }, + { + "epoch": 4.628349514563107, + "grad_norm": 0.22723358869552612, + "learning_rate": 3.838562030034338e-05, + "loss": 0.006, + "step": 59590 + }, + { + "epoch": 4.629126213592233, + "grad_norm": 0.2586221992969513, + "learning_rate": 3.836953850868447e-05, + "loss": 0.0061, + "step": 59600 + }, + { + "epoch": 4.629902912621359, + "grad_norm": 0.11433076858520508, + "learning_rate": 3.835345798891533e-05, + "loss": 0.0089, + "step": 59610 + }, + { + "epoch": 4.630679611650486, + "grad_norm": 0.26147574186325073, + "learning_rate": 3.833737874279452e-05, + "loss": 0.0044, + "step": 59620 + }, + { + "epoch": 4.631456310679612, + "grad_norm": 0.2104010432958603, + "learning_rate": 3.8321300772080433e-05, + "loss": 0.0054, + "step": 59630 + }, + { + "epoch": 4.632233009708738, + "grad_norm": 0.24601416289806366, + "learning_rate": 3.8305224078531345e-05, + "loss": 0.0082, + "step": 59640 + }, + { + "epoch": 4.633009708737864, + "grad_norm": 0.11951612681150436, + "learning_rate": 3.828914866390537e-05, + "loss": 0.0049, + "step": 59650 + }, + { + "epoch": 4.633786407766991, + "grad_norm": 0.2874595522880554, + "learning_rate": 3.827307452996048e-05, + "loss": 0.0059, + "step": 59660 + }, + { + "epoch": 4.634563106796117, + "grad_norm": 0.2581667900085449, + "learning_rate": 3.825700167845454e-05, + "loss": 0.0061, + "step": 59670 + }, + { + "epoch": 4.635339805825243, + "grad_norm": 0.5825666785240173, + "learning_rate": 3.8240930111145214e-05, + "loss": 0.0074, + "step": 59680 + }, + { + "epoch": 4.636116504854369, + "grad_norm": 0.22065766155719757, + "learning_rate": 3.8224859829790116e-05, + "loss": 0.0066, + "step": 59690 + }, + { + "epoch": 4.636893203883496, + "grad_norm": 0.23353509604930878, + "learning_rate": 3.820879083614664e-05, + "loss": 0.0071, + "step": 59700 + }, + { + "epoch": 4.637669902912622, + "grad_norm": 0.3198179602622986, + "learning_rate": 3.819272313197208e-05, + "loss": 0.0071, + "step": 59710 + }, + { + "epoch": 4.638446601941748, + "grad_norm": 0.49689197540283203, + "learning_rate": 3.817665671902355e-05, + "loss": 0.0068, + "step": 59720 + }, + { + "epoch": 4.639223300970873, + "grad_norm": 0.300439715385437, + "learning_rate": 3.816059159905808e-05, + "loss": 0.0068, + "step": 59730 + }, + { + "epoch": 4.64, + "grad_norm": 0.20142917335033417, + "learning_rate": 3.8144527773832476e-05, + "loss": 0.0058, + "step": 59740 + }, + { + "epoch": 4.640776699029126, + "grad_norm": 0.28598833084106445, + "learning_rate": 3.812846524510352e-05, + "loss": 0.0067, + "step": 59750 + }, + { + "epoch": 4.641553398058252, + "grad_norm": 0.28327852487564087, + "learning_rate": 3.811240401462775e-05, + "loss": 0.0075, + "step": 59760 + }, + { + "epoch": 4.642330097087378, + "grad_norm": 0.248565211892128, + "learning_rate": 3.809634408416162e-05, + "loss": 0.006, + "step": 59770 + }, + { + "epoch": 4.643106796116505, + "grad_norm": 0.4108726382255554, + "learning_rate": 3.8080285455461394e-05, + "loss": 0.0069, + "step": 59780 + }, + { + "epoch": 4.643883495145631, + "grad_norm": 0.21890662610530853, + "learning_rate": 3.806422813028323e-05, + "loss": 0.0053, + "step": 59790 + }, + { + "epoch": 4.644660194174757, + "grad_norm": 0.16893649101257324, + "learning_rate": 3.8048172110383114e-05, + "loss": 0.0041, + "step": 59800 + }, + { + "epoch": 4.645436893203883, + "grad_norm": 0.26460808515548706, + "learning_rate": 3.8032117397516944e-05, + "loss": 0.0062, + "step": 59810 + }, + { + "epoch": 4.64621359223301, + "grad_norm": 0.263801246881485, + "learning_rate": 3.801606399344042e-05, + "loss": 0.0051, + "step": 59820 + }, + { + "epoch": 4.646990291262136, + "grad_norm": 0.43972745537757874, + "learning_rate": 3.800001189990913e-05, + "loss": 0.0069, + "step": 59830 + }, + { + "epoch": 4.647766990291262, + "grad_norm": 0.16686506569385529, + "learning_rate": 3.798396111867847e-05, + "loss": 0.0065, + "step": 59840 + }, + { + "epoch": 4.648543689320388, + "grad_norm": 0.1319521814584732, + "learning_rate": 3.796791165150378e-05, + "loss": 0.0069, + "step": 59850 + }, + { + "epoch": 4.649320388349515, + "grad_norm": 0.2346303015947342, + "learning_rate": 3.795186350014018e-05, + "loss": 0.0058, + "step": 59860 + }, + { + "epoch": 4.650097087378641, + "grad_norm": 0.2042807787656784, + "learning_rate": 3.793581666634266e-05, + "loss": 0.0056, + "step": 59870 + }, + { + "epoch": 4.650873786407767, + "grad_norm": 0.17346031963825226, + "learning_rate": 3.791977115186609e-05, + "loss": 0.0061, + "step": 59880 + }, + { + "epoch": 4.651650485436893, + "grad_norm": 0.2807857394218445, + "learning_rate": 3.79037269584652e-05, + "loss": 0.0084, + "step": 59890 + }, + { + "epoch": 4.65242718446602, + "grad_norm": 0.3367103636264801, + "learning_rate": 3.7887684087894545e-05, + "loss": 0.007, + "step": 59900 + }, + { + "epoch": 4.653203883495146, + "grad_norm": 0.14761893451213837, + "learning_rate": 3.787164254190854e-05, + "loss": 0.0064, + "step": 59910 + }, + { + "epoch": 4.653980582524272, + "grad_norm": 0.20638303458690643, + "learning_rate": 3.7855602322261473e-05, + "loss": 0.0067, + "step": 59920 + }, + { + "epoch": 4.654757281553398, + "grad_norm": 0.29111751914024353, + "learning_rate": 3.783956343070746e-05, + "loss": 0.0056, + "step": 59930 + }, + { + "epoch": 4.655533980582525, + "grad_norm": 0.22199280560016632, + "learning_rate": 3.7823525869000526e-05, + "loss": 0.0052, + "step": 59940 + }, + { + "epoch": 4.65631067961165, + "grad_norm": 0.1094941720366478, + "learning_rate": 3.78074896388945e-05, + "loss": 0.0039, + "step": 59950 + }, + { + "epoch": 4.657087378640776, + "grad_norm": 0.31324297189712524, + "learning_rate": 3.7791454742143075e-05, + "loss": 0.007, + "step": 59960 + }, + { + "epoch": 4.6578640776699025, + "grad_norm": 0.19293276965618134, + "learning_rate": 3.7775421180499795e-05, + "loss": 0.0061, + "step": 59970 + }, + { + "epoch": 4.658640776699029, + "grad_norm": 0.18109804391860962, + "learning_rate": 3.7759388955718086e-05, + "loss": 0.0076, + "step": 59980 + }, + { + "epoch": 4.659417475728155, + "grad_norm": 0.13824518024921417, + "learning_rate": 3.774335806955117e-05, + "loss": 0.0065, + "step": 59990 + }, + { + "epoch": 4.660194174757281, + "grad_norm": 0.20559737086296082, + "learning_rate": 3.7727328523752215e-05, + "loss": 0.0054, + "step": 60000 + }, + { + "epoch": 4.6609708737864075, + "grad_norm": 0.38177183270454407, + "learning_rate": 3.771130032007415e-05, + "loss": 0.0051, + "step": 60010 + }, + { + "epoch": 4.661747572815534, + "grad_norm": 0.3914213478565216, + "learning_rate": 3.769527346026982e-05, + "loss": 0.006, + "step": 60020 + }, + { + "epoch": 4.66252427184466, + "grad_norm": 0.370945543050766, + "learning_rate": 3.767924794609188e-05, + "loss": 0.0071, + "step": 60030 + }, + { + "epoch": 4.663300970873786, + "grad_norm": 0.3958364725112915, + "learning_rate": 3.7663223779292854e-05, + "loss": 0.0072, + "step": 60040 + }, + { + "epoch": 4.6640776699029125, + "grad_norm": 0.14616833627223969, + "learning_rate": 3.764720096162512e-05, + "loss": 0.0057, + "step": 60050 + }, + { + "epoch": 4.664854368932039, + "grad_norm": 0.6583935022354126, + "learning_rate": 3.763117949484092e-05, + "loss": 0.0055, + "step": 60060 + }, + { + "epoch": 4.665631067961165, + "grad_norm": 0.5444225668907166, + "learning_rate": 3.7615159380692344e-05, + "loss": 0.0068, + "step": 60070 + }, + { + "epoch": 4.666407766990291, + "grad_norm": 0.0735616534948349, + "learning_rate": 3.759914062093132e-05, + "loss": 0.0034, + "step": 60080 + }, + { + "epoch": 4.6671844660194175, + "grad_norm": 0.15707118809223175, + "learning_rate": 3.7583123217309615e-05, + "loss": 0.0048, + "step": 60090 + }, + { + "epoch": 4.667961165048544, + "grad_norm": 0.16579607129096985, + "learning_rate": 3.7567107171578904e-05, + "loss": 0.0063, + "step": 60100 + }, + { + "epoch": 4.66873786407767, + "grad_norm": 0.1843990832567215, + "learning_rate": 3.755109248549066e-05, + "loss": 0.0048, + "step": 60110 + }, + { + "epoch": 4.669514563106796, + "grad_norm": 0.1763933300971985, + "learning_rate": 3.75350791607962e-05, + "loss": 0.007, + "step": 60120 + }, + { + "epoch": 4.6702912621359225, + "grad_norm": 0.2576678693294525, + "learning_rate": 3.751906719924676e-05, + "loss": 0.0071, + "step": 60130 + }, + { + "epoch": 4.671067961165049, + "grad_norm": 0.27278098464012146, + "learning_rate": 3.750305660259337e-05, + "loss": 0.0066, + "step": 60140 + }, + { + "epoch": 4.671844660194175, + "grad_norm": 0.10931548476219177, + "learning_rate": 3.748704737258693e-05, + "loss": 0.0054, + "step": 60150 + }, + { + "epoch": 4.672621359223301, + "grad_norm": 0.7947617769241333, + "learning_rate": 3.747103951097816e-05, + "loss": 0.0063, + "step": 60160 + }, + { + "epoch": 4.6733980582524275, + "grad_norm": 0.5688517689704895, + "learning_rate": 3.745503301951768e-05, + "loss": 0.0063, + "step": 60170 + }, + { + "epoch": 4.674174757281554, + "grad_norm": 0.18877427279949188, + "learning_rate": 3.743902789995592e-05, + "loss": 0.0062, + "step": 60180 + }, + { + "epoch": 4.67495145631068, + "grad_norm": 0.1411970555782318, + "learning_rate": 3.7423024154043183e-05, + "loss": 0.0076, + "step": 60190 + }, + { + "epoch": 4.675728155339806, + "grad_norm": 0.2946711480617523, + "learning_rate": 3.740702178352963e-05, + "loss": 0.0084, + "step": 60200 + }, + { + "epoch": 4.6765048543689325, + "grad_norm": 0.2037971466779709, + "learning_rate": 3.739102079016523e-05, + "loss": 0.0067, + "step": 60210 + }, + { + "epoch": 4.677281553398059, + "grad_norm": 0.19453299045562744, + "learning_rate": 3.7375021175699846e-05, + "loss": 0.0073, + "step": 60220 + }, + { + "epoch": 4.678058252427184, + "grad_norm": 0.2074899971485138, + "learning_rate": 3.735902294188317e-05, + "loss": 0.006, + "step": 60230 + }, + { + "epoch": 4.67883495145631, + "grad_norm": 0.2792617976665497, + "learning_rate": 3.734302609046473e-05, + "loss": 0.0049, + "step": 60240 + }, + { + "epoch": 4.679611650485437, + "grad_norm": 0.4225321114063263, + "learning_rate": 3.732703062319391e-05, + "loss": 0.0062, + "step": 60250 + }, + { + "epoch": 4.680388349514563, + "grad_norm": 0.11662029474973679, + "learning_rate": 3.731103654181999e-05, + "loss": 0.0062, + "step": 60260 + }, + { + "epoch": 4.681165048543689, + "grad_norm": 0.5346348881721497, + "learning_rate": 3.729504384809203e-05, + "loss": 0.0068, + "step": 60270 + }, + { + "epoch": 4.681941747572815, + "grad_norm": 0.3742848038673401, + "learning_rate": 3.727905254375898e-05, + "loss": 0.0047, + "step": 60280 + }, + { + "epoch": 4.682718446601942, + "grad_norm": 0.32002201676368713, + "learning_rate": 3.7263062630569614e-05, + "loss": 0.0057, + "step": 60290 + }, + { + "epoch": 4.683495145631068, + "grad_norm": 0.19288600981235504, + "learning_rate": 3.724707411027256e-05, + "loss": 0.005, + "step": 60300 + }, + { + "epoch": 4.684271844660194, + "grad_norm": 0.2208525538444519, + "learning_rate": 3.723108698461631e-05, + "loss": 0.0059, + "step": 60310 + }, + { + "epoch": 4.68504854368932, + "grad_norm": 0.20528362691402435, + "learning_rate": 3.72151012553492e-05, + "loss": 0.0041, + "step": 60320 + }, + { + "epoch": 4.685825242718447, + "grad_norm": 0.09165684133768082, + "learning_rate": 3.719911692421939e-05, + "loss": 0.0053, + "step": 60330 + }, + { + "epoch": 4.686601941747573, + "grad_norm": 0.3207225501537323, + "learning_rate": 3.71831339929749e-05, + "loss": 0.0043, + "step": 60340 + }, + { + "epoch": 4.687378640776699, + "grad_norm": 0.24487215280532837, + "learning_rate": 3.716715246336362e-05, + "loss": 0.0055, + "step": 60350 + }, + { + "epoch": 4.688155339805825, + "grad_norm": 0.15866707265377045, + "learning_rate": 3.715117233713324e-05, + "loss": 0.0042, + "step": 60360 + }, + { + "epoch": 4.688932038834952, + "grad_norm": 0.1458800584077835, + "learning_rate": 3.713519361603135e-05, + "loss": 0.0048, + "step": 60370 + }, + { + "epoch": 4.689708737864078, + "grad_norm": 0.2569110095500946, + "learning_rate": 3.711921630180532e-05, + "loss": 0.0047, + "step": 60380 + }, + { + "epoch": 4.690485436893204, + "grad_norm": 0.47269782423973083, + "learning_rate": 3.710324039620245e-05, + "loss": 0.0058, + "step": 60390 + }, + { + "epoch": 4.69126213592233, + "grad_norm": 0.1427420973777771, + "learning_rate": 3.708726590096982e-05, + "loss": 0.0047, + "step": 60400 + }, + { + "epoch": 4.692038834951457, + "grad_norm": 0.41207173466682434, + "learning_rate": 3.707129281785437e-05, + "loss": 0.007, + "step": 60410 + }, + { + "epoch": 4.692815533980583, + "grad_norm": 0.21802115440368652, + "learning_rate": 3.705532114860291e-05, + "loss": 0.0072, + "step": 60420 + }, + { + "epoch": 4.693592233009709, + "grad_norm": 0.28511685132980347, + "learning_rate": 3.703935089496207e-05, + "loss": 0.0056, + "step": 60430 + }, + { + "epoch": 4.694368932038835, + "grad_norm": 0.24731819331645966, + "learning_rate": 3.702338205867829e-05, + "loss": 0.0045, + "step": 60440 + }, + { + "epoch": 4.695145631067961, + "grad_norm": 0.4903281331062317, + "learning_rate": 3.7007414641497976e-05, + "loss": 0.0043, + "step": 60450 + }, + { + "epoch": 4.695922330097087, + "grad_norm": 0.46231573820114136, + "learning_rate": 3.699144864516726e-05, + "loss": 0.011, + "step": 60460 + }, + { + "epoch": 4.696699029126213, + "grad_norm": 0.23163141310214996, + "learning_rate": 3.6975484071432153e-05, + "loss": 0.0084, + "step": 60470 + }, + { + "epoch": 4.697475728155339, + "grad_norm": 0.3131406307220459, + "learning_rate": 3.695952092203854e-05, + "loss": 0.0056, + "step": 60480 + }, + { + "epoch": 4.698252427184466, + "grad_norm": 0.24751558899879456, + "learning_rate": 3.69435591987321e-05, + "loss": 0.0057, + "step": 60490 + }, + { + "epoch": 4.699029126213592, + "grad_norm": 0.16891610622406006, + "learning_rate": 3.6927598903258374e-05, + "loss": 0.0062, + "step": 60500 + }, + { + "epoch": 4.699805825242718, + "grad_norm": 0.23548078536987305, + "learning_rate": 3.691164003736281e-05, + "loss": 0.0045, + "step": 60510 + }, + { + "epoch": 4.700582524271844, + "grad_norm": 0.26241546869277954, + "learning_rate": 3.689568260279061e-05, + "loss": 0.0036, + "step": 60520 + }, + { + "epoch": 4.701359223300971, + "grad_norm": 0.462697297334671, + "learning_rate": 3.687972660128686e-05, + "loss": 0.0069, + "step": 60530 + }, + { + "epoch": 4.702135922330097, + "grad_norm": 0.20771418511867523, + "learning_rate": 3.686377203459648e-05, + "loss": 0.0081, + "step": 60540 + }, + { + "epoch": 4.702912621359223, + "grad_norm": 0.2933613359928131, + "learning_rate": 3.6847818904464226e-05, + "loss": 0.0058, + "step": 60550 + }, + { + "epoch": 4.703689320388349, + "grad_norm": 0.28792035579681396, + "learning_rate": 3.683186721263474e-05, + "loss": 0.005, + "step": 60560 + }, + { + "epoch": 4.704466019417476, + "grad_norm": 0.19220900535583496, + "learning_rate": 3.6815916960852435e-05, + "loss": 0.008, + "step": 60570 + }, + { + "epoch": 4.705242718446602, + "grad_norm": 0.11984918266534805, + "learning_rate": 3.679996815086165e-05, + "loss": 0.0052, + "step": 60580 + }, + { + "epoch": 4.706019417475728, + "grad_norm": 0.3717349171638489, + "learning_rate": 3.678402078440649e-05, + "loss": 0.0062, + "step": 60590 + }, + { + "epoch": 4.706796116504854, + "grad_norm": 0.3951554000377655, + "learning_rate": 3.676807486323096e-05, + "loss": 0.0053, + "step": 60600 + }, + { + "epoch": 4.707572815533981, + "grad_norm": 0.15517355501651764, + "learning_rate": 3.6752130389078864e-05, + "loss": 0.0047, + "step": 60610 + }, + { + "epoch": 4.708349514563107, + "grad_norm": 0.2787703275680542, + "learning_rate": 3.673618736369388e-05, + "loss": 0.0063, + "step": 60620 + }, + { + "epoch": 4.709126213592233, + "grad_norm": 0.34136414527893066, + "learning_rate": 3.6720245788819474e-05, + "loss": 0.0046, + "step": 60630 + }, + { + "epoch": 4.709902912621359, + "grad_norm": 0.18450748920440674, + "learning_rate": 3.6704305666199044e-05, + "loss": 0.007, + "step": 60640 + }, + { + "epoch": 4.710679611650486, + "grad_norm": 0.3956381380558014, + "learning_rate": 3.668836699757576e-05, + "loss": 0.0063, + "step": 60650 + }, + { + "epoch": 4.711456310679612, + "grad_norm": 0.2647354006767273, + "learning_rate": 3.667242978469265e-05, + "loss": 0.0099, + "step": 60660 + }, + { + "epoch": 4.712233009708738, + "grad_norm": 0.2698780298233032, + "learning_rate": 3.6656494029292575e-05, + "loss": 0.0048, + "step": 60670 + }, + { + "epoch": 4.713009708737864, + "grad_norm": 0.10947618633508682, + "learning_rate": 3.664055973311825e-05, + "loss": 0.0057, + "step": 60680 + }, + { + "epoch": 4.713786407766991, + "grad_norm": 0.4437128007411957, + "learning_rate": 3.662462689791221e-05, + "loss": 0.0057, + "step": 60690 + }, + { + "epoch": 4.714563106796117, + "grad_norm": 0.2838500440120697, + "learning_rate": 3.660869552541689e-05, + "loss": 0.0059, + "step": 60700 + }, + { + "epoch": 4.715339805825243, + "grad_norm": 0.08591734617948532, + "learning_rate": 3.659276561737448e-05, + "loss": 0.004, + "step": 60710 + }, + { + "epoch": 4.716116504854369, + "grad_norm": 0.3497638404369354, + "learning_rate": 3.657683717552708e-05, + "loss": 0.0079, + "step": 60720 + }, + { + "epoch": 4.716893203883496, + "grad_norm": 0.18766623735427856, + "learning_rate": 3.6560910201616574e-05, + "loss": 0.004, + "step": 60730 + }, + { + "epoch": 4.717669902912621, + "grad_norm": 0.3454679846763611, + "learning_rate": 3.654498469738472e-05, + "loss": 0.0058, + "step": 60740 + }, + { + "epoch": 4.718446601941747, + "grad_norm": 0.23635521531105042, + "learning_rate": 3.6529060664573126e-05, + "loss": 0.0043, + "step": 60750 + }, + { + "epoch": 4.7192233009708735, + "grad_norm": 0.14916236698627472, + "learning_rate": 3.6513138104923176e-05, + "loss": 0.0061, + "step": 60760 + }, + { + "epoch": 4.72, + "grad_norm": 0.3546792268753052, + "learning_rate": 3.6497217020176176e-05, + "loss": 0.0051, + "step": 60770 + }, + { + "epoch": 4.720776699029126, + "grad_norm": 0.33819857239723206, + "learning_rate": 3.648129741207323e-05, + "loss": 0.0075, + "step": 60780 + }, + { + "epoch": 4.721553398058252, + "grad_norm": 0.2383793741464615, + "learning_rate": 3.646537928235527e-05, + "loss": 0.0049, + "step": 60790 + }, + { + "epoch": 4.7223300970873785, + "grad_norm": 0.15890179574489594, + "learning_rate": 3.6449462632763075e-05, + "loss": 0.0106, + "step": 60800 + }, + { + "epoch": 4.723106796116505, + "grad_norm": 0.12764976918697357, + "learning_rate": 3.643354746503729e-05, + "loss": 0.0073, + "step": 60810 + }, + { + "epoch": 4.723883495145631, + "grad_norm": 0.12645412981510162, + "learning_rate": 3.6417633780918345e-05, + "loss": 0.0062, + "step": 60820 + }, + { + "epoch": 4.724660194174757, + "grad_norm": 0.1942753940820694, + "learning_rate": 3.640172158214655e-05, + "loss": 0.0059, + "step": 60830 + }, + { + "epoch": 4.7254368932038835, + "grad_norm": 0.21499696373939514, + "learning_rate": 3.6385810870462035e-05, + "loss": 0.0067, + "step": 60840 + }, + { + "epoch": 4.72621359223301, + "grad_norm": 0.46769842505455017, + "learning_rate": 3.636990164760479e-05, + "loss": 0.0069, + "step": 60850 + }, + { + "epoch": 4.726990291262136, + "grad_norm": 0.28340858221054077, + "learning_rate": 3.6353993915314596e-05, + "loss": 0.0043, + "step": 60860 + }, + { + "epoch": 4.727766990291262, + "grad_norm": 0.17570656538009644, + "learning_rate": 3.6338087675331124e-05, + "loss": 0.0067, + "step": 60870 + }, + { + "epoch": 4.7285436893203885, + "grad_norm": 0.30618909001350403, + "learning_rate": 3.632218292939383e-05, + "loss": 0.0044, + "step": 60880 + }, + { + "epoch": 4.729320388349515, + "grad_norm": 0.21447603404521942, + "learning_rate": 3.6306279679242033e-05, + "loss": 0.0089, + "step": 60890 + }, + { + "epoch": 4.730097087378641, + "grad_norm": 0.16684819757938385, + "learning_rate": 3.629037792661492e-05, + "loss": 0.0077, + "step": 60900 + }, + { + "epoch": 4.730873786407767, + "grad_norm": 0.2564035654067993, + "learning_rate": 3.627447767325147e-05, + "loss": 0.0066, + "step": 60910 + }, + { + "epoch": 4.7316504854368935, + "grad_norm": 0.2977530062198639, + "learning_rate": 3.62585789208905e-05, + "loss": 0.0058, + "step": 60920 + }, + { + "epoch": 4.73242718446602, + "grad_norm": 0.136533722281456, + "learning_rate": 3.624268167127068e-05, + "loss": 0.0044, + "step": 60930 + }, + { + "epoch": 4.733203883495146, + "grad_norm": 0.43549492955207825, + "learning_rate": 3.622678592613051e-05, + "loss": 0.0051, + "step": 60940 + }, + { + "epoch": 4.733980582524272, + "grad_norm": 0.2797870934009552, + "learning_rate": 3.6210891687208304e-05, + "loss": 0.0049, + "step": 60950 + }, + { + "epoch": 4.734757281553398, + "grad_norm": 0.31943100690841675, + "learning_rate": 3.619499895624228e-05, + "loss": 0.006, + "step": 60960 + }, + { + "epoch": 4.735533980582524, + "grad_norm": 0.3467440605163574, + "learning_rate": 3.61791077349704e-05, + "loss": 0.0066, + "step": 60970 + }, + { + "epoch": 4.73631067961165, + "grad_norm": 0.35513150691986084, + "learning_rate": 3.616321802513053e-05, + "loss": 0.0053, + "step": 60980 + }, + { + "epoch": 4.737087378640776, + "grad_norm": 0.23075136542320251, + "learning_rate": 3.6147329828460325e-05, + "loss": 0.0057, + "step": 60990 + }, + { + "epoch": 4.737864077669903, + "grad_norm": 0.0925961285829544, + "learning_rate": 3.6131443146697306e-05, + "loss": 0.0042, + "step": 61000 + }, + { + "epoch": 4.738640776699029, + "grad_norm": 0.22376792132854462, + "learning_rate": 3.611555798157879e-05, + "loss": 0.0041, + "step": 61010 + }, + { + "epoch": 4.739417475728155, + "grad_norm": 0.13525988161563873, + "learning_rate": 3.6099674334842e-05, + "loss": 0.0062, + "step": 61020 + }, + { + "epoch": 4.740194174757281, + "grad_norm": 0.12238038331270218, + "learning_rate": 3.6083792208223934e-05, + "loss": 0.0115, + "step": 61030 + }, + { + "epoch": 4.740970873786408, + "grad_norm": 0.11550012230873108, + "learning_rate": 3.606791160346142e-05, + "loss": 0.0037, + "step": 61040 + }, + { + "epoch": 4.741747572815534, + "grad_norm": 0.19024108350276947, + "learning_rate": 3.6052032522291134e-05, + "loss": 0.0061, + "step": 61050 + }, + { + "epoch": 4.74252427184466, + "grad_norm": 0.14554962515830994, + "learning_rate": 3.603615496644962e-05, + "loss": 0.0041, + "step": 61060 + }, + { + "epoch": 4.743300970873786, + "grad_norm": 0.11887984722852707, + "learning_rate": 3.60202789376732e-05, + "loss": 0.0058, + "step": 61070 + }, + { + "epoch": 4.744077669902913, + "grad_norm": 0.28702443838119507, + "learning_rate": 3.6004404437698036e-05, + "loss": 0.0069, + "step": 61080 + }, + { + "epoch": 4.744854368932039, + "grad_norm": 0.25472939014434814, + "learning_rate": 3.5988531468260184e-05, + "loss": 0.0066, + "step": 61090 + }, + { + "epoch": 4.745631067961165, + "grad_norm": 0.11835392564535141, + "learning_rate": 3.597266003109546e-05, + "loss": 0.0051, + "step": 61100 + }, + { + "epoch": 4.746407766990291, + "grad_norm": 0.2376217097043991, + "learning_rate": 3.5956790127939557e-05, + "loss": 0.0052, + "step": 61110 + }, + { + "epoch": 4.747184466019418, + "grad_norm": 0.1861591786146164, + "learning_rate": 3.594092176052796e-05, + "loss": 0.0064, + "step": 61120 + }, + { + "epoch": 4.747961165048544, + "grad_norm": 0.17717242240905762, + "learning_rate": 3.592505493059603e-05, + "loss": 0.0051, + "step": 61130 + }, + { + "epoch": 4.74873786407767, + "grad_norm": 0.1555185168981552, + "learning_rate": 3.5909189639878905e-05, + "loss": 0.0054, + "step": 61140 + }, + { + "epoch": 4.749514563106796, + "grad_norm": 0.13761265575885773, + "learning_rate": 3.5893325890111626e-05, + "loss": 0.004, + "step": 61150 + }, + { + "epoch": 4.750291262135923, + "grad_norm": 0.2548893690109253, + "learning_rate": 3.5877463683029034e-05, + "loss": 0.0062, + "step": 61160 + }, + { + "epoch": 4.751067961165049, + "grad_norm": 0.4121254086494446, + "learning_rate": 3.586160302036578e-05, + "loss": 0.0058, + "step": 61170 + }, + { + "epoch": 4.751844660194175, + "grad_norm": 0.4295835494995117, + "learning_rate": 3.584574390385636e-05, + "loss": 0.0056, + "step": 61180 + }, + { + "epoch": 4.752621359223301, + "grad_norm": 0.25869283080101013, + "learning_rate": 3.582988633523511e-05, + "loss": 0.0055, + "step": 61190 + }, + { + "epoch": 4.753398058252428, + "grad_norm": 0.4135473966598511, + "learning_rate": 3.581403031623616e-05, + "loss": 0.0084, + "step": 61200 + }, + { + "epoch": 4.754174757281554, + "grad_norm": 0.32695886492729187, + "learning_rate": 3.579817584859354e-05, + "loss": 0.0068, + "step": 61210 + }, + { + "epoch": 4.75495145631068, + "grad_norm": 0.21492509543895721, + "learning_rate": 3.578232293404107e-05, + "loss": 0.0046, + "step": 61220 + }, + { + "epoch": 4.755728155339806, + "grad_norm": 0.21146805584430695, + "learning_rate": 3.5766471574312385e-05, + "loss": 0.0065, + "step": 61230 + }, + { + "epoch": 4.756504854368932, + "grad_norm": 0.16027279198169708, + "learning_rate": 3.5750621771140964e-05, + "loss": 0.004, + "step": 61240 + }, + { + "epoch": 4.757281553398058, + "grad_norm": 0.14842845499515533, + "learning_rate": 3.5734773526260115e-05, + "loss": 0.0064, + "step": 61250 + }, + { + "epoch": 4.758058252427184, + "grad_norm": 0.17499373853206635, + "learning_rate": 3.5718926841402996e-05, + "loss": 0.0059, + "step": 61260 + }, + { + "epoch": 4.75883495145631, + "grad_norm": 0.19956748187541962, + "learning_rate": 3.5703081718302555e-05, + "loss": 0.0055, + "step": 61270 + }, + { + "epoch": 4.759611650485437, + "grad_norm": 0.23425345122814178, + "learning_rate": 3.568723815869162e-05, + "loss": 0.0059, + "step": 61280 + }, + { + "epoch": 4.760388349514563, + "grad_norm": 0.10382137447595596, + "learning_rate": 3.5671396164302777e-05, + "loss": 0.0042, + "step": 61290 + }, + { + "epoch": 4.761165048543689, + "grad_norm": 0.23844560980796814, + "learning_rate": 3.565555573686853e-05, + "loss": 0.0066, + "step": 61300 + }, + { + "epoch": 4.761941747572815, + "grad_norm": 0.4363933801651001, + "learning_rate": 3.563971687812113e-05, + "loss": 0.0053, + "step": 61310 + }, + { + "epoch": 4.762718446601942, + "grad_norm": 0.2467333972454071, + "learning_rate": 3.5623879589792706e-05, + "loss": 0.0045, + "step": 61320 + }, + { + "epoch": 4.763495145631068, + "grad_norm": 0.2510601580142975, + "learning_rate": 3.5608043873615185e-05, + "loss": 0.0085, + "step": 61330 + }, + { + "epoch": 4.764271844660194, + "grad_norm": 0.2866288125514984, + "learning_rate": 3.559220973132035e-05, + "loss": 0.006, + "step": 61340 + }, + { + "epoch": 4.76504854368932, + "grad_norm": 0.1620088517665863, + "learning_rate": 3.557637716463981e-05, + "loss": 0.0053, + "step": 61350 + }, + { + "epoch": 4.765825242718447, + "grad_norm": 0.3498842716217041, + "learning_rate": 3.556054617530497e-05, + "loss": 0.0062, + "step": 61360 + }, + { + "epoch": 4.766601941747573, + "grad_norm": 0.4211069345474243, + "learning_rate": 3.5544716765047106e-05, + "loss": 0.0063, + "step": 61370 + }, + { + "epoch": 4.767378640776699, + "grad_norm": 0.2662079334259033, + "learning_rate": 3.5528888935597274e-05, + "loss": 0.0074, + "step": 61380 + }, + { + "epoch": 4.768155339805825, + "grad_norm": 0.37207353115081787, + "learning_rate": 3.551306268868636e-05, + "loss": 0.0059, + "step": 61390 + }, + { + "epoch": 4.768932038834952, + "grad_norm": 0.14747317135334015, + "learning_rate": 3.5497238026045166e-05, + "loss": 0.0049, + "step": 61400 + }, + { + "epoch": 4.769708737864078, + "grad_norm": 0.27875471115112305, + "learning_rate": 3.548141494940421e-05, + "loss": 0.0062, + "step": 61410 + }, + { + "epoch": 4.770485436893204, + "grad_norm": 0.2375178039073944, + "learning_rate": 3.546559346049388e-05, + "loss": 0.0061, + "step": 61420 + }, + { + "epoch": 4.77126213592233, + "grad_norm": 0.3819501996040344, + "learning_rate": 3.544977356104441e-05, + "loss": 0.006, + "step": 61430 + }, + { + "epoch": 4.772038834951457, + "grad_norm": 0.25076407194137573, + "learning_rate": 3.543395525278582e-05, + "loss": 0.0057, + "step": 61440 + }, + { + "epoch": 4.772815533980583, + "grad_norm": 0.13529449701309204, + "learning_rate": 3.5418138537447985e-05, + "loss": 0.0071, + "step": 61450 + }, + { + "epoch": 4.773592233009708, + "grad_norm": 0.39935725927352905, + "learning_rate": 3.5402323416760574e-05, + "loss": 0.0065, + "step": 61460 + }, + { + "epoch": 4.7743689320388345, + "grad_norm": 0.12493962794542313, + "learning_rate": 3.538650989245316e-05, + "loss": 0.0039, + "step": 61470 + }, + { + "epoch": 4.775145631067961, + "grad_norm": 0.21765738725662231, + "learning_rate": 3.537069796625504e-05, + "loss": 0.0062, + "step": 61480 + }, + { + "epoch": 4.775922330097087, + "grad_norm": 0.2861090302467346, + "learning_rate": 3.53548876398954e-05, + "loss": 0.0085, + "step": 61490 + }, + { + "epoch": 4.776699029126213, + "grad_norm": 0.1834803968667984, + "learning_rate": 3.5339078915103214e-05, + "loss": 0.0042, + "step": 61500 + }, + { + "epoch": 4.7774757281553395, + "grad_norm": 0.21410396695137024, + "learning_rate": 3.532327179360733e-05, + "loss": 0.0059, + "step": 61510 + }, + { + "epoch": 4.778252427184466, + "grad_norm": 0.2957595884799957, + "learning_rate": 3.530746627713636e-05, + "loss": 0.0041, + "step": 61520 + }, + { + "epoch": 4.779029126213592, + "grad_norm": 0.25864994525909424, + "learning_rate": 3.5291662367418805e-05, + "loss": 0.0071, + "step": 61530 + }, + { + "epoch": 4.779805825242718, + "grad_norm": 0.12158842384815216, + "learning_rate": 3.527586006618293e-05, + "loss": 0.0047, + "step": 61540 + }, + { + "epoch": 4.7805825242718445, + "grad_norm": 0.18114827573299408, + "learning_rate": 3.526005937515687e-05, + "loss": 0.0076, + "step": 61550 + }, + { + "epoch": 4.781359223300971, + "grad_norm": 0.2295651137828827, + "learning_rate": 3.524426029606856e-05, + "loss": 0.0039, + "step": 61560 + }, + { + "epoch": 4.782135922330097, + "grad_norm": 0.26534944772720337, + "learning_rate": 3.5228462830645746e-05, + "loss": 0.0051, + "step": 61570 + }, + { + "epoch": 4.782912621359223, + "grad_norm": 0.11734126508235931, + "learning_rate": 3.5212666980616036e-05, + "loss": 0.0051, + "step": 61580 + }, + { + "epoch": 4.7836893203883495, + "grad_norm": 0.3283289074897766, + "learning_rate": 3.519687274770681e-05, + "loss": 0.0041, + "step": 61590 + }, + { + "epoch": 4.784466019417476, + "grad_norm": 0.2530145049095154, + "learning_rate": 3.518108013364535e-05, + "loss": 0.0064, + "step": 61600 + }, + { + "epoch": 4.785242718446602, + "grad_norm": 0.22029900550842285, + "learning_rate": 3.516528914015868e-05, + "loss": 0.0054, + "step": 61610 + }, + { + "epoch": 4.786019417475728, + "grad_norm": 0.11704923957586288, + "learning_rate": 3.5149499768973694e-05, + "loss": 0.0064, + "step": 61620 + }, + { + "epoch": 4.7867961165048545, + "grad_norm": 0.4798972010612488, + "learning_rate": 3.513371202181708e-05, + "loss": 0.0052, + "step": 61630 + }, + { + "epoch": 4.787572815533981, + "grad_norm": 0.2407003492116928, + "learning_rate": 3.511792590041537e-05, + "loss": 0.005, + "step": 61640 + }, + { + "epoch": 4.788349514563107, + "grad_norm": 0.32459089159965515, + "learning_rate": 3.510214140649488e-05, + "loss": 0.0062, + "step": 61650 + }, + { + "epoch": 4.789126213592233, + "grad_norm": 0.390509694814682, + "learning_rate": 3.508635854178184e-05, + "loss": 0.0051, + "step": 61660 + }, + { + "epoch": 4.7899029126213595, + "grad_norm": 0.09802159667015076, + "learning_rate": 3.50705773080022e-05, + "loss": 0.0063, + "step": 61670 + }, + { + "epoch": 4.790679611650486, + "grad_norm": 0.9220086932182312, + "learning_rate": 3.505479770688178e-05, + "loss": 0.0051, + "step": 61680 + }, + { + "epoch": 4.791456310679612, + "grad_norm": 0.1227421760559082, + "learning_rate": 3.503901974014622e-05, + "loss": 0.005, + "step": 61690 + }, + { + "epoch": 4.792233009708738, + "grad_norm": 0.15032657980918884, + "learning_rate": 3.502324340952095e-05, + "loss": 0.0073, + "step": 61700 + }, + { + "epoch": 4.7930097087378645, + "grad_norm": 0.34663304686546326, + "learning_rate": 3.500746871673126e-05, + "loss": 0.0065, + "step": 61710 + }, + { + "epoch": 4.793786407766991, + "grad_norm": 0.20654062926769257, + "learning_rate": 3.499169566350227e-05, + "loss": 0.0055, + "step": 61720 + }, + { + "epoch": 4.794563106796117, + "grad_norm": 0.4142490327358246, + "learning_rate": 3.4975924251558876e-05, + "loss": 0.005, + "step": 61730 + }, + { + "epoch": 4.795339805825243, + "grad_norm": 0.20728372037410736, + "learning_rate": 3.496015448262582e-05, + "loss": 0.0039, + "step": 61740 + }, + { + "epoch": 4.796116504854369, + "grad_norm": 0.5041105151176453, + "learning_rate": 3.494438635842764e-05, + "loss": 0.0061, + "step": 61750 + }, + { + "epoch": 4.796893203883495, + "grad_norm": 0.26089540123939514, + "learning_rate": 3.492861988068874e-05, + "loss": 0.0047, + "step": 61760 + }, + { + "epoch": 4.797669902912621, + "grad_norm": 0.2329234480857849, + "learning_rate": 3.491285505113332e-05, + "loss": 0.0066, + "step": 61770 + }, + { + "epoch": 4.798446601941747, + "grad_norm": 0.258517861366272, + "learning_rate": 3.4897091871485376e-05, + "loss": 0.0075, + "step": 61780 + }, + { + "epoch": 4.799223300970874, + "grad_norm": 0.37973618507385254, + "learning_rate": 3.488133034346875e-05, + "loss": 0.0076, + "step": 61790 + }, + { + "epoch": 4.8, + "grad_norm": 0.2826817035675049, + "learning_rate": 3.486557046880713e-05, + "loss": 0.0067, + "step": 61800 + }, + { + "epoch": 4.800776699029126, + "grad_norm": 0.7872323989868164, + "learning_rate": 3.484981224922395e-05, + "loss": 0.0046, + "step": 61810 + }, + { + "epoch": 4.801553398058252, + "grad_norm": 0.14968238770961761, + "learning_rate": 3.4834055686442544e-05, + "loss": 0.0062, + "step": 61820 + }, + { + "epoch": 4.802330097087379, + "grad_norm": 0.2607087194919586, + "learning_rate": 3.4818300782186e-05, + "loss": 0.0053, + "step": 61830 + }, + { + "epoch": 4.803106796116505, + "grad_norm": 0.12012400478124619, + "learning_rate": 3.4802547538177236e-05, + "loss": 0.0046, + "step": 61840 + }, + { + "epoch": 4.803883495145631, + "grad_norm": 0.3595500886440277, + "learning_rate": 3.478679595613905e-05, + "loss": 0.0053, + "step": 61850 + }, + { + "epoch": 4.804660194174757, + "grad_norm": 0.18592621386051178, + "learning_rate": 3.4771046037793975e-05, + "loss": 0.0052, + "step": 61860 + }, + { + "epoch": 4.805436893203884, + "grad_norm": 0.3025428056716919, + "learning_rate": 3.4755297784864416e-05, + "loss": 0.0085, + "step": 61870 + }, + { + "epoch": 4.80621359223301, + "grad_norm": 0.19573980569839478, + "learning_rate": 3.473955119907257e-05, + "loss": 0.0045, + "step": 61880 + }, + { + "epoch": 4.806990291262136, + "grad_norm": 0.2911219894886017, + "learning_rate": 3.472380628214046e-05, + "loss": 0.006, + "step": 61890 + }, + { + "epoch": 4.807766990291262, + "grad_norm": 0.3472442626953125, + "learning_rate": 3.4708063035789906e-05, + "loss": 0.0065, + "step": 61900 + }, + { + "epoch": 4.808543689320389, + "grad_norm": 0.3826909363269806, + "learning_rate": 3.4692321461742614e-05, + "loss": 0.0061, + "step": 61910 + }, + { + "epoch": 4.809320388349515, + "grad_norm": 0.26867517828941345, + "learning_rate": 3.4676581561720023e-05, + "loss": 0.0068, + "step": 61920 + }, + { + "epoch": 4.810097087378641, + "grad_norm": 0.17953810095787048, + "learning_rate": 3.466084333744344e-05, + "loss": 0.0045, + "step": 61930 + }, + { + "epoch": 4.810873786407767, + "grad_norm": 0.3143218457698822, + "learning_rate": 3.4645106790633964e-05, + "loss": 0.0049, + "step": 61940 + }, + { + "epoch": 4.8116504854368936, + "grad_norm": 0.2485594004392624, + "learning_rate": 3.4629371923012526e-05, + "loss": 0.0059, + "step": 61950 + }, + { + "epoch": 4.81242718446602, + "grad_norm": 0.5581115484237671, + "learning_rate": 3.4613638736299856e-05, + "loss": 0.0051, + "step": 61960 + }, + { + "epoch": 4.813203883495145, + "grad_norm": 0.4393523633480072, + "learning_rate": 3.459790723221652e-05, + "loss": 0.004, + "step": 61970 + }, + { + "epoch": 4.813980582524271, + "grad_norm": 0.35791805386543274, + "learning_rate": 3.45821774124829e-05, + "loss": 0.0068, + "step": 61980 + }, + { + "epoch": 4.814757281553398, + "grad_norm": 0.1857270449399948, + "learning_rate": 3.456644927881918e-05, + "loss": 0.0042, + "step": 61990 + }, + { + "epoch": 4.815533980582524, + "grad_norm": 0.19989502429962158, + "learning_rate": 3.4550722832945354e-05, + "loss": 0.0048, + "step": 62000 + }, + { + "epoch": 4.81631067961165, + "grad_norm": 0.2575717568397522, + "learning_rate": 3.453499807658126e-05, + "loss": 0.0056, + "step": 62010 + }, + { + "epoch": 4.817087378640776, + "grad_norm": 0.12215528637170792, + "learning_rate": 3.451927501144653e-05, + "loss": 0.0071, + "step": 62020 + }, + { + "epoch": 4.817864077669903, + "grad_norm": 0.1713676154613495, + "learning_rate": 3.450355363926059e-05, + "loss": 0.0066, + "step": 62030 + }, + { + "epoch": 4.818640776699029, + "grad_norm": 0.2521074712276459, + "learning_rate": 3.448783396174273e-05, + "loss": 0.0062, + "step": 62040 + }, + { + "epoch": 4.819417475728155, + "grad_norm": 0.16008512675762177, + "learning_rate": 3.4472115980612034e-05, + "loss": 0.0064, + "step": 62050 + }, + { + "epoch": 4.820194174757281, + "grad_norm": 0.3664238750934601, + "learning_rate": 3.4456399697587396e-05, + "loss": 0.0055, + "step": 62060 + }, + { + "epoch": 4.820970873786408, + "grad_norm": 0.5508674383163452, + "learning_rate": 3.4440685114387506e-05, + "loss": 0.0058, + "step": 62070 + }, + { + "epoch": 4.821747572815534, + "grad_norm": 0.4398137629032135, + "learning_rate": 3.442497223273091e-05, + "loss": 0.0076, + "step": 62080 + }, + { + "epoch": 4.82252427184466, + "grad_norm": 0.07574682682752609, + "learning_rate": 3.4409261054335926e-05, + "loss": 0.0063, + "step": 62090 + }, + { + "epoch": 4.823300970873786, + "grad_norm": 0.6564043164253235, + "learning_rate": 3.439355158092069e-05, + "loss": 0.0067, + "step": 62100 + }, + { + "epoch": 4.824077669902913, + "grad_norm": 0.18166279792785645, + "learning_rate": 3.437784381420322e-05, + "loss": 0.0071, + "step": 62110 + }, + { + "epoch": 4.824854368932039, + "grad_norm": 0.243688702583313, + "learning_rate": 3.436213775590126e-05, + "loss": 0.0068, + "step": 62120 + }, + { + "epoch": 4.825631067961165, + "grad_norm": 0.09980934113264084, + "learning_rate": 3.4346433407732405e-05, + "loss": 0.0044, + "step": 62130 + }, + { + "epoch": 4.826407766990291, + "grad_norm": 0.1279917061328888, + "learning_rate": 3.433073077141405e-05, + "loss": 0.0058, + "step": 62140 + }, + { + "epoch": 4.827184466019418, + "grad_norm": 0.19497039914131165, + "learning_rate": 3.431502984866343e-05, + "loss": 0.0036, + "step": 62150 + }, + { + "epoch": 4.827961165048544, + "grad_norm": 0.1306454837322235, + "learning_rate": 3.429933064119753e-05, + "loss": 0.0073, + "step": 62160 + }, + { + "epoch": 4.82873786407767, + "grad_norm": 0.1509060114622116, + "learning_rate": 3.4283633150733255e-05, + "loss": 0.0049, + "step": 62170 + }, + { + "epoch": 4.829514563106796, + "grad_norm": 0.5027254819869995, + "learning_rate": 3.4267937378987225e-05, + "loss": 0.0054, + "step": 62180 + }, + { + "epoch": 4.830291262135923, + "grad_norm": 0.2070339173078537, + "learning_rate": 3.425224332767592e-05, + "loss": 0.0045, + "step": 62190 + }, + { + "epoch": 4.831067961165049, + "grad_norm": 0.19939075410366058, + "learning_rate": 3.4236550998515595e-05, + "loss": 0.0078, + "step": 62200 + }, + { + "epoch": 4.831844660194175, + "grad_norm": 0.1173505187034607, + "learning_rate": 3.422086039322235e-05, + "loss": 0.0054, + "step": 62210 + }, + { + "epoch": 4.832621359223301, + "grad_norm": 0.3486510217189789, + "learning_rate": 3.420517151351209e-05, + "loss": 0.0041, + "step": 62220 + }, + { + "epoch": 4.833398058252428, + "grad_norm": 0.742729127407074, + "learning_rate": 3.418948436110053e-05, + "loss": 0.0085, + "step": 62230 + }, + { + "epoch": 4.834174757281554, + "grad_norm": 0.4990170896053314, + "learning_rate": 3.417379893770319e-05, + "loss": 0.0066, + "step": 62240 + }, + { + "epoch": 4.834951456310679, + "grad_norm": 0.21517440676689148, + "learning_rate": 3.41581152450354e-05, + "loss": 0.0049, + "step": 62250 + }, + { + "epoch": 4.8357281553398055, + "grad_norm": 0.29421889781951904, + "learning_rate": 3.414243328481231e-05, + "loss": 0.0064, + "step": 62260 + }, + { + "epoch": 4.836504854368932, + "grad_norm": 0.1603432148694992, + "learning_rate": 3.412675305874888e-05, + "loss": 0.0049, + "step": 62270 + }, + { + "epoch": 4.837281553398058, + "grad_norm": 0.4359583854675293, + "learning_rate": 3.4111074568559876e-05, + "loss": 0.0055, + "step": 62280 + }, + { + "epoch": 4.838058252427184, + "grad_norm": 0.28320422768592834, + "learning_rate": 3.4095397815959846e-05, + "loss": 0.0061, + "step": 62290 + }, + { + "epoch": 4.8388349514563105, + "grad_norm": 0.19968435168266296, + "learning_rate": 3.407972280266322e-05, + "loss": 0.0056, + "step": 62300 + }, + { + "epoch": 4.839611650485437, + "grad_norm": 0.13796411454677582, + "learning_rate": 3.4064049530384166e-05, + "loss": 0.0043, + "step": 62310 + }, + { + "epoch": 4.840388349514563, + "grad_norm": 0.1396304965019226, + "learning_rate": 3.40483780008367e-05, + "loss": 0.0067, + "step": 62320 + }, + { + "epoch": 4.841165048543689, + "grad_norm": 0.15690433979034424, + "learning_rate": 3.4032708215734635e-05, + "loss": 0.0038, + "step": 62330 + }, + { + "epoch": 4.8419417475728155, + "grad_norm": 0.24051551520824432, + "learning_rate": 3.4017040176791584e-05, + "loss": 0.0063, + "step": 62340 + }, + { + "epoch": 4.842718446601942, + "grad_norm": 0.22415785491466522, + "learning_rate": 3.4001373885720976e-05, + "loss": 0.0067, + "step": 62350 + }, + { + "epoch": 4.843495145631068, + "grad_norm": 0.2862996459007263, + "learning_rate": 3.398570934423608e-05, + "loss": 0.0059, + "step": 62360 + }, + { + "epoch": 4.844271844660194, + "grad_norm": 0.15541522204875946, + "learning_rate": 3.397004655404994e-05, + "loss": 0.0054, + "step": 62370 + }, + { + "epoch": 4.8450485436893205, + "grad_norm": 0.2380509078502655, + "learning_rate": 3.395438551687541e-05, + "loss": 0.0078, + "step": 62380 + }, + { + "epoch": 4.845825242718447, + "grad_norm": 0.14429926872253418, + "learning_rate": 3.393872623442515e-05, + "loss": 0.0037, + "step": 62390 + }, + { + "epoch": 4.846601941747573, + "grad_norm": 0.16766321659088135, + "learning_rate": 3.392306870841164e-05, + "loss": 0.0042, + "step": 62400 + }, + { + "epoch": 4.847378640776699, + "grad_norm": 0.452372282743454, + "learning_rate": 3.390741294054715e-05, + "loss": 0.0037, + "step": 62410 + }, + { + "epoch": 4.8481553398058255, + "grad_norm": 0.17104384303092957, + "learning_rate": 3.38917589325438e-05, + "loss": 0.0056, + "step": 62420 + }, + { + "epoch": 4.848932038834952, + "grad_norm": 0.23353898525238037, + "learning_rate": 3.387610668611349e-05, + "loss": 0.0089, + "step": 62430 + }, + { + "epoch": 4.849708737864078, + "grad_norm": 0.20453304052352905, + "learning_rate": 3.38604562029679e-05, + "loss": 0.0061, + "step": 62440 + }, + { + "epoch": 4.850485436893204, + "grad_norm": 0.12025334686040878, + "learning_rate": 3.384480748481855e-05, + "loss": 0.0035, + "step": 62450 + }, + { + "epoch": 4.8512621359223305, + "grad_norm": 0.21133562922477722, + "learning_rate": 3.382916053337676e-05, + "loss": 0.0049, + "step": 62460 + }, + { + "epoch": 4.852038834951456, + "grad_norm": 0.14748500287532806, + "learning_rate": 3.381351535035366e-05, + "loss": 0.0064, + "step": 62470 + }, + { + "epoch": 4.852815533980582, + "grad_norm": 0.17354647815227509, + "learning_rate": 3.379787193746019e-05, + "loss": 0.0071, + "step": 62480 + }, + { + "epoch": 4.853592233009708, + "grad_norm": 0.46218135952949524, + "learning_rate": 3.378223029640709e-05, + "loss": 0.005, + "step": 62490 + }, + { + "epoch": 4.854368932038835, + "grad_norm": 0.12115689367055893, + "learning_rate": 3.376659042890489e-05, + "loss": 0.0095, + "step": 62500 + }, + { + "epoch": 4.855145631067961, + "grad_norm": 0.1353004276752472, + "learning_rate": 3.375095233666397e-05, + "loss": 0.0052, + "step": 62510 + }, + { + "epoch": 4.855922330097087, + "grad_norm": 0.242057204246521, + "learning_rate": 3.373531602139447e-05, + "loss": 0.0051, + "step": 62520 + }, + { + "epoch": 4.856699029126213, + "grad_norm": 0.44980424642562866, + "learning_rate": 3.371968148480635e-05, + "loss": 0.0068, + "step": 62530 + }, + { + "epoch": 4.85747572815534, + "grad_norm": 0.3418298363685608, + "learning_rate": 3.370404872860937e-05, + "loss": 0.0083, + "step": 62540 + }, + { + "epoch": 4.858252427184466, + "grad_norm": 0.2567998766899109, + "learning_rate": 3.3688417754513145e-05, + "loss": 0.0048, + "step": 62550 + }, + { + "epoch": 4.859029126213592, + "grad_norm": 0.3270380198955536, + "learning_rate": 3.367278856422703e-05, + "loss": 0.0066, + "step": 62560 + }, + { + "epoch": 4.859805825242718, + "grad_norm": 0.2575841546058655, + "learning_rate": 3.365716115946021e-05, + "loss": 0.0074, + "step": 62570 + }, + { + "epoch": 4.860582524271845, + "grad_norm": 0.1947934329509735, + "learning_rate": 3.3641535541921666e-05, + "loss": 0.0035, + "step": 62580 + }, + { + "epoch": 4.861359223300971, + "grad_norm": 0.15491320192813873, + "learning_rate": 3.36259117133202e-05, + "loss": 0.0067, + "step": 62590 + }, + { + "epoch": 4.862135922330097, + "grad_norm": 0.38637956976890564, + "learning_rate": 3.36102896753644e-05, + "loss": 0.0066, + "step": 62600 + }, + { + "epoch": 4.862912621359223, + "grad_norm": 0.30005934834480286, + "learning_rate": 3.35946694297627e-05, + "loss": 0.0048, + "step": 62610 + }, + { + "epoch": 4.86368932038835, + "grad_norm": 0.7967183589935303, + "learning_rate": 3.357905097822328e-05, + "loss": 0.0043, + "step": 62620 + }, + { + "epoch": 4.864466019417476, + "grad_norm": 0.3378066420555115, + "learning_rate": 3.356343432245416e-05, + "loss": 0.0071, + "step": 62630 + }, + { + "epoch": 4.865242718446602, + "grad_norm": 0.43987837433815, + "learning_rate": 3.354781946416315e-05, + "loss": 0.0054, + "step": 62640 + }, + { + "epoch": 4.866019417475728, + "grad_norm": 0.2004174143075943, + "learning_rate": 3.3532206405057864e-05, + "loss": 0.0079, + "step": 62650 + }, + { + "epoch": 4.8667961165048546, + "grad_norm": 0.09428676962852478, + "learning_rate": 3.351659514684573e-05, + "loss": 0.0069, + "step": 62660 + }, + { + "epoch": 4.867572815533981, + "grad_norm": 0.618643581867218, + "learning_rate": 3.350098569123395e-05, + "loss": 0.0069, + "step": 62670 + }, + { + "epoch": 4.868349514563107, + "grad_norm": 0.2544431984424591, + "learning_rate": 3.348537803992958e-05, + "loss": 0.007, + "step": 62680 + }, + { + "epoch": 4.869126213592233, + "grad_norm": 0.7295322418212891, + "learning_rate": 3.346977219463945e-05, + "loss": 0.0061, + "step": 62690 + }, + { + "epoch": 4.8699029126213595, + "grad_norm": 0.3453660309314728, + "learning_rate": 3.345416815707017e-05, + "loss": 0.0061, + "step": 62700 + }, + { + "epoch": 4.870679611650486, + "grad_norm": 0.17397664487361908, + "learning_rate": 3.3438565928928185e-05, + "loss": 0.0053, + "step": 62710 + }, + { + "epoch": 4.871456310679612, + "grad_norm": 0.1173095777630806, + "learning_rate": 3.3422965511919736e-05, + "loss": 0.0046, + "step": 62720 + }, + { + "epoch": 4.872233009708738, + "grad_norm": 0.1184968501329422, + "learning_rate": 3.3407366907750846e-05, + "loss": 0.0055, + "step": 62730 + }, + { + "epoch": 4.8730097087378645, + "grad_norm": 0.19490133225917816, + "learning_rate": 3.339177011812737e-05, + "loss": 0.0066, + "step": 62740 + }, + { + "epoch": 4.87378640776699, + "grad_norm": 0.3904924988746643, + "learning_rate": 3.3376175144754935e-05, + "loss": 0.007, + "step": 62750 + }, + { + "epoch": 4.874563106796116, + "grad_norm": 0.10630641132593155, + "learning_rate": 3.3360581989339004e-05, + "loss": 0.0052, + "step": 62760 + }, + { + "epoch": 4.875339805825242, + "grad_norm": 0.09603501856327057, + "learning_rate": 3.3344990653584804e-05, + "loss": 0.0053, + "step": 62770 + }, + { + "epoch": 4.876116504854369, + "grad_norm": 0.3931886851787567, + "learning_rate": 3.332940113919739e-05, + "loss": 0.0049, + "step": 62780 + }, + { + "epoch": 4.876893203883495, + "grad_norm": 0.2237115353345871, + "learning_rate": 3.331381344788161e-05, + "loss": 0.0047, + "step": 62790 + }, + { + "epoch": 4.877669902912621, + "grad_norm": 0.44425737857818604, + "learning_rate": 3.3298227581342076e-05, + "loss": 0.0049, + "step": 62800 + }, + { + "epoch": 4.878446601941747, + "grad_norm": 0.18255716562271118, + "learning_rate": 3.328264354128329e-05, + "loss": 0.0064, + "step": 62810 + }, + { + "epoch": 4.879223300970874, + "grad_norm": 0.3065962791442871, + "learning_rate": 3.3267061329409465e-05, + "loss": 0.005, + "step": 62820 + }, + { + "epoch": 4.88, + "grad_norm": 0.1998215764760971, + "learning_rate": 3.325148094742466e-05, + "loss": 0.0058, + "step": 62830 + }, + { + "epoch": 4.880776699029126, + "grad_norm": 0.1921125203371048, + "learning_rate": 3.323590239703271e-05, + "loss": 0.0063, + "step": 62840 + }, + { + "epoch": 4.881553398058252, + "grad_norm": 0.2578534781932831, + "learning_rate": 3.322032567993728e-05, + "loss": 0.0062, + "step": 62850 + }, + { + "epoch": 4.882330097087379, + "grad_norm": 0.24266138672828674, + "learning_rate": 3.3204750797841774e-05, + "loss": 0.0066, + "step": 62860 + }, + { + "epoch": 4.883106796116505, + "grad_norm": 0.26169681549072266, + "learning_rate": 3.31891777524495e-05, + "loss": 0.0045, + "step": 62870 + }, + { + "epoch": 4.883883495145631, + "grad_norm": 0.18920442461967468, + "learning_rate": 3.317360654546347e-05, + "loss": 0.0045, + "step": 62880 + }, + { + "epoch": 4.884660194174757, + "grad_norm": 0.4043499231338501, + "learning_rate": 3.3158037178586534e-05, + "loss": 0.0069, + "step": 62890 + }, + { + "epoch": 4.885436893203884, + "grad_norm": 0.17860160768032074, + "learning_rate": 3.3142469653521316e-05, + "loss": 0.0038, + "step": 62900 + }, + { + "epoch": 4.88621359223301, + "grad_norm": 0.13480030000209808, + "learning_rate": 3.3126903971970284e-05, + "loss": 0.0066, + "step": 62910 + }, + { + "epoch": 4.886990291262136, + "grad_norm": 0.13524487614631653, + "learning_rate": 3.311134013563564e-05, + "loss": 0.006, + "step": 62920 + }, + { + "epoch": 4.887766990291262, + "grad_norm": 0.28521618247032166, + "learning_rate": 3.309577814621947e-05, + "loss": 0.004, + "step": 62930 + }, + { + "epoch": 4.888543689320389, + "grad_norm": 0.3797505497932434, + "learning_rate": 3.308021800542359e-05, + "loss": 0.0054, + "step": 62940 + }, + { + "epoch": 4.889320388349515, + "grad_norm": 0.29020777344703674, + "learning_rate": 3.306465971494963e-05, + "loss": 0.0056, + "step": 62950 + }, + { + "epoch": 4.890097087378641, + "grad_norm": 0.30850550532341003, + "learning_rate": 3.304910327649901e-05, + "loss": 0.0071, + "step": 62960 + }, + { + "epoch": 4.890873786407767, + "grad_norm": 0.378834992647171, + "learning_rate": 3.303354869177297e-05, + "loss": 0.0069, + "step": 62970 + }, + { + "epoch": 4.891650485436893, + "grad_norm": 0.2340499460697174, + "learning_rate": 3.301799596247255e-05, + "loss": 0.0054, + "step": 62980 + }, + { + "epoch": 4.892427184466019, + "grad_norm": 0.12493182718753815, + "learning_rate": 3.300244509029854e-05, + "loss": 0.0035, + "step": 62990 + }, + { + "epoch": 4.893203883495145, + "grad_norm": 0.4320890009403229, + "learning_rate": 3.2986896076951585e-05, + "loss": 0.0058, + "step": 63000 + }, + { + "epoch": 4.8939805825242715, + "grad_norm": 0.19580568373203278, + "learning_rate": 3.297134892413211e-05, + "loss": 0.0079, + "step": 63010 + }, + { + "epoch": 4.894757281553398, + "grad_norm": 0.2597958743572235, + "learning_rate": 3.295580363354031e-05, + "loss": 0.0039, + "step": 63020 + }, + { + "epoch": 4.895533980582524, + "grad_norm": 0.2972811162471771, + "learning_rate": 3.2940260206876194e-05, + "loss": 0.0042, + "step": 63030 + }, + { + "epoch": 4.89631067961165, + "grad_norm": 0.27599549293518066, + "learning_rate": 3.292471864583958e-05, + "loss": 0.0049, + "step": 63040 + }, + { + "epoch": 4.8970873786407765, + "grad_norm": 0.08189857006072998, + "learning_rate": 3.290917895213004e-05, + "loss": 0.0059, + "step": 63050 + }, + { + "epoch": 4.897864077669903, + "grad_norm": 0.25799626111984253, + "learning_rate": 3.289364112744701e-05, + "loss": 0.0048, + "step": 63060 + }, + { + "epoch": 4.898640776699029, + "grad_norm": 0.09249560534954071, + "learning_rate": 3.287810517348967e-05, + "loss": 0.0041, + "step": 63070 + }, + { + "epoch": 4.899417475728155, + "grad_norm": 0.4364120364189148, + "learning_rate": 3.2862571091957014e-05, + "loss": 0.007, + "step": 63080 + }, + { + "epoch": 4.9001941747572815, + "grad_norm": 0.30737319588661194, + "learning_rate": 3.284703888454781e-05, + "loss": 0.0065, + "step": 63090 + }, + { + "epoch": 4.900970873786408, + "grad_norm": 0.18293723464012146, + "learning_rate": 3.2831508552960646e-05, + "loss": 0.0065, + "step": 63100 + }, + { + "epoch": 4.901747572815534, + "grad_norm": 0.2535199820995331, + "learning_rate": 3.281598009889387e-05, + "loss": 0.0039, + "step": 63110 + }, + { + "epoch": 4.90252427184466, + "grad_norm": 0.7934587001800537, + "learning_rate": 3.28004535240457e-05, + "loss": 0.0053, + "step": 63120 + }, + { + "epoch": 4.9033009708737865, + "grad_norm": 0.5108607411384583, + "learning_rate": 3.278492883011407e-05, + "loss": 0.0061, + "step": 63130 + }, + { + "epoch": 4.904077669902913, + "grad_norm": 0.3616960942745209, + "learning_rate": 3.276940601879675e-05, + "loss": 0.0055, + "step": 63140 + }, + { + "epoch": 4.904854368932039, + "grad_norm": 0.4508245289325714, + "learning_rate": 3.275388509179127e-05, + "loss": 0.0058, + "step": 63150 + }, + { + "epoch": 4.905631067961165, + "grad_norm": 0.4798317551612854, + "learning_rate": 3.2738366050794985e-05, + "loss": 0.0051, + "step": 63160 + }, + { + "epoch": 4.9064077669902915, + "grad_norm": 0.3367137014865875, + "learning_rate": 3.2722848897505034e-05, + "loss": 0.0058, + "step": 63170 + }, + { + "epoch": 4.907184466019418, + "grad_norm": 0.16546012461185455, + "learning_rate": 3.2707333633618356e-05, + "loss": 0.0041, + "step": 63180 + }, + { + "epoch": 4.907961165048544, + "grad_norm": 0.2569309175014496, + "learning_rate": 3.269182026083167e-05, + "loss": 0.0059, + "step": 63190 + }, + { + "epoch": 4.90873786407767, + "grad_norm": 0.2028174251317978, + "learning_rate": 3.2676308780841516e-05, + "loss": 0.0077, + "step": 63200 + }, + { + "epoch": 4.9095145631067965, + "grad_norm": 0.42463305592536926, + "learning_rate": 3.266079919534416e-05, + "loss": 0.0059, + "step": 63210 + }, + { + "epoch": 4.910291262135923, + "grad_norm": 0.27066269516944885, + "learning_rate": 3.264529150603575e-05, + "loss": 0.0075, + "step": 63220 + }, + { + "epoch": 4.911067961165049, + "grad_norm": 0.4441174268722534, + "learning_rate": 3.262978571461217e-05, + "loss": 0.0059, + "step": 63230 + }, + { + "epoch": 4.911844660194175, + "grad_norm": 0.10385548323392868, + "learning_rate": 3.26142818227691e-05, + "loss": 0.0045, + "step": 63240 + }, + { + "epoch": 4.9126213592233015, + "grad_norm": 0.23685529828071594, + "learning_rate": 3.259877983220203e-05, + "loss": 0.006, + "step": 63250 + }, + { + "epoch": 4.913398058252427, + "grad_norm": 0.19598747789859772, + "learning_rate": 3.2583279744606246e-05, + "loss": 0.0053, + "step": 63260 + }, + { + "epoch": 4.914174757281553, + "grad_norm": 0.18812210857868195, + "learning_rate": 3.25677815616768e-05, + "loss": 0.0034, + "step": 63270 + }, + { + "epoch": 4.914951456310679, + "grad_norm": 0.19682252407073975, + "learning_rate": 3.255228528510856e-05, + "loss": 0.0043, + "step": 63280 + }, + { + "epoch": 4.915728155339806, + "grad_norm": 0.16577008366584778, + "learning_rate": 3.253679091659617e-05, + "loss": 0.0048, + "step": 63290 + }, + { + "epoch": 4.916504854368932, + "grad_norm": 0.233177050948143, + "learning_rate": 3.252129845783404e-05, + "loss": 0.004, + "step": 63300 + }, + { + "epoch": 4.917281553398058, + "grad_norm": 0.17843535542488098, + "learning_rate": 3.2505807910516456e-05, + "loss": 0.0081, + "step": 63310 + }, + { + "epoch": 4.918058252427184, + "grad_norm": 0.11894863843917847, + "learning_rate": 3.249031927633742e-05, + "loss": 0.0063, + "step": 63320 + }, + { + "epoch": 4.9188349514563106, + "grad_norm": 0.14702166616916656, + "learning_rate": 3.2474832556990737e-05, + "loss": 0.0042, + "step": 63330 + }, + { + "epoch": 4.919611650485437, + "grad_norm": 0.2464374303817749, + "learning_rate": 3.245934775417002e-05, + "loss": 0.0047, + "step": 63340 + }, + { + "epoch": 4.920388349514563, + "grad_norm": 0.42194658517837524, + "learning_rate": 3.244386486956866e-05, + "loss": 0.0049, + "step": 63350 + }, + { + "epoch": 4.921165048543689, + "grad_norm": 0.26403510570526123, + "learning_rate": 3.2428383904879846e-05, + "loss": 0.0042, + "step": 63360 + }, + { + "epoch": 4.9219417475728156, + "grad_norm": 0.1184377372264862, + "learning_rate": 3.241290486179652e-05, + "loss": 0.0053, + "step": 63370 + }, + { + "epoch": 4.922718446601942, + "grad_norm": 0.2802605926990509, + "learning_rate": 3.239742774201151e-05, + "loss": 0.0073, + "step": 63380 + }, + { + "epoch": 4.923495145631068, + "grad_norm": 0.14829280972480774, + "learning_rate": 3.238195254721733e-05, + "loss": 0.0054, + "step": 63390 + }, + { + "epoch": 4.924271844660194, + "grad_norm": 0.22057810425758362, + "learning_rate": 3.236647927910633e-05, + "loss": 0.0067, + "step": 63400 + }, + { + "epoch": 4.9250485436893205, + "grad_norm": 0.43601298332214355, + "learning_rate": 3.2351007939370646e-05, + "loss": 0.0074, + "step": 63410 + }, + { + "epoch": 4.925825242718447, + "grad_norm": 0.6369355320930481, + "learning_rate": 3.2335538529702184e-05, + "loss": 0.0073, + "step": 63420 + }, + { + "epoch": 4.926601941747573, + "grad_norm": 0.20578664541244507, + "learning_rate": 3.232007105179268e-05, + "loss": 0.0066, + "step": 63430 + }, + { + "epoch": 4.927378640776699, + "grad_norm": 0.14850647747516632, + "learning_rate": 3.230460550733363e-05, + "loss": 0.0048, + "step": 63440 + }, + { + "epoch": 4.9281553398058255, + "grad_norm": 0.31648316979408264, + "learning_rate": 3.2289141898016315e-05, + "loss": 0.0047, + "step": 63450 + }, + { + "epoch": 4.928932038834952, + "grad_norm": 0.13315650820732117, + "learning_rate": 3.227368022553181e-05, + "loss": 0.003, + "step": 63460 + }, + { + "epoch": 4.929708737864078, + "grad_norm": 0.28942111134529114, + "learning_rate": 3.225822049157099e-05, + "loss": 0.0056, + "step": 63470 + }, + { + "epoch": 4.930485436893203, + "grad_norm": 0.21154476702213287, + "learning_rate": 3.2242762697824494e-05, + "loss": 0.0082, + "step": 63480 + }, + { + "epoch": 4.93126213592233, + "grad_norm": 0.35923880338668823, + "learning_rate": 3.222730684598279e-05, + "loss": 0.0052, + "step": 63490 + }, + { + "epoch": 4.932038834951456, + "grad_norm": 0.23190715909004211, + "learning_rate": 3.221185293773605e-05, + "loss": 0.0063, + "step": 63500 + }, + { + "epoch": 4.932815533980582, + "grad_norm": 0.06036071479320526, + "learning_rate": 3.2196400974774354e-05, + "loss": 0.0037, + "step": 63510 + }, + { + "epoch": 4.933592233009708, + "grad_norm": 0.19595612585544586, + "learning_rate": 3.2180950958787485e-05, + "loss": 0.0047, + "step": 63520 + }, + { + "epoch": 4.934368932038835, + "grad_norm": 0.45107147097587585, + "learning_rate": 3.2165502891465024e-05, + "loss": 0.0059, + "step": 63530 + }, + { + "epoch": 4.935145631067961, + "grad_norm": 0.26443371176719666, + "learning_rate": 3.215005677449636e-05, + "loss": 0.0046, + "step": 63540 + }, + { + "epoch": 4.935922330097087, + "grad_norm": 0.14490462839603424, + "learning_rate": 3.2134612609570644e-05, + "loss": 0.0063, + "step": 63550 + }, + { + "epoch": 4.936699029126213, + "grad_norm": 0.7915794253349304, + "learning_rate": 3.211917039837682e-05, + "loss": 0.0065, + "step": 63560 + }, + { + "epoch": 4.93747572815534, + "grad_norm": 0.17566369473934174, + "learning_rate": 3.210373014260365e-05, + "loss": 0.0062, + "step": 63570 + }, + { + "epoch": 4.938252427184466, + "grad_norm": 0.7490570545196533, + "learning_rate": 3.208829184393965e-05, + "loss": 0.0054, + "step": 63580 + }, + { + "epoch": 4.939029126213592, + "grad_norm": 0.1432313621044159, + "learning_rate": 3.207285550407312e-05, + "loss": 0.0052, + "step": 63590 + }, + { + "epoch": 4.939805825242718, + "grad_norm": 0.21996985375881195, + "learning_rate": 3.205742112469217e-05, + "loss": 0.0054, + "step": 63600 + }, + { + "epoch": 4.940582524271845, + "grad_norm": 0.24153199791908264, + "learning_rate": 3.204198870748465e-05, + "loss": 0.0043, + "step": 63610 + }, + { + "epoch": 4.941359223300971, + "grad_norm": 0.2083362191915512, + "learning_rate": 3.202655825413824e-05, + "loss": 0.0062, + "step": 63620 + }, + { + "epoch": 4.942135922330097, + "grad_norm": 0.40648430585861206, + "learning_rate": 3.2011129766340416e-05, + "loss": 0.0052, + "step": 63630 + }, + { + "epoch": 4.942912621359223, + "grad_norm": 0.1989009529352188, + "learning_rate": 3.1995703245778384e-05, + "loss": 0.0051, + "step": 63640 + }, + { + "epoch": 4.94368932038835, + "grad_norm": 0.31978416442871094, + "learning_rate": 3.1980278694139186e-05, + "loss": 0.0074, + "step": 63650 + }, + { + "epoch": 4.944466019417476, + "grad_norm": 0.26401686668395996, + "learning_rate": 3.196485611310962e-05, + "loss": 0.0062, + "step": 63660 + }, + { + "epoch": 4.945242718446602, + "grad_norm": 0.17651785910129547, + "learning_rate": 3.1949435504376257e-05, + "loss": 0.006, + "step": 63670 + }, + { + "epoch": 4.946019417475728, + "grad_norm": 0.1679908037185669, + "learning_rate": 3.1934016869625505e-05, + "loss": 0.0053, + "step": 63680 + }, + { + "epoch": 4.946796116504855, + "grad_norm": 0.11452740430831909, + "learning_rate": 3.191860021054349e-05, + "loss": 0.0063, + "step": 63690 + }, + { + "epoch": 4.947572815533981, + "grad_norm": 0.2801971137523651, + "learning_rate": 3.190318552881618e-05, + "loss": 0.0062, + "step": 63700 + }, + { + "epoch": 4.948349514563107, + "grad_norm": 0.31639131903648376, + "learning_rate": 3.1887772826129285e-05, + "loss": 0.0065, + "step": 63710 + }, + { + "epoch": 4.949126213592233, + "grad_norm": 0.227729931473732, + "learning_rate": 3.1872362104168335e-05, + "loss": 0.0075, + "step": 63720 + }, + { + "epoch": 4.94990291262136, + "grad_norm": 0.22526216506958008, + "learning_rate": 3.185695336461861e-05, + "loss": 0.0045, + "step": 63730 + }, + { + "epoch": 4.950679611650486, + "grad_norm": 0.5265501141548157, + "learning_rate": 3.184154660916519e-05, + "loss": 0.0053, + "step": 63740 + }, + { + "epoch": 4.951456310679612, + "grad_norm": 0.14697995781898499, + "learning_rate": 3.1826141839492906e-05, + "loss": 0.0048, + "step": 63750 + }, + { + "epoch": 4.9522330097087375, + "grad_norm": 0.19773556292057037, + "learning_rate": 3.181073905728644e-05, + "loss": 0.0049, + "step": 63760 + }, + { + "epoch": 4.953009708737864, + "grad_norm": 0.32772162556648254, + "learning_rate": 3.1795338264230214e-05, + "loss": 0.0048, + "step": 63770 + }, + { + "epoch": 4.95378640776699, + "grad_norm": 0.1822279691696167, + "learning_rate": 3.1779939462008426e-05, + "loss": 0.006, + "step": 63780 + }, + { + "epoch": 4.954563106796116, + "grad_norm": 0.1446007639169693, + "learning_rate": 3.1764542652305066e-05, + "loss": 0.0044, + "step": 63790 + }, + { + "epoch": 4.9553398058252425, + "grad_norm": 0.18643008172512054, + "learning_rate": 3.17491478368039e-05, + "loss": 0.0048, + "step": 63800 + }, + { + "epoch": 4.956116504854369, + "grad_norm": 0.3085951805114746, + "learning_rate": 3.173375501718847e-05, + "loss": 0.0079, + "step": 63810 + }, + { + "epoch": 4.956893203883495, + "grad_norm": 0.4855487644672394, + "learning_rate": 3.171836419514215e-05, + "loss": 0.0051, + "step": 63820 + }, + { + "epoch": 4.957669902912621, + "grad_norm": 0.42977944016456604, + "learning_rate": 3.170297537234803e-05, + "loss": 0.0048, + "step": 63830 + }, + { + "epoch": 4.9584466019417475, + "grad_norm": 0.18650923669338226, + "learning_rate": 3.168758855048901e-05, + "loss": 0.0045, + "step": 63840 + }, + { + "epoch": 4.959223300970874, + "grad_norm": 0.19638419151306152, + "learning_rate": 3.1672203731247785e-05, + "loss": 0.0061, + "step": 63850 + }, + { + "epoch": 4.96, + "grad_norm": 0.3807399570941925, + "learning_rate": 3.16568209163068e-05, + "loss": 0.0067, + "step": 63860 + }, + { + "epoch": 4.960776699029126, + "grad_norm": 0.1547357141971588, + "learning_rate": 3.1641440107348295e-05, + "loss": 0.0055, + "step": 63870 + }, + { + "epoch": 4.9615533980582525, + "grad_norm": 0.18012021481990814, + "learning_rate": 3.162606130605429e-05, + "loss": 0.0059, + "step": 63880 + }, + { + "epoch": 4.962330097087379, + "grad_norm": 0.39961501955986023, + "learning_rate": 3.16106845141066e-05, + "loss": 0.0051, + "step": 63890 + }, + { + "epoch": 4.963106796116505, + "grad_norm": 0.34630346298217773, + "learning_rate": 3.159530973318682e-05, + "loss": 0.007, + "step": 63900 + }, + { + "epoch": 4.963883495145631, + "grad_norm": 0.1810074746608734, + "learning_rate": 3.1579936964976286e-05, + "loss": 0.005, + "step": 63910 + }, + { + "epoch": 4.9646601941747575, + "grad_norm": 0.5504099726676941, + "learning_rate": 3.156456621115615e-05, + "loss": 0.0057, + "step": 63920 + }, + { + "epoch": 4.965436893203884, + "grad_norm": 0.19915272295475006, + "learning_rate": 3.154919747340734e-05, + "loss": 0.0051, + "step": 63930 + }, + { + "epoch": 4.96621359223301, + "grad_norm": 0.21091727912425995, + "learning_rate": 3.153383075341054e-05, + "loss": 0.0077, + "step": 63940 + }, + { + "epoch": 4.966990291262136, + "grad_norm": 0.2599128484725952, + "learning_rate": 3.1518466052846264e-05, + "loss": 0.0057, + "step": 63950 + }, + { + "epoch": 4.9677669902912625, + "grad_norm": 0.11512871831655502, + "learning_rate": 3.150310337339474e-05, + "loss": 0.0051, + "step": 63960 + }, + { + "epoch": 4.968543689320389, + "grad_norm": 0.2483569085597992, + "learning_rate": 3.148774271673603e-05, + "loss": 0.0078, + "step": 63970 + }, + { + "epoch": 4.969320388349515, + "grad_norm": 0.15591181814670563, + "learning_rate": 3.147238408454995e-05, + "loss": 0.0055, + "step": 63980 + }, + { + "epoch": 4.97009708737864, + "grad_norm": 0.21231870353221893, + "learning_rate": 3.1457027478516086e-05, + "loss": 0.0053, + "step": 63990 + }, + { + "epoch": 4.970873786407767, + "grad_norm": 0.24124379456043243, + "learning_rate": 3.144167290031382e-05, + "loss": 0.005, + "step": 64000 + }, + { + "epoch": 4.971650485436893, + "grad_norm": 0.13232144713401794, + "learning_rate": 3.142632035162227e-05, + "loss": 0.0076, + "step": 64010 + }, + { + "epoch": 4.972427184466019, + "grad_norm": 0.31446585059165955, + "learning_rate": 3.141096983412044e-05, + "loss": 0.005, + "step": 64020 + }, + { + "epoch": 4.973203883495145, + "grad_norm": 0.13684038817882538, + "learning_rate": 3.139562134948699e-05, + "loss": 0.0038, + "step": 64030 + }, + { + "epoch": 4.9739805825242716, + "grad_norm": 0.2787514328956604, + "learning_rate": 3.1380274899400424e-05, + "loss": 0.0067, + "step": 64040 + }, + { + "epoch": 4.974757281553398, + "grad_norm": 0.237760528922081, + "learning_rate": 3.1364930485539e-05, + "loss": 0.0071, + "step": 64050 + }, + { + "epoch": 4.975533980582524, + "grad_norm": 0.2057947963476181, + "learning_rate": 3.134958810958076e-05, + "loss": 0.006, + "step": 64060 + }, + { + "epoch": 4.97631067961165, + "grad_norm": 0.19070470333099365, + "learning_rate": 3.13342477732035e-05, + "loss": 0.0072, + "step": 64070 + }, + { + "epoch": 4.9770873786407765, + "grad_norm": 0.1751585155725479, + "learning_rate": 3.131890947808487e-05, + "loss": 0.0055, + "step": 64080 + }, + { + "epoch": 4.977864077669903, + "grad_norm": 0.22586266696453094, + "learning_rate": 3.13035732259022e-05, + "loss": 0.0058, + "step": 64090 + }, + { + "epoch": 4.978640776699029, + "grad_norm": 0.3249288499355316, + "learning_rate": 3.1288239018332656e-05, + "loss": 0.0046, + "step": 64100 + }, + { + "epoch": 4.979417475728155, + "grad_norm": 0.21274715662002563, + "learning_rate": 3.1272906857053164e-05, + "loss": 0.006, + "step": 64110 + }, + { + "epoch": 4.9801941747572815, + "grad_norm": 0.2566611170768738, + "learning_rate": 3.125757674374042e-05, + "loss": 0.0041, + "step": 64120 + }, + { + "epoch": 4.980970873786408, + "grad_norm": 0.26303812861442566, + "learning_rate": 3.124224868007088e-05, + "loss": 0.0052, + "step": 64130 + }, + { + "epoch": 4.981747572815534, + "grad_norm": 0.14801762998104095, + "learning_rate": 3.1226922667720854e-05, + "loss": 0.005, + "step": 64140 + }, + { + "epoch": 4.98252427184466, + "grad_norm": 0.2635174095630646, + "learning_rate": 3.121159870836633e-05, + "loss": 0.0054, + "step": 64150 + }, + { + "epoch": 4.9833009708737865, + "grad_norm": 0.16019101440906525, + "learning_rate": 3.119627680368312e-05, + "loss": 0.0055, + "step": 64160 + }, + { + "epoch": 4.984077669902913, + "grad_norm": 0.3345945179462433, + "learning_rate": 3.1180956955346816e-05, + "loss": 0.007, + "step": 64170 + }, + { + "epoch": 4.984854368932039, + "grad_norm": 0.5580815672874451, + "learning_rate": 3.116563916503276e-05, + "loss": 0.0063, + "step": 64180 + }, + { + "epoch": 4.985631067961165, + "grad_norm": 1.2338165044784546, + "learning_rate": 3.1150323434416095e-05, + "loss": 0.0054, + "step": 64190 + }, + { + "epoch": 4.9864077669902915, + "grad_norm": 0.3178693950176239, + "learning_rate": 3.1135009765171696e-05, + "loss": 0.0093, + "step": 64200 + }, + { + "epoch": 4.987184466019418, + "grad_norm": 0.20883585512638092, + "learning_rate": 3.111969815897429e-05, + "loss": 0.0051, + "step": 64210 + }, + { + "epoch": 4.987961165048544, + "grad_norm": 0.3521007299423218, + "learning_rate": 3.1104388617498295e-05, + "loss": 0.0077, + "step": 64220 + }, + { + "epoch": 4.98873786407767, + "grad_norm": 0.29686474800109863, + "learning_rate": 3.1089081142417966e-05, + "loss": 0.0062, + "step": 64230 + }, + { + "epoch": 4.9895145631067965, + "grad_norm": 0.3613947927951813, + "learning_rate": 3.1073775735407285e-05, + "loss": 0.006, + "step": 64240 + }, + { + "epoch": 4.990291262135923, + "grad_norm": 0.12121005356311798, + "learning_rate": 3.1058472398140035e-05, + "loss": 0.0074, + "step": 64250 + }, + { + "epoch": 4.991067961165049, + "grad_norm": 0.16554802656173706, + "learning_rate": 3.1043171132289736e-05, + "loss": 0.0053, + "step": 64260 + }, + { + "epoch": 4.991844660194174, + "grad_norm": 0.2425207644701004, + "learning_rate": 3.1027871939529774e-05, + "loss": 0.0049, + "step": 64270 + }, + { + "epoch": 4.992621359223301, + "grad_norm": 0.15645010769367218, + "learning_rate": 3.1012574821533214e-05, + "loss": 0.0046, + "step": 64280 + }, + { + "epoch": 4.993398058252427, + "grad_norm": 0.964953601360321, + "learning_rate": 3.099727977997291e-05, + "loss": 0.0066, + "step": 64290 + }, + { + "epoch": 4.994174757281553, + "grad_norm": 0.24783378839492798, + "learning_rate": 3.098198681652154e-05, + "loss": 0.0053, + "step": 64300 + }, + { + "epoch": 4.994951456310679, + "grad_norm": 0.2674401104450226, + "learning_rate": 3.0966695932851475e-05, + "loss": 0.0036, + "step": 64310 + }, + { + "epoch": 4.995728155339806, + "grad_norm": 0.26130762696266174, + "learning_rate": 3.0951407130634926e-05, + "loss": 0.0074, + "step": 64320 + }, + { + "epoch": 4.996504854368932, + "grad_norm": 0.12423847615718842, + "learning_rate": 3.0936120411543864e-05, + "loss": 0.0056, + "step": 64330 + }, + { + "epoch": 4.997281553398058, + "grad_norm": 0.2628175914287567, + "learning_rate": 3.0920835777250015e-05, + "loss": 0.0054, + "step": 64340 + }, + { + "epoch": 4.998058252427184, + "grad_norm": 0.2629261910915375, + "learning_rate": 3.090555322942488e-05, + "loss": 0.0069, + "step": 64350 + }, + { + "epoch": 4.998834951456311, + "grad_norm": 0.22335219383239746, + "learning_rate": 3.089027276973973e-05, + "loss": 0.0058, + "step": 64360 + }, + { + "epoch": 4.999611650485437, + "grad_norm": 0.1629759818315506, + "learning_rate": 3.0874994399865615e-05, + "loss": 0.0052, + "step": 64370 + }, + { + "epoch": 5.000388349514563, + "grad_norm": 0.21477241814136505, + "learning_rate": 3.085971812147337e-05, + "loss": 0.0053, + "step": 64380 + }, + { + "epoch": 5.001165048543689, + "grad_norm": 0.12476144731044769, + "learning_rate": 3.0844443936233557e-05, + "loss": 0.0038, + "step": 64390 + }, + { + "epoch": 5.001941747572816, + "grad_norm": 0.11170660704374313, + "learning_rate": 3.082917184581657e-05, + "loss": 0.0043, + "step": 64400 + }, + { + "epoch": 5.002718446601942, + "grad_norm": 0.4502290189266205, + "learning_rate": 3.0813901851892513e-05, + "loss": 0.0057, + "step": 64410 + }, + { + "epoch": 5.003495145631068, + "grad_norm": 0.14306746423244476, + "learning_rate": 3.0798633956131316e-05, + "loss": 0.0042, + "step": 64420 + }, + { + "epoch": 5.004271844660194, + "grad_norm": 0.16491682827472687, + "learning_rate": 3.078336816020264e-05, + "loss": 0.0078, + "step": 64430 + }, + { + "epoch": 5.005048543689321, + "grad_norm": 0.11078979074954987, + "learning_rate": 3.0768104465775935e-05, + "loss": 0.0041, + "step": 64440 + }, + { + "epoch": 5.005825242718447, + "grad_norm": 0.15433178842067719, + "learning_rate": 3.0752842874520396e-05, + "loss": 0.006, + "step": 64450 + }, + { + "epoch": 5.006601941747573, + "grad_norm": 0.4007244408130646, + "learning_rate": 3.073758338810503e-05, + "loss": 0.0073, + "step": 64460 + }, + { + "epoch": 5.007378640776699, + "grad_norm": 0.31968533992767334, + "learning_rate": 3.072232600819859e-05, + "loss": 0.0053, + "step": 64470 + }, + { + "epoch": 5.008155339805826, + "grad_norm": 0.385170578956604, + "learning_rate": 3.070707073646959e-05, + "loss": 0.0049, + "step": 64480 + }, + { + "epoch": 5.008932038834952, + "grad_norm": 0.22228048741817474, + "learning_rate": 3.069181757458633e-05, + "loss": 0.0049, + "step": 64490 + }, + { + "epoch": 5.009708737864078, + "grad_norm": 0.2567038834095001, + "learning_rate": 3.067656652421687e-05, + "loss": 0.0053, + "step": 64500 + }, + { + "epoch": 5.0104854368932035, + "grad_norm": 0.10387459397315979, + "learning_rate": 3.0661317587029025e-05, + "loss": 0.004, + "step": 64510 + }, + { + "epoch": 5.01126213592233, + "grad_norm": 0.1987447589635849, + "learning_rate": 3.064607076469043e-05, + "loss": 0.0056, + "step": 64520 + }, + { + "epoch": 5.012038834951456, + "grad_norm": 0.20151835680007935, + "learning_rate": 3.063082605886845e-05, + "loss": 0.0038, + "step": 64530 + }, + { + "epoch": 5.012815533980582, + "grad_norm": 0.09015870094299316, + "learning_rate": 3.061558347123019e-05, + "loss": 0.0054, + "step": 64540 + }, + { + "epoch": 5.0135922330097085, + "grad_norm": 0.18577007949352264, + "learning_rate": 3.060034300344259e-05, + "loss": 0.0049, + "step": 64550 + }, + { + "epoch": 5.014368932038835, + "grad_norm": 0.27403637766838074, + "learning_rate": 3.058510465717231e-05, + "loss": 0.0049, + "step": 64560 + }, + { + "epoch": 5.015145631067961, + "grad_norm": 0.29880502820014954, + "learning_rate": 3.0569868434085804e-05, + "loss": 0.0081, + "step": 64570 + }, + { + "epoch": 5.015922330097087, + "grad_norm": 0.2924043834209442, + "learning_rate": 3.055463433584924e-05, + "loss": 0.0052, + "step": 64580 + }, + { + "epoch": 5.0166990291262135, + "grad_norm": 0.3595779836177826, + "learning_rate": 3.0539402364128656e-05, + "loss": 0.0059, + "step": 64590 + }, + { + "epoch": 5.01747572815534, + "grad_norm": 0.18437708914279938, + "learning_rate": 3.052417252058977e-05, + "loss": 0.0064, + "step": 64600 + }, + { + "epoch": 5.018252427184466, + "grad_norm": 0.34633466601371765, + "learning_rate": 3.0508944806898092e-05, + "loss": 0.0086, + "step": 64610 + }, + { + "epoch": 5.019029126213592, + "grad_norm": 0.10156489908695221, + "learning_rate": 3.04937192247189e-05, + "loss": 0.0052, + "step": 64620 + }, + { + "epoch": 5.0198058252427185, + "grad_norm": 0.26933082938194275, + "learning_rate": 3.047849577571725e-05, + "loss": 0.0048, + "step": 64630 + }, + { + "epoch": 5.020582524271845, + "grad_norm": 0.2084563821554184, + "learning_rate": 3.046327446155795e-05, + "loss": 0.0091, + "step": 64640 + }, + { + "epoch": 5.021359223300971, + "grad_norm": 0.13767901062965393, + "learning_rate": 3.0448055283905587e-05, + "loss": 0.0069, + "step": 64650 + }, + { + "epoch": 5.022135922330097, + "grad_norm": 0.20885950326919556, + "learning_rate": 3.0432838244424494e-05, + "loss": 0.0054, + "step": 64660 + }, + { + "epoch": 5.0229126213592235, + "grad_norm": 0.15098120272159576, + "learning_rate": 3.04176233447788e-05, + "loss": 0.0061, + "step": 64670 + }, + { + "epoch": 5.02368932038835, + "grad_norm": 0.17055954039096832, + "learning_rate": 3.040241058663238e-05, + "loss": 0.0053, + "step": 64680 + }, + { + "epoch": 5.024466019417476, + "grad_norm": 0.6189903616905212, + "learning_rate": 3.038719997164887e-05, + "loss": 0.0062, + "step": 64690 + }, + { + "epoch": 5.025242718446602, + "grad_norm": 0.2502208948135376, + "learning_rate": 3.0371991501491683e-05, + "loss": 0.0054, + "step": 64700 + }, + { + "epoch": 5.0260194174757284, + "grad_norm": 0.11669906228780746, + "learning_rate": 3.0356785177823975e-05, + "loss": 0.0034, + "step": 64710 + }, + { + "epoch": 5.026796116504855, + "grad_norm": 0.13976946473121643, + "learning_rate": 3.0341581002308726e-05, + "loss": 0.004, + "step": 64720 + }, + { + "epoch": 5.027572815533981, + "grad_norm": 0.954639196395874, + "learning_rate": 3.0326378976608627e-05, + "loss": 0.0065, + "step": 64730 + }, + { + "epoch": 5.028349514563107, + "grad_norm": 0.24855268001556396, + "learning_rate": 3.0311179102386146e-05, + "loss": 0.0053, + "step": 64740 + }, + { + "epoch": 5.029126213592233, + "grad_norm": 0.8148065805435181, + "learning_rate": 3.0295981381303513e-05, + "loss": 0.007, + "step": 64750 + }, + { + "epoch": 5.029902912621359, + "grad_norm": 0.19213519990444183, + "learning_rate": 3.0280785815022727e-05, + "loss": 0.004, + "step": 64760 + }, + { + "epoch": 5.030679611650485, + "grad_norm": 0.23569895327091217, + "learning_rate": 3.0265592405205544e-05, + "loss": 0.007, + "step": 64770 + }, + { + "epoch": 5.031456310679611, + "grad_norm": 0.11899343878030777, + "learning_rate": 3.0250401153513517e-05, + "loss": 0.0061, + "step": 64780 + }, + { + "epoch": 5.0322330097087375, + "grad_norm": 0.25120794773101807, + "learning_rate": 3.0235212061607936e-05, + "loss": 0.0066, + "step": 64790 + }, + { + "epoch": 5.033009708737864, + "grad_norm": 0.18860064446926117, + "learning_rate": 3.0220025131149842e-05, + "loss": 0.0048, + "step": 64800 + }, + { + "epoch": 5.03378640776699, + "grad_norm": 0.20274846255779266, + "learning_rate": 3.020484036380007e-05, + "loss": 0.0054, + "step": 64810 + }, + { + "epoch": 5.034563106796116, + "grad_norm": 0.7148615717887878, + "learning_rate": 3.0189657761219188e-05, + "loss": 0.0063, + "step": 64820 + }, + { + "epoch": 5.0353398058252425, + "grad_norm": 0.12023554742336273, + "learning_rate": 3.017447732506753e-05, + "loss": 0.0051, + "step": 64830 + }, + { + "epoch": 5.036116504854369, + "grad_norm": 0.25427892804145813, + "learning_rate": 3.015929905700524e-05, + "loss": 0.0068, + "step": 64840 + }, + { + "epoch": 5.036893203883495, + "grad_norm": 0.1652221381664276, + "learning_rate": 3.014412295869218e-05, + "loss": 0.0058, + "step": 64850 + }, + { + "epoch": 5.037669902912621, + "grad_norm": 0.25476405024528503, + "learning_rate": 3.0128949031787977e-05, + "loss": 0.0084, + "step": 64860 + }, + { + "epoch": 5.0384466019417475, + "grad_norm": 0.2689645290374756, + "learning_rate": 3.0113777277952022e-05, + "loss": 0.0063, + "step": 64870 + }, + { + "epoch": 5.039223300970874, + "grad_norm": 0.3045889139175415, + "learning_rate": 3.0098607698843496e-05, + "loss": 0.0064, + "step": 64880 + }, + { + "epoch": 5.04, + "grad_norm": 0.1901988685131073, + "learning_rate": 3.0083440296121312e-05, + "loss": 0.0056, + "step": 64890 + }, + { + "epoch": 5.040776699029126, + "grad_norm": 0.27500107884407043, + "learning_rate": 3.0068275071444142e-05, + "loss": 0.0045, + "step": 64900 + }, + { + "epoch": 5.0415533980582525, + "grad_norm": 0.13015125691890717, + "learning_rate": 3.0053112026470442e-05, + "loss": 0.0049, + "step": 64910 + }, + { + "epoch": 5.042330097087379, + "grad_norm": 0.17383021116256714, + "learning_rate": 3.0037951162858434e-05, + "loss": 0.005, + "step": 64920 + }, + { + "epoch": 5.043106796116505, + "grad_norm": 0.13282202184200287, + "learning_rate": 3.002279248226607e-05, + "loss": 0.0048, + "step": 64930 + }, + { + "epoch": 5.043883495145631, + "grad_norm": 0.12285799533128738, + "learning_rate": 3.000763598635109e-05, + "loss": 0.0044, + "step": 64940 + }, + { + "epoch": 5.0446601941747575, + "grad_norm": 0.09673889726400375, + "learning_rate": 2.999248167677098e-05, + "loss": 0.0042, + "step": 64950 + }, + { + "epoch": 5.045436893203884, + "grad_norm": 0.374016135931015, + "learning_rate": 2.9977329555182976e-05, + "loss": 0.0043, + "step": 64960 + }, + { + "epoch": 5.04621359223301, + "grad_norm": 0.2471013218164444, + "learning_rate": 2.9962179623244124e-05, + "loss": 0.0038, + "step": 64970 + }, + { + "epoch": 5.046990291262136, + "grad_norm": 0.3008076548576355, + "learning_rate": 2.9947031882611175e-05, + "loss": 0.0045, + "step": 64980 + }, + { + "epoch": 5.0477669902912625, + "grad_norm": 0.2567932605743408, + "learning_rate": 2.9931886334940685e-05, + "loss": 0.0075, + "step": 64990 + }, + { + "epoch": 5.048543689320389, + "grad_norm": 0.27090033888816833, + "learning_rate": 2.9916742981888923e-05, + "loss": 0.0062, + "step": 65000 + }, + { + "epoch": 5.049320388349514, + "grad_norm": 0.22885148227214813, + "learning_rate": 2.990160182511196e-05, + "loss": 0.0043, + "step": 65010 + }, + { + "epoch": 5.05009708737864, + "grad_norm": 0.4281608462333679, + "learning_rate": 2.9886462866265574e-05, + "loss": 0.0029, + "step": 65020 + }, + { + "epoch": 5.050873786407767, + "grad_norm": 0.20938004553318024, + "learning_rate": 2.9871326107005392e-05, + "loss": 0.0058, + "step": 65030 + }, + { + "epoch": 5.051650485436893, + "grad_norm": 0.3995095491409302, + "learning_rate": 2.9856191548986735e-05, + "loss": 0.0067, + "step": 65040 + }, + { + "epoch": 5.052427184466019, + "grad_norm": 0.5413580536842346, + "learning_rate": 2.9841059193864672e-05, + "loss": 0.0047, + "step": 65050 + }, + { + "epoch": 5.053203883495145, + "grad_norm": 0.3309146463871002, + "learning_rate": 2.982592904329407e-05, + "loss": 0.0047, + "step": 65060 + }, + { + "epoch": 5.053980582524272, + "grad_norm": 0.16629546880722046, + "learning_rate": 2.981080109892954e-05, + "loss": 0.0047, + "step": 65070 + }, + { + "epoch": 5.054757281553398, + "grad_norm": 0.13731256127357483, + "learning_rate": 2.9795675362425434e-05, + "loss": 0.0051, + "step": 65080 + }, + { + "epoch": 5.055533980582524, + "grad_norm": 0.19144977629184723, + "learning_rate": 2.97805518354359e-05, + "loss": 0.0056, + "step": 65090 + }, + { + "epoch": 5.05631067961165, + "grad_norm": 0.10255236178636551, + "learning_rate": 2.9765430519614825e-05, + "loss": 0.0054, + "step": 65100 + }, + { + "epoch": 5.057087378640777, + "grad_norm": 0.16128656268119812, + "learning_rate": 2.975031141661584e-05, + "loss": 0.0057, + "step": 65110 + }, + { + "epoch": 5.057864077669903, + "grad_norm": 0.1568577140569687, + "learning_rate": 2.9735194528092347e-05, + "loss": 0.0074, + "step": 65120 + }, + { + "epoch": 5.058640776699029, + "grad_norm": 0.2041459083557129, + "learning_rate": 2.9720079855697525e-05, + "loss": 0.0071, + "step": 65130 + }, + { + "epoch": 5.059417475728155, + "grad_norm": 0.24162766337394714, + "learning_rate": 2.970496740108427e-05, + "loss": 0.0049, + "step": 65140 + }, + { + "epoch": 5.060194174757282, + "grad_norm": 0.1899825930595398, + "learning_rate": 2.968985716590527e-05, + "loss": 0.0041, + "step": 65150 + }, + { + "epoch": 5.060970873786408, + "grad_norm": 1.0353302955627441, + "learning_rate": 2.9674749151812953e-05, + "loss": 0.0046, + "step": 65160 + }, + { + "epoch": 5.061747572815534, + "grad_norm": 0.22992220520973206, + "learning_rate": 2.965964336045951e-05, + "loss": 0.0048, + "step": 65170 + }, + { + "epoch": 5.06252427184466, + "grad_norm": 0.11965043842792511, + "learning_rate": 2.9644539793496894e-05, + "loss": 0.0059, + "step": 65180 + }, + { + "epoch": 5.063300970873787, + "grad_norm": 0.4764552414417267, + "learning_rate": 2.9629438452576807e-05, + "loss": 0.0087, + "step": 65190 + }, + { + "epoch": 5.064077669902913, + "grad_norm": 0.5226620435714722, + "learning_rate": 2.9614339339350698e-05, + "loss": 0.004, + "step": 65200 + }, + { + "epoch": 5.064854368932039, + "grad_norm": 0.19295048713684082, + "learning_rate": 2.959924245546979e-05, + "loss": 0.006, + "step": 65210 + }, + { + "epoch": 5.065631067961165, + "grad_norm": 0.303930401802063, + "learning_rate": 2.9584147802585038e-05, + "loss": 0.0053, + "step": 65220 + }, + { + "epoch": 5.066407766990292, + "grad_norm": 0.6855619549751282, + "learning_rate": 2.956905538234721e-05, + "loss": 0.0054, + "step": 65230 + }, + { + "epoch": 5.067184466019418, + "grad_norm": 0.1308620423078537, + "learning_rate": 2.9553965196406764e-05, + "loss": 0.0043, + "step": 65240 + }, + { + "epoch": 5.067961165048544, + "grad_norm": 0.10101766884326935, + "learning_rate": 2.9538877246413943e-05, + "loss": 0.0048, + "step": 65250 + }, + { + "epoch": 5.06873786407767, + "grad_norm": 0.1257386952638626, + "learning_rate": 2.952379153401874e-05, + "loss": 0.0062, + "step": 65260 + }, + { + "epoch": 5.069514563106796, + "grad_norm": 0.09071632474660873, + "learning_rate": 2.9508708060870914e-05, + "loss": 0.0038, + "step": 65270 + }, + { + "epoch": 5.070291262135922, + "grad_norm": 0.14099355041980743, + "learning_rate": 2.9493626828619948e-05, + "loss": 0.0048, + "step": 65280 + }, + { + "epoch": 5.071067961165048, + "grad_norm": 0.1861339956521988, + "learning_rate": 2.9478547838915137e-05, + "loss": 0.0052, + "step": 65290 + }, + { + "epoch": 5.0718446601941745, + "grad_norm": 0.2618233263492584, + "learning_rate": 2.946347109340548e-05, + "loss": 0.0062, + "step": 65300 + }, + { + "epoch": 5.072621359223301, + "grad_norm": 0.7414228320121765, + "learning_rate": 2.944839659373976e-05, + "loss": 0.0048, + "step": 65310 + }, + { + "epoch": 5.073398058252427, + "grad_norm": 0.15013553202152252, + "learning_rate": 2.943332434156647e-05, + "loss": 0.0049, + "step": 65320 + }, + { + "epoch": 5.074174757281553, + "grad_norm": 0.18543103337287903, + "learning_rate": 2.9418254338533912e-05, + "loss": 0.0041, + "step": 65330 + }, + { + "epoch": 5.0749514563106795, + "grad_norm": 0.2437039464712143, + "learning_rate": 2.940318658629011e-05, + "loss": 0.0074, + "step": 65340 + }, + { + "epoch": 5.075728155339806, + "grad_norm": 0.2452116757631302, + "learning_rate": 2.938812108648287e-05, + "loss": 0.004, + "step": 65350 + }, + { + "epoch": 5.076504854368932, + "grad_norm": 0.17864017188549042, + "learning_rate": 2.937305784075972e-05, + "loss": 0.0072, + "step": 65360 + }, + { + "epoch": 5.077281553398058, + "grad_norm": 0.18608559668064117, + "learning_rate": 2.9357996850767937e-05, + "loss": 0.0047, + "step": 65370 + }, + { + "epoch": 5.0780582524271844, + "grad_norm": 0.1545078456401825, + "learning_rate": 2.9342938118154595e-05, + "loss": 0.0065, + "step": 65380 + }, + { + "epoch": 5.078834951456311, + "grad_norm": 0.16720792651176453, + "learning_rate": 2.9327881644566485e-05, + "loss": 0.0044, + "step": 65390 + }, + { + "epoch": 5.079611650485437, + "grad_norm": 0.2584558129310608, + "learning_rate": 2.9312827431650157e-05, + "loss": 0.0059, + "step": 65400 + }, + { + "epoch": 5.080388349514563, + "grad_norm": 0.0889689028263092, + "learning_rate": 2.92977754810519e-05, + "loss": 0.0044, + "step": 65410 + }, + { + "epoch": 5.0811650485436894, + "grad_norm": 0.45577171444892883, + "learning_rate": 2.928272579441781e-05, + "loss": 0.0062, + "step": 65420 + }, + { + "epoch": 5.081941747572816, + "grad_norm": 0.1442190706729889, + "learning_rate": 2.9267678373393676e-05, + "loss": 0.0042, + "step": 65430 + }, + { + "epoch": 5.082718446601942, + "grad_norm": 0.14154507219791412, + "learning_rate": 2.925263321962507e-05, + "loss": 0.0062, + "step": 65440 + }, + { + "epoch": 5.083495145631068, + "grad_norm": 0.36652815341949463, + "learning_rate": 2.9237590334757292e-05, + "loss": 0.0051, + "step": 65450 + }, + { + "epoch": 5.084271844660194, + "grad_norm": 0.2512737810611725, + "learning_rate": 2.922254972043543e-05, + "loss": 0.0042, + "step": 65460 + }, + { + "epoch": 5.085048543689321, + "grad_norm": 0.21531778573989868, + "learning_rate": 2.9207511378304265e-05, + "loss": 0.0039, + "step": 65470 + }, + { + "epoch": 5.085825242718447, + "grad_norm": 0.09715323150157928, + "learning_rate": 2.9192475310008428e-05, + "loss": 0.0037, + "step": 65480 + }, + { + "epoch": 5.086601941747573, + "grad_norm": 0.36418092250823975, + "learning_rate": 2.917744151719219e-05, + "loss": 0.0041, + "step": 65490 + }, + { + "epoch": 5.087378640776699, + "grad_norm": 0.27616825699806213, + "learning_rate": 2.916241000149964e-05, + "loss": 0.0056, + "step": 65500 + }, + { + "epoch": 5.088155339805825, + "grad_norm": 0.18207012116909027, + "learning_rate": 2.9147380764574622e-05, + "loss": 0.0045, + "step": 65510 + }, + { + "epoch": 5.088932038834951, + "grad_norm": 0.27737319469451904, + "learning_rate": 2.9132353808060674e-05, + "loss": 0.0052, + "step": 65520 + }, + { + "epoch": 5.089708737864077, + "grad_norm": 0.27915158867836, + "learning_rate": 2.9117329133601134e-05, + "loss": 0.0067, + "step": 65530 + }, + { + "epoch": 5.0904854368932035, + "grad_norm": 0.3558104932308197, + "learning_rate": 2.9102306742839093e-05, + "loss": 0.0044, + "step": 65540 + }, + { + "epoch": 5.09126213592233, + "grad_norm": 0.43587884306907654, + "learning_rate": 2.9087286637417375e-05, + "loss": 0.0052, + "step": 65550 + }, + { + "epoch": 5.092038834951456, + "grad_norm": 0.18317808210849762, + "learning_rate": 2.9072268818978537e-05, + "loss": 0.0066, + "step": 65560 + }, + { + "epoch": 5.092815533980582, + "grad_norm": 0.7335310578346252, + "learning_rate": 2.905725328916493e-05, + "loss": 0.0071, + "step": 65570 + }, + { + "epoch": 5.0935922330097085, + "grad_norm": 0.1721830815076828, + "learning_rate": 2.9042240049618608e-05, + "loss": 0.0045, + "step": 65580 + }, + { + "epoch": 5.094368932038835, + "grad_norm": 0.17280080914497375, + "learning_rate": 2.9027229101981413e-05, + "loss": 0.0033, + "step": 65590 + }, + { + "epoch": 5.095145631067961, + "grad_norm": 0.31915026903152466, + "learning_rate": 2.901222044789488e-05, + "loss": 0.0041, + "step": 65600 + }, + { + "epoch": 5.095922330097087, + "grad_norm": 0.4406963884830475, + "learning_rate": 2.89972140890004e-05, + "loss": 0.0049, + "step": 65610 + }, + { + "epoch": 5.0966990291262135, + "grad_norm": 0.1634802222251892, + "learning_rate": 2.8982210026938988e-05, + "loss": 0.0043, + "step": 65620 + }, + { + "epoch": 5.09747572815534, + "grad_norm": 0.3692518174648285, + "learning_rate": 2.896720826335151e-05, + "loss": 0.0039, + "step": 65630 + }, + { + "epoch": 5.098252427184466, + "grad_norm": 0.21657989919185638, + "learning_rate": 2.8952208799878493e-05, + "loss": 0.0045, + "step": 65640 + }, + { + "epoch": 5.099029126213592, + "grad_norm": 0.19925369322299957, + "learning_rate": 2.89372116381603e-05, + "loss": 0.0047, + "step": 65650 + }, + { + "epoch": 5.0998058252427185, + "grad_norm": 0.17655648291110992, + "learning_rate": 2.8922216779836932e-05, + "loss": 0.0043, + "step": 65660 + }, + { + "epoch": 5.100582524271845, + "grad_norm": 0.15305642783641815, + "learning_rate": 2.8907224226548285e-05, + "loss": 0.0056, + "step": 65670 + }, + { + "epoch": 5.101359223300971, + "grad_norm": 0.29154154658317566, + "learning_rate": 2.8892233979933868e-05, + "loss": 0.0065, + "step": 65680 + }, + { + "epoch": 5.102135922330097, + "grad_norm": 0.18710528314113617, + "learning_rate": 2.8877246041633026e-05, + "loss": 0.0062, + "step": 65690 + }, + { + "epoch": 5.1029126213592235, + "grad_norm": 0.17145168781280518, + "learning_rate": 2.886226041328478e-05, + "loss": 0.0063, + "step": 65700 + }, + { + "epoch": 5.10368932038835, + "grad_norm": 0.12938666343688965, + "learning_rate": 2.8847277096527957e-05, + "loss": 0.0051, + "step": 65710 + }, + { + "epoch": 5.104466019417476, + "grad_norm": 0.21261771023273468, + "learning_rate": 2.8832296093001105e-05, + "loss": 0.0069, + "step": 65720 + }, + { + "epoch": 5.105242718446602, + "grad_norm": 0.2936694622039795, + "learning_rate": 2.8817317404342547e-05, + "loss": 0.0035, + "step": 65730 + }, + { + "epoch": 5.1060194174757285, + "grad_norm": 0.1974988579750061, + "learning_rate": 2.880234103219029e-05, + "loss": 0.0068, + "step": 65740 + }, + { + "epoch": 5.106796116504855, + "grad_norm": 0.318000465631485, + "learning_rate": 2.878736697818215e-05, + "loss": 0.0057, + "step": 65750 + }, + { + "epoch": 5.107572815533981, + "grad_norm": 0.16140538454055786, + "learning_rate": 2.877239524395568e-05, + "loss": 0.006, + "step": 65760 + }, + { + "epoch": 5.108349514563106, + "grad_norm": 0.31539279222488403, + "learning_rate": 2.8757425831148134e-05, + "loss": 0.0065, + "step": 65770 + }, + { + "epoch": 5.109126213592233, + "grad_norm": 0.34177330136299133, + "learning_rate": 2.874245874139658e-05, + "loss": 0.0067, + "step": 65780 + }, + { + "epoch": 5.109902912621359, + "grad_norm": 0.1409856379032135, + "learning_rate": 2.8727493976337737e-05, + "loss": 0.0083, + "step": 65790 + }, + { + "epoch": 5.110679611650485, + "grad_norm": 0.22795072197914124, + "learning_rate": 2.8712531537608218e-05, + "loss": 0.0073, + "step": 65800 + }, + { + "epoch": 5.111456310679611, + "grad_norm": 0.11926282942295074, + "learning_rate": 2.8697571426844217e-05, + "loss": 0.0043, + "step": 65810 + }, + { + "epoch": 5.112233009708738, + "grad_norm": 0.14155493676662445, + "learning_rate": 2.86826136456818e-05, + "loss": 0.0052, + "step": 65820 + }, + { + "epoch": 5.113009708737864, + "grad_norm": 0.17061933875083923, + "learning_rate": 2.8667658195756685e-05, + "loss": 0.0037, + "step": 65830 + }, + { + "epoch": 5.11378640776699, + "grad_norm": 0.11166277527809143, + "learning_rate": 2.8652705078704412e-05, + "loss": 0.0045, + "step": 65840 + }, + { + "epoch": 5.114563106796116, + "grad_norm": 0.2424498200416565, + "learning_rate": 2.863775429616018e-05, + "loss": 0.0059, + "step": 65850 + }, + { + "epoch": 5.115339805825243, + "grad_norm": 0.3178021311759949, + "learning_rate": 2.862280584975906e-05, + "loss": 0.0055, + "step": 65860 + }, + { + "epoch": 5.116116504854369, + "grad_norm": 0.2044374942779541, + "learning_rate": 2.8607859741135735e-05, + "loss": 0.0042, + "step": 65870 + }, + { + "epoch": 5.116893203883495, + "grad_norm": 0.24952344596385956, + "learning_rate": 2.8592915971924723e-05, + "loss": 0.0047, + "step": 65880 + }, + { + "epoch": 5.117669902912621, + "grad_norm": 0.6694483757019043, + "learning_rate": 2.8577974543760212e-05, + "loss": 0.0037, + "step": 65890 + }, + { + "epoch": 5.118446601941748, + "grad_norm": 0.8320331573486328, + "learning_rate": 2.8563035458276222e-05, + "loss": 0.006, + "step": 65900 + }, + { + "epoch": 5.119223300970874, + "grad_norm": 0.10225194692611694, + "learning_rate": 2.8548098717106424e-05, + "loss": 0.0033, + "step": 65910 + }, + { + "epoch": 5.12, + "grad_norm": 0.25422996282577515, + "learning_rate": 2.85331643218843e-05, + "loss": 0.0044, + "step": 65920 + }, + { + "epoch": 5.120776699029126, + "grad_norm": 0.16631273925304413, + "learning_rate": 2.8518232274243046e-05, + "loss": 0.005, + "step": 65930 + }, + { + "epoch": 5.121553398058253, + "grad_norm": 0.3182899057865143, + "learning_rate": 2.8503302575815638e-05, + "loss": 0.0055, + "step": 65940 + }, + { + "epoch": 5.122330097087379, + "grad_norm": 0.11368296295404434, + "learning_rate": 2.8488375228234716e-05, + "loss": 0.0045, + "step": 65950 + }, + { + "epoch": 5.123106796116505, + "grad_norm": 0.1924804300069809, + "learning_rate": 2.8473450233132738e-05, + "loss": 0.0051, + "step": 65960 + }, + { + "epoch": 5.123883495145631, + "grad_norm": 0.37769830226898193, + "learning_rate": 2.8458527592141898e-05, + "loss": 0.0059, + "step": 65970 + }, + { + "epoch": 5.124660194174758, + "grad_norm": 0.2128756195306778, + "learning_rate": 2.8443607306894073e-05, + "loss": 0.0065, + "step": 65980 + }, + { + "epoch": 5.125436893203884, + "grad_norm": 0.3142711818218231, + "learning_rate": 2.8428689379020946e-05, + "loss": 0.0066, + "step": 65990 + }, + { + "epoch": 5.12621359223301, + "grad_norm": 0.10141260176897049, + "learning_rate": 2.841377381015391e-05, + "loss": 0.0049, + "step": 66000 + }, + { + "epoch": 5.126990291262136, + "grad_norm": 0.3089126646518707, + "learning_rate": 2.8398860601924144e-05, + "loss": 0.0059, + "step": 66010 + }, + { + "epoch": 5.127766990291262, + "grad_norm": 0.2682987153530121, + "learning_rate": 2.838394975596248e-05, + "loss": 0.0054, + "step": 66020 + }, + { + "epoch": 5.128543689320388, + "grad_norm": 0.2617766559123993, + "learning_rate": 2.83690412738996e-05, + "loss": 0.0068, + "step": 66030 + }, + { + "epoch": 5.129320388349514, + "grad_norm": 0.358472615480423, + "learning_rate": 2.835413515736579e-05, + "loss": 0.0051, + "step": 66040 + }, + { + "epoch": 5.1300970873786405, + "grad_norm": 0.328409880399704, + "learning_rate": 2.8339231407991267e-05, + "loss": 0.0065, + "step": 66050 + }, + { + "epoch": 5.130873786407767, + "grad_norm": 0.5978981256484985, + "learning_rate": 2.832433002740581e-05, + "loss": 0.0053, + "step": 66060 + }, + { + "epoch": 5.131650485436893, + "grad_norm": 0.2081165462732315, + "learning_rate": 2.8309431017239047e-05, + "loss": 0.0048, + "step": 66070 + }, + { + "epoch": 5.132427184466019, + "grad_norm": 0.36092421412467957, + "learning_rate": 2.829453437912028e-05, + "loss": 0.0071, + "step": 66080 + }, + { + "epoch": 5.1332038834951454, + "grad_norm": 0.22255705296993256, + "learning_rate": 2.8279640114678617e-05, + "loss": 0.0057, + "step": 66090 + }, + { + "epoch": 5.133980582524272, + "grad_norm": 0.24725869297981262, + "learning_rate": 2.8264748225542836e-05, + "loss": 0.0068, + "step": 66100 + }, + { + "epoch": 5.134757281553398, + "grad_norm": 0.11436779797077179, + "learning_rate": 2.8249858713341516e-05, + "loss": 0.006, + "step": 66110 + }, + { + "epoch": 5.135533980582524, + "grad_norm": 0.1832578033208847, + "learning_rate": 2.823497157970294e-05, + "loss": 0.006, + "step": 66120 + }, + { + "epoch": 5.13631067961165, + "grad_norm": 0.18598362803459167, + "learning_rate": 2.8220086826255164e-05, + "loss": 0.0041, + "step": 66130 + }, + { + "epoch": 5.137087378640777, + "grad_norm": 0.3494184613227844, + "learning_rate": 2.8205204454625934e-05, + "loss": 0.0047, + "step": 66140 + }, + { + "epoch": 5.137864077669903, + "grad_norm": 0.36873364448547363, + "learning_rate": 2.8190324466442798e-05, + "loss": 0.0052, + "step": 66150 + }, + { + "epoch": 5.138640776699029, + "grad_norm": 0.4415408968925476, + "learning_rate": 2.8175446863332965e-05, + "loss": 0.0057, + "step": 66160 + }, + { + "epoch": 5.139417475728155, + "grad_norm": 0.6104373335838318, + "learning_rate": 2.816057164692345e-05, + "loss": 0.0087, + "step": 66170 + }, + { + "epoch": 5.140194174757282, + "grad_norm": 0.5567962527275085, + "learning_rate": 2.8145698818840992e-05, + "loss": 0.0064, + "step": 66180 + }, + { + "epoch": 5.140970873786408, + "grad_norm": 0.5630232691764832, + "learning_rate": 2.8130828380712065e-05, + "loss": 0.0063, + "step": 66190 + }, + { + "epoch": 5.141747572815534, + "grad_norm": 0.3342423439025879, + "learning_rate": 2.811596033416285e-05, + "loss": 0.0073, + "step": 66200 + }, + { + "epoch": 5.14252427184466, + "grad_norm": 0.24985943734645844, + "learning_rate": 2.8101094680819307e-05, + "loss": 0.0055, + "step": 66210 + }, + { + "epoch": 5.143300970873787, + "grad_norm": 0.20509855449199677, + "learning_rate": 2.808623142230715e-05, + "loss": 0.0063, + "step": 66220 + }, + { + "epoch": 5.144077669902913, + "grad_norm": 0.5360904335975647, + "learning_rate": 2.8071370560251757e-05, + "loss": 0.0031, + "step": 66230 + }, + { + "epoch": 5.144854368932039, + "grad_norm": 0.14110858738422394, + "learning_rate": 2.805651209627831e-05, + "loss": 0.0052, + "step": 66240 + }, + { + "epoch": 5.145631067961165, + "grad_norm": 0.3210890591144562, + "learning_rate": 2.804165603201171e-05, + "loss": 0.0048, + "step": 66250 + }, + { + "epoch": 5.146407766990292, + "grad_norm": 0.09635642170906067, + "learning_rate": 2.802680236907661e-05, + "loss": 0.0074, + "step": 66260 + }, + { + "epoch": 5.147184466019418, + "grad_norm": 0.1530729979276657, + "learning_rate": 2.8011951109097344e-05, + "loss": 0.0049, + "step": 66270 + }, + { + "epoch": 5.147961165048543, + "grad_norm": 0.16667181253433228, + "learning_rate": 2.7997102253698072e-05, + "loss": 0.0058, + "step": 66280 + }, + { + "epoch": 5.1487378640776695, + "grad_norm": 0.09657607227563858, + "learning_rate": 2.7982255804502588e-05, + "loss": 0.005, + "step": 66290 + }, + { + "epoch": 5.149514563106796, + "grad_norm": 0.11925430595874786, + "learning_rate": 2.7967411763134516e-05, + "loss": 0.0044, + "step": 66300 + }, + { + "epoch": 5.150291262135922, + "grad_norm": 0.3332217335700989, + "learning_rate": 2.7952570131217166e-05, + "loss": 0.0048, + "step": 66310 + }, + { + "epoch": 5.151067961165048, + "grad_norm": 0.15394708514213562, + "learning_rate": 2.7937730910373618e-05, + "loss": 0.0048, + "step": 66320 + }, + { + "epoch": 5.1518446601941745, + "grad_norm": 0.33464303612709045, + "learning_rate": 2.7922894102226628e-05, + "loss": 0.0066, + "step": 66330 + }, + { + "epoch": 5.152621359223301, + "grad_norm": 0.648998498916626, + "learning_rate": 2.790805970839876e-05, + "loss": 0.0075, + "step": 66340 + }, + { + "epoch": 5.153398058252427, + "grad_norm": 0.24984320998191833, + "learning_rate": 2.789322773051225e-05, + "loss": 0.0056, + "step": 66350 + }, + { + "epoch": 5.154174757281553, + "grad_norm": 0.28078675270080566, + "learning_rate": 2.7878398170189124e-05, + "loss": 0.0049, + "step": 66360 + }, + { + "epoch": 5.1549514563106795, + "grad_norm": 0.16240812838077545, + "learning_rate": 2.7863571029051112e-05, + "loss": 0.0051, + "step": 66370 + }, + { + "epoch": 5.155728155339806, + "grad_norm": 0.18523386120796204, + "learning_rate": 2.784874630871971e-05, + "loss": 0.0053, + "step": 66380 + }, + { + "epoch": 5.156504854368932, + "grad_norm": 0.15737253427505493, + "learning_rate": 2.7833924010816086e-05, + "loss": 0.0051, + "step": 66390 + }, + { + "epoch": 5.157281553398058, + "grad_norm": 0.10472317039966583, + "learning_rate": 2.7819104136961226e-05, + "loss": 0.0039, + "step": 66400 + }, + { + "epoch": 5.1580582524271845, + "grad_norm": 0.14886516332626343, + "learning_rate": 2.780428668877577e-05, + "loss": 0.0072, + "step": 66410 + }, + { + "epoch": 5.158834951456311, + "grad_norm": 0.1426854133605957, + "learning_rate": 2.7789471667880152e-05, + "loss": 0.0045, + "step": 66420 + }, + { + "epoch": 5.159611650485437, + "grad_norm": 0.2389959692955017, + "learning_rate": 2.7774659075894523e-05, + "loss": 0.0055, + "step": 66430 + }, + { + "epoch": 5.160388349514563, + "grad_norm": 0.3703027069568634, + "learning_rate": 2.7759848914438778e-05, + "loss": 0.0033, + "step": 66440 + }, + { + "epoch": 5.1611650485436895, + "grad_norm": 0.2213735282421112, + "learning_rate": 2.7745041185132496e-05, + "loss": 0.0057, + "step": 66450 + }, + { + "epoch": 5.161941747572816, + "grad_norm": 0.15339788794517517, + "learning_rate": 2.773023588959506e-05, + "loss": 0.0078, + "step": 66460 + }, + { + "epoch": 5.162718446601942, + "grad_norm": 0.7033426761627197, + "learning_rate": 2.7715433029445555e-05, + "loss": 0.0037, + "step": 66470 + }, + { + "epoch": 5.163495145631068, + "grad_norm": 0.2607276439666748, + "learning_rate": 2.7700632606302766e-05, + "loss": 0.0082, + "step": 66480 + }, + { + "epoch": 5.1642718446601945, + "grad_norm": 0.3396335244178772, + "learning_rate": 2.768583462178528e-05, + "loss": 0.0054, + "step": 66490 + }, + { + "epoch": 5.165048543689321, + "grad_norm": 0.32448843121528625, + "learning_rate": 2.767103907751136e-05, + "loss": 0.0058, + "step": 66500 + }, + { + "epoch": 5.165825242718447, + "grad_norm": 0.19103601574897766, + "learning_rate": 2.7656245975099065e-05, + "loss": 0.0064, + "step": 66510 + }, + { + "epoch": 5.166601941747572, + "grad_norm": 0.49334225058555603, + "learning_rate": 2.7641455316166093e-05, + "loss": 0.0058, + "step": 66520 + }, + { + "epoch": 5.167378640776699, + "grad_norm": 0.11711540818214417, + "learning_rate": 2.7626667102329973e-05, + "loss": 0.006, + "step": 66530 + }, + { + "epoch": 5.168155339805825, + "grad_norm": 0.10558979958295822, + "learning_rate": 2.761188133520788e-05, + "loss": 0.0045, + "step": 66540 + }, + { + "epoch": 5.168932038834951, + "grad_norm": 0.19643306732177734, + "learning_rate": 2.7597098016416784e-05, + "loss": 0.0067, + "step": 66550 + }, + { + "epoch": 5.169708737864077, + "grad_norm": 0.2736835777759552, + "learning_rate": 2.7582317147573368e-05, + "loss": 0.0046, + "step": 66560 + }, + { + "epoch": 5.170485436893204, + "grad_norm": 0.3980993628501892, + "learning_rate": 2.7567538730294064e-05, + "loss": 0.0063, + "step": 66570 + }, + { + "epoch": 5.17126213592233, + "grad_norm": 0.19032563269138336, + "learning_rate": 2.7552762766194973e-05, + "loss": 0.0072, + "step": 66580 + }, + { + "epoch": 5.172038834951456, + "grad_norm": 0.3620630204677582, + "learning_rate": 2.753798925689202e-05, + "loss": 0.0043, + "step": 66590 + }, + { + "epoch": 5.172815533980582, + "grad_norm": 0.2044689953327179, + "learning_rate": 2.7523218204000774e-05, + "loss": 0.0044, + "step": 66600 + }, + { + "epoch": 5.173592233009709, + "grad_norm": 0.2633201777935028, + "learning_rate": 2.7508449609136612e-05, + "loss": 0.0056, + "step": 66610 + }, + { + "epoch": 5.174368932038835, + "grad_norm": 0.2805126905441284, + "learning_rate": 2.749368347391454e-05, + "loss": 0.0053, + "step": 66620 + }, + { + "epoch": 5.175145631067961, + "grad_norm": 0.10250665247440338, + "learning_rate": 2.7478919799949443e-05, + "loss": 0.0077, + "step": 66630 + }, + { + "epoch": 5.175922330097087, + "grad_norm": 0.22523382306098938, + "learning_rate": 2.7464158588855803e-05, + "loss": 0.0055, + "step": 66640 + }, + { + "epoch": 5.176699029126214, + "grad_norm": 0.26026639342308044, + "learning_rate": 2.7449399842247918e-05, + "loss": 0.0058, + "step": 66650 + }, + { + "epoch": 5.17747572815534, + "grad_norm": 0.20183375477790833, + "learning_rate": 2.7434643561739736e-05, + "loss": 0.0046, + "step": 66660 + }, + { + "epoch": 5.178252427184466, + "grad_norm": 0.175989031791687, + "learning_rate": 2.7419889748945016e-05, + "loss": 0.0057, + "step": 66670 + }, + { + "epoch": 5.179029126213592, + "grad_norm": 0.2075764685869217, + "learning_rate": 2.7405138405477203e-05, + "loss": 0.0071, + "step": 66680 + }, + { + "epoch": 5.179805825242719, + "grad_norm": 0.18847203254699707, + "learning_rate": 2.73903895329495e-05, + "loss": 0.0051, + "step": 66690 + }, + { + "epoch": 5.180582524271845, + "grad_norm": 0.2900073230266571, + "learning_rate": 2.737564313297478e-05, + "loss": 0.006, + "step": 66700 + }, + { + "epoch": 5.181359223300971, + "grad_norm": 0.1261233389377594, + "learning_rate": 2.736089920716572e-05, + "loss": 0.0055, + "step": 66710 + }, + { + "epoch": 5.182135922330097, + "grad_norm": 0.20148441195487976, + "learning_rate": 2.7346157757134706e-05, + "loss": 0.0077, + "step": 66720 + }, + { + "epoch": 5.182912621359224, + "grad_norm": 0.1688326895236969, + "learning_rate": 2.733141878449379e-05, + "loss": 0.0037, + "step": 66730 + }, + { + "epoch": 5.18368932038835, + "grad_norm": 0.2292134016752243, + "learning_rate": 2.7316682290854844e-05, + "loss": 0.0051, + "step": 66740 + }, + { + "epoch": 5.184466019417476, + "grad_norm": 0.19583836197853088, + "learning_rate": 2.7301948277829414e-05, + "loss": 0.0028, + "step": 66750 + }, + { + "epoch": 5.185242718446602, + "grad_norm": 0.1759532243013382, + "learning_rate": 2.7287216747028814e-05, + "loss": 0.0039, + "step": 66760 + }, + { + "epoch": 5.186019417475729, + "grad_norm": 0.10009574890136719, + "learning_rate": 2.7272487700064025e-05, + "loss": 0.0059, + "step": 66770 + }, + { + "epoch": 5.186796116504854, + "grad_norm": 0.11566805839538574, + "learning_rate": 2.725776113854583e-05, + "loss": 0.0049, + "step": 66780 + }, + { + "epoch": 5.18757281553398, + "grad_norm": 0.12468105554580688, + "learning_rate": 2.7243037064084664e-05, + "loss": 0.0043, + "step": 66790 + }, + { + "epoch": 5.1883495145631064, + "grad_norm": 0.2643592953681946, + "learning_rate": 2.7228315478290766e-05, + "loss": 0.0044, + "step": 66800 + }, + { + "epoch": 5.189126213592233, + "grad_norm": 0.30539825558662415, + "learning_rate": 2.721359638277402e-05, + "loss": 0.0059, + "step": 66810 + }, + { + "epoch": 5.189902912621359, + "grad_norm": 0.28773224353790283, + "learning_rate": 2.7198879779144148e-05, + "loss": 0.005, + "step": 66820 + }, + { + "epoch": 5.190679611650485, + "grad_norm": 0.38881805539131165, + "learning_rate": 2.7184165669010485e-05, + "loss": 0.0054, + "step": 66830 + }, + { + "epoch": 5.191456310679611, + "grad_norm": 0.1541731059551239, + "learning_rate": 2.716945405398218e-05, + "loss": 0.0041, + "step": 66840 + }, + { + "epoch": 5.192233009708738, + "grad_norm": 0.2104882448911667, + "learning_rate": 2.715474493566803e-05, + "loss": 0.0045, + "step": 66850 + }, + { + "epoch": 5.193009708737864, + "grad_norm": 0.21456630527973175, + "learning_rate": 2.7140038315676652e-05, + "loss": 0.0047, + "step": 66860 + }, + { + "epoch": 5.19378640776699, + "grad_norm": 0.27867090702056885, + "learning_rate": 2.7125334195616275e-05, + "loss": 0.0062, + "step": 66870 + }, + { + "epoch": 5.194563106796116, + "grad_norm": 0.281755268573761, + "learning_rate": 2.7110632577094997e-05, + "loss": 0.0069, + "step": 66880 + }, + { + "epoch": 5.195339805825243, + "grad_norm": 0.18171291053295135, + "learning_rate": 2.709593346172051e-05, + "loss": 0.0052, + "step": 66890 + }, + { + "epoch": 5.196116504854369, + "grad_norm": 0.1809656322002411, + "learning_rate": 2.708123685110032e-05, + "loss": 0.0059, + "step": 66900 + }, + { + "epoch": 5.196893203883495, + "grad_norm": 0.12960684299468994, + "learning_rate": 2.7066542746841588e-05, + "loss": 0.0073, + "step": 66910 + }, + { + "epoch": 5.197669902912621, + "grad_norm": 0.12264562398195267, + "learning_rate": 2.7051851150551267e-05, + "loss": 0.0052, + "step": 66920 + }, + { + "epoch": 5.198446601941748, + "grad_norm": 0.5685974359512329, + "learning_rate": 2.7037162063835996e-05, + "loss": 0.0054, + "step": 66930 + }, + { + "epoch": 5.199223300970874, + "grad_norm": 0.31384196877479553, + "learning_rate": 2.702247548830218e-05, + "loss": 0.006, + "step": 66940 + }, + { + "epoch": 5.2, + "grad_norm": 0.19565972685813904, + "learning_rate": 2.7007791425555883e-05, + "loss": 0.0049, + "step": 66950 + }, + { + "epoch": 5.200776699029126, + "grad_norm": 0.277785986661911, + "learning_rate": 2.6993109877202945e-05, + "loss": 0.0044, + "step": 66960 + }, + { + "epoch": 5.201553398058253, + "grad_norm": 0.13069167733192444, + "learning_rate": 2.6978430844848935e-05, + "loss": 0.006, + "step": 66970 + }, + { + "epoch": 5.202330097087379, + "grad_norm": 0.4968162775039673, + "learning_rate": 2.69637543300991e-05, + "loss": 0.0055, + "step": 66980 + }, + { + "epoch": 5.203106796116505, + "grad_norm": 0.20089103281497955, + "learning_rate": 2.694908033455848e-05, + "loss": 0.0044, + "step": 66990 + }, + { + "epoch": 5.203883495145631, + "grad_norm": 0.16405081748962402, + "learning_rate": 2.6934408859831733e-05, + "loss": 0.0042, + "step": 67000 + }, + { + "epoch": 5.204660194174758, + "grad_norm": 0.12088025361299515, + "learning_rate": 2.69197399075234e-05, + "loss": 0.0045, + "step": 67010 + }, + { + "epoch": 5.205436893203884, + "grad_norm": 0.25453871488571167, + "learning_rate": 2.6905073479237584e-05, + "loss": 0.0058, + "step": 67020 + }, + { + "epoch": 5.206213592233009, + "grad_norm": 0.35079023241996765, + "learning_rate": 2.6890409576578235e-05, + "loss": 0.006, + "step": 67030 + }, + { + "epoch": 5.2069902912621355, + "grad_norm": 0.2563827335834503, + "learning_rate": 2.687574820114892e-05, + "loss": 0.0077, + "step": 67040 + }, + { + "epoch": 5.207766990291262, + "grad_norm": 0.30408889055252075, + "learning_rate": 2.686108935455305e-05, + "loss": 0.0065, + "step": 67050 + }, + { + "epoch": 5.208543689320388, + "grad_norm": 0.2573874890804291, + "learning_rate": 2.6846433038393603e-05, + "loss": 0.0058, + "step": 67060 + }, + { + "epoch": 5.209320388349514, + "grad_norm": 0.10773293673992157, + "learning_rate": 2.6831779254273482e-05, + "loss": 0.0049, + "step": 67070 + }, + { + "epoch": 5.2100970873786405, + "grad_norm": 0.10075671225786209, + "learning_rate": 2.6817128003795122e-05, + "loss": 0.0038, + "step": 67080 + }, + { + "epoch": 5.210873786407767, + "grad_norm": 0.4075053334236145, + "learning_rate": 2.680247928856081e-05, + "loss": 0.0045, + "step": 67090 + }, + { + "epoch": 5.211650485436893, + "grad_norm": 0.2535038888454437, + "learning_rate": 2.6787833110172466e-05, + "loss": 0.0051, + "step": 67100 + }, + { + "epoch": 5.212427184466019, + "grad_norm": 0.17237824201583862, + "learning_rate": 2.6773189470231807e-05, + "loss": 0.0041, + "step": 67110 + }, + { + "epoch": 5.2132038834951455, + "grad_norm": 0.22010673582553864, + "learning_rate": 2.675854837034022e-05, + "loss": 0.0069, + "step": 67120 + }, + { + "epoch": 5.213980582524272, + "grad_norm": 0.258958101272583, + "learning_rate": 2.6743909812098833e-05, + "loss": 0.0049, + "step": 67130 + }, + { + "epoch": 5.214757281553398, + "grad_norm": 0.13461479544639587, + "learning_rate": 2.67292737971085e-05, + "loss": 0.0063, + "step": 67140 + }, + { + "epoch": 5.215533980582524, + "grad_norm": 0.22599472105503082, + "learning_rate": 2.671464032696982e-05, + "loss": 0.0062, + "step": 67150 + }, + { + "epoch": 5.2163106796116505, + "grad_norm": 0.5792844295501709, + "learning_rate": 2.6700009403283055e-05, + "loss": 0.0075, + "step": 67160 + }, + { + "epoch": 5.217087378640777, + "grad_norm": 0.18255141377449036, + "learning_rate": 2.6685381027648214e-05, + "loss": 0.0063, + "step": 67170 + }, + { + "epoch": 5.217864077669903, + "grad_norm": 0.23940588533878326, + "learning_rate": 2.6670755201665086e-05, + "loss": 0.0045, + "step": 67180 + }, + { + "epoch": 5.218640776699029, + "grad_norm": 0.3577898442745209, + "learning_rate": 2.6656131926933066e-05, + "loss": 0.0084, + "step": 67190 + }, + { + "epoch": 5.2194174757281555, + "grad_norm": 0.387012243270874, + "learning_rate": 2.664151120505136e-05, + "loss": 0.0044, + "step": 67200 + }, + { + "epoch": 5.220194174757282, + "grad_norm": 0.48462244868278503, + "learning_rate": 2.6626893037618873e-05, + "loss": 0.0056, + "step": 67210 + }, + { + "epoch": 5.220970873786408, + "grad_norm": 0.19576966762542725, + "learning_rate": 2.6612277426234243e-05, + "loss": 0.0053, + "step": 67220 + }, + { + "epoch": 5.221747572815534, + "grad_norm": 0.18166705965995789, + "learning_rate": 2.659766437249577e-05, + "loss": 0.0067, + "step": 67230 + }, + { + "epoch": 5.2225242718446605, + "grad_norm": 0.1724337786436081, + "learning_rate": 2.6583053878001557e-05, + "loss": 0.0058, + "step": 67240 + }, + { + "epoch": 5.223300970873787, + "grad_norm": 0.2689741253852844, + "learning_rate": 2.6568445944349323e-05, + "loss": 0.0064, + "step": 67250 + }, + { + "epoch": 5.224077669902913, + "grad_norm": 0.1914588361978531, + "learning_rate": 2.6553840573136657e-05, + "loss": 0.006, + "step": 67260 + }, + { + "epoch": 5.224854368932039, + "grad_norm": 0.13486050069332123, + "learning_rate": 2.6539237765960706e-05, + "loss": 0.0035, + "step": 67270 + }, + { + "epoch": 5.2256310679611655, + "grad_norm": 0.27905502915382385, + "learning_rate": 2.6524637524418465e-05, + "loss": 0.0067, + "step": 67280 + }, + { + "epoch": 5.226407766990291, + "grad_norm": 0.3120648264884949, + "learning_rate": 2.651003985010655e-05, + "loss": 0.007, + "step": 67290 + }, + { + "epoch": 5.227184466019417, + "grad_norm": 0.2290385216474533, + "learning_rate": 2.649544474462138e-05, + "loss": 0.0087, + "step": 67300 + }, + { + "epoch": 5.227961165048543, + "grad_norm": 0.4657120108604431, + "learning_rate": 2.6480852209559015e-05, + "loss": 0.0061, + "step": 67310 + }, + { + "epoch": 5.22873786407767, + "grad_norm": 0.2402624636888504, + "learning_rate": 2.6466262246515296e-05, + "loss": 0.0058, + "step": 67320 + }, + { + "epoch": 5.229514563106796, + "grad_norm": 0.16460859775543213, + "learning_rate": 2.6451674857085746e-05, + "loss": 0.0038, + "step": 67330 + }, + { + "epoch": 5.230291262135922, + "grad_norm": 0.3295614719390869, + "learning_rate": 2.6437090042865655e-05, + "loss": 0.0057, + "step": 67340 + }, + { + "epoch": 5.231067961165048, + "grad_norm": 1.3794044256210327, + "learning_rate": 2.6422507805449947e-05, + "loss": 0.0048, + "step": 67350 + }, + { + "epoch": 5.231844660194175, + "grad_norm": 0.26089146733283997, + "learning_rate": 2.6407928146433357e-05, + "loss": 0.0055, + "step": 67360 + }, + { + "epoch": 5.232621359223301, + "grad_norm": 0.33795273303985596, + "learning_rate": 2.639335106741026e-05, + "loss": 0.0053, + "step": 67370 + }, + { + "epoch": 5.233398058252427, + "grad_norm": 0.06765572726726532, + "learning_rate": 2.6378776569974796e-05, + "loss": 0.005, + "step": 67380 + }, + { + "epoch": 5.234174757281553, + "grad_norm": 0.1932181864976883, + "learning_rate": 2.6364204655720814e-05, + "loss": 0.0078, + "step": 67390 + }, + { + "epoch": 5.23495145631068, + "grad_norm": 0.14801423251628876, + "learning_rate": 2.6349635326241894e-05, + "loss": 0.0054, + "step": 67400 + }, + { + "epoch": 5.235728155339806, + "grad_norm": 0.5559052228927612, + "learning_rate": 2.633506858313127e-05, + "loss": 0.0057, + "step": 67410 + }, + { + "epoch": 5.236504854368932, + "grad_norm": 0.5765283107757568, + "learning_rate": 2.6320504427981975e-05, + "loss": 0.0052, + "step": 67420 + }, + { + "epoch": 5.237281553398058, + "grad_norm": 0.25457563996315, + "learning_rate": 2.6305942862386734e-05, + "loss": 0.0051, + "step": 67430 + }, + { + "epoch": 5.238058252427185, + "grad_norm": 0.262410968542099, + "learning_rate": 2.629138388793793e-05, + "loss": 0.0054, + "step": 67440 + }, + { + "epoch": 5.238834951456311, + "grad_norm": 0.17519822716712952, + "learning_rate": 2.6276827506227743e-05, + "loss": 0.0047, + "step": 67450 + }, + { + "epoch": 5.239611650485437, + "grad_norm": 0.23752981424331665, + "learning_rate": 2.6262273718848037e-05, + "loss": 0.0035, + "step": 67460 + }, + { + "epoch": 5.240388349514563, + "grad_norm": 0.11757592111825943, + "learning_rate": 2.6247722527390407e-05, + "loss": 0.0039, + "step": 67470 + }, + { + "epoch": 5.24116504854369, + "grad_norm": 0.13382206857204437, + "learning_rate": 2.6233173933446104e-05, + "loss": 0.0043, + "step": 67480 + }, + { + "epoch": 5.241941747572816, + "grad_norm": 0.3878365457057953, + "learning_rate": 2.6218627938606188e-05, + "loss": 0.0064, + "step": 67490 + }, + { + "epoch": 5.242718446601942, + "grad_norm": 0.19256079196929932, + "learning_rate": 2.620408454446135e-05, + "loss": 0.0045, + "step": 67500 + }, + { + "epoch": 5.243495145631068, + "grad_norm": 0.1586894690990448, + "learning_rate": 2.6189543752602046e-05, + "loss": 0.0045, + "step": 67510 + }, + { + "epoch": 5.244271844660195, + "grad_norm": 0.3154642581939697, + "learning_rate": 2.6175005564618444e-05, + "loss": 0.0045, + "step": 67520 + }, + { + "epoch": 5.24504854368932, + "grad_norm": 0.3271494209766388, + "learning_rate": 2.6160469982100428e-05, + "loss": 0.0046, + "step": 67530 + }, + { + "epoch": 5.245825242718446, + "grad_norm": 0.11044799536466599, + "learning_rate": 2.6145937006637555e-05, + "loss": 0.0059, + "step": 67540 + }, + { + "epoch": 5.246601941747572, + "grad_norm": 0.22239339351654053, + "learning_rate": 2.6131406639819166e-05, + "loss": 0.0068, + "step": 67550 + }, + { + "epoch": 5.247378640776699, + "grad_norm": 0.3262022137641907, + "learning_rate": 2.611687888323424e-05, + "loss": 0.0069, + "step": 67560 + }, + { + "epoch": 5.248155339805825, + "grad_norm": 0.5539337992668152, + "learning_rate": 2.610235373847153e-05, + "loss": 0.0057, + "step": 67570 + }, + { + "epoch": 5.248932038834951, + "grad_norm": 0.40186697244644165, + "learning_rate": 2.6087831207119494e-05, + "loss": 0.0046, + "step": 67580 + }, + { + "epoch": 5.249708737864077, + "grad_norm": 0.5764353275299072, + "learning_rate": 2.60733112907663e-05, + "loss": 0.0061, + "step": 67590 + }, + { + "epoch": 5.250485436893204, + "grad_norm": 0.27746298909187317, + "learning_rate": 2.6058793990999786e-05, + "loss": 0.0041, + "step": 67600 + }, + { + "epoch": 5.25126213592233, + "grad_norm": 0.16203586757183075, + "learning_rate": 2.6044279309407593e-05, + "loss": 0.0061, + "step": 67610 + }, + { + "epoch": 5.252038834951456, + "grad_norm": 0.3081413209438324, + "learning_rate": 2.6029767247576976e-05, + "loss": 0.0052, + "step": 67620 + }, + { + "epoch": 5.252815533980582, + "grad_norm": 0.14568807184696198, + "learning_rate": 2.601525780709497e-05, + "loss": 0.0045, + "step": 67630 + }, + { + "epoch": 5.253592233009709, + "grad_norm": 0.18913033604621887, + "learning_rate": 2.6000750989548318e-05, + "loss": 0.0035, + "step": 67640 + }, + { + "epoch": 5.254368932038835, + "grad_norm": 0.4879782795906067, + "learning_rate": 2.5986246796523462e-05, + "loss": 0.006, + "step": 67650 + }, + { + "epoch": 5.255145631067961, + "grad_norm": 0.18423134088516235, + "learning_rate": 2.5971745229606546e-05, + "loss": 0.0044, + "step": 67660 + }, + { + "epoch": 5.255922330097087, + "grad_norm": 0.11380894482135773, + "learning_rate": 2.5957246290383442e-05, + "loss": 0.0073, + "step": 67670 + }, + { + "epoch": 5.256699029126214, + "grad_norm": 0.12035533785820007, + "learning_rate": 2.594274998043975e-05, + "loss": 0.0045, + "step": 67680 + }, + { + "epoch": 5.25747572815534, + "grad_norm": 0.24814404547214508, + "learning_rate": 2.5928256301360733e-05, + "loss": 0.0054, + "step": 67690 + }, + { + "epoch": 5.258252427184466, + "grad_norm": 0.1520681530237198, + "learning_rate": 2.591376525473142e-05, + "loss": 0.006, + "step": 67700 + }, + { + "epoch": 5.259029126213592, + "grad_norm": 0.23260729014873505, + "learning_rate": 2.5899276842136523e-05, + "loss": 0.0054, + "step": 67710 + }, + { + "epoch": 5.259805825242719, + "grad_norm": 0.23548920452594757, + "learning_rate": 2.5884791065160495e-05, + "loss": 0.0052, + "step": 67720 + }, + { + "epoch": 5.260582524271845, + "grad_norm": 0.13839133083820343, + "learning_rate": 2.5870307925387436e-05, + "loss": 0.0076, + "step": 67730 + }, + { + "epoch": 5.261359223300971, + "grad_norm": 0.14978012442588806, + "learning_rate": 2.585582742440125e-05, + "loss": 0.0071, + "step": 67740 + }, + { + "epoch": 5.262135922330097, + "grad_norm": 0.18313923478126526, + "learning_rate": 2.5841349563785455e-05, + "loss": 0.0053, + "step": 67750 + }, + { + "epoch": 5.262912621359224, + "grad_norm": 0.15108618140220642, + "learning_rate": 2.5826874345123355e-05, + "loss": 0.0047, + "step": 67760 + }, + { + "epoch": 5.26368932038835, + "grad_norm": 0.11218876391649246, + "learning_rate": 2.581240176999792e-05, + "loss": 0.0074, + "step": 67770 + }, + { + "epoch": 5.264466019417476, + "grad_norm": 0.33212989568710327, + "learning_rate": 2.579793183999189e-05, + "loss": 0.0072, + "step": 67780 + }, + { + "epoch": 5.265242718446602, + "grad_norm": 0.5102304220199585, + "learning_rate": 2.5783464556687627e-05, + "loss": 0.0074, + "step": 67790 + }, + { + "epoch": 5.266019417475728, + "grad_norm": 0.43102848529815674, + "learning_rate": 2.576899992166729e-05, + "loss": 0.0055, + "step": 67800 + }, + { + "epoch": 5.266796116504854, + "grad_norm": 0.16706013679504395, + "learning_rate": 2.5754537936512668e-05, + "loss": 0.006, + "step": 67810 + }, + { + "epoch": 5.26757281553398, + "grad_norm": 0.25854459404945374, + "learning_rate": 2.574007860280535e-05, + "loss": 0.0051, + "step": 67820 + }, + { + "epoch": 5.2683495145631065, + "grad_norm": 0.22153495252132416, + "learning_rate": 2.5725621922126518e-05, + "loss": 0.0035, + "step": 67830 + }, + { + "epoch": 5.269126213592233, + "grad_norm": 0.12906353175640106, + "learning_rate": 2.571116789605722e-05, + "loss": 0.0065, + "step": 67840 + }, + { + "epoch": 5.269902912621359, + "grad_norm": 0.13617254793643951, + "learning_rate": 2.5696716526178054e-05, + "loss": 0.0038, + "step": 67850 + }, + { + "epoch": 5.270679611650485, + "grad_norm": 0.19624309241771698, + "learning_rate": 2.568226781406946e-05, + "loss": 0.0085, + "step": 67860 + }, + { + "epoch": 5.2714563106796115, + "grad_norm": 0.3237672448158264, + "learning_rate": 2.5667821761311474e-05, + "loss": 0.005, + "step": 67870 + }, + { + "epoch": 5.272233009708738, + "grad_norm": 0.45458391308784485, + "learning_rate": 2.5653378369483917e-05, + "loss": 0.0075, + "step": 67880 + }, + { + "epoch": 5.273009708737864, + "grad_norm": 0.14424635469913483, + "learning_rate": 2.5638937640166292e-05, + "loss": 0.004, + "step": 67890 + }, + { + "epoch": 5.27378640776699, + "grad_norm": 0.1507367044687271, + "learning_rate": 2.5624499574937842e-05, + "loss": 0.004, + "step": 67900 + }, + { + "epoch": 5.2745631067961165, + "grad_norm": 0.20183104276657104, + "learning_rate": 2.5610064175377456e-05, + "loss": 0.0046, + "step": 67910 + }, + { + "epoch": 5.275339805825243, + "grad_norm": 0.1713922917842865, + "learning_rate": 2.5595631443063782e-05, + "loss": 0.0054, + "step": 67920 + }, + { + "epoch": 5.276116504854369, + "grad_norm": 0.3662252426147461, + "learning_rate": 2.5581201379575177e-05, + "loss": 0.0048, + "step": 67930 + }, + { + "epoch": 5.276893203883495, + "grad_norm": 0.24092619121074677, + "learning_rate": 2.5566773986489655e-05, + "loss": 0.0049, + "step": 67940 + }, + { + "epoch": 5.2776699029126215, + "grad_norm": 0.12567515671253204, + "learning_rate": 2.5552349265385e-05, + "loss": 0.0065, + "step": 67950 + }, + { + "epoch": 5.278446601941748, + "grad_norm": 0.16103863716125488, + "learning_rate": 2.5537927217838675e-05, + "loss": 0.0065, + "step": 67960 + }, + { + "epoch": 5.279223300970874, + "grad_norm": 0.1708724945783615, + "learning_rate": 2.5523507845427864e-05, + "loss": 0.0041, + "step": 67970 + }, + { + "epoch": 5.28, + "grad_norm": 0.19287757575511932, + "learning_rate": 2.5509091149729413e-05, + "loss": 0.0035, + "step": 67980 + }, + { + "epoch": 5.2807766990291265, + "grad_norm": 0.2899777591228485, + "learning_rate": 2.5494677132319962e-05, + "loss": 0.0039, + "step": 67990 + }, + { + "epoch": 5.281553398058253, + "grad_norm": 0.18333810567855835, + "learning_rate": 2.548026579477575e-05, + "loss": 0.003, + "step": 68000 + }, + { + "epoch": 5.282330097087379, + "grad_norm": 0.4173068106174469, + "learning_rate": 2.5465857138672834e-05, + "loss": 0.0054, + "step": 68010 + }, + { + "epoch": 5.283106796116505, + "grad_norm": 0.19196118414402008, + "learning_rate": 2.5451451165586848e-05, + "loss": 0.0082, + "step": 68020 + }, + { + "epoch": 5.283883495145631, + "grad_norm": 0.1341579258441925, + "learning_rate": 2.543704787709329e-05, + "loss": 0.0049, + "step": 68030 + }, + { + "epoch": 5.284660194174757, + "grad_norm": 0.0940774530172348, + "learning_rate": 2.542264727476723e-05, + "loss": 0.0042, + "step": 68040 + }, + { + "epoch": 5.285436893203883, + "grad_norm": 0.1998567134141922, + "learning_rate": 2.5408249360183535e-05, + "loss": 0.0049, + "step": 68050 + }, + { + "epoch": 5.286213592233009, + "grad_norm": 0.18461202085018158, + "learning_rate": 2.5393854134916694e-05, + "loss": 0.0043, + "step": 68060 + }, + { + "epoch": 5.286990291262136, + "grad_norm": 0.40562281012535095, + "learning_rate": 2.5379461600540993e-05, + "loss": 0.0053, + "step": 68070 + }, + { + "epoch": 5.287766990291262, + "grad_norm": 0.5022984147071838, + "learning_rate": 2.5365071758630317e-05, + "loss": 0.0046, + "step": 68080 + }, + { + "epoch": 5.288543689320388, + "grad_norm": 0.31451094150543213, + "learning_rate": 2.535068461075839e-05, + "loss": 0.0059, + "step": 68090 + }, + { + "epoch": 5.289320388349514, + "grad_norm": 0.3060912787914276, + "learning_rate": 2.5336300158498516e-05, + "loss": 0.0042, + "step": 68100 + }, + { + "epoch": 5.290097087378641, + "grad_norm": 0.46982982754707336, + "learning_rate": 2.5321918403423793e-05, + "loss": 0.0052, + "step": 68110 + }, + { + "epoch": 5.290873786407767, + "grad_norm": 0.3842516243457794, + "learning_rate": 2.5307539347106957e-05, + "loss": 0.0043, + "step": 68120 + }, + { + "epoch": 5.291650485436893, + "grad_norm": 0.2430967539548874, + "learning_rate": 2.529316299112049e-05, + "loss": 0.0065, + "step": 68130 + }, + { + "epoch": 5.292427184466019, + "grad_norm": 0.08942961692810059, + "learning_rate": 2.5278789337036567e-05, + "loss": 0.006, + "step": 68140 + }, + { + "epoch": 5.293203883495146, + "grad_norm": 0.42598456144332886, + "learning_rate": 2.52644183864271e-05, + "loss": 0.0058, + "step": 68150 + }, + { + "epoch": 5.293980582524272, + "grad_norm": 0.24050851166248322, + "learning_rate": 2.525005014086363e-05, + "loss": 0.005, + "step": 68160 + }, + { + "epoch": 5.294757281553398, + "grad_norm": 0.2131638377904892, + "learning_rate": 2.5235684601917465e-05, + "loss": 0.0061, + "step": 68170 + }, + { + "epoch": 5.295533980582524, + "grad_norm": 0.41373658180236816, + "learning_rate": 2.5221321771159618e-05, + "loss": 0.0061, + "step": 68180 + }, + { + "epoch": 5.296310679611651, + "grad_norm": 0.18816736340522766, + "learning_rate": 2.5206961650160743e-05, + "loss": 0.0038, + "step": 68190 + }, + { + "epoch": 5.297087378640777, + "grad_norm": 0.1917935609817505, + "learning_rate": 2.519260424049129e-05, + "loss": 0.0074, + "step": 68200 + }, + { + "epoch": 5.297864077669903, + "grad_norm": 0.1858798861503601, + "learning_rate": 2.5178249543721304e-05, + "loss": 0.006, + "step": 68210 + }, + { + "epoch": 5.298640776699029, + "grad_norm": 0.27982816100120544, + "learning_rate": 2.5163897561420668e-05, + "loss": 0.0041, + "step": 68220 + }, + { + "epoch": 5.299417475728156, + "grad_norm": 0.12579990923404694, + "learning_rate": 2.514954829515883e-05, + "loss": 0.0047, + "step": 68230 + }, + { + "epoch": 5.300194174757282, + "grad_norm": 0.23802831768989563, + "learning_rate": 2.513520174650505e-05, + "loss": 0.0041, + "step": 68240 + }, + { + "epoch": 5.300970873786408, + "grad_norm": 0.3035009503364563, + "learning_rate": 2.5120857917028207e-05, + "loss": 0.0059, + "step": 68250 + }, + { + "epoch": 5.301747572815534, + "grad_norm": 0.42649728059768677, + "learning_rate": 2.5106516808296953e-05, + "loss": 0.0052, + "step": 68260 + }, + { + "epoch": 5.302524271844661, + "grad_norm": 0.25918659567832947, + "learning_rate": 2.5092178421879552e-05, + "loss": 0.0048, + "step": 68270 + }, + { + "epoch": 5.303300970873787, + "grad_norm": 0.3019393980503082, + "learning_rate": 2.5077842759344116e-05, + "loss": 0.0067, + "step": 68280 + }, + { + "epoch": 5.304077669902913, + "grad_norm": 0.19844602048397064, + "learning_rate": 2.5063509822258304e-05, + "loss": 0.005, + "step": 68290 + }, + { + "epoch": 5.304854368932038, + "grad_norm": 0.23442521691322327, + "learning_rate": 2.5049179612189576e-05, + "loss": 0.0047, + "step": 68300 + }, + { + "epoch": 5.305631067961165, + "grad_norm": 0.1099926009774208, + "learning_rate": 2.5034852130705043e-05, + "loss": 0.0055, + "step": 68310 + }, + { + "epoch": 5.306407766990291, + "grad_norm": 0.14614823460578918, + "learning_rate": 2.5020527379371572e-05, + "loss": 0.0047, + "step": 68320 + }, + { + "epoch": 5.307184466019417, + "grad_norm": 0.21105027198791504, + "learning_rate": 2.500620535975564e-05, + "loss": 0.0051, + "step": 68330 + }, + { + "epoch": 5.307961165048543, + "grad_norm": 0.2451939433813095, + "learning_rate": 2.4991886073423515e-05, + "loss": 0.0039, + "step": 68340 + }, + { + "epoch": 5.30873786407767, + "grad_norm": 0.121425099670887, + "learning_rate": 2.4977569521941134e-05, + "loss": 0.0057, + "step": 68350 + }, + { + "epoch": 5.309514563106796, + "grad_norm": 0.24050210416316986, + "learning_rate": 2.4963255706874143e-05, + "loss": 0.0056, + "step": 68360 + }, + { + "epoch": 5.310291262135922, + "grad_norm": 0.21604229509830475, + "learning_rate": 2.4948944629787853e-05, + "loss": 0.0063, + "step": 68370 + }, + { + "epoch": 5.311067961165048, + "grad_norm": 0.2867068350315094, + "learning_rate": 2.493463629224731e-05, + "loss": 0.0041, + "step": 68380 + }, + { + "epoch": 5.311844660194175, + "grad_norm": 0.13668043911457062, + "learning_rate": 2.492033069581728e-05, + "loss": 0.0036, + "step": 68390 + }, + { + "epoch": 5.312621359223301, + "grad_norm": 0.10698603093624115, + "learning_rate": 2.4906027842062137e-05, + "loss": 0.006, + "step": 68400 + }, + { + "epoch": 5.313398058252427, + "grad_norm": 0.3103550374507904, + "learning_rate": 2.48917277325461e-05, + "loss": 0.0051, + "step": 68410 + }, + { + "epoch": 5.314174757281553, + "grad_norm": 0.22898349165916443, + "learning_rate": 2.487743036883295e-05, + "loss": 0.0059, + "step": 68420 + }, + { + "epoch": 5.31495145631068, + "grad_norm": 0.2214798927307129, + "learning_rate": 2.486313575248626e-05, + "loss": 0.0057, + "step": 68430 + }, + { + "epoch": 5.315728155339806, + "grad_norm": 0.14297634363174438, + "learning_rate": 2.4848843885069235e-05, + "loss": 0.0055, + "step": 68440 + }, + { + "epoch": 5.316504854368932, + "grad_norm": 0.17969073355197906, + "learning_rate": 2.483455476814484e-05, + "loss": 0.0055, + "step": 68450 + }, + { + "epoch": 5.317281553398058, + "grad_norm": 0.2046515792608261, + "learning_rate": 2.4820268403275672e-05, + "loss": 0.008, + "step": 68460 + }, + { + "epoch": 5.318058252427185, + "grad_norm": 0.3102700412273407, + "learning_rate": 2.480598479202413e-05, + "loss": 0.0068, + "step": 68470 + }, + { + "epoch": 5.318834951456311, + "grad_norm": 0.3010816276073456, + "learning_rate": 2.4791703935952193e-05, + "loss": 0.0042, + "step": 68480 + }, + { + "epoch": 5.319611650485437, + "grad_norm": 0.8435407876968384, + "learning_rate": 2.477742583662163e-05, + "loss": 0.0043, + "step": 68490 + }, + { + "epoch": 5.320388349514563, + "grad_norm": 0.11772625893354416, + "learning_rate": 2.476315049559384e-05, + "loss": 0.0046, + "step": 68500 + }, + { + "epoch": 5.32116504854369, + "grad_norm": 0.1835699826478958, + "learning_rate": 2.4748877914429987e-05, + "loss": 0.0046, + "step": 68510 + }, + { + "epoch": 5.321941747572816, + "grad_norm": 0.1072593629360199, + "learning_rate": 2.4734608094690857e-05, + "loss": 0.005, + "step": 68520 + }, + { + "epoch": 5.322718446601942, + "grad_norm": 0.0888114720582962, + "learning_rate": 2.4720341037937006e-05, + "loss": 0.0054, + "step": 68530 + }, + { + "epoch": 5.3234951456310675, + "grad_norm": 0.2517341375350952, + "learning_rate": 2.4706076745728645e-05, + "loss": 0.0054, + "step": 68540 + }, + { + "epoch": 5.324271844660194, + "grad_norm": 0.1817018836736679, + "learning_rate": 2.469181521962573e-05, + "loss": 0.0047, + "step": 68550 + }, + { + "epoch": 5.32504854368932, + "grad_norm": 0.2822929620742798, + "learning_rate": 2.467755646118783e-05, + "loss": 0.0071, + "step": 68560 + }, + { + "epoch": 5.325825242718446, + "grad_norm": 0.15905137360095978, + "learning_rate": 2.46633004719743e-05, + "loss": 0.0046, + "step": 68570 + }, + { + "epoch": 5.3266019417475725, + "grad_norm": 0.2672750651836395, + "learning_rate": 2.4649047253544123e-05, + "loss": 0.0042, + "step": 68580 + }, + { + "epoch": 5.327378640776699, + "grad_norm": 0.19799982011318207, + "learning_rate": 2.4634796807456016e-05, + "loss": 0.0056, + "step": 68590 + }, + { + "epoch": 5.328155339805825, + "grad_norm": 0.3037695288658142, + "learning_rate": 2.46205491352684e-05, + "loss": 0.004, + "step": 68600 + }, + { + "epoch": 5.328932038834951, + "grad_norm": 0.16194544732570648, + "learning_rate": 2.4606304238539375e-05, + "loss": 0.0049, + "step": 68610 + }, + { + "epoch": 5.3297087378640775, + "grad_norm": 0.17186446487903595, + "learning_rate": 2.4592062118826753e-05, + "loss": 0.0082, + "step": 68620 + }, + { + "epoch": 5.330485436893204, + "grad_norm": 0.4566745162010193, + "learning_rate": 2.4577822777688004e-05, + "loss": 0.0054, + "step": 68630 + }, + { + "epoch": 5.33126213592233, + "grad_norm": 0.2337978333234787, + "learning_rate": 2.4563586216680347e-05, + "loss": 0.0052, + "step": 68640 + }, + { + "epoch": 5.332038834951456, + "grad_norm": 0.21853294968605042, + "learning_rate": 2.4549352437360622e-05, + "loss": 0.0082, + "step": 68650 + }, + { + "epoch": 5.3328155339805825, + "grad_norm": 0.4548909664154053, + "learning_rate": 2.4535121441285493e-05, + "loss": 0.0062, + "step": 68660 + }, + { + "epoch": 5.333592233009709, + "grad_norm": 0.3274337947368622, + "learning_rate": 2.4520893230011172e-05, + "loss": 0.0073, + "step": 68670 + }, + { + "epoch": 5.334368932038835, + "grad_norm": 0.2711142897605896, + "learning_rate": 2.450666780509368e-05, + "loss": 0.0059, + "step": 68680 + }, + { + "epoch": 5.335145631067961, + "grad_norm": 0.42139849066734314, + "learning_rate": 2.4492445168088656e-05, + "loss": 0.0035, + "step": 68690 + }, + { + "epoch": 5.3359223300970875, + "grad_norm": 0.3084416389465332, + "learning_rate": 2.447822532055149e-05, + "loss": 0.0107, + "step": 68700 + }, + { + "epoch": 5.336699029126214, + "grad_norm": 0.8697035908699036, + "learning_rate": 2.4464008264037212e-05, + "loss": 0.005, + "step": 68710 + }, + { + "epoch": 5.33747572815534, + "grad_norm": 0.5747854113578796, + "learning_rate": 2.4449794000100605e-05, + "loss": 0.0048, + "step": 68720 + }, + { + "epoch": 5.338252427184466, + "grad_norm": 0.27234840393066406, + "learning_rate": 2.4435582530296108e-05, + "loss": 0.0062, + "step": 68730 + }, + { + "epoch": 5.3390291262135925, + "grad_norm": 0.16411790251731873, + "learning_rate": 2.4421373856177887e-05, + "loss": 0.0056, + "step": 68740 + }, + { + "epoch": 5.339805825242719, + "grad_norm": 0.30576834082603455, + "learning_rate": 2.440716797929975e-05, + "loss": 0.0033, + "step": 68750 + }, + { + "epoch": 5.340582524271845, + "grad_norm": 0.2692307233810425, + "learning_rate": 2.439296490121526e-05, + "loss": 0.0059, + "step": 68760 + }, + { + "epoch": 5.341359223300971, + "grad_norm": 0.5596547722816467, + "learning_rate": 2.4378764623477608e-05, + "loss": 0.0071, + "step": 68770 + }, + { + "epoch": 5.3421359223300975, + "grad_norm": 0.13437488675117493, + "learning_rate": 2.436456714763974e-05, + "loss": 0.0051, + "step": 68780 + }, + { + "epoch": 5.342912621359224, + "grad_norm": 0.3936631381511688, + "learning_rate": 2.4350372475254262e-05, + "loss": 0.0052, + "step": 68790 + }, + { + "epoch": 5.34368932038835, + "grad_norm": 0.172512486577034, + "learning_rate": 2.433618060787351e-05, + "loss": 0.006, + "step": 68800 + }, + { + "epoch": 5.344466019417475, + "grad_norm": 0.19240637123584747, + "learning_rate": 2.432199154704944e-05, + "loss": 0.0071, + "step": 68810 + }, + { + "epoch": 5.345242718446602, + "grad_norm": 0.12289781868457794, + "learning_rate": 2.430780529433377e-05, + "loss": 0.0038, + "step": 68820 + }, + { + "epoch": 5.346019417475728, + "grad_norm": 0.18574056029319763, + "learning_rate": 2.42936218512779e-05, + "loss": 0.0053, + "step": 68830 + }, + { + "epoch": 5.346796116504854, + "grad_norm": 0.08879240602254868, + "learning_rate": 2.427944121943288e-05, + "loss": 0.0039, + "step": 68840 + }, + { + "epoch": 5.34757281553398, + "grad_norm": 0.05412331223487854, + "learning_rate": 2.4265263400349497e-05, + "loss": 0.0033, + "step": 68850 + }, + { + "epoch": 5.348349514563107, + "grad_norm": 0.3115653097629547, + "learning_rate": 2.4251088395578213e-05, + "loss": 0.005, + "step": 68860 + }, + { + "epoch": 5.349126213592233, + "grad_norm": 0.19314172863960266, + "learning_rate": 2.4236916206669214e-05, + "loss": 0.0042, + "step": 68870 + }, + { + "epoch": 5.349902912621359, + "grad_norm": 0.795753002166748, + "learning_rate": 2.4222746835172304e-05, + "loss": 0.0068, + "step": 68880 + }, + { + "epoch": 5.350679611650485, + "grad_norm": 0.3479883074760437, + "learning_rate": 2.4208580282637065e-05, + "loss": 0.0058, + "step": 68890 + }, + { + "epoch": 5.351456310679612, + "grad_norm": 0.35071176290512085, + "learning_rate": 2.4194416550612687e-05, + "loss": 0.004, + "step": 68900 + }, + { + "epoch": 5.352233009708738, + "grad_norm": 0.514319896697998, + "learning_rate": 2.418025564064812e-05, + "loss": 0.004, + "step": 68910 + }, + { + "epoch": 5.353009708737864, + "grad_norm": 0.19345536828041077, + "learning_rate": 2.4166097554291978e-05, + "loss": 0.0057, + "step": 68920 + }, + { + "epoch": 5.35378640776699, + "grad_norm": 0.23811201751232147, + "learning_rate": 2.4151942293092583e-05, + "loss": 0.0042, + "step": 68930 + }, + { + "epoch": 5.354563106796117, + "grad_norm": 0.17048195004463196, + "learning_rate": 2.41377898585979e-05, + "loss": 0.0042, + "step": 68940 + }, + { + "epoch": 5.355339805825243, + "grad_norm": 0.2694801688194275, + "learning_rate": 2.4123640252355652e-05, + "loss": 0.0048, + "step": 68950 + }, + { + "epoch": 5.356116504854369, + "grad_norm": 0.12789778411388397, + "learning_rate": 2.410949347591318e-05, + "loss": 0.0074, + "step": 68960 + }, + { + "epoch": 5.356893203883495, + "grad_norm": 0.09355945885181427, + "learning_rate": 2.4095349530817578e-05, + "loss": 0.0053, + "step": 68970 + }, + { + "epoch": 5.357669902912622, + "grad_norm": 0.5705152750015259, + "learning_rate": 2.4081208418615603e-05, + "loss": 0.0051, + "step": 68980 + }, + { + "epoch": 5.358446601941748, + "grad_norm": 0.17699646949768066, + "learning_rate": 2.4067070140853725e-05, + "loss": 0.0051, + "step": 68990 + }, + { + "epoch": 5.359223300970874, + "grad_norm": 0.896060585975647, + "learning_rate": 2.405293469907805e-05, + "loss": 0.0042, + "step": 69000 + }, + { + "epoch": 5.36, + "grad_norm": 0.2208680510520935, + "learning_rate": 2.4038802094834446e-05, + "loss": 0.0058, + "step": 69010 + }, + { + "epoch": 5.360776699029127, + "grad_norm": 0.2894987165927887, + "learning_rate": 2.4024672329668395e-05, + "loss": 0.004, + "step": 69020 + }, + { + "epoch": 5.361553398058253, + "grad_norm": 0.13922329246997833, + "learning_rate": 2.4010545405125125e-05, + "loss": 0.0049, + "step": 69030 + }, + { + "epoch": 5.362330097087378, + "grad_norm": 0.2999364733695984, + "learning_rate": 2.3996421322749528e-05, + "loss": 0.0055, + "step": 69040 + }, + { + "epoch": 5.363106796116504, + "grad_norm": 0.26416897773742676, + "learning_rate": 2.3982300084086224e-05, + "loss": 0.0063, + "step": 69050 + }, + { + "epoch": 5.363883495145631, + "grad_norm": 0.1570153385400772, + "learning_rate": 2.396818169067944e-05, + "loss": 0.0057, + "step": 69060 + }, + { + "epoch": 5.364660194174757, + "grad_norm": 0.28986525535583496, + "learning_rate": 2.3954066144073178e-05, + "loss": 0.0063, + "step": 69070 + }, + { + "epoch": 5.365436893203883, + "grad_norm": 0.3596283793449402, + "learning_rate": 2.39399534458111e-05, + "loss": 0.0037, + "step": 69080 + }, + { + "epoch": 5.366213592233009, + "grad_norm": 0.15851671993732452, + "learning_rate": 2.392584359743651e-05, + "loss": 0.0058, + "step": 69090 + }, + { + "epoch": 5.366990291262136, + "grad_norm": 0.1993558555841446, + "learning_rate": 2.3911736600492463e-05, + "loss": 0.0063, + "step": 69100 + }, + { + "epoch": 5.367766990291262, + "grad_norm": 0.13544782996177673, + "learning_rate": 2.3897632456521672e-05, + "loss": 0.0036, + "step": 69110 + }, + { + "epoch": 5.368543689320388, + "grad_norm": 0.36048611998558044, + "learning_rate": 2.388353116706658e-05, + "loss": 0.0054, + "step": 69120 + }, + { + "epoch": 5.369320388349514, + "grad_norm": 0.12081186473369598, + "learning_rate": 2.3869432733669232e-05, + "loss": 0.0053, + "step": 69130 + }, + { + "epoch": 5.370097087378641, + "grad_norm": 0.15675202012062073, + "learning_rate": 2.385533715787145e-05, + "loss": 0.0044, + "step": 69140 + }, + { + "epoch": 5.370873786407767, + "grad_norm": 0.19705137610435486, + "learning_rate": 2.3841244441214673e-05, + "loss": 0.0056, + "step": 69150 + }, + { + "epoch": 5.371650485436893, + "grad_norm": 0.14596675336360931, + "learning_rate": 2.3827154585240073e-05, + "loss": 0.0048, + "step": 69160 + }, + { + "epoch": 5.372427184466019, + "grad_norm": 0.25123322010040283, + "learning_rate": 2.3813067591488498e-05, + "loss": 0.0069, + "step": 69170 + }, + { + "epoch": 5.373203883495146, + "grad_norm": 0.17266088724136353, + "learning_rate": 2.3798983461500494e-05, + "loss": 0.0048, + "step": 69180 + }, + { + "epoch": 5.373980582524272, + "grad_norm": 0.4239639341831207, + "learning_rate": 2.3784902196816256e-05, + "loss": 0.0046, + "step": 69190 + }, + { + "epoch": 5.374757281553398, + "grad_norm": 0.19631187617778778, + "learning_rate": 2.3770823798975716e-05, + "loss": 0.0049, + "step": 69200 + }, + { + "epoch": 5.375533980582524, + "grad_norm": 0.22496894001960754, + "learning_rate": 2.3756748269518424e-05, + "loss": 0.0044, + "step": 69210 + }, + { + "epoch": 5.376310679611651, + "grad_norm": 0.21801863610744476, + "learning_rate": 2.374267560998371e-05, + "loss": 0.0054, + "step": 69220 + }, + { + "epoch": 5.377087378640777, + "grad_norm": 0.22429032623767853, + "learning_rate": 2.3728605821910477e-05, + "loss": 0.0045, + "step": 69230 + }, + { + "epoch": 5.377864077669903, + "grad_norm": 0.4239775836467743, + "learning_rate": 2.3714538906837453e-05, + "loss": 0.0053, + "step": 69240 + }, + { + "epoch": 5.378640776699029, + "grad_norm": 0.3183400332927704, + "learning_rate": 2.3700474866302913e-05, + "loss": 0.0044, + "step": 69250 + }, + { + "epoch": 5.379417475728156, + "grad_norm": 0.1607092022895813, + "learning_rate": 2.3686413701844917e-05, + "loss": 0.0048, + "step": 69260 + }, + { + "epoch": 5.380194174757282, + "grad_norm": 0.16571155190467834, + "learning_rate": 2.367235541500114e-05, + "loss": 0.004, + "step": 69270 + }, + { + "epoch": 5.380970873786408, + "grad_norm": 0.9006075263023376, + "learning_rate": 2.3658300007308993e-05, + "loss": 0.0053, + "step": 69280 + }, + { + "epoch": 5.381747572815534, + "grad_norm": 0.41143423318862915, + "learning_rate": 2.364424748030555e-05, + "loss": 0.0051, + "step": 69290 + }, + { + "epoch": 5.382524271844661, + "grad_norm": 0.23739153146743774, + "learning_rate": 2.3630197835527595e-05, + "loss": 0.0044, + "step": 69300 + }, + { + "epoch": 5.383300970873786, + "grad_norm": 0.1898777186870575, + "learning_rate": 2.3616151074511546e-05, + "loss": 0.0055, + "step": 69310 + }, + { + "epoch": 5.384077669902912, + "grad_norm": 0.27810508012771606, + "learning_rate": 2.3602107198793545e-05, + "loss": 0.0049, + "step": 69320 + }, + { + "epoch": 5.3848543689320385, + "grad_norm": 0.1908741295337677, + "learning_rate": 2.3588066209909427e-05, + "loss": 0.0057, + "step": 69330 + }, + { + "epoch": 5.385631067961165, + "grad_norm": 0.27311134338378906, + "learning_rate": 2.3574028109394662e-05, + "loss": 0.0045, + "step": 69340 + }, + { + "epoch": 5.386407766990291, + "grad_norm": 0.3584893047809601, + "learning_rate": 2.3559992898784454e-05, + "loss": 0.0085, + "step": 69350 + }, + { + "epoch": 5.387184466019417, + "grad_norm": 0.17808882892131805, + "learning_rate": 2.354596057961367e-05, + "loss": 0.0046, + "step": 69360 + }, + { + "epoch": 5.3879611650485435, + "grad_norm": 0.13669566810131073, + "learning_rate": 2.353193115341688e-05, + "loss": 0.0035, + "step": 69370 + }, + { + "epoch": 5.38873786407767, + "grad_norm": 0.3481382429599762, + "learning_rate": 2.3517904621728294e-05, + "loss": 0.0031, + "step": 69380 + }, + { + "epoch": 5.389514563106796, + "grad_norm": 0.14531934261322021, + "learning_rate": 2.3503880986081855e-05, + "loss": 0.0036, + "step": 69390 + }, + { + "epoch": 5.390291262135922, + "grad_norm": 0.1680549532175064, + "learning_rate": 2.3489860248011136e-05, + "loss": 0.0035, + "step": 69400 + }, + { + "epoch": 5.3910679611650485, + "grad_norm": 0.3971209228038788, + "learning_rate": 2.3475842409049464e-05, + "loss": 0.0043, + "step": 69410 + }, + { + "epoch": 5.391844660194175, + "grad_norm": 0.1944120079278946, + "learning_rate": 2.346182747072975e-05, + "loss": 0.0045, + "step": 69420 + }, + { + "epoch": 5.392621359223301, + "grad_norm": 0.16507285833358765, + "learning_rate": 2.3447815434584718e-05, + "loss": 0.0041, + "step": 69430 + }, + { + "epoch": 5.393398058252427, + "grad_norm": 0.1453857570886612, + "learning_rate": 2.3433806302146656e-05, + "loss": 0.0045, + "step": 69440 + }, + { + "epoch": 5.3941747572815535, + "grad_norm": 0.28850412368774414, + "learning_rate": 2.3419800074947612e-05, + "loss": 0.0045, + "step": 69450 + }, + { + "epoch": 5.39495145631068, + "grad_norm": 0.1242416724562645, + "learning_rate": 2.3405796754519248e-05, + "loss": 0.0051, + "step": 69460 + }, + { + "epoch": 5.395728155339806, + "grad_norm": 0.19866913557052612, + "learning_rate": 2.3391796342392992e-05, + "loss": 0.005, + "step": 69470 + }, + { + "epoch": 5.396504854368932, + "grad_norm": 0.27407729625701904, + "learning_rate": 2.337779884009984e-05, + "loss": 0.0036, + "step": 69480 + }, + { + "epoch": 5.3972815533980585, + "grad_norm": 0.2617986500263214, + "learning_rate": 2.3363804249170624e-05, + "loss": 0.0044, + "step": 69490 + }, + { + "epoch": 5.398058252427185, + "grad_norm": 0.237123504281044, + "learning_rate": 2.3349812571135703e-05, + "loss": 0.0056, + "step": 69500 + }, + { + "epoch": 5.398834951456311, + "grad_norm": 0.3634128272533417, + "learning_rate": 2.333582380752523e-05, + "loss": 0.0057, + "step": 69510 + }, + { + "epoch": 5.399611650485437, + "grad_norm": 0.5256853699684143, + "learning_rate": 2.3321837959868963e-05, + "loss": 0.0078, + "step": 69520 + }, + { + "epoch": 5.4003883495145635, + "grad_norm": 0.06841788440942764, + "learning_rate": 2.3307855029696383e-05, + "loss": 0.0045, + "step": 69530 + }, + { + "epoch": 5.40116504854369, + "grad_norm": 0.10355937480926514, + "learning_rate": 2.3293875018536648e-05, + "loss": 0.0026, + "step": 69540 + }, + { + "epoch": 5.401941747572815, + "grad_norm": 0.09543900936841965, + "learning_rate": 2.327989792791861e-05, + "loss": 0.0052, + "step": 69550 + }, + { + "epoch": 5.402718446601941, + "grad_norm": 0.39668160676956177, + "learning_rate": 2.326592375937074e-05, + "loss": 0.0059, + "step": 69560 + }, + { + "epoch": 5.403495145631068, + "grad_norm": 0.21019531786441803, + "learning_rate": 2.3251952514421248e-05, + "loss": 0.0054, + "step": 69570 + }, + { + "epoch": 5.404271844660194, + "grad_norm": 0.2824153006076813, + "learning_rate": 2.323798419459804e-05, + "loss": 0.0044, + "step": 69580 + }, + { + "epoch": 5.40504854368932, + "grad_norm": 0.9327188730239868, + "learning_rate": 2.3224018801428625e-05, + "loss": 0.0057, + "step": 69590 + }, + { + "epoch": 5.405825242718446, + "grad_norm": 0.10672017931938171, + "learning_rate": 2.3210056336440273e-05, + "loss": 0.0058, + "step": 69600 + }, + { + "epoch": 5.406601941747573, + "grad_norm": 0.23884335160255432, + "learning_rate": 2.3196096801159844e-05, + "loss": 0.0045, + "step": 69610 + }, + { + "epoch": 5.407378640776699, + "grad_norm": 0.2481101006269455, + "learning_rate": 2.3182140197114012e-05, + "loss": 0.0062, + "step": 69620 + }, + { + "epoch": 5.408155339805825, + "grad_norm": 0.19873377680778503, + "learning_rate": 2.3168186525828985e-05, + "loss": 0.0051, + "step": 69630 + }, + { + "epoch": 5.408932038834951, + "grad_norm": 0.2195536345243454, + "learning_rate": 2.3154235788830757e-05, + "loss": 0.0048, + "step": 69640 + }, + { + "epoch": 5.409708737864078, + "grad_norm": 0.3499765992164612, + "learning_rate": 2.3140287987644927e-05, + "loss": 0.0048, + "step": 69650 + }, + { + "epoch": 5.410485436893204, + "grad_norm": 0.244779571890831, + "learning_rate": 2.3126343123796833e-05, + "loss": 0.0062, + "step": 69660 + }, + { + "epoch": 5.41126213592233, + "grad_norm": 0.11089769005775452, + "learning_rate": 2.3112401198811418e-05, + "loss": 0.0042, + "step": 69670 + }, + { + "epoch": 5.412038834951456, + "grad_norm": 0.08924812078475952, + "learning_rate": 2.3098462214213418e-05, + "loss": 0.0035, + "step": 69680 + }, + { + "epoch": 5.412815533980583, + "grad_norm": 0.11162572354078293, + "learning_rate": 2.308452617152712e-05, + "loss": 0.0069, + "step": 69690 + }, + { + "epoch": 5.413592233009709, + "grad_norm": 0.2225692719221115, + "learning_rate": 2.3070593072276603e-05, + "loss": 0.0046, + "step": 69700 + }, + { + "epoch": 5.414368932038835, + "grad_norm": 0.2910420894622803, + "learning_rate": 2.3056662917985518e-05, + "loss": 0.0051, + "step": 69710 + }, + { + "epoch": 5.415145631067961, + "grad_norm": 0.6883571147918701, + "learning_rate": 2.3042735710177283e-05, + "loss": 0.0036, + "step": 69720 + }, + { + "epoch": 5.415922330097088, + "grad_norm": 0.3169547915458679, + "learning_rate": 2.302881145037492e-05, + "loss": 0.0058, + "step": 69730 + }, + { + "epoch": 5.416699029126214, + "grad_norm": 0.1840420812368393, + "learning_rate": 2.3014890140101192e-05, + "loss": 0.0044, + "step": 69740 + }, + { + "epoch": 5.41747572815534, + "grad_norm": 0.1740192472934723, + "learning_rate": 2.30009717808785e-05, + "loss": 0.0054, + "step": 69750 + }, + { + "epoch": 5.418252427184466, + "grad_norm": 0.2865894138813019, + "learning_rate": 2.2987056374228967e-05, + "loss": 0.0052, + "step": 69760 + }, + { + "epoch": 5.419029126213593, + "grad_norm": 0.3186577558517456, + "learning_rate": 2.297314392167431e-05, + "loss": 0.0052, + "step": 69770 + }, + { + "epoch": 5.419805825242719, + "grad_norm": 0.18881958723068237, + "learning_rate": 2.295923442473601e-05, + "loss": 0.0056, + "step": 69780 + }, + { + "epoch": 5.420582524271845, + "grad_norm": 0.11902853846549988, + "learning_rate": 2.294532788493519e-05, + "loss": 0.006, + "step": 69790 + }, + { + "epoch": 5.421359223300971, + "grad_norm": 0.21907180547714233, + "learning_rate": 2.293142430379262e-05, + "loss": 0.0039, + "step": 69800 + }, + { + "epoch": 5.422135922330098, + "grad_norm": 0.19510024785995483, + "learning_rate": 2.291752368282879e-05, + "loss": 0.0047, + "step": 69810 + }, + { + "epoch": 5.422912621359223, + "grad_norm": 0.17241992056369781, + "learning_rate": 2.290362602356385e-05, + "loss": 0.0058, + "step": 69820 + }, + { + "epoch": 5.423689320388349, + "grad_norm": 0.10119474679231644, + "learning_rate": 2.288973132751765e-05, + "loss": 0.0044, + "step": 69830 + }, + { + "epoch": 5.424466019417475, + "grad_norm": 0.44007954001426697, + "learning_rate": 2.287583959620965e-05, + "loss": 0.0056, + "step": 69840 + }, + { + "epoch": 5.425242718446602, + "grad_norm": 0.22702108323574066, + "learning_rate": 2.2861950831159074e-05, + "loss": 0.0033, + "step": 69850 + }, + { + "epoch": 5.426019417475728, + "grad_norm": 0.09526331722736359, + "learning_rate": 2.284806503388472e-05, + "loss": 0.0054, + "step": 69860 + }, + { + "epoch": 5.426796116504854, + "grad_norm": 0.1896718293428421, + "learning_rate": 2.283418220590518e-05, + "loss": 0.0053, + "step": 69870 + }, + { + "epoch": 5.42757281553398, + "grad_norm": 0.29182443022727966, + "learning_rate": 2.282030234873862e-05, + "loss": 0.0044, + "step": 69880 + }, + { + "epoch": 5.428349514563107, + "grad_norm": 0.11922818422317505, + "learning_rate": 2.2806425463902948e-05, + "loss": 0.0054, + "step": 69890 + }, + { + "epoch": 5.429126213592233, + "grad_norm": 0.15518486499786377, + "learning_rate": 2.279255155291568e-05, + "loss": 0.006, + "step": 69900 + }, + { + "epoch": 5.429902912621359, + "grad_norm": 0.1009276732802391, + "learning_rate": 2.277868061729409e-05, + "loss": 0.0049, + "step": 69910 + }, + { + "epoch": 5.430679611650485, + "grad_norm": 0.20646552741527557, + "learning_rate": 2.276481265855504e-05, + "loss": 0.0052, + "step": 69920 + }, + { + "epoch": 5.431456310679612, + "grad_norm": 0.14948523044586182, + "learning_rate": 2.2750947678215128e-05, + "loss": 0.0072, + "step": 69930 + }, + { + "epoch": 5.432233009708738, + "grad_norm": 0.1763138622045517, + "learning_rate": 2.2737085677790616e-05, + "loss": 0.0048, + "step": 69940 + }, + { + "epoch": 5.433009708737864, + "grad_norm": 0.10130912810564041, + "learning_rate": 2.2723226658797443e-05, + "loss": 0.0044, + "step": 69950 + }, + { + "epoch": 5.43378640776699, + "grad_norm": 0.2767033576965332, + "learning_rate": 2.270937062275117e-05, + "loss": 0.006, + "step": 69960 + }, + { + "epoch": 5.434563106796117, + "grad_norm": 0.13659949600696564, + "learning_rate": 2.2695517571167113e-05, + "loss": 0.0045, + "step": 69970 + }, + { + "epoch": 5.435339805825243, + "grad_norm": 0.1377447545528412, + "learning_rate": 2.268166750556019e-05, + "loss": 0.0054, + "step": 69980 + }, + { + "epoch": 5.436116504854369, + "grad_norm": 0.30200016498565674, + "learning_rate": 2.266782042744503e-05, + "loss": 0.0047, + "step": 69990 + }, + { + "epoch": 5.436893203883495, + "grad_norm": 0.2006092220544815, + "learning_rate": 2.2653976338335935e-05, + "loss": 0.0035, + "step": 70000 + }, + { + "epoch": 5.437669902912622, + "grad_norm": 0.31615176796913147, + "learning_rate": 2.2640135239746895e-05, + "loss": 0.0045, + "step": 70010 + }, + { + "epoch": 5.438446601941748, + "grad_norm": 0.14794275164604187, + "learning_rate": 2.2626297133191515e-05, + "loss": 0.0054, + "step": 70020 + }, + { + "epoch": 5.439223300970874, + "grad_norm": 0.19223672151565552, + "learning_rate": 2.261246202018312e-05, + "loss": 0.0066, + "step": 70030 + }, + { + "epoch": 5.44, + "grad_norm": 0.3897266387939453, + "learning_rate": 2.2598629902234718e-05, + "loss": 0.004, + "step": 70040 + }, + { + "epoch": 5.440776699029126, + "grad_norm": 0.2007625699043274, + "learning_rate": 2.258480078085894e-05, + "loss": 0.0058, + "step": 70050 + }, + { + "epoch": 5.441553398058252, + "grad_norm": 0.7508700489997864, + "learning_rate": 2.2570974657568126e-05, + "loss": 0.0052, + "step": 70060 + }, + { + "epoch": 5.442330097087378, + "grad_norm": 0.1584361493587494, + "learning_rate": 2.2557151533874286e-05, + "loss": 0.0054, + "step": 70070 + }, + { + "epoch": 5.4431067961165045, + "grad_norm": 0.2918003797531128, + "learning_rate": 2.2543331411289116e-05, + "loss": 0.0059, + "step": 70080 + }, + { + "epoch": 5.443883495145631, + "grad_norm": 0.1418270319700241, + "learning_rate": 2.252951429132392e-05, + "loss": 0.0056, + "step": 70090 + }, + { + "epoch": 5.444660194174757, + "grad_norm": 0.47171175479888916, + "learning_rate": 2.2515700175489762e-05, + "loss": 0.0063, + "step": 70100 + }, + { + "epoch": 5.445436893203883, + "grad_norm": 0.14464880526065826, + "learning_rate": 2.2501889065297288e-05, + "loss": 0.0051, + "step": 70110 + }, + { + "epoch": 5.4462135922330095, + "grad_norm": 0.2201145589351654, + "learning_rate": 2.2488080962256887e-05, + "loss": 0.0053, + "step": 70120 + }, + { + "epoch": 5.446990291262136, + "grad_norm": 0.09456927329301834, + "learning_rate": 2.2474275867878585e-05, + "loss": 0.0053, + "step": 70130 + }, + { + "epoch": 5.447766990291262, + "grad_norm": 0.14648224413394928, + "learning_rate": 2.246047378367211e-05, + "loss": 0.0037, + "step": 70140 + }, + { + "epoch": 5.448543689320388, + "grad_norm": 0.2176380306482315, + "learning_rate": 2.2446674711146798e-05, + "loss": 0.0038, + "step": 70150 + }, + { + "epoch": 5.4493203883495145, + "grad_norm": 0.10743379592895508, + "learning_rate": 2.2432878651811734e-05, + "loss": 0.0046, + "step": 70160 + }, + { + "epoch": 5.450097087378641, + "grad_norm": 0.09293827414512634, + "learning_rate": 2.2419085607175594e-05, + "loss": 0.0033, + "step": 70170 + }, + { + "epoch": 5.450873786407767, + "grad_norm": 0.38913920521736145, + "learning_rate": 2.240529557874678e-05, + "loss": 0.0061, + "step": 70180 + }, + { + "epoch": 5.451650485436893, + "grad_norm": 0.21764086186885834, + "learning_rate": 2.2391508568033358e-05, + "loss": 0.0035, + "step": 70190 + }, + { + "epoch": 5.4524271844660195, + "grad_norm": 0.27461737394332886, + "learning_rate": 2.237772457654307e-05, + "loss": 0.0039, + "step": 70200 + }, + { + "epoch": 5.453203883495146, + "grad_norm": 0.1788158267736435, + "learning_rate": 2.236394360578327e-05, + "loss": 0.004, + "step": 70210 + }, + { + "epoch": 5.453980582524272, + "grad_norm": 0.13825723528862, + "learning_rate": 2.2350165657261073e-05, + "loss": 0.0037, + "step": 70220 + }, + { + "epoch": 5.454757281553398, + "grad_norm": 0.21624569594860077, + "learning_rate": 2.2336390732483164e-05, + "loss": 0.0066, + "step": 70230 + }, + { + "epoch": 5.4555339805825245, + "grad_norm": 0.16279077529907227, + "learning_rate": 2.2322618832955976e-05, + "loss": 0.0035, + "step": 70240 + }, + { + "epoch": 5.456310679611651, + "grad_norm": 0.17583008110523224, + "learning_rate": 2.2308849960185578e-05, + "loss": 0.0056, + "step": 70250 + }, + { + "epoch": 5.457087378640777, + "grad_norm": 0.09926772117614746, + "learning_rate": 2.229508411567774e-05, + "loss": 0.0048, + "step": 70260 + }, + { + "epoch": 5.457864077669903, + "grad_norm": 0.22504960000514984, + "learning_rate": 2.2281321300937823e-05, + "loss": 0.0052, + "step": 70270 + }, + { + "epoch": 5.4586407766990295, + "grad_norm": 0.21580061316490173, + "learning_rate": 2.2267561517470937e-05, + "loss": 0.0048, + "step": 70280 + }, + { + "epoch": 5.459417475728156, + "grad_norm": 0.24873003363609314, + "learning_rate": 2.2253804766781845e-05, + "loss": 0.0042, + "step": 70290 + }, + { + "epoch": 5.460194174757282, + "grad_norm": 0.133745014667511, + "learning_rate": 2.2240051050374933e-05, + "loss": 0.0054, + "step": 70300 + }, + { + "epoch": 5.460970873786408, + "grad_norm": 0.16596361994743347, + "learning_rate": 2.22263003697543e-05, + "loss": 0.0052, + "step": 70310 + }, + { + "epoch": 5.461747572815534, + "grad_norm": 0.22917184233665466, + "learning_rate": 2.2212552726423702e-05, + "loss": 0.006, + "step": 70320 + }, + { + "epoch": 5.46252427184466, + "grad_norm": 0.19751447439193726, + "learning_rate": 2.2198808121886578e-05, + "loss": 0.0037, + "step": 70330 + }, + { + "epoch": 5.463300970873786, + "grad_norm": 0.18534815311431885, + "learning_rate": 2.218506655764598e-05, + "loss": 0.004, + "step": 70340 + }, + { + "epoch": 5.464077669902912, + "grad_norm": 0.1872367411851883, + "learning_rate": 2.2171328035204704e-05, + "loss": 0.004, + "step": 70350 + }, + { + "epoch": 5.464854368932039, + "grad_norm": 0.2756895124912262, + "learning_rate": 2.2157592556065133e-05, + "loss": 0.0071, + "step": 70360 + }, + { + "epoch": 5.465631067961165, + "grad_norm": 0.20309801399707794, + "learning_rate": 2.2143860121729382e-05, + "loss": 0.0048, + "step": 70370 + }, + { + "epoch": 5.466407766990291, + "grad_norm": 0.10602859407663345, + "learning_rate": 2.2130130733699206e-05, + "loss": 0.0035, + "step": 70380 + }, + { + "epoch": 5.467184466019417, + "grad_norm": 0.16149739921092987, + "learning_rate": 2.2116404393476052e-05, + "loss": 0.005, + "step": 70390 + }, + { + "epoch": 5.467961165048544, + "grad_norm": 0.1207858994603157, + "learning_rate": 2.2102681102560967e-05, + "loss": 0.0035, + "step": 70400 + }, + { + "epoch": 5.46873786407767, + "grad_norm": 0.9553532600402832, + "learning_rate": 2.208896086245476e-05, + "loss": 0.0056, + "step": 70410 + }, + { + "epoch": 5.469514563106796, + "grad_norm": 0.20146769285202026, + "learning_rate": 2.2075243674657808e-05, + "loss": 0.0067, + "step": 70420 + }, + { + "epoch": 5.470291262135922, + "grad_norm": 0.10297735780477524, + "learning_rate": 2.206152954067024e-05, + "loss": 0.0057, + "step": 70430 + }, + { + "epoch": 5.471067961165049, + "grad_norm": 0.24230904877185822, + "learning_rate": 2.2047818461991758e-05, + "loss": 0.0054, + "step": 70440 + }, + { + "epoch": 5.471844660194175, + "grad_norm": 0.3228150010108948, + "learning_rate": 2.2034110440121864e-05, + "loss": 0.0044, + "step": 70450 + }, + { + "epoch": 5.472621359223301, + "grad_norm": 0.38198044896125793, + "learning_rate": 2.2020405476559585e-05, + "loss": 0.0051, + "step": 70460 + }, + { + "epoch": 5.473398058252427, + "grad_norm": 0.49954715371131897, + "learning_rate": 2.2006703572803722e-05, + "loss": 0.0054, + "step": 70470 + }, + { + "epoch": 5.474174757281554, + "grad_norm": 0.19517293572425842, + "learning_rate": 2.1993004730352646e-05, + "loss": 0.0047, + "step": 70480 + }, + { + "epoch": 5.47495145631068, + "grad_norm": 0.25468823313713074, + "learning_rate": 2.1979308950704464e-05, + "loss": 0.0078, + "step": 70490 + }, + { + "epoch": 5.475728155339806, + "grad_norm": 0.13446427881717682, + "learning_rate": 2.1965616235356923e-05, + "loss": 0.0044, + "step": 70500 + }, + { + "epoch": 5.476504854368932, + "grad_norm": 1.1147911548614502, + "learning_rate": 2.1951926585807465e-05, + "loss": 0.004, + "step": 70510 + }, + { + "epoch": 5.4772815533980586, + "grad_norm": 0.437155157327652, + "learning_rate": 2.193824000355313e-05, + "loss": 0.0041, + "step": 70520 + }, + { + "epoch": 5.478058252427185, + "grad_norm": 0.13410836458206177, + "learning_rate": 2.192455649009067e-05, + "loss": 0.0034, + "step": 70530 + }, + { + "epoch": 5.478834951456311, + "grad_norm": 0.17390479147434235, + "learning_rate": 2.1910876046916523e-05, + "loss": 0.0061, + "step": 70540 + }, + { + "epoch": 5.479611650485437, + "grad_norm": 0.09934908151626587, + "learning_rate": 2.1897198675526725e-05, + "loss": 0.005, + "step": 70550 + }, + { + "epoch": 5.480388349514563, + "grad_norm": 0.1705949902534485, + "learning_rate": 2.1883524377417024e-05, + "loss": 0.0048, + "step": 70560 + }, + { + "epoch": 5.481165048543689, + "grad_norm": 0.34825143218040466, + "learning_rate": 2.1869853154082825e-05, + "loss": 0.0079, + "step": 70570 + }, + { + "epoch": 5.481941747572815, + "grad_norm": 0.2717203199863434, + "learning_rate": 2.1856185007019215e-05, + "loss": 0.0045, + "step": 70580 + }, + { + "epoch": 5.482718446601941, + "grad_norm": 0.23366864025592804, + "learning_rate": 2.1842519937720874e-05, + "loss": 0.0052, + "step": 70590 + }, + { + "epoch": 5.483495145631068, + "grad_norm": 0.08774442970752716, + "learning_rate": 2.182885794768224e-05, + "loss": 0.0055, + "step": 70600 + }, + { + "epoch": 5.484271844660194, + "grad_norm": 0.25077974796295166, + "learning_rate": 2.1815199038397326e-05, + "loss": 0.0047, + "step": 70610 + }, + { + "epoch": 5.48504854368932, + "grad_norm": 0.24858123064041138, + "learning_rate": 2.1801543211359893e-05, + "loss": 0.0044, + "step": 70620 + }, + { + "epoch": 5.485825242718446, + "grad_norm": 0.1595686972141266, + "learning_rate": 2.178789046806326e-05, + "loss": 0.0054, + "step": 70630 + }, + { + "epoch": 5.486601941747573, + "grad_norm": 0.12338491529226303, + "learning_rate": 2.1774240810000547e-05, + "loss": 0.0053, + "step": 70640 + }, + { + "epoch": 5.487378640776699, + "grad_norm": 0.09108693152666092, + "learning_rate": 2.17605942386644e-05, + "loss": 0.0037, + "step": 70650 + }, + { + "epoch": 5.488155339805825, + "grad_norm": 0.1939799040555954, + "learning_rate": 2.174695075554723e-05, + "loss": 0.0063, + "step": 70660 + }, + { + "epoch": 5.488932038834951, + "grad_norm": 0.1792116016149521, + "learning_rate": 2.173331036214103e-05, + "loss": 0.0037, + "step": 70670 + }, + { + "epoch": 5.489708737864078, + "grad_norm": 0.205010786652565, + "learning_rate": 2.171967305993753e-05, + "loss": 0.0048, + "step": 70680 + }, + { + "epoch": 5.490485436893204, + "grad_norm": 0.2271731048822403, + "learning_rate": 2.1706038850428024e-05, + "loss": 0.0036, + "step": 70690 + }, + { + "epoch": 5.49126213592233, + "grad_norm": 0.2992085814476013, + "learning_rate": 2.169240773510361e-05, + "loss": 0.0063, + "step": 70700 + }, + { + "epoch": 5.492038834951456, + "grad_norm": 0.2766699194908142, + "learning_rate": 2.1678779715454906e-05, + "loss": 0.005, + "step": 70710 + }, + { + "epoch": 5.492815533980583, + "grad_norm": 0.1565568596124649, + "learning_rate": 2.166515479297229e-05, + "loss": 0.0052, + "step": 70720 + }, + { + "epoch": 5.493592233009709, + "grad_norm": 0.21432362496852875, + "learning_rate": 2.1651532969145732e-05, + "loss": 0.0048, + "step": 70730 + }, + { + "epoch": 5.494368932038835, + "grad_norm": 0.09704770892858505, + "learning_rate": 2.1637914245464903e-05, + "loss": 0.0051, + "step": 70740 + }, + { + "epoch": 5.495145631067961, + "grad_norm": 0.30146095156669617, + "learning_rate": 2.1624298623419133e-05, + "loss": 0.0038, + "step": 70750 + }, + { + "epoch": 5.495922330097088, + "grad_norm": 0.15891486406326294, + "learning_rate": 2.1610686104497413e-05, + "loss": 0.0058, + "step": 70760 + }, + { + "epoch": 5.496699029126214, + "grad_norm": 0.2028503566980362, + "learning_rate": 2.1597076690188366e-05, + "loss": 0.0042, + "step": 70770 + }, + { + "epoch": 5.49747572815534, + "grad_norm": 0.1659373790025711, + "learning_rate": 2.1583470381980315e-05, + "loss": 0.0034, + "step": 70780 + }, + { + "epoch": 5.498252427184466, + "grad_norm": 0.2613812983036041, + "learning_rate": 2.156986718136123e-05, + "loss": 0.0048, + "step": 70790 + }, + { + "epoch": 5.499029126213593, + "grad_norm": 0.4057634472846985, + "learning_rate": 2.155626708981871e-05, + "loss": 0.0055, + "step": 70800 + }, + { + "epoch": 5.499805825242719, + "grad_norm": 0.38867029547691345, + "learning_rate": 2.1542670108840075e-05, + "loss": 0.0073, + "step": 70810 + }, + { + "epoch": 5.500582524271845, + "grad_norm": 0.3548274338245392, + "learning_rate": 2.152907623991221e-05, + "loss": 0.0059, + "step": 70820 + }, + { + "epoch": 5.5013592233009705, + "grad_norm": 0.27555492520332336, + "learning_rate": 2.1515485484521802e-05, + "loss": 0.0033, + "step": 70830 + }, + { + "epoch": 5.502135922330097, + "grad_norm": 0.3397141993045807, + "learning_rate": 2.150189784415506e-05, + "loss": 0.0042, + "step": 70840 + }, + { + "epoch": 5.502912621359223, + "grad_norm": 0.21040773391723633, + "learning_rate": 2.148831332029794e-05, + "loss": 0.0039, + "step": 70850 + }, + { + "epoch": 5.503689320388349, + "grad_norm": 0.1376645863056183, + "learning_rate": 2.147473191443599e-05, + "loss": 0.0046, + "step": 70860 + }, + { + "epoch": 5.5044660194174755, + "grad_norm": 0.12212605029344559, + "learning_rate": 2.1461153628054492e-05, + "loss": 0.0045, + "step": 70870 + }, + { + "epoch": 5.505242718446602, + "grad_norm": 0.15922288596630096, + "learning_rate": 2.1447578462638284e-05, + "loss": 0.0074, + "step": 70880 + }, + { + "epoch": 5.506019417475728, + "grad_norm": 0.16451700031757355, + "learning_rate": 2.143400641967201e-05, + "loss": 0.0055, + "step": 70890 + }, + { + "epoch": 5.506796116504854, + "grad_norm": 0.3554143011569977, + "learning_rate": 2.1420437500639823e-05, + "loss": 0.0065, + "step": 70900 + }, + { + "epoch": 5.5075728155339805, + "grad_norm": 0.14976665377616882, + "learning_rate": 2.1406871707025646e-05, + "loss": 0.0055, + "step": 70910 + }, + { + "epoch": 5.508349514563107, + "grad_norm": 0.3927727937698364, + "learning_rate": 2.1393309040312963e-05, + "loss": 0.0049, + "step": 70920 + }, + { + "epoch": 5.509126213592233, + "grad_norm": 0.09106413275003433, + "learning_rate": 2.137974950198502e-05, + "loss": 0.005, + "step": 70930 + }, + { + "epoch": 5.509902912621359, + "grad_norm": 0.1368388533592224, + "learning_rate": 2.1366193093524617e-05, + "loss": 0.0062, + "step": 70940 + }, + { + "epoch": 5.5106796116504855, + "grad_norm": 0.24904055893421173, + "learning_rate": 2.1352639816414288e-05, + "loss": 0.004, + "step": 70950 + }, + { + "epoch": 5.511456310679612, + "grad_norm": 0.21626929938793182, + "learning_rate": 2.1339089672136202e-05, + "loss": 0.0048, + "step": 70960 + }, + { + "epoch": 5.512233009708738, + "grad_norm": 0.15626585483551025, + "learning_rate": 2.1325542662172193e-05, + "loss": 0.0047, + "step": 70970 + }, + { + "epoch": 5.513009708737864, + "grad_norm": 0.1407000720500946, + "learning_rate": 2.1311998788003712e-05, + "loss": 0.006, + "step": 70980 + }, + { + "epoch": 5.5137864077669905, + "grad_norm": 0.11357142776250839, + "learning_rate": 2.1298458051111907e-05, + "loss": 0.0064, + "step": 70990 + }, + { + "epoch": 5.514563106796117, + "grad_norm": 0.19175836443901062, + "learning_rate": 2.1284920452977602e-05, + "loss": 0.0049, + "step": 71000 + }, + { + "epoch": 5.515339805825243, + "grad_norm": 0.21556317806243896, + "learning_rate": 2.12713859950812e-05, + "loss": 0.0049, + "step": 71010 + }, + { + "epoch": 5.516116504854369, + "grad_norm": 0.26209720969200134, + "learning_rate": 2.1257854678902832e-05, + "loss": 0.006, + "step": 71020 + }, + { + "epoch": 5.5168932038834955, + "grad_norm": 0.2623541057109833, + "learning_rate": 2.124432650592227e-05, + "loss": 0.0053, + "step": 71030 + }, + { + "epoch": 5.517669902912622, + "grad_norm": 0.13869692385196686, + "learning_rate": 2.1230801477618937e-05, + "loss": 0.0047, + "step": 71040 + }, + { + "epoch": 5.518446601941747, + "grad_norm": 0.34018459916114807, + "learning_rate": 2.121727959547189e-05, + "loss": 0.005, + "step": 71050 + }, + { + "epoch": 5.519223300970873, + "grad_norm": 0.14092965424060822, + "learning_rate": 2.1203760860959886e-05, + "loss": 0.0044, + "step": 71060 + }, + { + "epoch": 5.52, + "grad_norm": 0.10325656086206436, + "learning_rate": 2.119024527556127e-05, + "loss": 0.0039, + "step": 71070 + }, + { + "epoch": 5.520776699029126, + "grad_norm": 0.4048587381839752, + "learning_rate": 2.117673284075415e-05, + "loss": 0.005, + "step": 71080 + }, + { + "epoch": 5.521553398058252, + "grad_norm": 0.2741873860359192, + "learning_rate": 2.1163223558016165e-05, + "loss": 0.005, + "step": 71090 + }, + { + "epoch": 5.522330097087378, + "grad_norm": 0.3506443500518799, + "learning_rate": 2.1149717428824722e-05, + "loss": 0.0031, + "step": 71100 + }, + { + "epoch": 5.523106796116505, + "grad_norm": 0.26962000131607056, + "learning_rate": 2.1136214454656784e-05, + "loss": 0.0058, + "step": 71110 + }, + { + "epoch": 5.523883495145631, + "grad_norm": 0.725080132484436, + "learning_rate": 2.1122714636989056e-05, + "loss": 0.0058, + "step": 71120 + }, + { + "epoch": 5.524660194174757, + "grad_norm": 0.2751829922199249, + "learning_rate": 2.1109217977297823e-05, + "loss": 0.0053, + "step": 71130 + }, + { + "epoch": 5.525436893203883, + "grad_norm": 0.1835833340883255, + "learning_rate": 2.1095724477059077e-05, + "loss": 0.0053, + "step": 71140 + }, + { + "epoch": 5.52621359223301, + "grad_norm": 0.3163740336894989, + "learning_rate": 2.1082234137748434e-05, + "loss": 0.0063, + "step": 71150 + }, + { + "epoch": 5.526990291262136, + "grad_norm": 0.23110629618167877, + "learning_rate": 2.106874696084122e-05, + "loss": 0.0051, + "step": 71160 + }, + { + "epoch": 5.527766990291262, + "grad_norm": 0.5946069955825806, + "learning_rate": 2.1055262947812315e-05, + "loss": 0.0041, + "step": 71170 + }, + { + "epoch": 5.528543689320388, + "grad_norm": 0.21454885601997375, + "learning_rate": 2.104178210013636e-05, + "loss": 0.0051, + "step": 71180 + }, + { + "epoch": 5.529320388349515, + "grad_norm": 0.4237723648548126, + "learning_rate": 2.1028304419287554e-05, + "loss": 0.0047, + "step": 71190 + }, + { + "epoch": 5.530097087378641, + "grad_norm": 0.11148542165756226, + "learning_rate": 2.101482990673982e-05, + "loss": 0.0042, + "step": 71200 + }, + { + "epoch": 5.530873786407767, + "grad_norm": 0.14605198800563812, + "learning_rate": 2.1001358563966707e-05, + "loss": 0.0032, + "step": 71210 + }, + { + "epoch": 5.531650485436893, + "grad_norm": 0.26972559094429016, + "learning_rate": 2.098789039244144e-05, + "loss": 0.007, + "step": 71220 + }, + { + "epoch": 5.5324271844660196, + "grad_norm": 0.17718644440174103, + "learning_rate": 2.097442539363685e-05, + "loss": 0.0045, + "step": 71230 + }, + { + "epoch": 5.533203883495146, + "grad_norm": 0.18980196118354797, + "learning_rate": 2.0960963569025448e-05, + "loss": 0.004, + "step": 71240 + }, + { + "epoch": 5.533980582524272, + "grad_norm": 0.22982360422611237, + "learning_rate": 2.0947504920079435e-05, + "loss": 0.004, + "step": 71250 + }, + { + "epoch": 5.534757281553398, + "grad_norm": 0.32775476574897766, + "learning_rate": 2.093404944827059e-05, + "loss": 0.007, + "step": 71260 + }, + { + "epoch": 5.5355339805825245, + "grad_norm": 0.16048002243041992, + "learning_rate": 2.0920597155070394e-05, + "loss": 0.0045, + "step": 71270 + }, + { + "epoch": 5.536310679611651, + "grad_norm": 0.13818620145320892, + "learning_rate": 2.090714804194997e-05, + "loss": 0.0045, + "step": 71280 + }, + { + "epoch": 5.537087378640777, + "grad_norm": 0.17915965616703033, + "learning_rate": 2.0893702110380114e-05, + "loss": 0.0051, + "step": 71290 + }, + { + "epoch": 5.537864077669903, + "grad_norm": 0.34198999404907227, + "learning_rate": 2.0880259361831216e-05, + "loss": 0.0044, + "step": 71300 + }, + { + "epoch": 5.5386407766990295, + "grad_norm": 0.34409502148628235, + "learning_rate": 2.0866819797773397e-05, + "loss": 0.0045, + "step": 71310 + }, + { + "epoch": 5.539417475728156, + "grad_norm": 0.1994125097990036, + "learning_rate": 2.085338341967634e-05, + "loss": 0.0059, + "step": 71320 + }, + { + "epoch": 5.540194174757282, + "grad_norm": 0.4027525782585144, + "learning_rate": 2.083995022900946e-05, + "loss": 0.0041, + "step": 71330 + }, + { + "epoch": 5.540970873786407, + "grad_norm": 0.18194212019443512, + "learning_rate": 2.0826520227241768e-05, + "loss": 0.0039, + "step": 71340 + }, + { + "epoch": 5.541747572815534, + "grad_norm": 0.34069499373435974, + "learning_rate": 2.081309341584199e-05, + "loss": 0.008, + "step": 71350 + }, + { + "epoch": 5.54252427184466, + "grad_norm": 0.20069612562656403, + "learning_rate": 2.0799669796278413e-05, + "loss": 0.0032, + "step": 71360 + }, + { + "epoch": 5.543300970873786, + "grad_norm": 0.4096057415008545, + "learning_rate": 2.078624937001906e-05, + "loss": 0.0063, + "step": 71370 + }, + { + "epoch": 5.544077669902912, + "grad_norm": 0.2154582291841507, + "learning_rate": 2.0772832138531544e-05, + "loss": 0.0059, + "step": 71380 + }, + { + "epoch": 5.544854368932039, + "grad_norm": 0.14218321442604065, + "learning_rate": 2.0759418103283158e-05, + "loss": 0.0045, + "step": 71390 + }, + { + "epoch": 5.545631067961165, + "grad_norm": 0.15254658460617065, + "learning_rate": 2.0746007265740846e-05, + "loss": 0.0038, + "step": 71400 + }, + { + "epoch": 5.546407766990291, + "grad_norm": 0.3299763798713684, + "learning_rate": 2.0732599627371215e-05, + "loss": 0.0088, + "step": 71410 + }, + { + "epoch": 5.547184466019417, + "grad_norm": 0.21285855770111084, + "learning_rate": 2.0719195189640468e-05, + "loss": 0.0036, + "step": 71420 + }, + { + "epoch": 5.547961165048544, + "grad_norm": 0.21449129283428192, + "learning_rate": 2.0705793954014528e-05, + "loss": 0.0094, + "step": 71430 + }, + { + "epoch": 5.54873786407767, + "grad_norm": 0.42410656809806824, + "learning_rate": 2.06923959219589e-05, + "loss": 0.0036, + "step": 71440 + }, + { + "epoch": 5.549514563106796, + "grad_norm": 0.24743323028087616, + "learning_rate": 2.0679001094938793e-05, + "loss": 0.004, + "step": 71450 + }, + { + "epoch": 5.550291262135922, + "grad_norm": 0.23921307921409607, + "learning_rate": 2.066560947441904e-05, + "loss": 0.0045, + "step": 71460 + }, + { + "epoch": 5.551067961165049, + "grad_norm": 0.09421955794095993, + "learning_rate": 2.0652221061864152e-05, + "loss": 0.0049, + "step": 71470 + }, + { + "epoch": 5.551844660194175, + "grad_norm": 0.19721536338329315, + "learning_rate": 2.0638835858738226e-05, + "loss": 0.0043, + "step": 71480 + }, + { + "epoch": 5.552621359223301, + "grad_norm": 0.11504783481359482, + "learning_rate": 2.0625453866505068e-05, + "loss": 0.0052, + "step": 71490 + }, + { + "epoch": 5.553398058252427, + "grad_norm": 0.1830441653728485, + "learning_rate": 2.0612075086628123e-05, + "loss": 0.0037, + "step": 71500 + }, + { + "epoch": 5.554174757281554, + "grad_norm": 0.15685348212718964, + "learning_rate": 2.059869952057045e-05, + "loss": 0.0039, + "step": 71510 + }, + { + "epoch": 5.55495145631068, + "grad_norm": 0.25159287452697754, + "learning_rate": 2.0585327169794793e-05, + "loss": 0.0035, + "step": 71520 + }, + { + "epoch": 5.555728155339806, + "grad_norm": 0.3999635577201843, + "learning_rate": 2.0571958035763534e-05, + "loss": 0.0026, + "step": 71530 + }, + { + "epoch": 5.556504854368932, + "grad_norm": 0.16292597353458405, + "learning_rate": 2.0558592119938715e-05, + "loss": 0.0038, + "step": 71540 + }, + { + "epoch": 5.557281553398059, + "grad_norm": 0.525381326675415, + "learning_rate": 2.054522942378198e-05, + "loss": 0.0042, + "step": 71550 + }, + { + "epoch": 5.558058252427184, + "grad_norm": 0.17290769517421722, + "learning_rate": 2.053186994875469e-05, + "loss": 0.004, + "step": 71560 + }, + { + "epoch": 5.55883495145631, + "grad_norm": 0.2862843871116638, + "learning_rate": 2.0518513696317775e-05, + "loss": 0.0042, + "step": 71570 + }, + { + "epoch": 5.5596116504854365, + "grad_norm": 0.4011848568916321, + "learning_rate": 2.050516066793188e-05, + "loss": 0.0044, + "step": 71580 + }, + { + "epoch": 5.560388349514563, + "grad_norm": 0.17063121497631073, + "learning_rate": 2.0491810865057276e-05, + "loss": 0.007, + "step": 71590 + }, + { + "epoch": 5.561165048543689, + "grad_norm": 0.20451374351978302, + "learning_rate": 2.0478464289153882e-05, + "loss": 0.0048, + "step": 71600 + }, + { + "epoch": 5.561941747572815, + "grad_norm": 0.2673647701740265, + "learning_rate": 2.0465120941681232e-05, + "loss": 0.0082, + "step": 71610 + }, + { + "epoch": 5.5627184466019415, + "grad_norm": 0.26019465923309326, + "learning_rate": 2.0451780824098566e-05, + "loss": 0.0049, + "step": 71620 + }, + { + "epoch": 5.563495145631068, + "grad_norm": 0.2793848514556885, + "learning_rate": 2.043844393786471e-05, + "loss": 0.0041, + "step": 71630 + }, + { + "epoch": 5.564271844660194, + "grad_norm": 0.7556408643722534, + "learning_rate": 2.0425110284438193e-05, + "loss": 0.0072, + "step": 71640 + }, + { + "epoch": 5.56504854368932, + "grad_norm": 0.7079247236251831, + "learning_rate": 2.041177986527712e-05, + "loss": 0.0047, + "step": 71650 + }, + { + "epoch": 5.5658252427184465, + "grad_norm": 0.26765456795692444, + "learning_rate": 2.0398452681839347e-05, + "loss": 0.0044, + "step": 71660 + }, + { + "epoch": 5.566601941747573, + "grad_norm": 0.10725075006484985, + "learning_rate": 2.0385128735582272e-05, + "loss": 0.0088, + "step": 71670 + }, + { + "epoch": 5.567378640776699, + "grad_norm": 0.05264506861567497, + "learning_rate": 2.0371808027963003e-05, + "loss": 0.0046, + "step": 71680 + }, + { + "epoch": 5.568155339805825, + "grad_norm": 0.1633751094341278, + "learning_rate": 2.035849056043825e-05, + "loss": 0.0064, + "step": 71690 + }, + { + "epoch": 5.5689320388349515, + "grad_norm": 0.3345487713813782, + "learning_rate": 2.03451763344644e-05, + "loss": 0.0043, + "step": 71700 + }, + { + "epoch": 5.569708737864078, + "grad_norm": 0.3309847116470337, + "learning_rate": 2.0331865351497482e-05, + "loss": 0.0062, + "step": 71710 + }, + { + "epoch": 5.570485436893204, + "grad_norm": 0.21233773231506348, + "learning_rate": 2.0318557612993185e-05, + "loss": 0.0032, + "step": 71720 + }, + { + "epoch": 5.57126213592233, + "grad_norm": 0.36726489663124084, + "learning_rate": 2.0305253120406782e-05, + "loss": 0.0049, + "step": 71730 + }, + { + "epoch": 5.5720388349514565, + "grad_norm": 0.1816578358411789, + "learning_rate": 2.029195187519325e-05, + "loss": 0.0054, + "step": 71740 + }, + { + "epoch": 5.572815533980583, + "grad_norm": 0.5042437314987183, + "learning_rate": 2.0278653878807217e-05, + "loss": 0.0056, + "step": 71750 + }, + { + "epoch": 5.573592233009709, + "grad_norm": 0.24720466136932373, + "learning_rate": 2.0265359132702892e-05, + "loss": 0.0047, + "step": 71760 + }, + { + "epoch": 5.574368932038835, + "grad_norm": 0.06559149920940399, + "learning_rate": 2.0252067638334182e-05, + "loss": 0.0039, + "step": 71770 + }, + { + "epoch": 5.5751456310679615, + "grad_norm": 0.10631874948740005, + "learning_rate": 2.023877939715464e-05, + "loss": 0.0048, + "step": 71780 + }, + { + "epoch": 5.575922330097088, + "grad_norm": 0.5228469371795654, + "learning_rate": 2.0225494410617448e-05, + "loss": 0.0053, + "step": 71790 + }, + { + "epoch": 5.576699029126214, + "grad_norm": 0.20236018300056458, + "learning_rate": 2.02122126801754e-05, + "loss": 0.0066, + "step": 71800 + }, + { + "epoch": 5.57747572815534, + "grad_norm": 0.5252739191055298, + "learning_rate": 2.0198934207281018e-05, + "loss": 0.0043, + "step": 71810 + }, + { + "epoch": 5.5782524271844665, + "grad_norm": 0.36645108461380005, + "learning_rate": 2.0185658993386357e-05, + "loss": 0.0066, + "step": 71820 + }, + { + "epoch": 5.579029126213593, + "grad_norm": 0.21205709874629974, + "learning_rate": 2.0172387039943226e-05, + "loss": 0.0064, + "step": 71830 + }, + { + "epoch": 5.579805825242718, + "grad_norm": 0.20852503180503845, + "learning_rate": 2.0159118348402968e-05, + "loss": 0.0056, + "step": 71840 + }, + { + "epoch": 5.580582524271844, + "grad_norm": 0.17845331132411957, + "learning_rate": 2.0145852920216697e-05, + "loss": 0.006, + "step": 71850 + }, + { + "epoch": 5.581359223300971, + "grad_norm": 0.363627552986145, + "learning_rate": 2.0132590756835045e-05, + "loss": 0.0052, + "step": 71860 + }, + { + "epoch": 5.582135922330097, + "grad_norm": 0.15188470482826233, + "learning_rate": 2.0119331859708386e-05, + "loss": 0.0041, + "step": 71870 + }, + { + "epoch": 5.582912621359223, + "grad_norm": 0.18962405622005463, + "learning_rate": 2.0106076230286654e-05, + "loss": 0.0055, + "step": 71880 + }, + { + "epoch": 5.583689320388349, + "grad_norm": 0.22737987339496613, + "learning_rate": 2.0092823870019488e-05, + "loss": 0.0054, + "step": 71890 + }, + { + "epoch": 5.584466019417476, + "grad_norm": 0.4716254472732544, + "learning_rate": 2.0079574780356115e-05, + "loss": 0.0046, + "step": 71900 + }, + { + "epoch": 5.585242718446602, + "grad_norm": 0.1660766899585724, + "learning_rate": 2.0066328962745486e-05, + "loss": 0.0048, + "step": 71910 + }, + { + "epoch": 5.586019417475728, + "grad_norm": 0.15293210744857788, + "learning_rate": 2.00530864186361e-05, + "loss": 0.0037, + "step": 71920 + }, + { + "epoch": 5.586796116504854, + "grad_norm": 0.39028823375701904, + "learning_rate": 2.003984714947618e-05, + "loss": 0.005, + "step": 71930 + }, + { + "epoch": 5.5875728155339806, + "grad_norm": 0.32582345604896545, + "learning_rate": 2.0026611156713515e-05, + "loss": 0.0048, + "step": 71940 + }, + { + "epoch": 5.588349514563107, + "grad_norm": 0.09110553562641144, + "learning_rate": 2.0013378441795584e-05, + "loss": 0.007, + "step": 71950 + }, + { + "epoch": 5.589126213592233, + "grad_norm": 0.1474377065896988, + "learning_rate": 2.0000149006169506e-05, + "loss": 0.0051, + "step": 71960 + }, + { + "epoch": 5.589902912621359, + "grad_norm": 0.4082360565662384, + "learning_rate": 1.9986922851282048e-05, + "loss": 0.0061, + "step": 71970 + }, + { + "epoch": 5.5906796116504855, + "grad_norm": 0.34352385997772217, + "learning_rate": 1.997369997857956e-05, + "loss": 0.006, + "step": 71980 + }, + { + "epoch": 5.591456310679612, + "grad_norm": 0.11339441686868668, + "learning_rate": 1.9960480389508103e-05, + "loss": 0.0074, + "step": 71990 + }, + { + "epoch": 5.592233009708738, + "grad_norm": 0.12359363585710526, + "learning_rate": 1.9947264085513363e-05, + "loss": 0.0053, + "step": 72000 + }, + { + "epoch": 5.593009708737864, + "grad_norm": 0.18183191120624542, + "learning_rate": 1.993405106804062e-05, + "loss": 0.0026, + "step": 72010 + }, + { + "epoch": 5.5937864077669905, + "grad_norm": 0.20531485974788666, + "learning_rate": 1.9920841338534867e-05, + "loss": 0.005, + "step": 72020 + }, + { + "epoch": 5.594563106796117, + "grad_norm": 0.5755119323730469, + "learning_rate": 1.9907634898440646e-05, + "loss": 0.0037, + "step": 72030 + }, + { + "epoch": 5.595339805825243, + "grad_norm": 0.10292436927556992, + "learning_rate": 1.989443174920227e-05, + "loss": 0.0047, + "step": 72040 + }, + { + "epoch": 5.596116504854369, + "grad_norm": 0.4022420644760132, + "learning_rate": 1.988123189226355e-05, + "loss": 0.0064, + "step": 72050 + }, + { + "epoch": 5.596893203883495, + "grad_norm": 0.08659900724887848, + "learning_rate": 1.986803532906805e-05, + "loss": 0.0039, + "step": 72060 + }, + { + "epoch": 5.597669902912621, + "grad_norm": 0.2185819000005722, + "learning_rate": 1.985484206105889e-05, + "loss": 0.0047, + "step": 72070 + }, + { + "epoch": 5.598446601941747, + "grad_norm": 0.13699544966220856, + "learning_rate": 1.9841652089678887e-05, + "loss": 0.0042, + "step": 72080 + }, + { + "epoch": 5.599223300970873, + "grad_norm": 0.40940091013908386, + "learning_rate": 1.9828465416370435e-05, + "loss": 0.0043, + "step": 72090 + }, + { + "epoch": 5.6, + "grad_norm": 0.23526746034622192, + "learning_rate": 1.9815282042575685e-05, + "loss": 0.0035, + "step": 72100 + }, + { + "epoch": 5.600776699029126, + "grad_norm": 0.2630750238895416, + "learning_rate": 1.9802101969736285e-05, + "loss": 0.0079, + "step": 72110 + }, + { + "epoch": 5.601553398058252, + "grad_norm": 0.22585731744766235, + "learning_rate": 1.9788925199293633e-05, + "loss": 0.0045, + "step": 72120 + }, + { + "epoch": 5.602330097087378, + "grad_norm": 0.09021949768066406, + "learning_rate": 1.977575173268868e-05, + "loss": 0.0042, + "step": 72130 + }, + { + "epoch": 5.603106796116505, + "grad_norm": 0.2595090866088867, + "learning_rate": 1.9762581571362094e-05, + "loss": 0.0053, + "step": 72140 + }, + { + "epoch": 5.603883495145631, + "grad_norm": 0.1628202497959137, + "learning_rate": 1.9749414716754112e-05, + "loss": 0.0058, + "step": 72150 + }, + { + "epoch": 5.604660194174757, + "grad_norm": 0.6795814633369446, + "learning_rate": 1.973625117030465e-05, + "loss": 0.0041, + "step": 72160 + }, + { + "epoch": 5.605436893203883, + "grad_norm": 0.1830703765153885, + "learning_rate": 1.972309093345326e-05, + "loss": 0.0073, + "step": 72170 + }, + { + "epoch": 5.60621359223301, + "grad_norm": 0.2312757968902588, + "learning_rate": 1.970993400763915e-05, + "loss": 0.0028, + "step": 72180 + }, + { + "epoch": 5.606990291262136, + "grad_norm": 0.20677149295806885, + "learning_rate": 1.969678039430109e-05, + "loss": 0.0043, + "step": 72190 + }, + { + "epoch": 5.607766990291262, + "grad_norm": 0.24263818562030792, + "learning_rate": 1.9683630094877564e-05, + "loss": 0.0052, + "step": 72200 + }, + { + "epoch": 5.608543689320388, + "grad_norm": 0.2582593858242035, + "learning_rate": 1.9670483110806696e-05, + "loss": 0.0043, + "step": 72210 + }, + { + "epoch": 5.609320388349515, + "grad_norm": 0.33586812019348145, + "learning_rate": 1.9657339443526172e-05, + "loss": 0.0068, + "step": 72220 + }, + { + "epoch": 5.610097087378641, + "grad_norm": 0.7490130066871643, + "learning_rate": 1.964419909447339e-05, + "loss": 0.0041, + "step": 72230 + }, + { + "epoch": 5.610873786407767, + "grad_norm": 1.2031521797180176, + "learning_rate": 1.963106206508536e-05, + "loss": 0.005, + "step": 72240 + }, + { + "epoch": 5.611650485436893, + "grad_norm": 0.13848556578159332, + "learning_rate": 1.9617928356798738e-05, + "loss": 0.0043, + "step": 72250 + }, + { + "epoch": 5.61242718446602, + "grad_norm": 0.15286296606063843, + "learning_rate": 1.960479797104977e-05, + "loss": 0.0038, + "step": 72260 + }, + { + "epoch": 5.613203883495146, + "grad_norm": 0.10982449352741241, + "learning_rate": 1.9591670909274418e-05, + "loss": 0.0053, + "step": 72270 + }, + { + "epoch": 5.613980582524272, + "grad_norm": 0.35503846406936646, + "learning_rate": 1.9578547172908184e-05, + "loss": 0.0054, + "step": 72280 + }, + { + "epoch": 5.614757281553398, + "grad_norm": 0.17074616253376007, + "learning_rate": 1.9565426763386325e-05, + "loss": 0.0036, + "step": 72290 + }, + { + "epoch": 5.615533980582525, + "grad_norm": 0.20521599054336548, + "learning_rate": 1.9552309682143622e-05, + "loss": 0.0061, + "step": 72300 + }, + { + "epoch": 5.616310679611651, + "grad_norm": 0.1273922622203827, + "learning_rate": 1.9539195930614574e-05, + "loss": 0.0051, + "step": 72310 + }, + { + "epoch": 5.617087378640777, + "grad_norm": 0.5426204800605774, + "learning_rate": 1.952608551023324e-05, + "loss": 0.0039, + "step": 72320 + }, + { + "epoch": 5.617864077669903, + "grad_norm": 0.23458266258239746, + "learning_rate": 1.9512978422433397e-05, + "loss": 0.0034, + "step": 72330 + }, + { + "epoch": 5.61864077669903, + "grad_norm": 0.1897735446691513, + "learning_rate": 1.9499874668648376e-05, + "loss": 0.0054, + "step": 72340 + }, + { + "epoch": 5.619417475728155, + "grad_norm": 0.29198217391967773, + "learning_rate": 1.9486774250311213e-05, + "loss": 0.0058, + "step": 72350 + }, + { + "epoch": 5.620194174757281, + "grad_norm": 0.48439139127731323, + "learning_rate": 1.9473677168854537e-05, + "loss": 0.0045, + "step": 72360 + }, + { + "epoch": 5.6209708737864075, + "grad_norm": 1.3813987970352173, + "learning_rate": 1.946058342571065e-05, + "loss": 0.0044, + "step": 72370 + }, + { + "epoch": 5.621747572815534, + "grad_norm": 0.45497116446495056, + "learning_rate": 1.9447493022311424e-05, + "loss": 0.0053, + "step": 72380 + }, + { + "epoch": 5.62252427184466, + "grad_norm": 0.6406657099723816, + "learning_rate": 1.9434405960088442e-05, + "loss": 0.0052, + "step": 72390 + }, + { + "epoch": 5.623300970873786, + "grad_norm": 0.809800922870636, + "learning_rate": 1.9421322240472857e-05, + "loss": 0.0062, + "step": 72400 + }, + { + "epoch": 5.6240776699029125, + "grad_norm": 0.3971099555492401, + "learning_rate": 1.9408241864895494e-05, + "loss": 0.0046, + "step": 72410 + }, + { + "epoch": 5.624854368932039, + "grad_norm": 0.4044993817806244, + "learning_rate": 1.939516483478681e-05, + "loss": 0.0035, + "step": 72420 + }, + { + "epoch": 5.625631067961165, + "grad_norm": 0.12784597277641296, + "learning_rate": 1.9382091151576902e-05, + "loss": 0.0033, + "step": 72430 + }, + { + "epoch": 5.626407766990291, + "grad_norm": 0.26932746171951294, + "learning_rate": 1.9369020816695454e-05, + "loss": 0.007, + "step": 72440 + }, + { + "epoch": 5.6271844660194175, + "grad_norm": 0.46917954087257385, + "learning_rate": 1.935595383157184e-05, + "loss": 0.0053, + "step": 72450 + }, + { + "epoch": 5.627961165048544, + "grad_norm": 0.6201710104942322, + "learning_rate": 1.9342890197635057e-05, + "loss": 0.0034, + "step": 72460 + }, + { + "epoch": 5.62873786407767, + "grad_norm": 0.15729093551635742, + "learning_rate": 1.9329829916313685e-05, + "loss": 0.0077, + "step": 72470 + }, + { + "epoch": 5.629514563106796, + "grad_norm": 0.23682741820812225, + "learning_rate": 1.9316772989036013e-05, + "loss": 0.0042, + "step": 72480 + }, + { + "epoch": 5.6302912621359225, + "grad_norm": 0.13061843812465668, + "learning_rate": 1.930371941722991e-05, + "loss": 0.0044, + "step": 72490 + }, + { + "epoch": 5.631067961165049, + "grad_norm": 0.21334528923034668, + "learning_rate": 1.929066920232292e-05, + "loss": 0.0041, + "step": 72500 + }, + { + "epoch": 5.631844660194175, + "grad_norm": 0.3017542064189911, + "learning_rate": 1.9277622345742152e-05, + "loss": 0.0039, + "step": 72510 + }, + { + "epoch": 5.632621359223301, + "grad_norm": 0.06043602153658867, + "learning_rate": 1.926457884891444e-05, + "loss": 0.0048, + "step": 72520 + }, + { + "epoch": 5.6333980582524275, + "grad_norm": 0.7468801736831665, + "learning_rate": 1.9251538713266147e-05, + "loss": 0.0038, + "step": 72530 + }, + { + "epoch": 5.634174757281554, + "grad_norm": 0.3051735758781433, + "learning_rate": 1.9238501940223358e-05, + "loss": 0.0063, + "step": 72540 + }, + { + "epoch": 5.63495145631068, + "grad_norm": 0.3783198893070221, + "learning_rate": 1.922546853121175e-05, + "loss": 0.004, + "step": 72550 + }, + { + "epoch": 5.635728155339806, + "grad_norm": 0.5065764784812927, + "learning_rate": 1.9212438487656647e-05, + "loss": 0.0051, + "step": 72560 + }, + { + "epoch": 5.636504854368932, + "grad_norm": 0.8284860849380493, + "learning_rate": 1.9199411810982965e-05, + "loss": 0.0042, + "step": 72570 + }, + { + "epoch": 5.637281553398058, + "grad_norm": 0.08159002661705017, + "learning_rate": 1.918638850261532e-05, + "loss": 0.0026, + "step": 72580 + }, + { + "epoch": 5.638058252427184, + "grad_norm": 0.4509413242340088, + "learning_rate": 1.9173368563977885e-05, + "loss": 0.004, + "step": 72590 + }, + { + "epoch": 5.63883495145631, + "grad_norm": 0.18193796277046204, + "learning_rate": 1.9160351996494513e-05, + "loss": 0.0042, + "step": 72600 + }, + { + "epoch": 5.6396116504854366, + "grad_norm": 0.3760741949081421, + "learning_rate": 1.9147338801588677e-05, + "loss": 0.0048, + "step": 72610 + }, + { + "epoch": 5.640388349514563, + "grad_norm": 0.4123588502407074, + "learning_rate": 1.913432898068351e-05, + "loss": 0.0036, + "step": 72620 + }, + { + "epoch": 5.641165048543689, + "grad_norm": 0.08437094837427139, + "learning_rate": 1.912132253520169e-05, + "loss": 0.0037, + "step": 72630 + }, + { + "epoch": 5.641941747572815, + "grad_norm": 0.26910459995269775, + "learning_rate": 1.910831946656564e-05, + "loss": 0.0061, + "step": 72640 + }, + { + "epoch": 5.6427184466019416, + "grad_norm": 0.18945351243019104, + "learning_rate": 1.909531977619731e-05, + "loss": 0.004, + "step": 72650 + }, + { + "epoch": 5.643495145631068, + "grad_norm": 0.08586166799068451, + "learning_rate": 1.908232346551834e-05, + "loss": 0.0058, + "step": 72660 + }, + { + "epoch": 5.644271844660194, + "grad_norm": 0.2593387961387634, + "learning_rate": 1.9069330535949993e-05, + "loss": 0.0048, + "step": 72670 + }, + { + "epoch": 5.64504854368932, + "grad_norm": 0.2129310816526413, + "learning_rate": 1.9056340988913167e-05, + "loss": 0.0031, + "step": 72680 + }, + { + "epoch": 5.6458252427184465, + "grad_norm": 0.23294852674007416, + "learning_rate": 1.9043354825828357e-05, + "loss": 0.0043, + "step": 72690 + }, + { + "epoch": 5.646601941747573, + "grad_norm": 0.10789647698402405, + "learning_rate": 1.9030372048115724e-05, + "loss": 0.0036, + "step": 72700 + }, + { + "epoch": 5.647378640776699, + "grad_norm": 0.13537995517253876, + "learning_rate": 1.9017392657195056e-05, + "loss": 0.0031, + "step": 72710 + }, + { + "epoch": 5.648155339805825, + "grad_norm": 0.47548893094062805, + "learning_rate": 1.9004416654485724e-05, + "loss": 0.0053, + "step": 72720 + }, + { + "epoch": 5.6489320388349515, + "grad_norm": 0.17763184010982513, + "learning_rate": 1.8991444041406785e-05, + "loss": 0.0036, + "step": 72730 + }, + { + "epoch": 5.649708737864078, + "grad_norm": 0.3518480062484741, + "learning_rate": 1.8978474819376912e-05, + "loss": 0.0057, + "step": 72740 + }, + { + "epoch": 5.650485436893204, + "grad_norm": 0.14979198575019836, + "learning_rate": 1.8965508989814406e-05, + "loss": 0.0039, + "step": 72750 + }, + { + "epoch": 5.65126213592233, + "grad_norm": 0.1858118325471878, + "learning_rate": 1.8952546554137158e-05, + "loss": 0.0044, + "step": 72760 + }, + { + "epoch": 5.6520388349514565, + "grad_norm": 0.24596871435642242, + "learning_rate": 1.8939587513762757e-05, + "loss": 0.0049, + "step": 72770 + }, + { + "epoch": 5.652815533980583, + "grad_norm": 0.37209951877593994, + "learning_rate": 1.892663187010835e-05, + "loss": 0.006, + "step": 72780 + }, + { + "epoch": 5.653592233009709, + "grad_norm": 0.43226996064186096, + "learning_rate": 1.891367962459077e-05, + "loss": 0.0045, + "step": 72790 + }, + { + "epoch": 5.654368932038835, + "grad_norm": 0.3717838227748871, + "learning_rate": 1.890073077862644e-05, + "loss": 0.0044, + "step": 72800 + }, + { + "epoch": 5.6551456310679615, + "grad_norm": 0.24723488092422485, + "learning_rate": 1.888778533363146e-05, + "loss": 0.0041, + "step": 72810 + }, + { + "epoch": 5.655922330097088, + "grad_norm": 0.099437415599823, + "learning_rate": 1.8874843291021478e-05, + "loss": 0.0041, + "step": 72820 + }, + { + "epoch": 5.656699029126214, + "grad_norm": 0.19962264597415924, + "learning_rate": 1.8861904652211853e-05, + "loss": 0.0032, + "step": 72830 + }, + { + "epoch": 5.65747572815534, + "grad_norm": 0.5913354754447937, + "learning_rate": 1.8848969418617507e-05, + "loss": 0.0056, + "step": 72840 + }, + { + "epoch": 5.658252427184466, + "grad_norm": 0.134756937623024, + "learning_rate": 1.8836037591653044e-05, + "loss": 0.0034, + "step": 72850 + }, + { + "epoch": 5.659029126213592, + "grad_norm": 0.16210199892520905, + "learning_rate": 1.882310917273262e-05, + "loss": 0.0057, + "step": 72860 + }, + { + "epoch": 5.659805825242718, + "grad_norm": 0.5812098979949951, + "learning_rate": 1.881018416327014e-05, + "loss": 0.0053, + "step": 72870 + }, + { + "epoch": 5.660582524271844, + "grad_norm": 0.36083173751831055, + "learning_rate": 1.8797262564679014e-05, + "loss": 0.0044, + "step": 72880 + }, + { + "epoch": 5.661359223300971, + "grad_norm": 0.2994403541088104, + "learning_rate": 1.8784344378372355e-05, + "loss": 0.0035, + "step": 72890 + }, + { + "epoch": 5.662135922330097, + "grad_norm": 0.20852026343345642, + "learning_rate": 1.8771429605762843e-05, + "loss": 0.0027, + "step": 72900 + }, + { + "epoch": 5.662912621359223, + "grad_norm": 0.1670835167169571, + "learning_rate": 1.8758518248262835e-05, + "loss": 0.0053, + "step": 72910 + }, + { + "epoch": 5.663689320388349, + "grad_norm": 0.25604015588760376, + "learning_rate": 1.8745610307284305e-05, + "loss": 0.005, + "step": 72920 + }, + { + "epoch": 5.664466019417476, + "grad_norm": 0.23605510592460632, + "learning_rate": 1.8732705784238862e-05, + "loss": 0.0051, + "step": 72930 + }, + { + "epoch": 5.665242718446602, + "grad_norm": 0.5386414527893066, + "learning_rate": 1.8719804680537682e-05, + "loss": 0.0039, + "step": 72940 + }, + { + "epoch": 5.666019417475728, + "grad_norm": 0.28368228673934937, + "learning_rate": 1.870690699759164e-05, + "loss": 0.0051, + "step": 72950 + }, + { + "epoch": 5.666796116504854, + "grad_norm": 0.3352631628513336, + "learning_rate": 1.8694012736811213e-05, + "loss": 0.0062, + "step": 72960 + }, + { + "epoch": 5.667572815533981, + "grad_norm": 0.24614930152893066, + "learning_rate": 1.8681121899606463e-05, + "loss": 0.0058, + "step": 72970 + }, + { + "epoch": 5.668349514563107, + "grad_norm": 0.14663904905319214, + "learning_rate": 1.8668234487387144e-05, + "loss": 0.0042, + "step": 72980 + }, + { + "epoch": 5.669126213592233, + "grad_norm": 0.16663134098052979, + "learning_rate": 1.8655350501562592e-05, + "loss": 0.0047, + "step": 72990 + }, + { + "epoch": 5.669902912621359, + "grad_norm": 0.4374237358570099, + "learning_rate": 1.8642469943541796e-05, + "loss": 0.0045, + "step": 73000 + }, + { + "epoch": 5.670679611650486, + "grad_norm": 0.11435345560312271, + "learning_rate": 1.8629592814733328e-05, + "loss": 0.0041, + "step": 73010 + }, + { + "epoch": 5.671456310679612, + "grad_norm": 0.21478797495365143, + "learning_rate": 1.8616719116545438e-05, + "loss": 0.0069, + "step": 73020 + }, + { + "epoch": 5.672233009708738, + "grad_norm": 0.4233507513999939, + "learning_rate": 1.8603848850385952e-05, + "loss": 0.0039, + "step": 73030 + }, + { + "epoch": 5.673009708737864, + "grad_norm": 0.1342729777097702, + "learning_rate": 1.859098201766236e-05, + "loss": 0.0051, + "step": 73040 + }, + { + "epoch": 5.673786407766991, + "grad_norm": 0.27717792987823486, + "learning_rate": 1.8578118619781726e-05, + "loss": 0.0073, + "step": 73050 + }, + { + "epoch": 5.674563106796117, + "grad_norm": 0.1260887235403061, + "learning_rate": 1.856525865815083e-05, + "loss": 0.0062, + "step": 73060 + }, + { + "epoch": 5.675339805825242, + "grad_norm": 0.09567240625619888, + "learning_rate": 1.8552402134175966e-05, + "loss": 0.005, + "step": 73070 + }, + { + "epoch": 5.6761165048543685, + "grad_norm": 0.2803598642349243, + "learning_rate": 1.8539549049263143e-05, + "loss": 0.0034, + "step": 73080 + }, + { + "epoch": 5.676893203883495, + "grad_norm": 0.17062851786613464, + "learning_rate": 1.8526699404817917e-05, + "loss": 0.0047, + "step": 73090 + }, + { + "epoch": 5.677669902912621, + "grad_norm": 0.13961003720760345, + "learning_rate": 1.8513853202245546e-05, + "loss": 0.0056, + "step": 73100 + }, + { + "epoch": 5.678446601941747, + "grad_norm": 0.5102270841598511, + "learning_rate": 1.850101044295081e-05, + "loss": 0.0057, + "step": 73110 + }, + { + "epoch": 5.6792233009708735, + "grad_norm": 0.10360295325517654, + "learning_rate": 1.848817112833825e-05, + "loss": 0.0036, + "step": 73120 + }, + { + "epoch": 5.68, + "grad_norm": 0.10040480643510818, + "learning_rate": 1.8475335259811905e-05, + "loss": 0.0052, + "step": 73130 + }, + { + "epoch": 5.680776699029126, + "grad_norm": 0.39983972907066345, + "learning_rate": 1.8462502838775508e-05, + "loss": 0.0044, + "step": 73140 + }, + { + "epoch": 5.681553398058252, + "grad_norm": 0.32633349299430847, + "learning_rate": 1.8449673866632378e-05, + "loss": 0.0058, + "step": 73150 + }, + { + "epoch": 5.6823300970873785, + "grad_norm": 0.31776130199432373, + "learning_rate": 1.8436848344785473e-05, + "loss": 0.0044, + "step": 73160 + }, + { + "epoch": 5.683106796116505, + "grad_norm": 0.13140378892421722, + "learning_rate": 1.8424026274637378e-05, + "loss": 0.0052, + "step": 73170 + }, + { + "epoch": 5.683883495145631, + "grad_norm": 0.2682684361934662, + "learning_rate": 1.8411207657590312e-05, + "loss": 0.0045, + "step": 73180 + }, + { + "epoch": 5.684660194174757, + "grad_norm": 0.06844862550497055, + "learning_rate": 1.839839249504607e-05, + "loss": 0.0045, + "step": 73190 + }, + { + "epoch": 5.6854368932038835, + "grad_norm": 0.17995472252368927, + "learning_rate": 1.8385580788406108e-05, + "loss": 0.0047, + "step": 73200 + }, + { + "epoch": 5.68621359223301, + "grad_norm": 0.19062206149101257, + "learning_rate": 1.837277253907152e-05, + "loss": 0.0041, + "step": 73210 + }, + { + "epoch": 5.686990291262136, + "grad_norm": 0.20122312009334564, + "learning_rate": 1.8359967748442957e-05, + "loss": 0.0043, + "step": 73220 + }, + { + "epoch": 5.687766990291262, + "grad_norm": 0.0840652659535408, + "learning_rate": 1.8347166417920774e-05, + "loss": 0.005, + "step": 73230 + }, + { + "epoch": 5.6885436893203885, + "grad_norm": 0.09914775937795639, + "learning_rate": 1.833436854890484e-05, + "loss": 0.0057, + "step": 73240 + }, + { + "epoch": 5.689320388349515, + "grad_norm": 0.5326141715049744, + "learning_rate": 1.8321574142794796e-05, + "loss": 0.0094, + "step": 73250 + }, + { + "epoch": 5.690097087378641, + "grad_norm": 0.3156586289405823, + "learning_rate": 1.8308783200989748e-05, + "loss": 0.0042, + "step": 73260 + }, + { + "epoch": 5.690873786407767, + "grad_norm": 0.3291442096233368, + "learning_rate": 1.829599572488855e-05, + "loss": 0.0097, + "step": 73270 + }, + { + "epoch": 5.6916504854368934, + "grad_norm": 0.34456196427345276, + "learning_rate": 1.828321171588957e-05, + "loss": 0.0071, + "step": 73280 + }, + { + "epoch": 5.69242718446602, + "grad_norm": 0.13995859026908875, + "learning_rate": 1.827043117539089e-05, + "loss": 0.0053, + "step": 73290 + }, + { + "epoch": 5.693203883495146, + "grad_norm": 0.22947533428668976, + "learning_rate": 1.8257654104790116e-05, + "loss": 0.0051, + "step": 73300 + }, + { + "epoch": 5.693980582524272, + "grad_norm": 0.14562830328941345, + "learning_rate": 1.82448805054846e-05, + "loss": 0.0055, + "step": 73310 + }, + { + "epoch": 5.694757281553398, + "grad_norm": 0.24625779688358307, + "learning_rate": 1.823211037887119e-05, + "loss": 0.005, + "step": 73320 + }, + { + "epoch": 5.695533980582525, + "grad_norm": 0.9776873588562012, + "learning_rate": 1.821934372634645e-05, + "loss": 0.0039, + "step": 73330 + }, + { + "epoch": 5.696310679611651, + "grad_norm": 0.15118686854839325, + "learning_rate": 1.8206580549306472e-05, + "loss": 0.0042, + "step": 73340 + }, + { + "epoch": 5.697087378640777, + "grad_norm": 0.2403930425643921, + "learning_rate": 1.819382084914707e-05, + "loss": 0.0047, + "step": 73350 + }, + { + "epoch": 5.6978640776699025, + "grad_norm": 0.4400739371776581, + "learning_rate": 1.8181064627263577e-05, + "loss": 0.0037, + "step": 73360 + }, + { + "epoch": 5.698640776699029, + "grad_norm": 0.22626835107803345, + "learning_rate": 1.8168311885051022e-05, + "loss": 0.0048, + "step": 73370 + }, + { + "epoch": 5.699417475728155, + "grad_norm": 0.22392934560775757, + "learning_rate": 1.815556262390402e-05, + "loss": 0.0033, + "step": 73380 + }, + { + "epoch": 5.700194174757281, + "grad_norm": 0.09391508996486664, + "learning_rate": 1.8142816845216826e-05, + "loss": 0.0051, + "step": 73390 + }, + { + "epoch": 5.7009708737864075, + "grad_norm": 0.24723437428474426, + "learning_rate": 1.8130074550383276e-05, + "loss": 0.0041, + "step": 73400 + }, + { + "epoch": 5.701747572815534, + "grad_norm": 0.6624657511711121, + "learning_rate": 1.8117335740796848e-05, + "loss": 0.0072, + "step": 73410 + }, + { + "epoch": 5.70252427184466, + "grad_norm": 0.12164795398712158, + "learning_rate": 1.810460041785067e-05, + "loss": 0.004, + "step": 73420 + }, + { + "epoch": 5.703300970873786, + "grad_norm": 0.6691501140594482, + "learning_rate": 1.809186858293742e-05, + "loss": 0.0058, + "step": 73430 + }, + { + "epoch": 5.7040776699029125, + "grad_norm": 0.19459719955921173, + "learning_rate": 1.8079140237449448e-05, + "loss": 0.0039, + "step": 73440 + }, + { + "epoch": 5.704854368932039, + "grad_norm": 0.09711919724941254, + "learning_rate": 1.8066415382778707e-05, + "loss": 0.005, + "step": 73450 + }, + { + "epoch": 5.705631067961165, + "grad_norm": 0.1851174682378769, + "learning_rate": 1.8053694020316786e-05, + "loss": 0.0061, + "step": 73460 + }, + { + "epoch": 5.706407766990291, + "grad_norm": 0.7685784697532654, + "learning_rate": 1.8040976151454837e-05, + "loss": 0.0047, + "step": 73470 + }, + { + "epoch": 5.7071844660194175, + "grad_norm": 0.1626299023628235, + "learning_rate": 1.8028261777583704e-05, + "loss": 0.0059, + "step": 73480 + }, + { + "epoch": 5.707961165048544, + "grad_norm": 0.85196453332901, + "learning_rate": 1.8015550900093758e-05, + "loss": 0.0037, + "step": 73490 + }, + { + "epoch": 5.70873786407767, + "grad_norm": 0.41783806681632996, + "learning_rate": 1.800284352037512e-05, + "loss": 0.0039, + "step": 73500 + }, + { + "epoch": 5.709514563106796, + "grad_norm": 0.935681939125061, + "learning_rate": 1.799013963981738e-05, + "loss": 0.0035, + "step": 73510 + }, + { + "epoch": 5.7102912621359225, + "grad_norm": 0.20788271725177765, + "learning_rate": 1.797743925980987e-05, + "loss": 0.0032, + "step": 73520 + }, + { + "epoch": 5.711067961165049, + "grad_norm": 0.2200711965560913, + "learning_rate": 1.7964742381741435e-05, + "loss": 0.0036, + "step": 73530 + }, + { + "epoch": 5.711844660194175, + "grad_norm": 0.24234522879123688, + "learning_rate": 1.7952049007000627e-05, + "loss": 0.0036, + "step": 73540 + }, + { + "epoch": 5.712621359223301, + "grad_norm": 0.6022831797599792, + "learning_rate": 1.7939359136975535e-05, + "loss": 0.0038, + "step": 73550 + }, + { + "epoch": 5.7133980582524275, + "grad_norm": 0.4416830241680145, + "learning_rate": 1.7926672773053932e-05, + "loss": 0.0051, + "step": 73560 + }, + { + "epoch": 5.714174757281554, + "grad_norm": 0.1535177379846573, + "learning_rate": 1.791398991662317e-05, + "loss": 0.0061, + "step": 73570 + }, + { + "epoch": 5.714951456310679, + "grad_norm": 0.14645333588123322, + "learning_rate": 1.7901310569070244e-05, + "loss": 0.0064, + "step": 73580 + }, + { + "epoch": 5.715728155339805, + "grad_norm": 0.16451561450958252, + "learning_rate": 1.7888634731781724e-05, + "loss": 0.0023, + "step": 73590 + }, + { + "epoch": 5.716504854368932, + "grad_norm": 0.09651559591293335, + "learning_rate": 1.7875962406143844e-05, + "loss": 0.0056, + "step": 73600 + }, + { + "epoch": 5.717281553398058, + "grad_norm": 0.35261765122413635, + "learning_rate": 1.7863293593542403e-05, + "loss": 0.0055, + "step": 73610 + }, + { + "epoch": 5.718058252427184, + "grad_norm": 0.45527979731559753, + "learning_rate": 1.7850628295362853e-05, + "loss": 0.006, + "step": 73620 + }, + { + "epoch": 5.71883495145631, + "grad_norm": 0.06823711097240448, + "learning_rate": 1.7837966512990263e-05, + "loss": 0.0047, + "step": 73630 + }, + { + "epoch": 5.719611650485437, + "grad_norm": 0.2940475046634674, + "learning_rate": 1.782530824780931e-05, + "loss": 0.0029, + "step": 73640 + }, + { + "epoch": 5.720388349514563, + "grad_norm": 0.33635544776916504, + "learning_rate": 1.7812653501204263e-05, + "loss": 0.005, + "step": 73650 + }, + { + "epoch": 5.721165048543689, + "grad_norm": 0.3229898512363434, + "learning_rate": 1.780000227455903e-05, + "loss": 0.0057, + "step": 73660 + }, + { + "epoch": 5.721941747572815, + "grad_norm": 0.25655731558799744, + "learning_rate": 1.778735456925716e-05, + "loss": 0.0064, + "step": 73670 + }, + { + "epoch": 5.722718446601942, + "grad_norm": 0.4753132462501526, + "learning_rate": 1.7774710386681737e-05, + "loss": 0.0043, + "step": 73680 + }, + { + "epoch": 5.723495145631068, + "grad_norm": 0.15128710865974426, + "learning_rate": 1.7762069728215545e-05, + "loss": 0.0049, + "step": 73690 + }, + { + "epoch": 5.724271844660194, + "grad_norm": 0.4177230894565582, + "learning_rate": 1.7749432595240932e-05, + "loss": 0.0057, + "step": 73700 + }, + { + "epoch": 5.72504854368932, + "grad_norm": 0.21455952525138855, + "learning_rate": 1.773679898913991e-05, + "loss": 0.0049, + "step": 73710 + }, + { + "epoch": 5.725825242718447, + "grad_norm": 0.08289605379104614, + "learning_rate": 1.7724168911294015e-05, + "loss": 0.0032, + "step": 73720 + }, + { + "epoch": 5.726601941747573, + "grad_norm": 0.483632355928421, + "learning_rate": 1.7711542363084504e-05, + "loss": 0.0068, + "step": 73730 + }, + { + "epoch": 5.727378640776699, + "grad_norm": 0.2837047874927521, + "learning_rate": 1.7698919345892158e-05, + "loss": 0.0046, + "step": 73740 + }, + { + "epoch": 5.728155339805825, + "grad_norm": 0.40073519945144653, + "learning_rate": 1.768629986109743e-05, + "loss": 0.0041, + "step": 73750 + }, + { + "epoch": 5.728932038834952, + "grad_norm": 0.22356432676315308, + "learning_rate": 1.7673683910080357e-05, + "loss": 0.005, + "step": 73760 + }, + { + "epoch": 5.729708737864078, + "grad_norm": 0.2236374318599701, + "learning_rate": 1.7661071494220638e-05, + "loss": 0.0044, + "step": 73770 + }, + { + "epoch": 5.730485436893204, + "grad_norm": 0.27051541209220886, + "learning_rate": 1.7648462614897498e-05, + "loss": 0.0047, + "step": 73780 + }, + { + "epoch": 5.73126213592233, + "grad_norm": 0.16805385053157806, + "learning_rate": 1.7635857273489863e-05, + "loss": 0.0046, + "step": 73790 + }, + { + "epoch": 5.732038834951457, + "grad_norm": 0.15722379088401794, + "learning_rate": 1.7623255471376198e-05, + "loss": 0.0063, + "step": 73800 + }, + { + "epoch": 5.732815533980583, + "grad_norm": 0.12315690517425537, + "learning_rate": 1.7610657209934646e-05, + "loss": 0.0071, + "step": 73810 + }, + { + "epoch": 5.733592233009709, + "grad_norm": 0.11348598450422287, + "learning_rate": 1.7598062490542915e-05, + "loss": 0.0047, + "step": 73820 + }, + { + "epoch": 5.734368932038835, + "grad_norm": 0.23080094158649445, + "learning_rate": 1.7585471314578377e-05, + "loss": 0.0048, + "step": 73830 + }, + { + "epoch": 5.735145631067962, + "grad_norm": 0.16026251018047333, + "learning_rate": 1.7572883683417945e-05, + "loss": 0.0042, + "step": 73840 + }, + { + "epoch": 5.735922330097088, + "grad_norm": 0.4183599352836609, + "learning_rate": 1.7560299598438212e-05, + "loss": 0.0048, + "step": 73850 + }, + { + "epoch": 5.736699029126213, + "grad_norm": 0.10958994179964066, + "learning_rate": 1.7547719061015334e-05, + "loss": 0.005, + "step": 73860 + }, + { + "epoch": 5.7374757281553395, + "grad_norm": 0.20515723526477814, + "learning_rate": 1.75351420725251e-05, + "loss": 0.0066, + "step": 73870 + }, + { + "epoch": 5.738252427184466, + "grad_norm": 0.366514652967453, + "learning_rate": 1.7522568634342923e-05, + "loss": 0.0063, + "step": 73880 + }, + { + "epoch": 5.739029126213592, + "grad_norm": 0.19771690666675568, + "learning_rate": 1.7509998747843825e-05, + "loss": 0.0061, + "step": 73890 + }, + { + "epoch": 5.739805825242718, + "grad_norm": 0.22630782425403595, + "learning_rate": 1.7497432414402403e-05, + "loss": 0.006, + "step": 73900 + }, + { + "epoch": 5.7405825242718445, + "grad_norm": 0.10916502773761749, + "learning_rate": 1.7484869635392903e-05, + "loss": 0.0056, + "step": 73910 + }, + { + "epoch": 5.741359223300971, + "grad_norm": 0.15346765518188477, + "learning_rate": 1.747231041218919e-05, + "loss": 0.0045, + "step": 73920 + }, + { + "epoch": 5.742135922330097, + "grad_norm": 0.16274741291999817, + "learning_rate": 1.7459754746164692e-05, + "loss": 0.0045, + "step": 73930 + }, + { + "epoch": 5.742912621359223, + "grad_norm": 0.31282979249954224, + "learning_rate": 1.744720263869249e-05, + "loss": 0.0046, + "step": 73940 + }, + { + "epoch": 5.7436893203883495, + "grad_norm": 0.13367275893688202, + "learning_rate": 1.7434654091145263e-05, + "loss": 0.0038, + "step": 73950 + }, + { + "epoch": 5.744466019417476, + "grad_norm": 0.22836792469024658, + "learning_rate": 1.742210910489532e-05, + "loss": 0.0049, + "step": 73960 + }, + { + "epoch": 5.745242718446602, + "grad_norm": 0.1232609674334526, + "learning_rate": 1.7409567681314532e-05, + "loss": 0.0038, + "step": 73970 + }, + { + "epoch": 5.746019417475728, + "grad_norm": 0.14790835976600647, + "learning_rate": 1.7397029821774434e-05, + "loss": 0.0059, + "step": 73980 + }, + { + "epoch": 5.7467961165048544, + "grad_norm": 0.13190492987632751, + "learning_rate": 1.7384495527646126e-05, + "loss": 0.0051, + "step": 73990 + }, + { + "epoch": 5.747572815533981, + "grad_norm": 0.1173044741153717, + "learning_rate": 1.737196480030035e-05, + "loss": 0.0047, + "step": 74000 + }, + { + "epoch": 5.748349514563107, + "grad_norm": 0.4010501205921173, + "learning_rate": 1.7359437641107447e-05, + "loss": 0.0047, + "step": 74010 + }, + { + "epoch": 5.749126213592233, + "grad_norm": 0.1799708753824234, + "learning_rate": 1.7346914051437387e-05, + "loss": 0.0043, + "step": 74020 + }, + { + "epoch": 5.749902912621359, + "grad_norm": 0.1676599383354187, + "learning_rate": 1.733439403265969e-05, + "loss": 0.0049, + "step": 74030 + }, + { + "epoch": 5.750679611650486, + "grad_norm": 0.12244655191898346, + "learning_rate": 1.732187758614357e-05, + "loss": 0.0025, + "step": 74040 + }, + { + "epoch": 5.751456310679612, + "grad_norm": 0.33766353130340576, + "learning_rate": 1.730936471325777e-05, + "loss": 0.0063, + "step": 74050 + }, + { + "epoch": 5.752233009708738, + "grad_norm": 0.21426169574260712, + "learning_rate": 1.7296855415370718e-05, + "loss": 0.0037, + "step": 74060 + }, + { + "epoch": 5.753009708737864, + "grad_norm": 0.3517928421497345, + "learning_rate": 1.728434969385035e-05, + "loss": 0.0063, + "step": 74070 + }, + { + "epoch": 5.75378640776699, + "grad_norm": 0.15172548592090607, + "learning_rate": 1.7271847550064346e-05, + "loss": 0.0059, + "step": 74080 + }, + { + "epoch": 5.754563106796116, + "grad_norm": 0.13442964851856232, + "learning_rate": 1.725934898537987e-05, + "loss": 0.0052, + "step": 74090 + }, + { + "epoch": 5.755339805825242, + "grad_norm": 0.24583083391189575, + "learning_rate": 1.7246854001163783e-05, + "loss": 0.0031, + "step": 74100 + }, + { + "epoch": 5.7561165048543685, + "grad_norm": 0.17166993021965027, + "learning_rate": 1.7234362598782476e-05, + "loss": 0.0058, + "step": 74110 + }, + { + "epoch": 5.756893203883495, + "grad_norm": 0.21764227747917175, + "learning_rate": 1.722187477960202e-05, + "loss": 0.0054, + "step": 74120 + }, + { + "epoch": 5.757669902912621, + "grad_norm": 0.1289799064397812, + "learning_rate": 1.7209390544988057e-05, + "loss": 0.0054, + "step": 74130 + }, + { + "epoch": 5.758446601941747, + "grad_norm": 0.3371277153491974, + "learning_rate": 1.7196909896305856e-05, + "loss": 0.0057, + "step": 74140 + }, + { + "epoch": 5.7592233009708735, + "grad_norm": 0.13160422444343567, + "learning_rate": 1.718443283492026e-05, + "loss": 0.0042, + "step": 74150 + }, + { + "epoch": 5.76, + "grad_norm": 0.06987173110246658, + "learning_rate": 1.7171959362195746e-05, + "loss": 0.0031, + "step": 74160 + }, + { + "epoch": 5.760776699029126, + "grad_norm": 0.27888718247413635, + "learning_rate": 1.715948947949642e-05, + "loss": 0.0055, + "step": 74170 + }, + { + "epoch": 5.761553398058252, + "grad_norm": 0.12625613808631897, + "learning_rate": 1.714702318818593e-05, + "loss": 0.003, + "step": 74180 + }, + { + "epoch": 5.7623300970873785, + "grad_norm": 0.16352032124996185, + "learning_rate": 1.7134560489627592e-05, + "loss": 0.0057, + "step": 74190 + }, + { + "epoch": 5.763106796116505, + "grad_norm": 0.23704107105731964, + "learning_rate": 1.712210138518431e-05, + "loss": 0.0038, + "step": 74200 + }, + { + "epoch": 5.763883495145631, + "grad_norm": 0.16102801263332367, + "learning_rate": 1.71096458762186e-05, + "loss": 0.0037, + "step": 74210 + }, + { + "epoch": 5.764660194174757, + "grad_norm": 0.17524847388267517, + "learning_rate": 1.7097193964092552e-05, + "loss": 0.0041, + "step": 74220 + }, + { + "epoch": 5.7654368932038835, + "grad_norm": 0.2186160832643509, + "learning_rate": 1.708474565016792e-05, + "loss": 0.0054, + "step": 74230 + }, + { + "epoch": 5.76621359223301, + "grad_norm": 0.52439284324646, + "learning_rate": 1.7072300935806e-05, + "loss": 0.0039, + "step": 74240 + }, + { + "epoch": 5.766990291262136, + "grad_norm": 0.1643819808959961, + "learning_rate": 1.705985982236776e-05, + "loss": 0.0045, + "step": 74250 + }, + { + "epoch": 5.767766990291262, + "grad_norm": 0.254881888628006, + "learning_rate": 1.7047422311213697e-05, + "loss": 0.0054, + "step": 74260 + }, + { + "epoch": 5.7685436893203885, + "grad_norm": 0.13924409449100494, + "learning_rate": 1.703498840370401e-05, + "loss": 0.0045, + "step": 74270 + }, + { + "epoch": 5.769320388349515, + "grad_norm": 0.19705699384212494, + "learning_rate": 1.702255810119842e-05, + "loss": 0.0027, + "step": 74280 + }, + { + "epoch": 5.770097087378641, + "grad_norm": 0.33345112204551697, + "learning_rate": 1.7010131405056305e-05, + "loss": 0.0056, + "step": 74290 + }, + { + "epoch": 5.770873786407767, + "grad_norm": 0.9862732291221619, + "learning_rate": 1.6997708316636613e-05, + "loss": 0.0032, + "step": 74300 + }, + { + "epoch": 5.7716504854368935, + "grad_norm": 0.7183122634887695, + "learning_rate": 1.6985288837297932e-05, + "loss": 0.0059, + "step": 74310 + }, + { + "epoch": 5.77242718446602, + "grad_norm": 0.4024583697319031, + "learning_rate": 1.6972872968398394e-05, + "loss": 0.0053, + "step": 74320 + }, + { + "epoch": 5.773203883495146, + "grad_norm": 0.10617725551128387, + "learning_rate": 1.6960460711295845e-05, + "loss": 0.0031, + "step": 74330 + }, + { + "epoch": 5.773980582524272, + "grad_norm": 0.18604782223701477, + "learning_rate": 1.6948052067347623e-05, + "loss": 0.0038, + "step": 74340 + }, + { + "epoch": 5.7747572815533985, + "grad_norm": 0.15966592729091644, + "learning_rate": 1.693564703791075e-05, + "loss": 0.0041, + "step": 74350 + }, + { + "epoch": 5.775533980582525, + "grad_norm": 0.07107516378164291, + "learning_rate": 1.6923245624341793e-05, + "loss": 0.0039, + "step": 74360 + }, + { + "epoch": 5.77631067961165, + "grad_norm": 0.1507304310798645, + "learning_rate": 1.6910847827996962e-05, + "loss": 0.0038, + "step": 74370 + }, + { + "epoch": 5.777087378640776, + "grad_norm": 0.22230476140975952, + "learning_rate": 1.689845365023206e-05, + "loss": 0.0053, + "step": 74380 + }, + { + "epoch": 5.777864077669903, + "grad_norm": 1.5478755235671997, + "learning_rate": 1.6886063092402505e-05, + "loss": 0.0059, + "step": 74390 + }, + { + "epoch": 5.778640776699029, + "grad_norm": 0.5301185250282288, + "learning_rate": 1.6873676155863317e-05, + "loss": 0.0053, + "step": 74400 + }, + { + "epoch": 5.779417475728155, + "grad_norm": 0.1631004959344864, + "learning_rate": 1.6861292841969085e-05, + "loss": 0.0046, + "step": 74410 + }, + { + "epoch": 5.780194174757281, + "grad_norm": 0.19545361399650574, + "learning_rate": 1.684891315207406e-05, + "loss": 0.005, + "step": 74420 + }, + { + "epoch": 5.780970873786408, + "grad_norm": 0.2649168074131012, + "learning_rate": 1.6836537087532034e-05, + "loss": 0.0053, + "step": 74430 + }, + { + "epoch": 5.781747572815534, + "grad_norm": 0.2790067195892334, + "learning_rate": 1.6824164649696465e-05, + "loss": 0.004, + "step": 74440 + }, + { + "epoch": 5.78252427184466, + "grad_norm": 0.1742616444826126, + "learning_rate": 1.681179583992034e-05, + "loss": 0.0038, + "step": 74450 + }, + { + "epoch": 5.783300970873786, + "grad_norm": 0.2158963531255722, + "learning_rate": 1.679943065955636e-05, + "loss": 0.0048, + "step": 74460 + }, + { + "epoch": 5.784077669902913, + "grad_norm": 0.10752803087234497, + "learning_rate": 1.678706910995671e-05, + "loss": 0.0033, + "step": 74470 + }, + { + "epoch": 5.784854368932039, + "grad_norm": 0.2832319438457489, + "learning_rate": 1.6774711192473262e-05, + "loss": 0.0046, + "step": 74480 + }, + { + "epoch": 5.785631067961165, + "grad_norm": 0.20122180879116058, + "learning_rate": 1.676235690845743e-05, + "loss": 0.0038, + "step": 74490 + }, + { + "epoch": 5.786407766990291, + "grad_norm": 0.34596845507621765, + "learning_rate": 1.6750006259260285e-05, + "loss": 0.0087, + "step": 74500 + }, + { + "epoch": 5.787184466019418, + "grad_norm": 0.2086760252714157, + "learning_rate": 1.673765924623244e-05, + "loss": 0.0033, + "step": 74510 + }, + { + "epoch": 5.787961165048544, + "grad_norm": 0.5349002480506897, + "learning_rate": 1.67253158707242e-05, + "loss": 0.0035, + "step": 74520 + }, + { + "epoch": 5.78873786407767, + "grad_norm": 0.5538179278373718, + "learning_rate": 1.671297613408537e-05, + "loss": 0.0036, + "step": 74530 + }, + { + "epoch": 5.789514563106796, + "grad_norm": 0.1687629520893097, + "learning_rate": 1.6700640037665444e-05, + "loss": 0.0052, + "step": 74540 + }, + { + "epoch": 5.790291262135923, + "grad_norm": 0.29811933636665344, + "learning_rate": 1.668830758281344e-05, + "loss": 0.0049, + "step": 74550 + }, + { + "epoch": 5.791067961165049, + "grad_norm": 0.2591138184070587, + "learning_rate": 1.6675978770878052e-05, + "loss": 0.0051, + "step": 74560 + }, + { + "epoch": 5.791844660194175, + "grad_norm": 0.24148724973201752, + "learning_rate": 1.6663653603207506e-05, + "loss": 0.0066, + "step": 74570 + }, + { + "epoch": 5.792621359223301, + "grad_norm": 0.4629233479499817, + "learning_rate": 1.6651332081149685e-05, + "loss": 0.0067, + "step": 74580 + }, + { + "epoch": 5.793398058252427, + "grad_norm": 0.20280848443508148, + "learning_rate": 1.6639014206052046e-05, + "loss": 0.0056, + "step": 74590 + }, + { + "epoch": 5.794174757281553, + "grad_norm": 0.4607279300689697, + "learning_rate": 1.6626699979261657e-05, + "loss": 0.0039, + "step": 74600 + }, + { + "epoch": 5.794951456310679, + "grad_norm": 0.15410923957824707, + "learning_rate": 1.6614389402125197e-05, + "loss": 0.0044, + "step": 74610 + }, + { + "epoch": 5.7957281553398055, + "grad_norm": 0.1437387317419052, + "learning_rate": 1.66020824759889e-05, + "loss": 0.0032, + "step": 74620 + }, + { + "epoch": 5.796504854368932, + "grad_norm": 0.5782057046890259, + "learning_rate": 1.6589779202198675e-05, + "loss": 0.0047, + "step": 74630 + }, + { + "epoch": 5.797281553398058, + "grad_norm": 0.1388140469789505, + "learning_rate": 1.6577479582099924e-05, + "loss": 0.0043, + "step": 74640 + }, + { + "epoch": 5.798058252427184, + "grad_norm": 0.12533800303936005, + "learning_rate": 1.6565183617037788e-05, + "loss": 0.0035, + "step": 74650 + }, + { + "epoch": 5.7988349514563104, + "grad_norm": 0.18736664950847626, + "learning_rate": 1.6552891308356894e-05, + "loss": 0.0062, + "step": 74660 + }, + { + "epoch": 5.799611650485437, + "grad_norm": 0.3181251585483551, + "learning_rate": 1.6540602657401526e-05, + "loss": 0.0042, + "step": 74670 + }, + { + "epoch": 5.800388349514563, + "grad_norm": 0.3724377453327179, + "learning_rate": 1.6528317665515537e-05, + "loss": 0.0033, + "step": 74680 + }, + { + "epoch": 5.801165048543689, + "grad_norm": 0.2265477031469345, + "learning_rate": 1.6516036334042416e-05, + "loss": 0.0043, + "step": 74690 + }, + { + "epoch": 5.8019417475728154, + "grad_norm": 0.18381470441818237, + "learning_rate": 1.650375866432519e-05, + "loss": 0.0048, + "step": 74700 + }, + { + "epoch": 5.802718446601942, + "grad_norm": 0.26919180154800415, + "learning_rate": 1.649148465770659e-05, + "loss": 0.0076, + "step": 74710 + }, + { + "epoch": 5.803495145631068, + "grad_norm": 0.15456941723823547, + "learning_rate": 1.647921431552883e-05, + "loss": 0.0055, + "step": 74720 + }, + { + "epoch": 5.804271844660194, + "grad_norm": 0.4841707944869995, + "learning_rate": 1.646694763913381e-05, + "loss": 0.0056, + "step": 74730 + }, + { + "epoch": 5.80504854368932, + "grad_norm": 0.23396067321300507, + "learning_rate": 1.6454684629862966e-05, + "loss": 0.0027, + "step": 74740 + }, + { + "epoch": 5.805825242718447, + "grad_norm": 0.08742174506187439, + "learning_rate": 1.6442425289057388e-05, + "loss": 0.003, + "step": 74750 + }, + { + "epoch": 5.806601941747573, + "grad_norm": 0.14959730207920074, + "learning_rate": 1.6430169618057718e-05, + "loss": 0.0046, + "step": 74760 + }, + { + "epoch": 5.807378640776699, + "grad_norm": 0.26150384545326233, + "learning_rate": 1.641791761820422e-05, + "loss": 0.0033, + "step": 74770 + }, + { + "epoch": 5.808155339805825, + "grad_norm": 0.09518764168024063, + "learning_rate": 1.640566929083676e-05, + "loss": 0.0058, + "step": 74780 + }, + { + "epoch": 5.808932038834952, + "grad_norm": 0.16960468888282776, + "learning_rate": 1.6393424637294823e-05, + "loss": 0.0043, + "step": 74790 + }, + { + "epoch": 5.809708737864078, + "grad_norm": 0.44780293107032776, + "learning_rate": 1.638118365891742e-05, + "loss": 0.005, + "step": 74800 + }, + { + "epoch": 5.810485436893204, + "grad_norm": 0.19198812544345856, + "learning_rate": 1.636894635704323e-05, + "loss": 0.0038, + "step": 74810 + }, + { + "epoch": 5.81126213592233, + "grad_norm": 0.20887693762779236, + "learning_rate": 1.6356712733010517e-05, + "loss": 0.006, + "step": 74820 + }, + { + "epoch": 5.812038834951457, + "grad_norm": 0.24888084828853607, + "learning_rate": 1.6344482788157096e-05, + "loss": 0.0042, + "step": 74830 + }, + { + "epoch": 5.812815533980583, + "grad_norm": 2.1018285751342773, + "learning_rate": 1.6332256523820444e-05, + "loss": 0.0069, + "step": 74840 + }, + { + "epoch": 5.813592233009709, + "grad_norm": 0.24409830570220947, + "learning_rate": 1.6320033941337597e-05, + "loss": 0.0044, + "step": 74850 + }, + { + "epoch": 5.814368932038835, + "grad_norm": 0.13674676418304443, + "learning_rate": 1.6307815042045216e-05, + "loss": 0.0042, + "step": 74860 + }, + { + "epoch": 5.815145631067961, + "grad_norm": 0.41597577929496765, + "learning_rate": 1.6295599827279513e-05, + "loss": 0.0051, + "step": 74870 + }, + { + "epoch": 5.815922330097087, + "grad_norm": 0.44086095690727234, + "learning_rate": 1.6283388298376356e-05, + "loss": 0.0043, + "step": 74880 + }, + { + "epoch": 5.816699029126213, + "grad_norm": 0.6433510780334473, + "learning_rate": 1.6271180456671125e-05, + "loss": 0.0091, + "step": 74890 + }, + { + "epoch": 5.8174757281553395, + "grad_norm": 0.23058666288852692, + "learning_rate": 1.625897630349893e-05, + "loss": 0.0045, + "step": 74900 + }, + { + "epoch": 5.818252427184466, + "grad_norm": 0.1291574239730835, + "learning_rate": 1.6246775840194335e-05, + "loss": 0.0044, + "step": 74910 + }, + { + "epoch": 5.819029126213592, + "grad_norm": 0.13870038092136383, + "learning_rate": 1.6234579068091604e-05, + "loss": 0.0047, + "step": 74920 + }, + { + "epoch": 5.819805825242718, + "grad_norm": 0.11001831293106079, + "learning_rate": 1.6222385988524524e-05, + "loss": 0.0042, + "step": 74930 + }, + { + "epoch": 5.8205825242718445, + "grad_norm": 0.16868005692958832, + "learning_rate": 1.6210196602826544e-05, + "loss": 0.0042, + "step": 74940 + }, + { + "epoch": 5.821359223300971, + "grad_norm": 0.15808425843715668, + "learning_rate": 1.6198010912330646e-05, + "loss": 0.0041, + "step": 74950 + }, + { + "epoch": 5.822135922330097, + "grad_norm": 0.09932199865579605, + "learning_rate": 1.6185828918369456e-05, + "loss": 0.0034, + "step": 74960 + }, + { + "epoch": 5.822912621359223, + "grad_norm": 0.5477966070175171, + "learning_rate": 1.617365062227517e-05, + "loss": 0.0045, + "step": 74970 + }, + { + "epoch": 5.8236893203883495, + "grad_norm": 0.1503005474805832, + "learning_rate": 1.616147602537962e-05, + "loss": 0.0048, + "step": 74980 + }, + { + "epoch": 5.824466019417476, + "grad_norm": 0.13380728662014008, + "learning_rate": 1.614930512901415e-05, + "loss": 0.0035, + "step": 74990 + }, + { + "epoch": 5.825242718446602, + "grad_norm": 0.24064573645591736, + "learning_rate": 1.6137137934509804e-05, + "loss": 0.0057, + "step": 75000 + } + ], + "logging_steps": 10, + "max_steps": 100000, + "num_input_tokens_seen": 0, + "num_train_epochs": 8, + "save_steps": 25000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 32, + "trial_name": null, + "trial_params": null +}