diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,54063 @@ +{ + "best_global_step": 5390, + "best_metric": 0.9517125210555868, + "best_model_checkpoint": "/mnt/data2/weizhi/checkpoints/unifilter_llava_qwen3_0.6b_instruct_caption_interleaved_joint_mse_loss_siglip2_384_mmtoken_144_400k_data/checkpoint-5390", + "epoch": 10.0, + "eval_steps": 500, + "global_step": 7700, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0012987012987012987, + "grad_norm": 10720.807150415143, + "learning_rate": 0.0, + "loss": 62.9062, + "step": 1 + }, + { + "epoch": 0.0025974025974025974, + "grad_norm": 10358.352292753721, + "learning_rate": 8.658008658008659e-08, + "loss": 60.0312, + "step": 2 + }, + { + "epoch": 0.003896103896103896, + "grad_norm": 10380.856026776622, + "learning_rate": 1.7316017316017318e-07, + "loss": 60.1562, + "step": 3 + }, + { + "epoch": 0.005194805194805195, + "grad_norm": 10389.035543263068, + "learning_rate": 2.597402597402598e-07, + "loss": 60.0, + "step": 4 + }, + { + "epoch": 0.006493506493506494, + "grad_norm": 10646.044548098334, + "learning_rate": 3.4632034632034636e-07, + "loss": 62.6875, + "step": 5 + }, + { + "epoch": 0.007792207792207792, + "grad_norm": 10396.426127289324, + "learning_rate": 4.329004329004329e-07, + "loss": 59.9688, + "step": 6 + }, + { + "epoch": 0.00909090909090909, + "grad_norm": 10202.716484265176, + "learning_rate": 5.194805194805196e-07, + "loss": 58.5938, + "step": 7 + }, + { + "epoch": 0.01038961038961039, + "grad_norm": 10244.755936325682, + "learning_rate": 6.060606060606061e-07, + "loss": 58.5, + "step": 8 + }, + { + "epoch": 0.011688311688311689, + "grad_norm": 9910.499012963817, + "learning_rate": 6.926406926406927e-07, + "loss": 54.9375, + "step": 9 + }, + { + "epoch": 0.012987012987012988, + "grad_norm": 9655.382715356187, + "learning_rate": 7.792207792207792e-07, + "loss": 53.2188, + "step": 10 + }, + { + "epoch": 0.014285714285714285, + "grad_norm": 8366.070750027962, + "learning_rate": 8.658008658008658e-07, + "loss": 41.3125, + "step": 11 + }, + { + "epoch": 0.015584415584415584, + "grad_norm": 8271.37355662847, + "learning_rate": 9.523809523809525e-07, + "loss": 41.0625, + "step": 12 + }, + { + "epoch": 0.016883116883116882, + "grad_norm": 8142.11882141837, + "learning_rate": 1.0389610389610392e-06, + "loss": 39.5312, + "step": 13 + }, + { + "epoch": 0.01818181818181818, + "grad_norm": 7750.990345516432, + "learning_rate": 1.1255411255411256e-06, + "loss": 36.2812, + "step": 14 + }, + { + "epoch": 0.01948051948051948, + "grad_norm": 3572.6307258785323, + "learning_rate": 1.2121212121212122e-06, + "loss": 10.75, + "step": 15 + }, + { + "epoch": 0.02077922077922078, + "grad_norm": 3045.2367518702345, + "learning_rate": 1.2987012987012986e-06, + "loss": 8.9609, + "step": 16 + }, + { + "epoch": 0.02207792207792208, + "grad_norm": 2762.9909418984716, + "learning_rate": 1.3852813852813854e-06, + "loss": 7.6367, + "step": 17 + }, + { + "epoch": 0.023376623376623377, + "grad_norm": 2553.5598370805455, + "learning_rate": 1.471861471861472e-06, + "loss": 6.7109, + "step": 18 + }, + { + "epoch": 0.024675324675324677, + "grad_norm": 1556.5904168168804, + "learning_rate": 1.5584415584415584e-06, + "loss": 3.6602, + "step": 19 + }, + { + "epoch": 0.025974025974025976, + "grad_norm": 1171.953666080925, + "learning_rate": 1.6450216450216453e-06, + "loss": 2.7402, + "step": 20 + }, + { + "epoch": 0.02727272727272727, + "grad_norm": 1665.5653112178582, + "learning_rate": 1.7316017316017317e-06, + "loss": 3.8594, + "step": 21 + }, + { + "epoch": 0.02857142857142857, + "grad_norm": 6649.436010361148, + "learning_rate": 1.8181818181818183e-06, + "loss": 45.9375, + "step": 22 + }, + { + "epoch": 0.02987012987012987, + "grad_norm": 6797.14758571221, + "learning_rate": 1.904761904761905e-06, + "loss": 51.4375, + "step": 23 + }, + { + "epoch": 0.03116883116883117, + "grad_norm": 6215.109052988516, + "learning_rate": 1.9913419913419915e-06, + "loss": 44.2812, + "step": 24 + }, + { + "epoch": 0.032467532467532464, + "grad_norm": 5542.896487261591, + "learning_rate": 2.0779220779220784e-06, + "loss": 36.4375, + "step": 25 + }, + { + "epoch": 0.033766233766233764, + "grad_norm": 4557.443850154783, + "learning_rate": 2.1645021645021648e-06, + "loss": 26.4531, + "step": 26 + }, + { + "epoch": 0.03506493506493506, + "grad_norm": 3452.6910170029596, + "learning_rate": 2.251082251082251e-06, + "loss": 16.6328, + "step": 27 + }, + { + "epoch": 0.03636363636363636, + "grad_norm": 2408.1362419141533, + "learning_rate": 2.337662337662338e-06, + "loss": 9.0859, + "step": 28 + }, + { + "epoch": 0.03766233766233766, + "grad_norm": 1550.249552249401, + "learning_rate": 2.4242424242424244e-06, + "loss": 4.8398, + "step": 29 + }, + { + "epoch": 0.03896103896103896, + "grad_norm": 527.3959964476995, + "learning_rate": 2.510822510822511e-06, + "loss": 1.7617, + "step": 30 + }, + { + "epoch": 0.04025974025974026, + "grad_norm": 454.13282160972943, + "learning_rate": 2.597402597402597e-06, + "loss": 1.6602, + "step": 31 + }, + { + "epoch": 0.04155844155844156, + "grad_norm": 1296.6760987399068, + "learning_rate": 2.6839826839826844e-06, + "loss": 4.1602, + "step": 32 + }, + { + "epoch": 0.04285714285714286, + "grad_norm": 1768.7696083858784, + "learning_rate": 2.770562770562771e-06, + "loss": 6.8047, + "step": 33 + }, + { + "epoch": 0.04415584415584416, + "grad_norm": 1947.7736027768512, + "learning_rate": 2.8571428571428573e-06, + "loss": 8.1016, + "step": 34 + }, + { + "epoch": 0.045454545454545456, + "grad_norm": 2038.044943897444, + "learning_rate": 2.943722943722944e-06, + "loss": 8.8633, + "step": 35 + }, + { + "epoch": 0.046753246753246755, + "grad_norm": 1915.8308124815248, + "learning_rate": 3.0303030303030305e-06, + "loss": 8.2539, + "step": 36 + }, + { + "epoch": 0.048051948051948054, + "grad_norm": 1546.649126107092, + "learning_rate": 3.116883116883117e-06, + "loss": 5.9414, + "step": 37 + }, + { + "epoch": 0.04935064935064935, + "grad_norm": 1168.3231863626445, + "learning_rate": 3.2034632034632033e-06, + "loss": 4.1387, + "step": 38 + }, + { + "epoch": 0.05064935064935065, + "grad_norm": 805.0018102160935, + "learning_rate": 3.2900432900432905e-06, + "loss": 2.6094, + "step": 39 + }, + { + "epoch": 0.05194805194805195, + "grad_norm": 399.0115601579254, + "learning_rate": 3.376623376623377e-06, + "loss": 1.6641, + "step": 40 + }, + { + "epoch": 0.053246753246753244, + "grad_norm": 31.721573697124857, + "learning_rate": 3.4632034632034634e-06, + "loss": 1.2231, + "step": 41 + }, + { + "epoch": 0.05454545454545454, + "grad_norm": 353.4354760245915, + "learning_rate": 3.54978354978355e-06, + "loss": 1.5156, + "step": 42 + }, + { + "epoch": 0.05584415584415584, + "grad_norm": 651.8308805252252, + "learning_rate": 3.6363636363636366e-06, + "loss": 2.1289, + "step": 43 + }, + { + "epoch": 0.05714285714285714, + "grad_norm": 827.519418776379, + "learning_rate": 3.722943722943723e-06, + "loss": 2.7266, + "step": 44 + }, + { + "epoch": 0.05844155844155844, + "grad_norm": 983.9140977926254, + "learning_rate": 3.80952380952381e-06, + "loss": 3.2217, + "step": 45 + }, + { + "epoch": 0.05974025974025974, + "grad_norm": 975.4148942429965, + "learning_rate": 3.896103896103897e-06, + "loss": 3.1309, + "step": 46 + }, + { + "epoch": 0.06103896103896104, + "grad_norm": 840.9852548818782, + "learning_rate": 3.982683982683983e-06, + "loss": 2.8105, + "step": 47 + }, + { + "epoch": 0.06233766233766234, + "grad_norm": 646.8283085417288, + "learning_rate": 4.0692640692640695e-06, + "loss": 2.1211, + "step": 48 + }, + { + "epoch": 0.06363636363636363, + "grad_norm": 335.4029097458437, + "learning_rate": 4.155844155844157e-06, + "loss": 1.4834, + "step": 49 + }, + { + "epoch": 0.06493506493506493, + "grad_norm": 46.09595872925314, + "learning_rate": 4.242424242424243e-06, + "loss": 1.2246, + "step": 50 + }, + { + "epoch": 0.06623376623376623, + "grad_norm": 192.21020347490924, + "learning_rate": 4.3290043290043295e-06, + "loss": 1.4062, + "step": 51 + }, + { + "epoch": 0.06753246753246753, + "grad_norm": 346.7389066537953, + "learning_rate": 4.415584415584416e-06, + "loss": 1.46, + "step": 52 + }, + { + "epoch": 0.06883116883116883, + "grad_norm": 498.8304345124628, + "learning_rate": 4.502164502164502e-06, + "loss": 1.8623, + "step": 53 + }, + { + "epoch": 0.07012987012987013, + "grad_norm": 650.2768378617453, + "learning_rate": 4.5887445887445896e-06, + "loss": 2.1621, + "step": 54 + }, + { + "epoch": 0.07142857142857142, + "grad_norm": 575.912354294176, + "learning_rate": 4.675324675324676e-06, + "loss": 2.083, + "step": 55 + }, + { + "epoch": 0.07272727272727272, + "grad_norm": 461.4243264344223, + "learning_rate": 4.761904761904762e-06, + "loss": 1.7979, + "step": 56 + }, + { + "epoch": 0.07402597402597402, + "grad_norm": 308.6692689753203, + "learning_rate": 4.848484848484849e-06, + "loss": 1.498, + "step": 57 + }, + { + "epoch": 0.07532467532467532, + "grad_norm": 58.1750617476112, + "learning_rate": 4.935064935064935e-06, + "loss": 1.3369, + "step": 58 + }, + { + "epoch": 0.07662337662337662, + "grad_norm": 81.12260548267551, + "learning_rate": 5.021645021645022e-06, + "loss": 1.3096, + "step": 59 + }, + { + "epoch": 0.07792207792207792, + "grad_norm": 213.29552396003405, + "learning_rate": 5.108225108225109e-06, + "loss": 1.418, + "step": 60 + }, + { + "epoch": 0.07922077922077922, + "grad_norm": 386.2023600422629, + "learning_rate": 5.194805194805194e-06, + "loss": 1.5547, + "step": 61 + }, + { + "epoch": 0.08051948051948052, + "grad_norm": 423.76965199770956, + "learning_rate": 5.281385281385282e-06, + "loss": 1.6992, + "step": 62 + }, + { + "epoch": 0.08181818181818182, + "grad_norm": 367.18097956894354, + "learning_rate": 5.367965367965369e-06, + "loss": 1.6426, + "step": 63 + }, + { + "epoch": 0.08311688311688312, + "grad_norm": 224.05864878639184, + "learning_rate": 5.4545454545454545e-06, + "loss": 1.3452, + "step": 64 + }, + { + "epoch": 0.08441558441558442, + "grad_norm": 175.47813519288266, + "learning_rate": 5.541125541125542e-06, + "loss": 1.3047, + "step": 65 + }, + { + "epoch": 0.08571428571428572, + "grad_norm": 62.72826431747621, + "learning_rate": 5.627705627705629e-06, + "loss": 1.1787, + "step": 66 + }, + { + "epoch": 0.08701298701298701, + "grad_norm": 39.63558558237661, + "learning_rate": 5.7142857142857145e-06, + "loss": 1.2783, + "step": 67 + }, + { + "epoch": 0.08831168831168831, + "grad_norm": 233.51131021475425, + "learning_rate": 5.800865800865802e-06, + "loss": 1.3086, + "step": 68 + }, + { + "epoch": 0.08961038961038961, + "grad_norm": 231.42565507323454, + "learning_rate": 5.887445887445888e-06, + "loss": 1.3184, + "step": 69 + }, + { + "epoch": 0.09090909090909091, + "grad_norm": 268.1817212443567, + "learning_rate": 5.9740259740259746e-06, + "loss": 1.3096, + "step": 70 + }, + { + "epoch": 0.09220779220779221, + "grad_norm": 182.24921991617364, + "learning_rate": 6.060606060606061e-06, + "loss": 1.2871, + "step": 71 + }, + { + "epoch": 0.09350649350649351, + "grad_norm": 139.05631762790634, + "learning_rate": 6.147186147186147e-06, + "loss": 1.2725, + "step": 72 + }, + { + "epoch": 0.09480519480519481, + "grad_norm": 36.20941180937137, + "learning_rate": 6.233766233766234e-06, + "loss": 1.1914, + "step": 73 + }, + { + "epoch": 0.09610389610389611, + "grad_norm": 32.13572836955703, + "learning_rate": 6.320346320346321e-06, + "loss": 1.1455, + "step": 74 + }, + { + "epoch": 0.09740259740259741, + "grad_norm": 60.91882750446555, + "learning_rate": 6.406926406926407e-06, + "loss": 1.1777, + "step": 75 + }, + { + "epoch": 0.0987012987012987, + "grad_norm": 97.08382469021808, + "learning_rate": 6.493506493506494e-06, + "loss": 1.1221, + "step": 76 + }, + { + "epoch": 0.1, + "grad_norm": 200.40075736076088, + "learning_rate": 6.580086580086581e-06, + "loss": 1.1509, + "step": 77 + }, + { + "epoch": 0.1012987012987013, + "grad_norm": 161.70525082062508, + "learning_rate": 6.666666666666667e-06, + "loss": 1.2202, + "step": 78 + }, + { + "epoch": 0.1025974025974026, + "grad_norm": 97.71740279307001, + "learning_rate": 6.753246753246754e-06, + "loss": 1.1392, + "step": 79 + }, + { + "epoch": 0.1038961038961039, + "grad_norm": 79.49711423991204, + "learning_rate": 6.839826839826841e-06, + "loss": 1.1885, + "step": 80 + }, + { + "epoch": 0.10519480519480519, + "grad_norm": 47.85018851352415, + "learning_rate": 6.926406926406927e-06, + "loss": 1.0801, + "step": 81 + }, + { + "epoch": 0.10649350649350649, + "grad_norm": 20.188420940203518, + "learning_rate": 7.012987012987014e-06, + "loss": 1.1089, + "step": 82 + }, + { + "epoch": 0.10779220779220779, + "grad_norm": 152.56136365738027, + "learning_rate": 7.0995670995671e-06, + "loss": 1.1938, + "step": 83 + }, + { + "epoch": 0.10909090909090909, + "grad_norm": 156.02611927092914, + "learning_rate": 7.186147186147187e-06, + "loss": 1.1709, + "step": 84 + }, + { + "epoch": 0.11038961038961038, + "grad_norm": 112.31558573496567, + "learning_rate": 7.272727272727273e-06, + "loss": 1.0566, + "step": 85 + }, + { + "epoch": 0.11168831168831168, + "grad_norm": 91.25979609786367, + "learning_rate": 7.3593073593073596e-06, + "loss": 1.1104, + "step": 86 + }, + { + "epoch": 0.11298701298701298, + "grad_norm": 17.518569425031234, + "learning_rate": 7.445887445887446e-06, + "loss": 1.0957, + "step": 87 + }, + { + "epoch": 0.11428571428571428, + "grad_norm": 30.917195684582293, + "learning_rate": 7.532467532467533e-06, + "loss": 1.1133, + "step": 88 + }, + { + "epoch": 0.11558441558441558, + "grad_norm": 47.172698298778656, + "learning_rate": 7.61904761904762e-06, + "loss": 1.0078, + "step": 89 + }, + { + "epoch": 0.11688311688311688, + "grad_norm": 108.74612288637029, + "learning_rate": 7.705627705627707e-06, + "loss": 1.1006, + "step": 90 + }, + { + "epoch": 0.11818181818181818, + "grad_norm": 92.32599239121079, + "learning_rate": 7.792207792207793e-06, + "loss": 1.0947, + "step": 91 + }, + { + "epoch": 0.11948051948051948, + "grad_norm": 30.364245305910632, + "learning_rate": 7.87878787878788e-06, + "loss": 1.0171, + "step": 92 + }, + { + "epoch": 0.12077922077922078, + "grad_norm": 22.951503431876194, + "learning_rate": 7.965367965367966e-06, + "loss": 1.0669, + "step": 93 + }, + { + "epoch": 0.12207792207792208, + "grad_norm": 54.40182065372885, + "learning_rate": 8.051948051948052e-06, + "loss": 1.0425, + "step": 94 + }, + { + "epoch": 0.12337662337662338, + "grad_norm": 14.016114737404767, + "learning_rate": 8.138528138528139e-06, + "loss": 1.0322, + "step": 95 + }, + { + "epoch": 0.12467532467532468, + "grad_norm": 43.51259027462141, + "learning_rate": 8.225108225108225e-06, + "loss": 1.0215, + "step": 96 + }, + { + "epoch": 0.12597402597402596, + "grad_norm": 90.40588787774273, + "learning_rate": 8.311688311688313e-06, + "loss": 1.0771, + "step": 97 + }, + { + "epoch": 0.12727272727272726, + "grad_norm": 14.843618487209248, + "learning_rate": 8.398268398268398e-06, + "loss": 1.0933, + "step": 98 + }, + { + "epoch": 0.12857142857142856, + "grad_norm": 20.406200707521272, + "learning_rate": 8.484848484848486e-06, + "loss": 0.9883, + "step": 99 + }, + { + "epoch": 0.12987012987012986, + "grad_norm": 34.59497428751552, + "learning_rate": 8.571428571428571e-06, + "loss": 0.9443, + "step": 100 + }, + { + "epoch": 0.13116883116883116, + "grad_norm": 27.47616064645028, + "learning_rate": 8.658008658008659e-06, + "loss": 0.9824, + "step": 101 + }, + { + "epoch": 0.13246753246753246, + "grad_norm": 48.51131344524254, + "learning_rate": 8.744588744588745e-06, + "loss": 0.9692, + "step": 102 + }, + { + "epoch": 0.13376623376623376, + "grad_norm": 81.91179255177273, + "learning_rate": 8.831168831168832e-06, + "loss": 1.0137, + "step": 103 + }, + { + "epoch": 0.13506493506493505, + "grad_norm": 36.36379836233388, + "learning_rate": 8.917748917748918e-06, + "loss": 1.0601, + "step": 104 + }, + { + "epoch": 0.13636363636363635, + "grad_norm": 19.05641174813331, + "learning_rate": 9.004329004329005e-06, + "loss": 0.939, + "step": 105 + }, + { + "epoch": 0.13766233766233765, + "grad_norm": 18.950634335892424, + "learning_rate": 9.090909090909091e-06, + "loss": 0.9561, + "step": 106 + }, + { + "epoch": 0.13896103896103895, + "grad_norm": 88.27232308314414, + "learning_rate": 9.177489177489179e-06, + "loss": 1.0098, + "step": 107 + }, + { + "epoch": 0.14025974025974025, + "grad_norm": 16.284184807039168, + "learning_rate": 9.264069264069266e-06, + "loss": 0.9634, + "step": 108 + }, + { + "epoch": 0.14155844155844155, + "grad_norm": 45.81033066826727, + "learning_rate": 9.350649350649352e-06, + "loss": 0.9351, + "step": 109 + }, + { + "epoch": 0.14285714285714285, + "grad_norm": 48.51850567196589, + "learning_rate": 9.437229437229438e-06, + "loss": 0.9951, + "step": 110 + }, + { + "epoch": 0.14415584415584415, + "grad_norm": 30.562274899892678, + "learning_rate": 9.523809523809525e-06, + "loss": 0.9116, + "step": 111 + }, + { + "epoch": 0.14545454545454545, + "grad_norm": 41.42937702680897, + "learning_rate": 9.610389610389611e-06, + "loss": 0.9092, + "step": 112 + }, + { + "epoch": 0.14675324675324675, + "grad_norm": 65.9798965305908, + "learning_rate": 9.696969696969698e-06, + "loss": 0.9663, + "step": 113 + }, + { + "epoch": 0.14805194805194805, + "grad_norm": 39.61877272939499, + "learning_rate": 9.783549783549784e-06, + "loss": 0.8892, + "step": 114 + }, + { + "epoch": 0.14935064935064934, + "grad_norm": 42.565983780990265, + "learning_rate": 9.87012987012987e-06, + "loss": 0.8896, + "step": 115 + }, + { + "epoch": 0.15064935064935064, + "grad_norm": 19.10169366110021, + "learning_rate": 9.956709956709958e-06, + "loss": 0.8545, + "step": 116 + }, + { + "epoch": 0.15194805194805194, + "grad_norm": 13.919453915143253, + "learning_rate": 1.0043290043290043e-05, + "loss": 0.8242, + "step": 117 + }, + { + "epoch": 0.15324675324675324, + "grad_norm": 16.337246198969538, + "learning_rate": 1.012987012987013e-05, + "loss": 0.874, + "step": 118 + }, + { + "epoch": 0.15454545454545454, + "grad_norm": 31.527651634543712, + "learning_rate": 1.0216450216450218e-05, + "loss": 0.814, + "step": 119 + }, + { + "epoch": 0.15584415584415584, + "grad_norm": 27.4700003203198, + "learning_rate": 1.0303030303030304e-05, + "loss": 0.8667, + "step": 120 + }, + { + "epoch": 0.15714285714285714, + "grad_norm": 23.181247490013792, + "learning_rate": 1.0389610389610389e-05, + "loss": 0.8477, + "step": 121 + }, + { + "epoch": 0.15844155844155844, + "grad_norm": 14.37608152891486, + "learning_rate": 1.0476190476190477e-05, + "loss": 0.8677, + "step": 122 + }, + { + "epoch": 0.15974025974025974, + "grad_norm": 39.805812518253376, + "learning_rate": 1.0562770562770563e-05, + "loss": 0.8545, + "step": 123 + }, + { + "epoch": 0.16103896103896104, + "grad_norm": 57.30626395901855, + "learning_rate": 1.064935064935065e-05, + "loss": 0.8633, + "step": 124 + }, + { + "epoch": 0.16233766233766234, + "grad_norm": 20.693385330866473, + "learning_rate": 1.0735930735930738e-05, + "loss": 0.8008, + "step": 125 + }, + { + "epoch": 0.16363636363636364, + "grad_norm": 18.18314507704879, + "learning_rate": 1.0822510822510823e-05, + "loss": 0.7847, + "step": 126 + }, + { + "epoch": 0.16493506493506493, + "grad_norm": 85.41976167905393, + "learning_rate": 1.0909090909090909e-05, + "loss": 0.7749, + "step": 127 + }, + { + "epoch": 0.16623376623376623, + "grad_norm": 68.57282323348859, + "learning_rate": 1.0995670995670997e-05, + "loss": 0.7725, + "step": 128 + }, + { + "epoch": 0.16753246753246753, + "grad_norm": 43.872815604592944, + "learning_rate": 1.1082251082251083e-05, + "loss": 0.7241, + "step": 129 + }, + { + "epoch": 0.16883116883116883, + "grad_norm": 39.45059557380628, + "learning_rate": 1.116883116883117e-05, + "loss": 0.6816, + "step": 130 + }, + { + "epoch": 0.17012987012987013, + "grad_norm": 55.22573369871098, + "learning_rate": 1.1255411255411258e-05, + "loss": 0.7847, + "step": 131 + }, + { + "epoch": 0.17142857142857143, + "grad_norm": 68.88683721104424, + "learning_rate": 1.1341991341991343e-05, + "loss": 0.7153, + "step": 132 + }, + { + "epoch": 0.17272727272727273, + "grad_norm": 23.131668114474106, + "learning_rate": 1.1428571428571429e-05, + "loss": 0.7046, + "step": 133 + }, + { + "epoch": 0.17402597402597403, + "grad_norm": 57.46759833522016, + "learning_rate": 1.1515151515151517e-05, + "loss": 0.6587, + "step": 134 + }, + { + "epoch": 0.17532467532467533, + "grad_norm": 131.00292038268805, + "learning_rate": 1.1601731601731604e-05, + "loss": 0.6812, + "step": 135 + }, + { + "epoch": 0.17662337662337663, + "grad_norm": 29.382565241614262, + "learning_rate": 1.1688311688311688e-05, + "loss": 0.6267, + "step": 136 + }, + { + "epoch": 0.17792207792207793, + "grad_norm": 49.47773790801355, + "learning_rate": 1.1774891774891776e-05, + "loss": 0.6333, + "step": 137 + }, + { + "epoch": 0.17922077922077922, + "grad_norm": 126.29402109517045, + "learning_rate": 1.1861471861471863e-05, + "loss": 0.7354, + "step": 138 + }, + { + "epoch": 0.18051948051948052, + "grad_norm": 100.39414861510818, + "learning_rate": 1.1948051948051949e-05, + "loss": 0.6865, + "step": 139 + }, + { + "epoch": 0.18181818181818182, + "grad_norm": 19.289589967213548, + "learning_rate": 1.2034632034632037e-05, + "loss": 0.6401, + "step": 140 + }, + { + "epoch": 0.18311688311688312, + "grad_norm": 117.08028302333717, + "learning_rate": 1.2121212121212122e-05, + "loss": 0.626, + "step": 141 + }, + { + "epoch": 0.18441558441558442, + "grad_norm": 84.46993260438248, + "learning_rate": 1.2207792207792208e-05, + "loss": 0.6348, + "step": 142 + }, + { + "epoch": 0.18571428571428572, + "grad_norm": 16.96323101727435, + "learning_rate": 1.2294372294372295e-05, + "loss": 0.5728, + "step": 143 + }, + { + "epoch": 0.18701298701298702, + "grad_norm": 38.78475461253976, + "learning_rate": 1.2380952380952383e-05, + "loss": 0.7202, + "step": 144 + }, + { + "epoch": 0.18831168831168832, + "grad_norm": 68.33348939043026, + "learning_rate": 1.2467532467532468e-05, + "loss": 0.5469, + "step": 145 + }, + { + "epoch": 0.18961038961038962, + "grad_norm": 37.25871821648793, + "learning_rate": 1.2554112554112554e-05, + "loss": 0.5752, + "step": 146 + }, + { + "epoch": 0.19090909090909092, + "grad_norm": 33.12070075557407, + "learning_rate": 1.2640692640692642e-05, + "loss": 0.5535, + "step": 147 + }, + { + "epoch": 0.19220779220779222, + "grad_norm": 31.36574456068529, + "learning_rate": 1.2727272727272728e-05, + "loss": 0.4751, + "step": 148 + }, + { + "epoch": 0.19350649350649352, + "grad_norm": 60.14163096924735, + "learning_rate": 1.2813852813852813e-05, + "loss": 0.5176, + "step": 149 + }, + { + "epoch": 0.19480519480519481, + "grad_norm": 23.417243195729277, + "learning_rate": 1.2900432900432901e-05, + "loss": 0.573, + "step": 150 + }, + { + "epoch": 0.1961038961038961, + "grad_norm": 35.61627972747619, + "learning_rate": 1.2987012987012988e-05, + "loss": 0.5354, + "step": 151 + }, + { + "epoch": 0.1974025974025974, + "grad_norm": 63.22276778560939, + "learning_rate": 1.3073593073593074e-05, + "loss": 0.5435, + "step": 152 + }, + { + "epoch": 0.1987012987012987, + "grad_norm": 16.362250668568652, + "learning_rate": 1.3160173160173162e-05, + "loss": 0.45, + "step": 153 + }, + { + "epoch": 0.2, + "grad_norm": 33.720815446957424, + "learning_rate": 1.3246753246753249e-05, + "loss": 0.4871, + "step": 154 + }, + { + "epoch": 0.2012987012987013, + "grad_norm": 22.984946651143698, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.564, + "step": 155 + }, + { + "epoch": 0.2025974025974026, + "grad_norm": 34.01964296628851, + "learning_rate": 1.3419913419913421e-05, + "loss": 0.5217, + "step": 156 + }, + { + "epoch": 0.2038961038961039, + "grad_norm": 70.34549032470866, + "learning_rate": 1.3506493506493508e-05, + "loss": 0.425, + "step": 157 + }, + { + "epoch": 0.2051948051948052, + "grad_norm": 41.628088264757636, + "learning_rate": 1.3593073593073594e-05, + "loss": 0.4751, + "step": 158 + }, + { + "epoch": 0.2064935064935065, + "grad_norm": 71.58388401342985, + "learning_rate": 1.3679653679653682e-05, + "loss": 0.5376, + "step": 159 + }, + { + "epoch": 0.2077922077922078, + "grad_norm": 24.92145709044748, + "learning_rate": 1.3766233766233767e-05, + "loss": 0.446, + "step": 160 + }, + { + "epoch": 0.20909090909090908, + "grad_norm": 15.272685340465467, + "learning_rate": 1.3852813852813853e-05, + "loss": 0.429, + "step": 161 + }, + { + "epoch": 0.21038961038961038, + "grad_norm": 78.69358599292192, + "learning_rate": 1.3939393939393942e-05, + "loss": 0.4419, + "step": 162 + }, + { + "epoch": 0.21168831168831168, + "grad_norm": 76.04124708009267, + "learning_rate": 1.4025974025974028e-05, + "loss": 0.3596, + "step": 163 + }, + { + "epoch": 0.21298701298701297, + "grad_norm": 51.15044325639031, + "learning_rate": 1.4112554112554113e-05, + "loss": 0.4277, + "step": 164 + }, + { + "epoch": 0.21428571428571427, + "grad_norm": 134.38625071515878, + "learning_rate": 1.41991341991342e-05, + "loss": 0.4299, + "step": 165 + }, + { + "epoch": 0.21558441558441557, + "grad_norm": 24.0261223521635, + "learning_rate": 1.4285714285714287e-05, + "loss": 0.3784, + "step": 166 + }, + { + "epoch": 0.21688311688311687, + "grad_norm": 110.82644702237121, + "learning_rate": 1.4372294372294374e-05, + "loss": 0.3916, + "step": 167 + }, + { + "epoch": 0.21818181818181817, + "grad_norm": 55.96401705709766, + "learning_rate": 1.4458874458874458e-05, + "loss": 0.4531, + "step": 168 + }, + { + "epoch": 0.21948051948051947, + "grad_norm": 20.954147346275924, + "learning_rate": 1.4545454545454546e-05, + "loss": 0.4167, + "step": 169 + }, + { + "epoch": 0.22077922077922077, + "grad_norm": 102.41265934248479, + "learning_rate": 1.4632034632034633e-05, + "loss": 0.4473, + "step": 170 + }, + { + "epoch": 0.22207792207792207, + "grad_norm": 23.65526497103673, + "learning_rate": 1.4718614718614719e-05, + "loss": 0.4451, + "step": 171 + }, + { + "epoch": 0.22337662337662337, + "grad_norm": 98.76628856431871, + "learning_rate": 1.4805194805194807e-05, + "loss": 0.488, + "step": 172 + }, + { + "epoch": 0.22467532467532467, + "grad_norm": 32.633134769892166, + "learning_rate": 1.4891774891774892e-05, + "loss": 0.4395, + "step": 173 + }, + { + "epoch": 0.22597402597402597, + "grad_norm": 25.98719636462927, + "learning_rate": 1.4978354978354978e-05, + "loss": 0.4587, + "step": 174 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 73.24855578902226, + "learning_rate": 1.5064935064935066e-05, + "loss": 0.3511, + "step": 175 + }, + { + "epoch": 0.22857142857142856, + "grad_norm": 176.54573099066891, + "learning_rate": 1.5151515151515153e-05, + "loss": 0.6411, + "step": 176 + }, + { + "epoch": 0.22987012987012986, + "grad_norm": 143.67644154569467, + "learning_rate": 1.523809523809524e-05, + "loss": 0.4054, + "step": 177 + }, + { + "epoch": 0.23116883116883116, + "grad_norm": 134.5362757987968, + "learning_rate": 1.5324675324675326e-05, + "loss": 0.4009, + "step": 178 + }, + { + "epoch": 0.23246753246753246, + "grad_norm": 29.757340271995517, + "learning_rate": 1.5411255411255414e-05, + "loss": 0.3885, + "step": 179 + }, + { + "epoch": 0.23376623376623376, + "grad_norm": 114.37847595808283, + "learning_rate": 1.54978354978355e-05, + "loss": 0.3896, + "step": 180 + }, + { + "epoch": 0.23506493506493506, + "grad_norm": 148.95294665091507, + "learning_rate": 1.5584415584415587e-05, + "loss": 0.4186, + "step": 181 + }, + { + "epoch": 0.23636363636363636, + "grad_norm": 46.188183192005695, + "learning_rate": 1.567099567099567e-05, + "loss": 0.356, + "step": 182 + }, + { + "epoch": 0.23766233766233766, + "grad_norm": 133.27865876901436, + "learning_rate": 1.575757575757576e-05, + "loss": 0.4081, + "step": 183 + }, + { + "epoch": 0.23896103896103896, + "grad_norm": 39.789769289853666, + "learning_rate": 1.5844155844155847e-05, + "loss": 0.3496, + "step": 184 + }, + { + "epoch": 0.24025974025974026, + "grad_norm": 75.66170658676721, + "learning_rate": 1.5930735930735932e-05, + "loss": 0.4397, + "step": 185 + }, + { + "epoch": 0.24155844155844156, + "grad_norm": 73.32117171904788, + "learning_rate": 1.6017316017316017e-05, + "loss": 0.3625, + "step": 186 + }, + { + "epoch": 0.24285714285714285, + "grad_norm": 27.664112756001327, + "learning_rate": 1.6103896103896105e-05, + "loss": 0.334, + "step": 187 + }, + { + "epoch": 0.24415584415584415, + "grad_norm": 14.213747063587444, + "learning_rate": 1.6190476190476193e-05, + "loss": 0.4453, + "step": 188 + }, + { + "epoch": 0.24545454545454545, + "grad_norm": 87.4599330249861, + "learning_rate": 1.6277056277056278e-05, + "loss": 0.2906, + "step": 189 + }, + { + "epoch": 0.24675324675324675, + "grad_norm": 28.147266757577384, + "learning_rate": 1.6363636363636366e-05, + "loss": 0.3701, + "step": 190 + }, + { + "epoch": 0.24805194805194805, + "grad_norm": 68.6076640437454, + "learning_rate": 1.645021645021645e-05, + "loss": 0.3678, + "step": 191 + }, + { + "epoch": 0.24935064935064935, + "grad_norm": 92.2755677207207, + "learning_rate": 1.653679653679654e-05, + "loss": 0.417, + "step": 192 + }, + { + "epoch": 0.2506493506493506, + "grad_norm": 64.51269677633739, + "learning_rate": 1.6623376623376627e-05, + "loss": 0.3351, + "step": 193 + }, + { + "epoch": 0.2519480519480519, + "grad_norm": 34.48103525093349, + "learning_rate": 1.670995670995671e-05, + "loss": 0.4133, + "step": 194 + }, + { + "epoch": 0.2532467532467532, + "grad_norm": 36.154771366654764, + "learning_rate": 1.6796536796536796e-05, + "loss": 0.3417, + "step": 195 + }, + { + "epoch": 0.2545454545454545, + "grad_norm": 66.3422507624312, + "learning_rate": 1.6883116883116884e-05, + "loss": 0.3516, + "step": 196 + }, + { + "epoch": 0.2558441558441558, + "grad_norm": 12.472590287206508, + "learning_rate": 1.6969696969696972e-05, + "loss": 0.3297, + "step": 197 + }, + { + "epoch": 0.2571428571428571, + "grad_norm": 32.205191085990464, + "learning_rate": 1.7056277056277057e-05, + "loss": 0.3667, + "step": 198 + }, + { + "epoch": 0.2584415584415584, + "grad_norm": 43.082357749594756, + "learning_rate": 1.7142857142857142e-05, + "loss": 0.3284, + "step": 199 + }, + { + "epoch": 0.2597402597402597, + "grad_norm": 27.06458163718653, + "learning_rate": 1.722943722943723e-05, + "loss": 0.3384, + "step": 200 + }, + { + "epoch": 0.261038961038961, + "grad_norm": 96.4796045164573, + "learning_rate": 1.7316017316017318e-05, + "loss": 0.3085, + "step": 201 + }, + { + "epoch": 0.2623376623376623, + "grad_norm": 72.23575441975777, + "learning_rate": 1.7402597402597403e-05, + "loss": 0.3416, + "step": 202 + }, + { + "epoch": 0.2636363636363636, + "grad_norm": 62.2398460583449, + "learning_rate": 1.748917748917749e-05, + "loss": 0.3232, + "step": 203 + }, + { + "epoch": 0.2649350649350649, + "grad_norm": 42.02907854090804, + "learning_rate": 1.7575757575757576e-05, + "loss": 0.3936, + "step": 204 + }, + { + "epoch": 0.2662337662337662, + "grad_norm": 103.81884963639824, + "learning_rate": 1.7662337662337664e-05, + "loss": 0.3925, + "step": 205 + }, + { + "epoch": 0.2675324675324675, + "grad_norm": 118.15611869149033, + "learning_rate": 1.7748917748917752e-05, + "loss": 0.364, + "step": 206 + }, + { + "epoch": 0.2688311688311688, + "grad_norm": 94.11300442747371, + "learning_rate": 1.7835497835497836e-05, + "loss": 0.3947, + "step": 207 + }, + { + "epoch": 0.2701298701298701, + "grad_norm": 130.56943032700846, + "learning_rate": 1.792207792207792e-05, + "loss": 0.3309, + "step": 208 + }, + { + "epoch": 0.2714285714285714, + "grad_norm": 21.124525444550596, + "learning_rate": 1.800865800865801e-05, + "loss": 0.3223, + "step": 209 + }, + { + "epoch": 0.2727272727272727, + "grad_norm": 144.567588965618, + "learning_rate": 1.8095238095238097e-05, + "loss": 0.4329, + "step": 210 + }, + { + "epoch": 0.274025974025974, + "grad_norm": 71.98436737087185, + "learning_rate": 1.8181818181818182e-05, + "loss": 0.3882, + "step": 211 + }, + { + "epoch": 0.2753246753246753, + "grad_norm": 99.16571244628557, + "learning_rate": 1.826839826839827e-05, + "loss": 0.3727, + "step": 212 + }, + { + "epoch": 0.2766233766233766, + "grad_norm": 42.91854142760058, + "learning_rate": 1.8354978354978358e-05, + "loss": 0.323, + "step": 213 + }, + { + "epoch": 0.2779220779220779, + "grad_norm": 16.01506992686912, + "learning_rate": 1.8441558441558443e-05, + "loss": 0.3694, + "step": 214 + }, + { + "epoch": 0.2792207792207792, + "grad_norm": 96.06358809025875, + "learning_rate": 1.852813852813853e-05, + "loss": 0.3538, + "step": 215 + }, + { + "epoch": 0.2805194805194805, + "grad_norm": 21.733402946505567, + "learning_rate": 1.8614718614718616e-05, + "loss": 0.3512, + "step": 216 + }, + { + "epoch": 0.2818181818181818, + "grad_norm": 77.3957286200667, + "learning_rate": 1.8701298701298704e-05, + "loss": 0.4294, + "step": 217 + }, + { + "epoch": 0.2831168831168831, + "grad_norm": 11.780166360987728, + "learning_rate": 1.8787878787878792e-05, + "loss": 0.2932, + "step": 218 + }, + { + "epoch": 0.2844155844155844, + "grad_norm": 106.0516811813496, + "learning_rate": 1.8874458874458877e-05, + "loss": 0.2942, + "step": 219 + }, + { + "epoch": 0.2857142857142857, + "grad_norm": 14.71450344414706, + "learning_rate": 1.896103896103896e-05, + "loss": 0.3529, + "step": 220 + }, + { + "epoch": 0.287012987012987, + "grad_norm": 78.78108416469452, + "learning_rate": 1.904761904761905e-05, + "loss": 0.2903, + "step": 221 + }, + { + "epoch": 0.2883116883116883, + "grad_norm": 45.564011367188705, + "learning_rate": 1.9134199134199138e-05, + "loss": 0.3528, + "step": 222 + }, + { + "epoch": 0.2896103896103896, + "grad_norm": 48.67013146959329, + "learning_rate": 1.9220779220779222e-05, + "loss": 0.3085, + "step": 223 + }, + { + "epoch": 0.2909090909090909, + "grad_norm": 75.53564683851923, + "learning_rate": 1.9307359307359307e-05, + "loss": 0.2761, + "step": 224 + }, + { + "epoch": 0.2922077922077922, + "grad_norm": 72.66653252081262, + "learning_rate": 1.9393939393939395e-05, + "loss": 0.4083, + "step": 225 + }, + { + "epoch": 0.2935064935064935, + "grad_norm": 29.889796729047184, + "learning_rate": 1.9480519480519483e-05, + "loss": 0.3339, + "step": 226 + }, + { + "epoch": 0.2948051948051948, + "grad_norm": 65.39881382303452, + "learning_rate": 1.9567099567099568e-05, + "loss": 0.3245, + "step": 227 + }, + { + "epoch": 0.2961038961038961, + "grad_norm": 31.332252014470317, + "learning_rate": 1.9653679653679656e-05, + "loss": 0.3019, + "step": 228 + }, + { + "epoch": 0.2974025974025974, + "grad_norm": 65.01672578871172, + "learning_rate": 1.974025974025974e-05, + "loss": 0.3615, + "step": 229 + }, + { + "epoch": 0.2987012987012987, + "grad_norm": 17.414861846970332, + "learning_rate": 1.982683982683983e-05, + "loss": 0.2363, + "step": 230 + }, + { + "epoch": 0.3, + "grad_norm": 20.948116976728777, + "learning_rate": 1.9913419913419917e-05, + "loss": 0.3379, + "step": 231 + }, + { + "epoch": 0.3012987012987013, + "grad_norm": 104.78213337513078, + "learning_rate": 2e-05, + "loss": 0.3136, + "step": 232 + }, + { + "epoch": 0.3025974025974026, + "grad_norm": 48.23019529782554, + "learning_rate": 1.99999991154043e-05, + "loss": 0.3802, + "step": 233 + }, + { + "epoch": 0.3038961038961039, + "grad_norm": 29.465243339570065, + "learning_rate": 1.999999646161735e-05, + "loss": 0.2441, + "step": 234 + }, + { + "epoch": 0.3051948051948052, + "grad_norm": 12.313237830317007, + "learning_rate": 1.999999203863963e-05, + "loss": 0.2643, + "step": 235 + }, + { + "epoch": 0.3064935064935065, + "grad_norm": 60.79228224804833, + "learning_rate": 1.9999985846471914e-05, + "loss": 0.2983, + "step": 236 + }, + { + "epoch": 0.3077922077922078, + "grad_norm": 37.38334686568596, + "learning_rate": 1.99999778851153e-05, + "loss": 0.3529, + "step": 237 + }, + { + "epoch": 0.3090909090909091, + "grad_norm": 39.02054069609199, + "learning_rate": 1.9999968154571198e-05, + "loss": 0.2499, + "step": 238 + }, + { + "epoch": 0.3103896103896104, + "grad_norm": 37.98071302783246, + "learning_rate": 1.9999956654841325e-05, + "loss": 0.321, + "step": 239 + }, + { + "epoch": 0.3116883116883117, + "grad_norm": 69.26428495717747, + "learning_rate": 1.9999943385927722e-05, + "loss": 0.2827, + "step": 240 + }, + { + "epoch": 0.312987012987013, + "grad_norm": 87.53374640993022, + "learning_rate": 1.9999928347832733e-05, + "loss": 0.27, + "step": 241 + }, + { + "epoch": 0.3142857142857143, + "grad_norm": 79.76327570386641, + "learning_rate": 1.9999911540559013e-05, + "loss": 0.3606, + "step": 242 + }, + { + "epoch": 0.3155844155844156, + "grad_norm": 32.58071411196653, + "learning_rate": 1.9999892964109543e-05, + "loss": 0.2335, + "step": 243 + }, + { + "epoch": 0.3168831168831169, + "grad_norm": 15.369035890393954, + "learning_rate": 1.999987261848761e-05, + "loss": 0.3094, + "step": 244 + }, + { + "epoch": 0.3181818181818182, + "grad_norm": 31.147823333112676, + "learning_rate": 1.999985050369681e-05, + "loss": 0.2823, + "step": 245 + }, + { + "epoch": 0.3194805194805195, + "grad_norm": 17.782234623691206, + "learning_rate": 1.999982661974106e-05, + "loss": 0.1859, + "step": 246 + }, + { + "epoch": 0.3207792207792208, + "grad_norm": 55.67206066827065, + "learning_rate": 1.999980096662458e-05, + "loss": 0.2815, + "step": 247 + }, + { + "epoch": 0.3220779220779221, + "grad_norm": 16.973715316337124, + "learning_rate": 1.999977354435191e-05, + "loss": 0.3002, + "step": 248 + }, + { + "epoch": 0.3233766233766234, + "grad_norm": 59.857068747989956, + "learning_rate": 1.9999744352927903e-05, + "loss": 0.2988, + "step": 249 + }, + { + "epoch": 0.3246753246753247, + "grad_norm": 45.142722725758475, + "learning_rate": 1.999971339235772e-05, + "loss": 0.2841, + "step": 250 + }, + { + "epoch": 0.32597402597402597, + "grad_norm": 97.76972076046634, + "learning_rate": 1.9999680662646844e-05, + "loss": 0.3213, + "step": 251 + }, + { + "epoch": 0.32727272727272727, + "grad_norm": 36.5508236781459, + "learning_rate": 1.9999646163801064e-05, + "loss": 0.2782, + "step": 252 + }, + { + "epoch": 0.32857142857142857, + "grad_norm": 45.15868492068755, + "learning_rate": 1.9999609895826482e-05, + "loss": 0.2874, + "step": 253 + }, + { + "epoch": 0.32987012987012987, + "grad_norm": 117.16951855377603, + "learning_rate": 1.999957185872951e-05, + "loss": 0.3203, + "step": 254 + }, + { + "epoch": 0.33116883116883117, + "grad_norm": 69.87844951644402, + "learning_rate": 1.9999532052516888e-05, + "loss": 0.26, + "step": 255 + }, + { + "epoch": 0.33246753246753247, + "grad_norm": 99.05881090406892, + "learning_rate": 1.999949047719565e-05, + "loss": 0.2679, + "step": 256 + }, + { + "epoch": 0.33376623376623377, + "grad_norm": 58.86427207149895, + "learning_rate": 1.999944713277316e-05, + "loss": 0.3134, + "step": 257 + }, + { + "epoch": 0.33506493506493507, + "grad_norm": 111.9257526829768, + "learning_rate": 1.9999402019257076e-05, + "loss": 0.2913, + "step": 258 + }, + { + "epoch": 0.33636363636363636, + "grad_norm": 135.94910283255075, + "learning_rate": 1.999935513665538e-05, + "loss": 0.3204, + "step": 259 + }, + { + "epoch": 0.33766233766233766, + "grad_norm": 45.83643545938854, + "learning_rate": 1.9999306484976374e-05, + "loss": 0.2682, + "step": 260 + }, + { + "epoch": 0.33896103896103896, + "grad_norm": 198.6090495358003, + "learning_rate": 1.9999256064228665e-05, + "loss": 0.3408, + "step": 261 + }, + { + "epoch": 0.34025974025974026, + "grad_norm": 10.885985394690003, + "learning_rate": 1.999920387442117e-05, + "loss": 0.2547, + "step": 262 + }, + { + "epoch": 0.34155844155844156, + "grad_norm": 211.64062919785835, + "learning_rate": 1.9999149915563117e-05, + "loss": 0.3921, + "step": 263 + }, + { + "epoch": 0.34285714285714286, + "grad_norm": 57.80139217084719, + "learning_rate": 1.9999094187664066e-05, + "loss": 0.2865, + "step": 264 + }, + { + "epoch": 0.34415584415584416, + "grad_norm": 138.6901372389843, + "learning_rate": 1.9999036690733865e-05, + "loss": 0.2721, + "step": 265 + }, + { + "epoch": 0.34545454545454546, + "grad_norm": 131.1132238875295, + "learning_rate": 1.999897742478269e-05, + "loss": 0.3601, + "step": 266 + }, + { + "epoch": 0.34675324675324676, + "grad_norm": 40.36321345792584, + "learning_rate": 1.9998916389821025e-05, + "loss": 0.2998, + "step": 267 + }, + { + "epoch": 0.34805194805194806, + "grad_norm": 173.6274331057185, + "learning_rate": 1.999885358585967e-05, + "loss": 0.3088, + "step": 268 + }, + { + "epoch": 0.34935064935064936, + "grad_norm": 26.476696271622036, + "learning_rate": 1.9998789012909733e-05, + "loss": 0.2721, + "step": 269 + }, + { + "epoch": 0.35064935064935066, + "grad_norm": 105.94141728978543, + "learning_rate": 1.9998722670982644e-05, + "loss": 0.2544, + "step": 270 + }, + { + "epoch": 0.35194805194805195, + "grad_norm": 108.94310525529112, + "learning_rate": 1.9998654560090138e-05, + "loss": 0.3691, + "step": 271 + }, + { + "epoch": 0.35324675324675325, + "grad_norm": 83.53052183930616, + "learning_rate": 1.999858468024426e-05, + "loss": 0.2562, + "step": 272 + }, + { + "epoch": 0.35454545454545455, + "grad_norm": 75.37016053862132, + "learning_rate": 1.999851303145738e-05, + "loss": 0.2705, + "step": 273 + }, + { + "epoch": 0.35584415584415585, + "grad_norm": 30.86498721044062, + "learning_rate": 1.9998439613742172e-05, + "loss": 0.2797, + "step": 274 + }, + { + "epoch": 0.35714285714285715, + "grad_norm": 85.9172960294771, + "learning_rate": 1.9998364427111623e-05, + "loss": 0.2776, + "step": 275 + }, + { + "epoch": 0.35844155844155845, + "grad_norm": 37.746222023808684, + "learning_rate": 1.9998287471579036e-05, + "loss": 0.2339, + "step": 276 + }, + { + "epoch": 0.35974025974025975, + "grad_norm": 78.86236098947677, + "learning_rate": 1.9998208747158025e-05, + "loss": 0.2671, + "step": 277 + }, + { + "epoch": 0.36103896103896105, + "grad_norm": 8.99150816275129, + "learning_rate": 1.999812825386252e-05, + "loss": 0.2222, + "step": 278 + }, + { + "epoch": 0.36233766233766235, + "grad_norm": 84.15541112403344, + "learning_rate": 1.999804599170676e-05, + "loss": 0.2238, + "step": 279 + }, + { + "epoch": 0.36363636363636365, + "grad_norm": 52.67607377656539, + "learning_rate": 1.99979619607053e-05, + "loss": 0.2441, + "step": 280 + }, + { + "epoch": 0.36493506493506495, + "grad_norm": 44.55994088863868, + "learning_rate": 1.9997876160873005e-05, + "loss": 0.2095, + "step": 281 + }, + { + "epoch": 0.36623376623376624, + "grad_norm": 139.1941186193604, + "learning_rate": 1.9997788592225058e-05, + "loss": 0.3022, + "step": 282 + }, + { + "epoch": 0.36753246753246754, + "grad_norm": 42.24531055189884, + "learning_rate": 1.9997699254776947e-05, + "loss": 0.2231, + "step": 283 + }, + { + "epoch": 0.36883116883116884, + "grad_norm": 149.94579605524396, + "learning_rate": 1.999760814854448e-05, + "loss": 0.313, + "step": 284 + }, + { + "epoch": 0.37012987012987014, + "grad_norm": 16.447875349988436, + "learning_rate": 1.999751527354378e-05, + "loss": 0.2909, + "step": 285 + }, + { + "epoch": 0.37142857142857144, + "grad_norm": 139.2521172142892, + "learning_rate": 1.9997420629791267e-05, + "loss": 0.3182, + "step": 286 + }, + { + "epoch": 0.37272727272727274, + "grad_norm": 69.55456525641276, + "learning_rate": 1.99973242173037e-05, + "loss": 0.2501, + "step": 287 + }, + { + "epoch": 0.37402597402597404, + "grad_norm": 74.75221296936995, + "learning_rate": 1.999722603609812e-05, + "loss": 0.299, + "step": 288 + }, + { + "epoch": 0.37532467532467534, + "grad_norm": 84.160656332238, + "learning_rate": 1.9997126086191908e-05, + "loss": 0.2172, + "step": 289 + }, + { + "epoch": 0.37662337662337664, + "grad_norm": 42.17597101080939, + "learning_rate": 1.9997024367602745e-05, + "loss": 0.306, + "step": 290 + }, + { + "epoch": 0.37792207792207794, + "grad_norm": 127.97572056097177, + "learning_rate": 1.9996920880348623e-05, + "loss": 0.2745, + "step": 291 + }, + { + "epoch": 0.37922077922077924, + "grad_norm": 28.80867727672075, + "learning_rate": 1.999681562444786e-05, + "loss": 0.2366, + "step": 292 + }, + { + "epoch": 0.38051948051948054, + "grad_norm": 45.81134485253665, + "learning_rate": 1.9996708599919067e-05, + "loss": 0.3213, + "step": 293 + }, + { + "epoch": 0.38181818181818183, + "grad_norm": 44.88701010041816, + "learning_rate": 1.9996599806781184e-05, + "loss": 0.2479, + "step": 294 + }, + { + "epoch": 0.38311688311688313, + "grad_norm": 13.135180599692552, + "learning_rate": 1.999648924505346e-05, + "loss": 0.229, + "step": 295 + }, + { + "epoch": 0.38441558441558443, + "grad_norm": 42.15350946209576, + "learning_rate": 1.9996376914755452e-05, + "loss": 0.2576, + "step": 296 + }, + { + "epoch": 0.38571428571428573, + "grad_norm": 51.814938039155926, + "learning_rate": 1.999626281590704e-05, + "loss": 0.2479, + "step": 297 + }, + { + "epoch": 0.38701298701298703, + "grad_norm": 68.44865017531019, + "learning_rate": 1.99961469485284e-05, + "loss": 0.2227, + "step": 298 + }, + { + "epoch": 0.38831168831168833, + "grad_norm": 68.93246393928258, + "learning_rate": 1.9996029312640036e-05, + "loss": 0.2654, + "step": 299 + }, + { + "epoch": 0.38961038961038963, + "grad_norm": 48.32702361764178, + "learning_rate": 1.9995909908262764e-05, + "loss": 0.2314, + "step": 300 + }, + { + "epoch": 0.39090909090909093, + "grad_norm": 62.94793349952177, + "learning_rate": 1.9995788735417708e-05, + "loss": 0.2749, + "step": 301 + }, + { + "epoch": 0.3922077922077922, + "grad_norm": 12.841162553661828, + "learning_rate": 1.9995665794126297e-05, + "loss": 0.2416, + "step": 302 + }, + { + "epoch": 0.3935064935064935, + "grad_norm": 34.513218335411125, + "learning_rate": 1.999554108441029e-05, + "loss": 0.2577, + "step": 303 + }, + { + "epoch": 0.3948051948051948, + "grad_norm": 18.759377427282566, + "learning_rate": 1.999541460629175e-05, + "loss": 0.2494, + "step": 304 + }, + { + "epoch": 0.3961038961038961, + "grad_norm": 22.122390864942506, + "learning_rate": 1.999528635979305e-05, + "loss": 0.2076, + "step": 305 + }, + { + "epoch": 0.3974025974025974, + "grad_norm": 16.284357212909352, + "learning_rate": 1.9995156344936882e-05, + "loss": 0.21, + "step": 306 + }, + { + "epoch": 0.3987012987012987, + "grad_norm": 42.025939605822245, + "learning_rate": 1.9995024561746247e-05, + "loss": 0.2148, + "step": 307 + }, + { + "epoch": 0.4, + "grad_norm": 12.414577342786552, + "learning_rate": 1.9994891010244463e-05, + "loss": 0.2749, + "step": 308 + }, + { + "epoch": 0.4012987012987013, + "grad_norm": 46.07767592805509, + "learning_rate": 1.9994755690455154e-05, + "loss": 0.381, + "step": 309 + }, + { + "epoch": 0.4025974025974026, + "grad_norm": 55.36177026292617, + "learning_rate": 1.9994618602402255e-05, + "loss": 0.2316, + "step": 310 + }, + { + "epoch": 0.4038961038961039, + "grad_norm": 14.412235378476225, + "learning_rate": 1.9994479746110034e-05, + "loss": 0.2235, + "step": 311 + }, + { + "epoch": 0.4051948051948052, + "grad_norm": 82.11689948705244, + "learning_rate": 1.9994339121603044e-05, + "loss": 0.2728, + "step": 312 + }, + { + "epoch": 0.4064935064935065, + "grad_norm": 19.24795567711312, + "learning_rate": 1.9994196728906174e-05, + "loss": 0.2416, + "step": 313 + }, + { + "epoch": 0.4077922077922078, + "grad_norm": 57.34737786489921, + "learning_rate": 1.9994052568044612e-05, + "loss": 0.2256, + "step": 314 + }, + { + "epoch": 0.4090909090909091, + "grad_norm": 10.375395715860822, + "learning_rate": 1.999390663904386e-05, + "loss": 0.2474, + "step": 315 + }, + { + "epoch": 0.4103896103896104, + "grad_norm": 47.79221626148441, + "learning_rate": 1.9993758941929738e-05, + "loss": 0.2042, + "step": 316 + }, + { + "epoch": 0.4116883116883117, + "grad_norm": 16.481966639718856, + "learning_rate": 1.999360947672838e-05, + "loss": 0.1937, + "step": 317 + }, + { + "epoch": 0.412987012987013, + "grad_norm": 33.731589226099665, + "learning_rate": 1.999345824346622e-05, + "loss": 0.2695, + "step": 318 + }, + { + "epoch": 0.4142857142857143, + "grad_norm": 14.991503846882116, + "learning_rate": 1.9993305242170024e-05, + "loss": 0.2207, + "step": 319 + }, + { + "epoch": 0.4155844155844156, + "grad_norm": 36.62047097589661, + "learning_rate": 1.9993150472866856e-05, + "loss": 0.2142, + "step": 320 + }, + { + "epoch": 0.41688311688311686, + "grad_norm": 21.117519171885323, + "learning_rate": 1.99929939355841e-05, + "loss": 0.2122, + "step": 321 + }, + { + "epoch": 0.41818181818181815, + "grad_norm": 47.774816438482965, + "learning_rate": 1.9992835630349446e-05, + "loss": 0.213, + "step": 322 + }, + { + "epoch": 0.41948051948051945, + "grad_norm": 30.13086040675973, + "learning_rate": 1.9992675557190904e-05, + "loss": 0.2296, + "step": 323 + }, + { + "epoch": 0.42077922077922075, + "grad_norm": 18.81597909199686, + "learning_rate": 1.9992513716136796e-05, + "loss": 0.24, + "step": 324 + }, + { + "epoch": 0.42207792207792205, + "grad_norm": 31.83278406165478, + "learning_rate": 1.9992350107215755e-05, + "loss": 0.1943, + "step": 325 + }, + { + "epoch": 0.42337662337662335, + "grad_norm": 36.76673943555565, + "learning_rate": 1.9992184730456722e-05, + "loss": 0.2441, + "step": 326 + }, + { + "epoch": 0.42467532467532465, + "grad_norm": 27.67651224550881, + "learning_rate": 1.9992017585888957e-05, + "loss": 0.2314, + "step": 327 + }, + { + "epoch": 0.42597402597402595, + "grad_norm": 50.40301805787046, + "learning_rate": 1.999184867354203e-05, + "loss": 0.2265, + "step": 328 + }, + { + "epoch": 0.42727272727272725, + "grad_norm": 16.325510976977448, + "learning_rate": 1.9991677993445832e-05, + "loss": 0.2297, + "step": 329 + }, + { + "epoch": 0.42857142857142855, + "grad_norm": 11.542023301394718, + "learning_rate": 1.999150554563055e-05, + "loss": 0.2156, + "step": 330 + }, + { + "epoch": 0.42987012987012985, + "grad_norm": 19.068103058613488, + "learning_rate": 1.9991331330126698e-05, + "loss": 0.2117, + "step": 331 + }, + { + "epoch": 0.43116883116883115, + "grad_norm": 22.777050620150703, + "learning_rate": 1.9991155346965098e-05, + "loss": 0.2515, + "step": 332 + }, + { + "epoch": 0.43246753246753245, + "grad_norm": 40.05942532705973, + "learning_rate": 1.9990977596176885e-05, + "loss": 0.2561, + "step": 333 + }, + { + "epoch": 0.43376623376623374, + "grad_norm": 10.570224208519463, + "learning_rate": 1.9990798077793504e-05, + "loss": 0.1804, + "step": 334 + }, + { + "epoch": 0.43506493506493504, + "grad_norm": 21.589352826783298, + "learning_rate": 1.999061679184672e-05, + "loss": 0.1925, + "step": 335 + }, + { + "epoch": 0.43636363636363634, + "grad_norm": 36.38577013390246, + "learning_rate": 1.9990433738368603e-05, + "loss": 0.2426, + "step": 336 + }, + { + "epoch": 0.43766233766233764, + "grad_norm": 41.81081196165464, + "learning_rate": 1.9990248917391538e-05, + "loss": 0.204, + "step": 337 + }, + { + "epoch": 0.43896103896103894, + "grad_norm": 44.17414376815948, + "learning_rate": 1.999006232894822e-05, + "loss": 0.2164, + "step": 338 + }, + { + "epoch": 0.44025974025974024, + "grad_norm": 51.503777239137094, + "learning_rate": 1.998987397307167e-05, + "loss": 0.2402, + "step": 339 + }, + { + "epoch": 0.44155844155844154, + "grad_norm": 21.73421038672189, + "learning_rate": 1.9989683849795203e-05, + "loss": 0.1407, + "step": 340 + }, + { + "epoch": 0.44285714285714284, + "grad_norm": 9.21324225518647, + "learning_rate": 1.998949195915246e-05, + "loss": 0.2543, + "step": 341 + }, + { + "epoch": 0.44415584415584414, + "grad_norm": 70.85195802657755, + "learning_rate": 1.9989298301177387e-05, + "loss": 0.2119, + "step": 342 + }, + { + "epoch": 0.44545454545454544, + "grad_norm": 13.745130122190988, + "learning_rate": 1.998910287590425e-05, + "loss": 0.2297, + "step": 343 + }, + { + "epoch": 0.44675324675324674, + "grad_norm": 107.49679565699802, + "learning_rate": 1.9988905683367616e-05, + "loss": 0.2668, + "step": 344 + }, + { + "epoch": 0.44805194805194803, + "grad_norm": 23.78009955785302, + "learning_rate": 1.998870672360238e-05, + "loss": 0.2468, + "step": 345 + }, + { + "epoch": 0.44935064935064933, + "grad_norm": 103.31857344800754, + "learning_rate": 1.998850599664374e-05, + "loss": 0.2604, + "step": 346 + }, + { + "epoch": 0.45064935064935063, + "grad_norm": 12.539990018351418, + "learning_rate": 1.9988303502527203e-05, + "loss": 0.2229, + "step": 347 + }, + { + "epoch": 0.45194805194805193, + "grad_norm": 30.183064736516037, + "learning_rate": 1.99880992412886e-05, + "loss": 0.2164, + "step": 348 + }, + { + "epoch": 0.45324675324675323, + "grad_norm": 63.705573819843565, + "learning_rate": 1.998789321296407e-05, + "loss": 0.2397, + "step": 349 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 29.111881723301774, + "learning_rate": 1.9987685417590056e-05, + "loss": 0.1615, + "step": 350 + }, + { + "epoch": 0.45584415584415583, + "grad_norm": 69.24866519222537, + "learning_rate": 1.9987475855203325e-05, + "loss": 0.1987, + "step": 351 + }, + { + "epoch": 0.45714285714285713, + "grad_norm": 10.059401883425322, + "learning_rate": 1.9987264525840956e-05, + "loss": 0.2097, + "step": 352 + }, + { + "epoch": 0.4584415584415584, + "grad_norm": 78.75165643978794, + "learning_rate": 1.9987051429540335e-05, + "loss": 0.2346, + "step": 353 + }, + { + "epoch": 0.4597402597402597, + "grad_norm": 34.33106554040758, + "learning_rate": 1.998683656633916e-05, + "loss": 0.1857, + "step": 354 + }, + { + "epoch": 0.461038961038961, + "grad_norm": 62.879828752318566, + "learning_rate": 1.9986619936275445e-05, + "loss": 0.2144, + "step": 355 + }, + { + "epoch": 0.4623376623376623, + "grad_norm": 92.10628571198244, + "learning_rate": 1.998640153938752e-05, + "loss": 0.2672, + "step": 356 + }, + { + "epoch": 0.4636363636363636, + "grad_norm": 42.52711833098638, + "learning_rate": 1.998618137571402e-05, + "loss": 0.1788, + "step": 357 + }, + { + "epoch": 0.4649350649350649, + "grad_norm": 123.72610085884061, + "learning_rate": 1.9985959445293897e-05, + "loss": 0.2628, + "step": 358 + }, + { + "epoch": 0.4662337662337662, + "grad_norm": 34.7858439510532, + "learning_rate": 1.9985735748166414e-05, + "loss": 0.2012, + "step": 359 + }, + { + "epoch": 0.4675324675324675, + "grad_norm": 99.60842368281325, + "learning_rate": 1.9985510284371147e-05, + "loss": 0.2363, + "step": 360 + }, + { + "epoch": 0.4688311688311688, + "grad_norm": 14.0358684787841, + "learning_rate": 1.998528305394799e-05, + "loss": 0.207, + "step": 361 + }, + { + "epoch": 0.4701298701298701, + "grad_norm": 99.50613763321189, + "learning_rate": 1.9985054056937135e-05, + "loss": 0.238, + "step": 362 + }, + { + "epoch": 0.4714285714285714, + "grad_norm": 24.298727165971755, + "learning_rate": 1.9984823293379106e-05, + "loss": 0.2822, + "step": 363 + }, + { + "epoch": 0.4727272727272727, + "grad_norm": 101.13105122286787, + "learning_rate": 1.9984590763314722e-05, + "loss": 0.2638, + "step": 364 + }, + { + "epoch": 0.474025974025974, + "grad_norm": 5.772926654483309, + "learning_rate": 1.998435646678513e-05, + "loss": 0.2274, + "step": 365 + }, + { + "epoch": 0.4753246753246753, + "grad_norm": 73.82893489374572, + "learning_rate": 1.998412040383177e-05, + "loss": 0.2241, + "step": 366 + }, + { + "epoch": 0.4766233766233766, + "grad_norm": 27.179885199763863, + "learning_rate": 1.9983882574496415e-05, + "loss": 0.2133, + "step": 367 + }, + { + "epoch": 0.4779220779220779, + "grad_norm": 7.15199648677049, + "learning_rate": 1.9983642978821138e-05, + "loss": 0.202, + "step": 368 + }, + { + "epoch": 0.4792207792207792, + "grad_norm": 14.502701230404893, + "learning_rate": 1.998340161684833e-05, + "loss": 0.2576, + "step": 369 + }, + { + "epoch": 0.4805194805194805, + "grad_norm": 15.455489151350587, + "learning_rate": 1.9983158488620694e-05, + "loss": 0.2295, + "step": 370 + }, + { + "epoch": 0.4818181818181818, + "grad_norm": 10.599773960458354, + "learning_rate": 1.998291359418124e-05, + "loss": 0.1816, + "step": 371 + }, + { + "epoch": 0.4831168831168831, + "grad_norm": 15.488388305758818, + "learning_rate": 1.9982666933573295e-05, + "loss": 0.1885, + "step": 372 + }, + { + "epoch": 0.4844155844155844, + "grad_norm": 11.188404757685845, + "learning_rate": 1.99824185068405e-05, + "loss": 0.2092, + "step": 373 + }, + { + "epoch": 0.4857142857142857, + "grad_norm": 27.947982913865005, + "learning_rate": 1.9982168314026805e-05, + "loss": 0.2201, + "step": 374 + }, + { + "epoch": 0.487012987012987, + "grad_norm": 47.48773358641774, + "learning_rate": 1.9981916355176473e-05, + "loss": 0.1713, + "step": 375 + }, + { + "epoch": 0.4883116883116883, + "grad_norm": 7.894914121058864, + "learning_rate": 1.9981662630334084e-05, + "loss": 0.2014, + "step": 376 + }, + { + "epoch": 0.4896103896103896, + "grad_norm": 42.05095774346449, + "learning_rate": 1.9981407139544525e-05, + "loss": 0.1802, + "step": 377 + }, + { + "epoch": 0.4909090909090909, + "grad_norm": 11.254349709904607, + "learning_rate": 1.9981149882852996e-05, + "loss": 0.1945, + "step": 378 + }, + { + "epoch": 0.4922077922077922, + "grad_norm": 43.64676166772699, + "learning_rate": 1.9980890860305015e-05, + "loss": 0.184, + "step": 379 + }, + { + "epoch": 0.4935064935064935, + "grad_norm": 14.269254450046978, + "learning_rate": 1.9980630071946398e-05, + "loss": 0.1666, + "step": 380 + }, + { + "epoch": 0.4948051948051948, + "grad_norm": 26.113610384434143, + "learning_rate": 1.9980367517823295e-05, + "loss": 0.1648, + "step": 381 + }, + { + "epoch": 0.4961038961038961, + "grad_norm": 12.719152283956205, + "learning_rate": 1.998010319798215e-05, + "loss": 0.2101, + "step": 382 + }, + { + "epoch": 0.4974025974025974, + "grad_norm": 19.78537616456142, + "learning_rate": 1.9979837112469725e-05, + "loss": 0.2345, + "step": 383 + }, + { + "epoch": 0.4987012987012987, + "grad_norm": 15.769193886204457, + "learning_rate": 1.9979569261333106e-05, + "loss": 0.2817, + "step": 384 + }, + { + "epoch": 0.5, + "grad_norm": 34.29706290794183, + "learning_rate": 1.9979299644619667e-05, + "loss": 0.2291, + "step": 385 + }, + { + "epoch": 0.5012987012987012, + "grad_norm": 56.03651330668273, + "learning_rate": 1.997902826237712e-05, + "loss": 0.1566, + "step": 386 + }, + { + "epoch": 0.5025974025974026, + "grad_norm": 21.876156969775828, + "learning_rate": 1.9978755114653468e-05, + "loss": 0.222, + "step": 387 + }, + { + "epoch": 0.5038961038961038, + "grad_norm": 48.68658039718079, + "learning_rate": 1.9978480201497043e-05, + "loss": 0.1696, + "step": 388 + }, + { + "epoch": 0.5051948051948052, + "grad_norm": 8.282599567222974, + "learning_rate": 1.997820352295648e-05, + "loss": 0.1925, + "step": 389 + }, + { + "epoch": 0.5064935064935064, + "grad_norm": 38.56799582572708, + "learning_rate": 1.997792507908073e-05, + "loss": 0.1751, + "step": 390 + }, + { + "epoch": 0.5077922077922078, + "grad_norm": 34.93944162997997, + "learning_rate": 1.9977644869919053e-05, + "loss": 0.1918, + "step": 391 + }, + { + "epoch": 0.509090909090909, + "grad_norm": 63.38658657847159, + "learning_rate": 1.9977362895521025e-05, + "loss": 0.1977, + "step": 392 + }, + { + "epoch": 0.5103896103896104, + "grad_norm": 22.57610044784801, + "learning_rate": 1.997707915593653e-05, + "loss": 0.1863, + "step": 393 + }, + { + "epoch": 0.5116883116883116, + "grad_norm": 18.929022359591233, + "learning_rate": 1.9976793651215775e-05, + "loss": 0.1782, + "step": 394 + }, + { + "epoch": 0.512987012987013, + "grad_norm": 73.8239058115728, + "learning_rate": 1.9976506381409257e-05, + "loss": 0.2279, + "step": 395 + }, + { + "epoch": 0.5142857142857142, + "grad_norm": 20.92491173177454, + "learning_rate": 1.9976217346567813e-05, + "loss": 0.1563, + "step": 396 + }, + { + "epoch": 0.5155844155844156, + "grad_norm": 89.51351901583811, + "learning_rate": 1.9975926546742573e-05, + "loss": 0.1825, + "step": 397 + }, + { + "epoch": 0.5168831168831168, + "grad_norm": 16.548382078098587, + "learning_rate": 1.9975633981984983e-05, + "loss": 0.1739, + "step": 398 + }, + { + "epoch": 0.5181818181818182, + "grad_norm": 91.40219067426291, + "learning_rate": 1.9975339652346807e-05, + "loss": 0.1738, + "step": 399 + }, + { + "epoch": 0.5194805194805194, + "grad_norm": 5.604495787008308, + "learning_rate": 1.997504355788012e-05, + "loss": 0.1614, + "step": 400 + }, + { + "epoch": 0.5207792207792208, + "grad_norm": 125.19899834264794, + "learning_rate": 1.9974745698637298e-05, + "loss": 0.2463, + "step": 401 + }, + { + "epoch": 0.522077922077922, + "grad_norm": 13.072225046359895, + "learning_rate": 1.9974446074671046e-05, + "loss": 0.1771, + "step": 402 + }, + { + "epoch": 0.5233766233766234, + "grad_norm": 123.69617115955462, + "learning_rate": 1.9974144686034372e-05, + "loss": 0.2845, + "step": 403 + }, + { + "epoch": 0.5246753246753246, + "grad_norm": 48.43374994953599, + "learning_rate": 1.997384153278059e-05, + "loss": 0.2789, + "step": 404 + }, + { + "epoch": 0.525974025974026, + "grad_norm": 152.3461835730947, + "learning_rate": 1.9973536614963344e-05, + "loss": 0.2904, + "step": 405 + }, + { + "epoch": 0.5272727272727272, + "grad_norm": 46.441632007431544, + "learning_rate": 1.9973229932636577e-05, + "loss": 0.1972, + "step": 406 + }, + { + "epoch": 0.5285714285714286, + "grad_norm": 77.12472361195998, + "learning_rate": 1.997292148585454e-05, + "loss": 0.2524, + "step": 407 + }, + { + "epoch": 0.5298701298701298, + "grad_norm": 114.15351189517159, + "learning_rate": 1.9972611274671815e-05, + "loss": 0.1897, + "step": 408 + }, + { + "epoch": 0.5311688311688312, + "grad_norm": 43.6781610487255, + "learning_rate": 1.997229929914327e-05, + "loss": 0.2461, + "step": 409 + }, + { + "epoch": 0.5324675324675324, + "grad_norm": 103.90438609981665, + "learning_rate": 1.9971985559324116e-05, + "loss": 0.1875, + "step": 410 + }, + { + "epoch": 0.5337662337662338, + "grad_norm": 26.68374400259148, + "learning_rate": 1.9971670055269844e-05, + "loss": 0.1887, + "step": 411 + }, + { + "epoch": 0.535064935064935, + "grad_norm": 89.56774068501532, + "learning_rate": 1.997135278703628e-05, + "loss": 0.2402, + "step": 412 + }, + { + "epoch": 0.5363636363636364, + "grad_norm": 53.08284239481677, + "learning_rate": 1.9971033754679555e-05, + "loss": 0.2265, + "step": 413 + }, + { + "epoch": 0.5376623376623376, + "grad_norm": 49.33029326552398, + "learning_rate": 1.9970712958256113e-05, + "loss": 0.2023, + "step": 414 + }, + { + "epoch": 0.538961038961039, + "grad_norm": 75.07270755233593, + "learning_rate": 1.9970390397822705e-05, + "loss": 0.2588, + "step": 415 + }, + { + "epoch": 0.5402597402597402, + "grad_norm": 19.72115294604134, + "learning_rate": 1.9970066073436402e-05, + "loss": 0.1581, + "step": 416 + }, + { + "epoch": 0.5415584415584416, + "grad_norm": 44.51044962350022, + "learning_rate": 1.996973998515458e-05, + "loss": 0.2433, + "step": 417 + }, + { + "epoch": 0.5428571428571428, + "grad_norm": 72.17519943472482, + "learning_rate": 1.9969412133034932e-05, + "loss": 0.1875, + "step": 418 + }, + { + "epoch": 0.5441558441558442, + "grad_norm": 37.05509827451577, + "learning_rate": 1.9969082517135463e-05, + "loss": 0.2113, + "step": 419 + }, + { + "epoch": 0.5454545454545454, + "grad_norm": 107.12910337590046, + "learning_rate": 1.9968751137514485e-05, + "loss": 0.2247, + "step": 420 + }, + { + "epoch": 0.5467532467532468, + "grad_norm": 15.997046983620065, + "learning_rate": 1.9968417994230628e-05, + "loss": 0.2083, + "step": 421 + }, + { + "epoch": 0.548051948051948, + "grad_norm": 95.99023161843348, + "learning_rate": 1.996808308734283e-05, + "loss": 0.1868, + "step": 422 + }, + { + "epoch": 0.5493506493506494, + "grad_norm": 86.80446963051531, + "learning_rate": 1.996774641691034e-05, + "loss": 0.2469, + "step": 423 + }, + { + "epoch": 0.5506493506493506, + "grad_norm": 80.83292917901012, + "learning_rate": 1.9967407982992727e-05, + "loss": 0.1855, + "step": 424 + }, + { + "epoch": 0.551948051948052, + "grad_norm": 89.0899190734713, + "learning_rate": 1.996706778564986e-05, + "loss": 0.2543, + "step": 425 + }, + { + "epoch": 0.5532467532467532, + "grad_norm": 18.663521249046187, + "learning_rate": 1.9966725824941933e-05, + "loss": 0.1786, + "step": 426 + }, + { + "epoch": 0.5545454545454546, + "grad_norm": 84.28048901469901, + "learning_rate": 1.9966382100929444e-05, + "loss": 0.2098, + "step": 427 + }, + { + "epoch": 0.5558441558441558, + "grad_norm": 11.14102223137212, + "learning_rate": 1.9966036613673202e-05, + "loss": 0.1907, + "step": 428 + }, + { + "epoch": 0.5571428571428572, + "grad_norm": 51.61399436779026, + "learning_rate": 1.9965689363234325e-05, + "loss": 0.1811, + "step": 429 + }, + { + "epoch": 0.5584415584415584, + "grad_norm": 32.206760015932666, + "learning_rate": 1.9965340349674262e-05, + "loss": 0.1739, + "step": 430 + }, + { + "epoch": 0.5597402597402598, + "grad_norm": 29.18420941144133, + "learning_rate": 1.996498957305475e-05, + "loss": 0.1817, + "step": 431 + }, + { + "epoch": 0.561038961038961, + "grad_norm": 52.99853013557087, + "learning_rate": 1.9964637033437852e-05, + "loss": 0.168, + "step": 432 + }, + { + "epoch": 0.5623376623376624, + "grad_norm": 9.704284490978337, + "learning_rate": 1.9964282730885935e-05, + "loss": 0.1926, + "step": 433 + }, + { + "epoch": 0.5636363636363636, + "grad_norm": 75.25545546415982, + "learning_rate": 1.9963926665461682e-05, + "loss": 0.2117, + "step": 434 + }, + { + "epoch": 0.564935064935065, + "grad_norm": 40.63381822704264, + "learning_rate": 1.9963568837228096e-05, + "loss": 0.1661, + "step": 435 + }, + { + "epoch": 0.5662337662337662, + "grad_norm": 78.20576197693924, + "learning_rate": 1.9963209246248473e-05, + "loss": 0.1833, + "step": 436 + }, + { + "epoch": 0.5675324675324676, + "grad_norm": 38.20653410589377, + "learning_rate": 1.996284789258644e-05, + "loss": 0.2299, + "step": 437 + }, + { + "epoch": 0.5688311688311688, + "grad_norm": 12.384745946961369, + "learning_rate": 1.9962484776305927e-05, + "loss": 0.1971, + "step": 438 + }, + { + "epoch": 0.5701298701298702, + "grad_norm": 65.46554803043327, + "learning_rate": 1.9962119897471166e-05, + "loss": 0.187, + "step": 439 + }, + { + "epoch": 0.5714285714285714, + "grad_norm": 53.51225680349537, + "learning_rate": 1.9961753256146725e-05, + "loss": 0.19, + "step": 440 + }, + { + "epoch": 0.5727272727272728, + "grad_norm": 59.96699784362543, + "learning_rate": 1.9961384852397456e-05, + "loss": 0.2152, + "step": 441 + }, + { + "epoch": 0.574025974025974, + "grad_norm": 73.6635257387468, + "learning_rate": 1.9961014686288546e-05, + "loss": 0.1743, + "step": 442 + }, + { + "epoch": 0.5753246753246753, + "grad_norm": 18.732141862234705, + "learning_rate": 1.9960642757885482e-05, + "loss": 0.194, + "step": 443 + }, + { + "epoch": 0.5766233766233766, + "grad_norm": 117.36601381454275, + "learning_rate": 1.9960269067254068e-05, + "loss": 0.2004, + "step": 444 + }, + { + "epoch": 0.577922077922078, + "grad_norm": 20.002361554573596, + "learning_rate": 1.9959893614460413e-05, + "loss": 0.157, + "step": 445 + }, + { + "epoch": 0.5792207792207792, + "grad_norm": 77.89193574273702, + "learning_rate": 1.9959516399570945e-05, + "loss": 0.1696, + "step": 446 + }, + { + "epoch": 0.5805194805194805, + "grad_norm": 17.304221975335977, + "learning_rate": 1.9959137422652396e-05, + "loss": 0.1945, + "step": 447 + }, + { + "epoch": 0.5818181818181818, + "grad_norm": 50.07411457160169, + "learning_rate": 1.9958756683771815e-05, + "loss": 0.1646, + "step": 448 + }, + { + "epoch": 0.5831168831168831, + "grad_norm": 15.723993579225022, + "learning_rate": 1.9958374182996566e-05, + "loss": 0.1986, + "step": 449 + }, + { + "epoch": 0.5844155844155844, + "grad_norm": 10.3825976680003, + "learning_rate": 1.995798992039432e-05, + "loss": 0.1811, + "step": 450 + }, + { + "epoch": 0.5857142857142857, + "grad_norm": 45.709831567494106, + "learning_rate": 1.995760389603306e-05, + "loss": 0.1593, + "step": 451 + }, + { + "epoch": 0.587012987012987, + "grad_norm": 13.47744956437052, + "learning_rate": 1.995721610998108e-05, + "loss": 0.2144, + "step": 452 + }, + { + "epoch": 0.5883116883116883, + "grad_norm": 58.66145198728581, + "learning_rate": 1.9956826562306982e-05, + "loss": 0.2766, + "step": 453 + }, + { + "epoch": 0.5896103896103896, + "grad_norm": 16.736373560472014, + "learning_rate": 1.9956435253079696e-05, + "loss": 0.2084, + "step": 454 + }, + { + "epoch": 0.5909090909090909, + "grad_norm": 37.16190980726263, + "learning_rate": 1.995604218236844e-05, + "loss": 0.167, + "step": 455 + }, + { + "epoch": 0.5922077922077922, + "grad_norm": 16.10434020072141, + "learning_rate": 1.9955647350242765e-05, + "loss": 0.1788, + "step": 456 + }, + { + "epoch": 0.5935064935064935, + "grad_norm": 6.641735835261696, + "learning_rate": 1.995525075677252e-05, + "loss": 0.1738, + "step": 457 + }, + { + "epoch": 0.5948051948051948, + "grad_norm": 9.604073086331422, + "learning_rate": 1.995485240202787e-05, + "loss": 0.1597, + "step": 458 + }, + { + "epoch": 0.5961038961038961, + "grad_norm": 11.534719023036493, + "learning_rate": 1.995445228607929e-05, + "loss": 0.1631, + "step": 459 + }, + { + "epoch": 0.5974025974025974, + "grad_norm": 13.909158883230313, + "learning_rate": 1.9954050408997576e-05, + "loss": 0.1774, + "step": 460 + }, + { + "epoch": 0.5987012987012987, + "grad_norm": 13.344832274406173, + "learning_rate": 1.9953646770853817e-05, + "loss": 0.1886, + "step": 461 + }, + { + "epoch": 0.6, + "grad_norm": 11.974221029303115, + "learning_rate": 1.9953241371719433e-05, + "loss": 0.1892, + "step": 462 + }, + { + "epoch": 0.6012987012987013, + "grad_norm": 11.097881859973011, + "learning_rate": 1.995283421166614e-05, + "loss": 0.1871, + "step": 463 + }, + { + "epoch": 0.6025974025974026, + "grad_norm": 8.874981820097384, + "learning_rate": 1.995242529076598e-05, + "loss": 0.1526, + "step": 464 + }, + { + "epoch": 0.6038961038961039, + "grad_norm": 10.71933853139675, + "learning_rate": 1.995201460909129e-05, + "loss": 0.1542, + "step": 465 + }, + { + "epoch": 0.6051948051948052, + "grad_norm": 34.038517627188085, + "learning_rate": 1.9951602166714735e-05, + "loss": 0.1635, + "step": 466 + }, + { + "epoch": 0.6064935064935065, + "grad_norm": 26.853111838138847, + "learning_rate": 1.9951187963709277e-05, + "loss": 0.1677, + "step": 467 + }, + { + "epoch": 0.6077922077922078, + "grad_norm": 90.19310247096878, + "learning_rate": 1.9950772000148206e-05, + "loss": 0.1915, + "step": 468 + }, + { + "epoch": 0.6090909090909091, + "grad_norm": 19.003678909222888, + "learning_rate": 1.9950354276105103e-05, + "loss": 0.1537, + "step": 469 + }, + { + "epoch": 0.6103896103896104, + "grad_norm": 54.37516029684389, + "learning_rate": 1.9949934791653882e-05, + "loss": 0.183, + "step": 470 + }, + { + "epoch": 0.6116883116883117, + "grad_norm": 41.11543100663795, + "learning_rate": 1.994951354686875e-05, + "loss": 0.1298, + "step": 471 + }, + { + "epoch": 0.612987012987013, + "grad_norm": 48.328833712630704, + "learning_rate": 1.9949090541824233e-05, + "loss": 0.1559, + "step": 472 + }, + { + "epoch": 0.6142857142857143, + "grad_norm": 79.79632307282004, + "learning_rate": 1.9948665776595177e-05, + "loss": 0.1913, + "step": 473 + }, + { + "epoch": 0.6155844155844156, + "grad_norm": 30.39692507563288, + "learning_rate": 1.994823925125672e-05, + "loss": 0.1931, + "step": 474 + }, + { + "epoch": 0.6168831168831169, + "grad_norm": 97.89942131635709, + "learning_rate": 1.9947810965884334e-05, + "loss": 0.2196, + "step": 475 + }, + { + "epoch": 0.6181818181818182, + "grad_norm": 8.885811822863387, + "learning_rate": 1.994738092055378e-05, + "loss": 0.1688, + "step": 476 + }, + { + "epoch": 0.6194805194805195, + "grad_norm": 109.70302770053642, + "learning_rate": 1.994694911534115e-05, + "loss": 0.1907, + "step": 477 + }, + { + "epoch": 0.6207792207792208, + "grad_norm": 49.656687922365776, + "learning_rate": 1.9946515550322832e-05, + "loss": 0.1816, + "step": 478 + }, + { + "epoch": 0.6220779220779221, + "grad_norm": 138.29056828762432, + "learning_rate": 1.9946080225575537e-05, + "loss": 0.2087, + "step": 479 + }, + { + "epoch": 0.6233766233766234, + "grad_norm": 62.53640768689606, + "learning_rate": 1.994564314117628e-05, + "loss": 0.2147, + "step": 480 + }, + { + "epoch": 0.6246753246753247, + "grad_norm": 80.98691672081918, + "learning_rate": 1.9945204297202388e-05, + "loss": 0.2023, + "step": 481 + }, + { + "epoch": 0.625974025974026, + "grad_norm": 87.7121734590738, + "learning_rate": 1.9944763693731505e-05, + "loss": 0.1813, + "step": 482 + }, + { + "epoch": 0.6272727272727273, + "grad_norm": 17.32311154204881, + "learning_rate": 1.9944321330841578e-05, + "loss": 0.1684, + "step": 483 + }, + { + "epoch": 0.6285714285714286, + "grad_norm": 97.55256950370938, + "learning_rate": 1.994387720861087e-05, + "loss": 0.2527, + "step": 484 + }, + { + "epoch": 0.6298701298701299, + "grad_norm": 39.99695109963275, + "learning_rate": 1.9943431327117955e-05, + "loss": 0.1575, + "step": 485 + }, + { + "epoch": 0.6311688311688312, + "grad_norm": 93.36269692394168, + "learning_rate": 1.9942983686441723e-05, + "loss": 0.1854, + "step": 486 + }, + { + "epoch": 0.6324675324675325, + "grad_norm": 66.46751625980806, + "learning_rate": 1.9942534286661364e-05, + "loss": 0.1537, + "step": 487 + }, + { + "epoch": 0.6337662337662338, + "grad_norm": 39.517770508426416, + "learning_rate": 1.9942083127856388e-05, + "loss": 0.1484, + "step": 488 + }, + { + "epoch": 0.6350649350649351, + "grad_norm": 80.29251655504584, + "learning_rate": 1.9941630210106615e-05, + "loss": 0.2429, + "step": 489 + }, + { + "epoch": 0.6363636363636364, + "grad_norm": 33.301222253592506, + "learning_rate": 1.994117553349217e-05, + "loss": 0.1424, + "step": 490 + }, + { + "epoch": 0.6376623376623377, + "grad_norm": 91.98175527401935, + "learning_rate": 1.99407190980935e-05, + "loss": 0.2166, + "step": 491 + }, + { + "epoch": 0.638961038961039, + "grad_norm": 69.03274575326861, + "learning_rate": 1.9940260903991353e-05, + "loss": 0.2418, + "step": 492 + }, + { + "epoch": 0.6402597402597403, + "grad_norm": 84.87218562596402, + "learning_rate": 1.993980095126679e-05, + "loss": 0.1664, + "step": 493 + }, + { + "epoch": 0.6415584415584416, + "grad_norm": 65.12618329485339, + "learning_rate": 1.9939339240001196e-05, + "loss": 0.1571, + "step": 494 + }, + { + "epoch": 0.6428571428571429, + "grad_norm": 18.360760723322016, + "learning_rate": 1.9938875770276243e-05, + "loss": 0.1577, + "step": 495 + }, + { + "epoch": 0.6441558441558441, + "grad_norm": 91.76032118212522, + "learning_rate": 1.9938410542173937e-05, + "loss": 0.1943, + "step": 496 + }, + { + "epoch": 0.6454545454545455, + "grad_norm": 8.147951999055795, + "learning_rate": 1.993794355577658e-05, + "loss": 0.1891, + "step": 497 + }, + { + "epoch": 0.6467532467532467, + "grad_norm": 40.0602047748309, + "learning_rate": 1.9937474811166796e-05, + "loss": 0.1639, + "step": 498 + }, + { + "epoch": 0.6480519480519481, + "grad_norm": 57.838044346686395, + "learning_rate": 1.9937004308427515e-05, + "loss": 0.1572, + "step": 499 + }, + { + "epoch": 0.6493506493506493, + "grad_norm": 34.818495491091824, + "learning_rate": 1.993653204764197e-05, + "loss": 0.1948, + "step": 500 + }, + { + "epoch": 0.6506493506493507, + "grad_norm": 43.19723456498562, + "learning_rate": 1.9936058028893722e-05, + "loss": 0.1566, + "step": 501 + }, + { + "epoch": 0.6519480519480519, + "grad_norm": 21.620726102769005, + "learning_rate": 1.993558225226663e-05, + "loss": 0.1663, + "step": 502 + }, + { + "epoch": 0.6532467532467533, + "grad_norm": 18.930599709952975, + "learning_rate": 1.9935104717844867e-05, + "loss": 0.2852, + "step": 503 + }, + { + "epoch": 0.6545454545454545, + "grad_norm": 111.63674149025304, + "learning_rate": 1.9934625425712924e-05, + "loss": 0.1994, + "step": 504 + }, + { + "epoch": 0.6558441558441559, + "grad_norm": 14.189915483568207, + "learning_rate": 1.9934144375955587e-05, + "loss": 0.1494, + "step": 505 + }, + { + "epoch": 0.6571428571428571, + "grad_norm": 89.71603391842015, + "learning_rate": 1.9933661568657973e-05, + "loss": 0.1801, + "step": 506 + }, + { + "epoch": 0.6584415584415585, + "grad_norm": 26.040836828809613, + "learning_rate": 1.9933177003905494e-05, + "loss": 0.1857, + "step": 507 + }, + { + "epoch": 0.6597402597402597, + "grad_norm": 42.29571086128108, + "learning_rate": 1.9932690681783876e-05, + "loss": 0.1717, + "step": 508 + }, + { + "epoch": 0.6610389610389611, + "grad_norm": 51.72771505553675, + "learning_rate": 1.9932202602379167e-05, + "loss": 0.1843, + "step": 509 + }, + { + "epoch": 0.6623376623376623, + "grad_norm": 8.434690297890086, + "learning_rate": 1.9931712765777714e-05, + "loss": 0.1841, + "step": 510 + }, + { + "epoch": 0.6636363636363637, + "grad_norm": 33.59082603333855, + "learning_rate": 1.9931221172066175e-05, + "loss": 0.164, + "step": 511 + }, + { + "epoch": 0.6649350649350649, + "grad_norm": 51.66655529077922, + "learning_rate": 1.9930727821331524e-05, + "loss": 0.1742, + "step": 512 + }, + { + "epoch": 0.6662337662337663, + "grad_norm": 39.558636599914635, + "learning_rate": 1.993023271366105e-05, + "loss": 0.1407, + "step": 513 + }, + { + "epoch": 0.6675324675324675, + "grad_norm": 34.945195636054684, + "learning_rate": 1.9929735849142336e-05, + "loss": 0.1272, + "step": 514 + }, + { + "epoch": 0.6688311688311688, + "grad_norm": 14.304537093138835, + "learning_rate": 1.9929237227863298e-05, + "loss": 0.145, + "step": 515 + }, + { + "epoch": 0.6701298701298701, + "grad_norm": 49.01878747564461, + "learning_rate": 1.9928736849912144e-05, + "loss": 0.1693, + "step": 516 + }, + { + "epoch": 0.6714285714285714, + "grad_norm": 39.60937777847132, + "learning_rate": 1.9928234715377406e-05, + "loss": 0.1028, + "step": 517 + }, + { + "epoch": 0.6727272727272727, + "grad_norm": 61.07831282350039, + "learning_rate": 1.9927730824347915e-05, + "loss": 0.1485, + "step": 518 + }, + { + "epoch": 0.674025974025974, + "grad_norm": 35.03687547480847, + "learning_rate": 1.9927225176912826e-05, + "loss": 0.1686, + "step": 519 + }, + { + "epoch": 0.6753246753246753, + "grad_norm": 27.250968450689413, + "learning_rate": 1.992671777316159e-05, + "loss": 0.1161, + "step": 520 + }, + { + "epoch": 0.6766233766233766, + "grad_norm": 52.41672176850581, + "learning_rate": 1.992620861318398e-05, + "loss": 0.1735, + "step": 521 + }, + { + "epoch": 0.6779220779220779, + "grad_norm": 20.082471407349914, + "learning_rate": 1.992569769707008e-05, + "loss": 0.1763, + "step": 522 + }, + { + "epoch": 0.6792207792207792, + "grad_norm": 17.912033161148972, + "learning_rate": 1.992518502491028e-05, + "loss": 0.2004, + "step": 523 + }, + { + "epoch": 0.6805194805194805, + "grad_norm": 54.3986189701084, + "learning_rate": 1.9924670596795275e-05, + "loss": 0.146, + "step": 524 + }, + { + "epoch": 0.6818181818181818, + "grad_norm": 9.32192156743206, + "learning_rate": 1.992415441281608e-05, + "loss": 0.2008, + "step": 525 + }, + { + "epoch": 0.6831168831168831, + "grad_norm": 56.89727955973234, + "learning_rate": 1.9923636473064024e-05, + "loss": 0.1665, + "step": 526 + }, + { + "epoch": 0.6844155844155844, + "grad_norm": 17.163519089208478, + "learning_rate": 1.9923116777630733e-05, + "loss": 0.1461, + "step": 527 + }, + { + "epoch": 0.6857142857142857, + "grad_norm": 25.557807353572496, + "learning_rate": 1.9922595326608153e-05, + "loss": 0.2065, + "step": 528 + }, + { + "epoch": 0.687012987012987, + "grad_norm": 50.65815295797874, + "learning_rate": 1.9922072120088537e-05, + "loss": 0.1535, + "step": 529 + }, + { + "epoch": 0.6883116883116883, + "grad_norm": 30.77979456213415, + "learning_rate": 1.9921547158164457e-05, + "loss": 0.1393, + "step": 530 + }, + { + "epoch": 0.6896103896103896, + "grad_norm": 37.48507542075375, + "learning_rate": 1.992102044092878e-05, + "loss": 0.1545, + "step": 531 + }, + { + "epoch": 0.6909090909090909, + "grad_norm": 17.150236259024012, + "learning_rate": 1.99204919684747e-05, + "loss": 0.1433, + "step": 532 + }, + { + "epoch": 0.6922077922077922, + "grad_norm": 20.74529551400219, + "learning_rate": 1.9919961740895706e-05, + "loss": 0.2366, + "step": 533 + }, + { + "epoch": 0.6935064935064935, + "grad_norm": 36.64856087487405, + "learning_rate": 1.9919429758285616e-05, + "loss": 0.1393, + "step": 534 + }, + { + "epoch": 0.6948051948051948, + "grad_norm": 15.633529251631662, + "learning_rate": 1.991889602073854e-05, + "loss": 0.1485, + "step": 535 + }, + { + "epoch": 0.6961038961038961, + "grad_norm": 56.71819474548904, + "learning_rate": 1.991836052834891e-05, + "loss": 0.1614, + "step": 536 + }, + { + "epoch": 0.6974025974025974, + "grad_norm": 51.28394542839698, + "learning_rate": 1.9917823281211456e-05, + "loss": 0.1593, + "step": 537 + }, + { + "epoch": 0.6987012987012987, + "grad_norm": 28.4880864987389, + "learning_rate": 1.991728427942124e-05, + "loss": 0.1574, + "step": 538 + }, + { + "epoch": 0.7, + "grad_norm": 98.11891973233577, + "learning_rate": 1.9916743523073615e-05, + "loss": 0.1624, + "step": 539 + }, + { + "epoch": 0.7012987012987013, + "grad_norm": 15.826903468409686, + "learning_rate": 1.9916201012264255e-05, + "loss": 0.1508, + "step": 540 + }, + { + "epoch": 0.7025974025974026, + "grad_norm": 61.09664553616612, + "learning_rate": 1.9915656747089134e-05, + "loss": 0.1872, + "step": 541 + }, + { + "epoch": 0.7038961038961039, + "grad_norm": 30.76535279453644, + "learning_rate": 1.9915110727644547e-05, + "loss": 0.1208, + "step": 542 + }, + { + "epoch": 0.7051948051948052, + "grad_norm": 8.730574265887507, + "learning_rate": 1.9914562954027093e-05, + "loss": 0.1362, + "step": 543 + }, + { + "epoch": 0.7064935064935065, + "grad_norm": 81.87892012935244, + "learning_rate": 1.991401342633369e-05, + "loss": 0.1302, + "step": 544 + }, + { + "epoch": 0.7077922077922078, + "grad_norm": 18.53783926499007, + "learning_rate": 1.9913462144661556e-05, + "loss": 0.201, + "step": 545 + }, + { + "epoch": 0.7090909090909091, + "grad_norm": 91.68506455423227, + "learning_rate": 1.991290910910822e-05, + "loss": 0.1696, + "step": 546 + }, + { + "epoch": 0.7103896103896103, + "grad_norm": 46.48789291116875, + "learning_rate": 1.991235431977153e-05, + "loss": 0.1685, + "step": 547 + }, + { + "epoch": 0.7116883116883117, + "grad_norm": 81.02765916941878, + "learning_rate": 1.9911797776749636e-05, + "loss": 0.1705, + "step": 548 + }, + { + "epoch": 0.712987012987013, + "grad_norm": 87.19104245976997, + "learning_rate": 1.9911239480141e-05, + "loss": 0.1629, + "step": 549 + }, + { + "epoch": 0.7142857142857143, + "grad_norm": 75.47019559409459, + "learning_rate": 1.9910679430044395e-05, + "loss": 0.1656, + "step": 550 + }, + { + "epoch": 0.7155844155844155, + "grad_norm": 85.80617689250113, + "learning_rate": 1.9910117626558908e-05, + "loss": 0.2033, + "step": 551 + }, + { + "epoch": 0.7168831168831169, + "grad_norm": 44.61188308906432, + "learning_rate": 1.9909554069783935e-05, + "loss": 0.2046, + "step": 552 + }, + { + "epoch": 0.7181818181818181, + "grad_norm": 77.94282015534596, + "learning_rate": 1.990898875981917e-05, + "loss": 0.1858, + "step": 553 + }, + { + "epoch": 0.7194805194805195, + "grad_norm": 8.920753799760181, + "learning_rate": 1.990842169676464e-05, + "loss": 0.1404, + "step": 554 + }, + { + "epoch": 0.7207792207792207, + "grad_norm": 58.618943392736426, + "learning_rate": 1.990785288072066e-05, + "loss": 0.1625, + "step": 555 + }, + { + "epoch": 0.7220779220779221, + "grad_norm": 26.360857951794195, + "learning_rate": 1.9907282311787863e-05, + "loss": 0.139, + "step": 556 + }, + { + "epoch": 0.7233766233766233, + "grad_norm": 20.333968976501236, + "learning_rate": 1.9906709990067198e-05, + "loss": 0.1742, + "step": 557 + }, + { + "epoch": 0.7246753246753247, + "grad_norm": 35.23208139765357, + "learning_rate": 1.9906135915659924e-05, + "loss": 0.1562, + "step": 558 + }, + { + "epoch": 0.7259740259740259, + "grad_norm": 15.323981282702494, + "learning_rate": 1.9905560088667597e-05, + "loss": 0.1234, + "step": 559 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 23.801480244635336, + "learning_rate": 1.99049825091921e-05, + "loss": 0.1611, + "step": 560 + }, + { + "epoch": 0.7285714285714285, + "grad_norm": 29.101533492721842, + "learning_rate": 1.990440317733561e-05, + "loss": 0.1278, + "step": 561 + }, + { + "epoch": 0.7298701298701299, + "grad_norm": 17.087900291532797, + "learning_rate": 1.9903822093200625e-05, + "loss": 0.1801, + "step": 562 + }, + { + "epoch": 0.7311688311688311, + "grad_norm": 55.294017582376235, + "learning_rate": 1.9903239256889954e-05, + "loss": 0.1793, + "step": 563 + }, + { + "epoch": 0.7324675324675325, + "grad_norm": 18.90280186985712, + "learning_rate": 1.990265466850671e-05, + "loss": 0.225, + "step": 564 + }, + { + "epoch": 0.7337662337662337, + "grad_norm": 11.612098641949512, + "learning_rate": 1.9902068328154312e-05, + "loss": 0.1354, + "step": 565 + }, + { + "epoch": 0.7350649350649351, + "grad_norm": 23.360511822415184, + "learning_rate": 1.9901480235936502e-05, + "loss": 0.2009, + "step": 566 + }, + { + "epoch": 0.7363636363636363, + "grad_norm": 19.257238092274658, + "learning_rate": 1.990089039195732e-05, + "loss": 0.1677, + "step": 567 + }, + { + "epoch": 0.7376623376623377, + "grad_norm": 16.903641985739473, + "learning_rate": 1.9900298796321123e-05, + "loss": 0.1775, + "step": 568 + }, + { + "epoch": 0.7389610389610389, + "grad_norm": 9.493416653503525, + "learning_rate": 1.9899705449132575e-05, + "loss": 0.1642, + "step": 569 + }, + { + "epoch": 0.7402597402597403, + "grad_norm": 28.880505691230947, + "learning_rate": 1.989911035049665e-05, + "loss": 0.129, + "step": 570 + }, + { + "epoch": 0.7415584415584415, + "grad_norm": 8.988301740971204, + "learning_rate": 1.9898513500518634e-05, + "loss": 0.1276, + "step": 571 + }, + { + "epoch": 0.7428571428571429, + "grad_norm": 18.917733299171633, + "learning_rate": 1.9897914899304123e-05, + "loss": 0.1557, + "step": 572 + }, + { + "epoch": 0.7441558441558441, + "grad_norm": 12.372298427949026, + "learning_rate": 1.9897314546959015e-05, + "loss": 0.1237, + "step": 573 + }, + { + "epoch": 0.7454545454545455, + "grad_norm": 10.671387605271942, + "learning_rate": 1.9896712443589526e-05, + "loss": 0.1992, + "step": 574 + }, + { + "epoch": 0.7467532467532467, + "grad_norm": 17.655395181204334, + "learning_rate": 1.989610858930218e-05, + "loss": 0.1439, + "step": 575 + }, + { + "epoch": 0.7480519480519481, + "grad_norm": 21.2124295171341, + "learning_rate": 1.9895502984203816e-05, + "loss": 0.1576, + "step": 576 + }, + { + "epoch": 0.7493506493506493, + "grad_norm": 37.78546591260509, + "learning_rate": 1.9894895628401568e-05, + "loss": 0.1568, + "step": 577 + }, + { + "epoch": 0.7506493506493507, + "grad_norm": 11.020902743761214, + "learning_rate": 1.9894286522002894e-05, + "loss": 0.1367, + "step": 578 + }, + { + "epoch": 0.7519480519480519, + "grad_norm": 34.849705985236646, + "learning_rate": 1.9893675665115556e-05, + "loss": 0.1486, + "step": 579 + }, + { + "epoch": 0.7532467532467533, + "grad_norm": 15.989780913449499, + "learning_rate": 1.9893063057847627e-05, + "loss": 0.1445, + "step": 580 + }, + { + "epoch": 0.7545454545454545, + "grad_norm": 32.61804814977021, + "learning_rate": 1.989244870030749e-05, + "loss": 0.1826, + "step": 581 + }, + { + "epoch": 0.7558441558441559, + "grad_norm": 12.673079174785206, + "learning_rate": 1.989183259260383e-05, + "loss": 0.1615, + "step": 582 + }, + { + "epoch": 0.7571428571428571, + "grad_norm": 9.455898914721462, + "learning_rate": 1.989121473484565e-05, + "loss": 0.1606, + "step": 583 + }, + { + "epoch": 0.7584415584415585, + "grad_norm": 20.37949303711785, + "learning_rate": 1.989059512714227e-05, + "loss": 0.1764, + "step": 584 + }, + { + "epoch": 0.7597402597402597, + "grad_norm": 21.911640692364358, + "learning_rate": 1.9889973769603298e-05, + "loss": 0.1512, + "step": 585 + }, + { + "epoch": 0.7610389610389611, + "grad_norm": 8.78648171678769, + "learning_rate": 1.9889350662338676e-05, + "loss": 0.1523, + "step": 586 + }, + { + "epoch": 0.7623376623376623, + "grad_norm": 10.295685512258341, + "learning_rate": 1.9888725805458632e-05, + "loss": 0.1636, + "step": 587 + }, + { + "epoch": 0.7636363636363637, + "grad_norm": 20.282074503243464, + "learning_rate": 1.9888099199073726e-05, + "loss": 0.2396, + "step": 588 + }, + { + "epoch": 0.7649350649350649, + "grad_norm": 59.978299654720956, + "learning_rate": 1.9887470843294808e-05, + "loss": 0.1844, + "step": 589 + }, + { + "epoch": 0.7662337662337663, + "grad_norm": 31.01266376588936, + "learning_rate": 1.988684073823305e-05, + "loss": 0.1256, + "step": 590 + }, + { + "epoch": 0.7675324675324675, + "grad_norm": 49.885453198939615, + "learning_rate": 1.988620888399993e-05, + "loss": 0.1587, + "step": 591 + }, + { + "epoch": 0.7688311688311689, + "grad_norm": 27.03823260356419, + "learning_rate": 1.9885575280707233e-05, + "loss": 0.1993, + "step": 592 + }, + { + "epoch": 0.7701298701298701, + "grad_norm": 33.442479299191355, + "learning_rate": 1.988493992846706e-05, + "loss": 0.1426, + "step": 593 + }, + { + "epoch": 0.7714285714285715, + "grad_norm": 21.90059083993139, + "learning_rate": 1.9884302827391813e-05, + "loss": 0.1859, + "step": 594 + }, + { + "epoch": 0.7727272727272727, + "grad_norm": 9.779630109760449, + "learning_rate": 1.988366397759421e-05, + "loss": 0.1641, + "step": 595 + }, + { + "epoch": 0.7740259740259741, + "grad_norm": 13.113292443487262, + "learning_rate": 1.9883023379187267e-05, + "loss": 0.1577, + "step": 596 + }, + { + "epoch": 0.7753246753246753, + "grad_norm": 7.321299372347641, + "learning_rate": 1.988238103228433e-05, + "loss": 0.1314, + "step": 597 + }, + { + "epoch": 0.7766233766233767, + "grad_norm": 28.957879438547128, + "learning_rate": 1.9881736936999035e-05, + "loss": 0.11, + "step": 598 + }, + { + "epoch": 0.7779220779220779, + "grad_norm": 26.9774206488646, + "learning_rate": 1.9881091093445343e-05, + "loss": 0.172, + "step": 599 + }, + { + "epoch": 0.7792207792207793, + "grad_norm": 16.761310250159685, + "learning_rate": 1.9880443501737504e-05, + "loss": 0.1888, + "step": 600 + }, + { + "epoch": 0.7805194805194805, + "grad_norm": 7.1729594861191535, + "learning_rate": 1.98797941619901e-05, + "loss": 0.1268, + "step": 601 + }, + { + "epoch": 0.7818181818181819, + "grad_norm": 7.892603999951699, + "learning_rate": 1.9879143074318005e-05, + "loss": 0.1283, + "step": 602 + }, + { + "epoch": 0.7831168831168831, + "grad_norm": 30.799860138769073, + "learning_rate": 1.9878490238836415e-05, + "loss": 0.1263, + "step": 603 + }, + { + "epoch": 0.7844155844155845, + "grad_norm": 12.68916043156298, + "learning_rate": 1.9877835655660825e-05, + "loss": 0.1356, + "step": 604 + }, + { + "epoch": 0.7857142857142857, + "grad_norm": 51.20969637699772, + "learning_rate": 1.987717932490704e-05, + "loss": 0.148, + "step": 605 + }, + { + "epoch": 0.787012987012987, + "grad_norm": 10.44425550749443, + "learning_rate": 1.9876521246691183e-05, + "loss": 0.1855, + "step": 606 + }, + { + "epoch": 0.7883116883116883, + "grad_norm": 32.435447416088316, + "learning_rate": 1.9875861421129678e-05, + "loss": 0.1552, + "step": 607 + }, + { + "epoch": 0.7896103896103897, + "grad_norm": 15.316728276249531, + "learning_rate": 1.9875199848339265e-05, + "loss": 0.1785, + "step": 608 + }, + { + "epoch": 0.7909090909090909, + "grad_norm": 23.09834465539185, + "learning_rate": 1.987453652843698e-05, + "loss": 0.1213, + "step": 609 + }, + { + "epoch": 0.7922077922077922, + "grad_norm": 40.40695725789211, + "learning_rate": 1.9873871461540186e-05, + "loss": 0.1838, + "step": 610 + }, + { + "epoch": 0.7935064935064935, + "grad_norm": 17.327242310449254, + "learning_rate": 1.987320464776654e-05, + "loss": 0.2049, + "step": 611 + }, + { + "epoch": 0.7948051948051948, + "grad_norm": 51.6776146869577, + "learning_rate": 1.987253608723402e-05, + "loss": 0.1893, + "step": 612 + }, + { + "epoch": 0.7961038961038961, + "grad_norm": 28.5855862641096, + "learning_rate": 1.98718657800609e-05, + "loss": 0.1533, + "step": 613 + }, + { + "epoch": 0.7974025974025974, + "grad_norm": 97.50807287653687, + "learning_rate": 1.987119372636578e-05, + "loss": 0.2336, + "step": 614 + }, + { + "epoch": 0.7987012987012987, + "grad_norm": 16.29479273782878, + "learning_rate": 1.9870519926267547e-05, + "loss": 0.1583, + "step": 615 + }, + { + "epoch": 0.8, + "grad_norm": 68.23448574289843, + "learning_rate": 1.9869844379885418e-05, + "loss": 0.1678, + "step": 616 + }, + { + "epoch": 0.8012987012987013, + "grad_norm": 7.680857281984, + "learning_rate": 1.9869167087338908e-05, + "loss": 0.1539, + "step": 617 + }, + { + "epoch": 0.8025974025974026, + "grad_norm": 45.61146434298359, + "learning_rate": 1.986848804874784e-05, + "loss": 0.1497, + "step": 618 + }, + { + "epoch": 0.8038961038961039, + "grad_norm": 28.935369298977367, + "learning_rate": 1.9867807264232354e-05, + "loss": 0.1462, + "step": 619 + }, + { + "epoch": 0.8051948051948052, + "grad_norm": 37.998232863089704, + "learning_rate": 1.986712473391289e-05, + "loss": 0.1597, + "step": 620 + }, + { + "epoch": 0.8064935064935065, + "grad_norm": 50.38271846048889, + "learning_rate": 1.9866440457910202e-05, + "loss": 0.1749, + "step": 621 + }, + { + "epoch": 0.8077922077922078, + "grad_norm": 29.468001049613033, + "learning_rate": 1.9865754436345353e-05, + "loss": 0.1308, + "step": 622 + }, + { + "epoch": 0.8090909090909091, + "grad_norm": 75.29998465866485, + "learning_rate": 1.9865066669339707e-05, + "loss": 0.1827, + "step": 623 + }, + { + "epoch": 0.8103896103896104, + "grad_norm": 48.4147345696491, + "learning_rate": 1.986437715701495e-05, + "loss": 0.142, + "step": 624 + }, + { + "epoch": 0.8116883116883117, + "grad_norm": 54.08588180410964, + "learning_rate": 1.986368589949307e-05, + "loss": 0.1428, + "step": 625 + }, + { + "epoch": 0.812987012987013, + "grad_norm": 16.804756626731724, + "learning_rate": 1.9862992896896358e-05, + "loss": 0.1768, + "step": 626 + }, + { + "epoch": 0.8142857142857143, + "grad_norm": 72.64356876857815, + "learning_rate": 1.986229814934743e-05, + "loss": 0.1646, + "step": 627 + }, + { + "epoch": 0.8155844155844156, + "grad_norm": 12.749864983755476, + "learning_rate": 1.9861601656969185e-05, + "loss": 0.1781, + "step": 628 + }, + { + "epoch": 0.8168831168831169, + "grad_norm": 59.9131456980836, + "learning_rate": 1.9860903419884857e-05, + "loss": 0.15, + "step": 629 + }, + { + "epoch": 0.8181818181818182, + "grad_norm": 53.164928500174355, + "learning_rate": 1.9860203438217972e-05, + "loss": 0.204, + "step": 630 + }, + { + "epoch": 0.8194805194805195, + "grad_norm": 95.50165320028005, + "learning_rate": 1.9859501712092374e-05, + "loss": 0.1964, + "step": 631 + }, + { + "epoch": 0.8207792207792208, + "grad_norm": 15.5450058443917, + "learning_rate": 1.985879824163221e-05, + "loss": 0.1488, + "step": 632 + }, + { + "epoch": 0.8220779220779221, + "grad_norm": 22.808122422769312, + "learning_rate": 1.985809302696194e-05, + "loss": 0.1288, + "step": 633 + }, + { + "epoch": 0.8233766233766234, + "grad_norm": 17.52250620136283, + "learning_rate": 1.9857386068206324e-05, + "loss": 0.1348, + "step": 634 + }, + { + "epoch": 0.8246753246753247, + "grad_norm": 24.52256800292182, + "learning_rate": 1.985667736549044e-05, + "loss": 0.1445, + "step": 635 + }, + { + "epoch": 0.825974025974026, + "grad_norm": 7.15365942905463, + "learning_rate": 1.9855966918939673e-05, + "loss": 0.0776, + "step": 636 + }, + { + "epoch": 0.8272727272727273, + "grad_norm": 19.115009513417853, + "learning_rate": 1.9855254728679714e-05, + "loss": 0.1443, + "step": 637 + }, + { + "epoch": 0.8285714285714286, + "grad_norm": 7.937625139284422, + "learning_rate": 1.9854540794836562e-05, + "loss": 0.1217, + "step": 638 + }, + { + "epoch": 0.8298701298701299, + "grad_norm": 28.5866087013663, + "learning_rate": 1.9853825117536522e-05, + "loss": 0.1262, + "step": 639 + }, + { + "epoch": 0.8311688311688312, + "grad_norm": 31.84697934629128, + "learning_rate": 1.9853107696906217e-05, + "loss": 0.1163, + "step": 640 + }, + { + "epoch": 0.8324675324675325, + "grad_norm": 14.625844823245757, + "learning_rate": 1.985238853307257e-05, + "loss": 0.158, + "step": 641 + }, + { + "epoch": 0.8337662337662337, + "grad_norm": 35.47420763254657, + "learning_rate": 1.9851667626162813e-05, + "loss": 0.2273, + "step": 642 + }, + { + "epoch": 0.8350649350649351, + "grad_norm": 33.645491797896135, + "learning_rate": 1.985094497630449e-05, + "loss": 0.1425, + "step": 643 + }, + { + "epoch": 0.8363636363636363, + "grad_norm": 33.99669500930693, + "learning_rate": 1.9850220583625453e-05, + "loss": 0.1371, + "step": 644 + }, + { + "epoch": 0.8376623376623377, + "grad_norm": 33.28443521220901, + "learning_rate": 1.9849494448253865e-05, + "loss": 0.1485, + "step": 645 + }, + { + "epoch": 0.8389610389610389, + "grad_norm": 41.41662540276414, + "learning_rate": 1.984876657031818e-05, + "loss": 0.1169, + "step": 646 + }, + { + "epoch": 0.8402597402597403, + "grad_norm": 5.462742550598085, + "learning_rate": 1.984803694994718e-05, + "loss": 0.1233, + "step": 647 + }, + { + "epoch": 0.8415584415584415, + "grad_norm": 54.13305254816279, + "learning_rate": 1.9847305587269958e-05, + "loss": 0.1505, + "step": 648 + }, + { + "epoch": 0.8428571428571429, + "grad_norm": 17.32049412255105, + "learning_rate": 1.984657248241589e-05, + "loss": 0.1987, + "step": 649 + }, + { + "epoch": 0.8441558441558441, + "grad_norm": 18.882987907633968, + "learning_rate": 1.9845837635514692e-05, + "loss": 0.116, + "step": 650 + }, + { + "epoch": 0.8454545454545455, + "grad_norm": 30.497096793926687, + "learning_rate": 1.984510104669636e-05, + "loss": 0.1192, + "step": 651 + }, + { + "epoch": 0.8467532467532467, + "grad_norm": 47.00604420880673, + "learning_rate": 1.9844362716091217e-05, + "loss": 0.152, + "step": 652 + }, + { + "epoch": 0.8480519480519481, + "grad_norm": 23.487846911723747, + "learning_rate": 1.9843622643829887e-05, + "loss": 0.1487, + "step": 653 + }, + { + "epoch": 0.8493506493506493, + "grad_norm": 10.326619581384175, + "learning_rate": 1.98428808300433e-05, + "loss": 0.1381, + "step": 654 + }, + { + "epoch": 0.8506493506493507, + "grad_norm": 8.435546602671815, + "learning_rate": 1.9842137274862702e-05, + "loss": 0.1451, + "step": 655 + }, + { + "epoch": 0.8519480519480519, + "grad_norm": 11.194913155547193, + "learning_rate": 1.9841391978419638e-05, + "loss": 0.1288, + "step": 656 + }, + { + "epoch": 0.8532467532467533, + "grad_norm": 15.954405276183662, + "learning_rate": 1.9840644940845968e-05, + "loss": 0.1496, + "step": 657 + }, + { + "epoch": 0.8545454545454545, + "grad_norm": 20.359113248832088, + "learning_rate": 1.9839896162273856e-05, + "loss": 0.1311, + "step": 658 + }, + { + "epoch": 0.8558441558441559, + "grad_norm": 25.793975017792526, + "learning_rate": 1.9839145642835775e-05, + "loss": 0.1424, + "step": 659 + }, + { + "epoch": 0.8571428571428571, + "grad_norm": 27.336171550372537, + "learning_rate": 1.9838393382664505e-05, + "loss": 0.1578, + "step": 660 + }, + { + "epoch": 0.8584415584415584, + "grad_norm": 45.60782646137502, + "learning_rate": 1.9837639381893137e-05, + "loss": 0.1213, + "step": 661 + }, + { + "epoch": 0.8597402597402597, + "grad_norm": 37.921644347031744, + "learning_rate": 1.983688364065507e-05, + "loss": 0.1362, + "step": 662 + }, + { + "epoch": 0.861038961038961, + "grad_norm": 50.974596272792716, + "learning_rate": 1.9836126159084005e-05, + "loss": 0.1913, + "step": 663 + }, + { + "epoch": 0.8623376623376623, + "grad_norm": 27.758683664019134, + "learning_rate": 1.9835366937313958e-05, + "loss": 0.1152, + "step": 664 + }, + { + "epoch": 0.8636363636363636, + "grad_norm": 68.74047081983201, + "learning_rate": 1.9834605975479248e-05, + "loss": 0.1306, + "step": 665 + }, + { + "epoch": 0.8649350649350649, + "grad_norm": 27.23215731554995, + "learning_rate": 1.9833843273714506e-05, + "loss": 0.222, + "step": 666 + }, + { + "epoch": 0.8662337662337662, + "grad_norm": 65.19642749841573, + "learning_rate": 1.9833078832154666e-05, + "loss": 0.1562, + "step": 667 + }, + { + "epoch": 0.8675324675324675, + "grad_norm": 22.83894301531483, + "learning_rate": 1.9832312650934973e-05, + "loss": 0.121, + "step": 668 + }, + { + "epoch": 0.8688311688311688, + "grad_norm": 40.89739209687402, + "learning_rate": 1.983154473019098e-05, + "loss": 0.1225, + "step": 669 + }, + { + "epoch": 0.8701298701298701, + "grad_norm": 9.1455758465833, + "learning_rate": 1.983077507005855e-05, + "loss": 0.1586, + "step": 670 + }, + { + "epoch": 0.8714285714285714, + "grad_norm": 14.31606682358203, + "learning_rate": 1.9830003670673843e-05, + "loss": 0.1042, + "step": 671 + }, + { + "epoch": 0.8727272727272727, + "grad_norm": 49.50977591970933, + "learning_rate": 1.9829230532173337e-05, + "loss": 0.1445, + "step": 672 + }, + { + "epoch": 0.874025974025974, + "grad_norm": 35.545418781745276, + "learning_rate": 1.982845565469382e-05, + "loss": 0.1316, + "step": 673 + }, + { + "epoch": 0.8753246753246753, + "grad_norm": 33.722351944367155, + "learning_rate": 1.9827679038372377e-05, + "loss": 0.1574, + "step": 674 + }, + { + "epoch": 0.8766233766233766, + "grad_norm": 11.296362373377589, + "learning_rate": 1.982690068334641e-05, + "loss": 0.131, + "step": 675 + }, + { + "epoch": 0.8779220779220779, + "grad_norm": 61.11953153741602, + "learning_rate": 1.9826120589753623e-05, + "loss": 0.0984, + "step": 676 + }, + { + "epoch": 0.8792207792207792, + "grad_norm": 11.848613020175549, + "learning_rate": 1.9825338757732028e-05, + "loss": 0.1511, + "step": 677 + }, + { + "epoch": 0.8805194805194805, + "grad_norm": 53.02888661920757, + "learning_rate": 1.9824555187419948e-05, + "loss": 0.1065, + "step": 678 + }, + { + "epoch": 0.8818181818181818, + "grad_norm": 34.76337426973743, + "learning_rate": 1.9823769878956013e-05, + "loss": 0.1263, + "step": 679 + }, + { + "epoch": 0.8831168831168831, + "grad_norm": 31.417686885685026, + "learning_rate": 1.9822982832479157e-05, + "loss": 0.1463, + "step": 680 + }, + { + "epoch": 0.8844155844155844, + "grad_norm": 60.83215026796979, + "learning_rate": 1.9822194048128624e-05, + "loss": 0.2667, + "step": 681 + }, + { + "epoch": 0.8857142857142857, + "grad_norm": 34.20255425601077, + "learning_rate": 1.9821403526043962e-05, + "loss": 0.1282, + "step": 682 + }, + { + "epoch": 0.887012987012987, + "grad_norm": 51.1382641958208, + "learning_rate": 1.9820611266365036e-05, + "loss": 0.1331, + "step": 683 + }, + { + "epoch": 0.8883116883116883, + "grad_norm": 29.55014494721064, + "learning_rate": 1.981981726923201e-05, + "loss": 0.1208, + "step": 684 + }, + { + "epoch": 0.8896103896103896, + "grad_norm": 38.44018352748228, + "learning_rate": 1.9819021534785353e-05, + "loss": 0.1295, + "step": 685 + }, + { + "epoch": 0.8909090909090909, + "grad_norm": 21.766300957515426, + "learning_rate": 1.9818224063165847e-05, + "loss": 0.1335, + "step": 686 + }, + { + "epoch": 0.8922077922077922, + "grad_norm": 59.16757731768089, + "learning_rate": 1.9817424854514583e-05, + "loss": 0.1591, + "step": 687 + }, + { + "epoch": 0.8935064935064935, + "grad_norm": 24.852227893816792, + "learning_rate": 1.9816623908972955e-05, + "loss": 0.1224, + "step": 688 + }, + { + "epoch": 0.8948051948051948, + "grad_norm": 96.54439968107333, + "learning_rate": 1.9815821226682668e-05, + "loss": 0.1555, + "step": 689 + }, + { + "epoch": 0.8961038961038961, + "grad_norm": 43.55966060722475, + "learning_rate": 1.9815016807785726e-05, + "loss": 0.217, + "step": 690 + }, + { + "epoch": 0.8974025974025974, + "grad_norm": 58.29026940212491, + "learning_rate": 1.981421065242445e-05, + "loss": 0.1513, + "step": 691 + }, + { + "epoch": 0.8987012987012987, + "grad_norm": 14.761101671922706, + "learning_rate": 1.9813402760741464e-05, + "loss": 0.1294, + "step": 692 + }, + { + "epoch": 0.9, + "grad_norm": 77.05533521057328, + "learning_rate": 1.9812593132879698e-05, + "loss": 0.1804, + "step": 693 + }, + { + "epoch": 0.9012987012987013, + "grad_norm": 14.58673697678054, + "learning_rate": 1.9811781768982392e-05, + "loss": 0.1426, + "step": 694 + }, + { + "epoch": 0.9025974025974026, + "grad_norm": 60.29920717981256, + "learning_rate": 1.9810968669193094e-05, + "loss": 0.155, + "step": 695 + }, + { + "epoch": 0.9038961038961039, + "grad_norm": 37.99068214869369, + "learning_rate": 1.981015383365565e-05, + "loss": 0.1338, + "step": 696 + }, + { + "epoch": 0.9051948051948052, + "grad_norm": 50.15270014034821, + "learning_rate": 1.9809337262514224e-05, + "loss": 0.1863, + "step": 697 + }, + { + "epoch": 0.9064935064935065, + "grad_norm": 45.694616585221965, + "learning_rate": 1.980851895591329e-05, + "loss": 0.157, + "step": 698 + }, + { + "epoch": 0.9077922077922078, + "grad_norm": 27.652319713058716, + "learning_rate": 1.980769891399761e-05, + "loss": 0.105, + "step": 699 + }, + { + "epoch": 0.9090909090909091, + "grad_norm": 38.2260778985031, + "learning_rate": 1.9806877136912273e-05, + "loss": 0.1395, + "step": 700 + }, + { + "epoch": 0.9103896103896104, + "grad_norm": 16.397904945631797, + "learning_rate": 1.9806053624802667e-05, + "loss": 0.1411, + "step": 701 + }, + { + "epoch": 0.9116883116883117, + "grad_norm": 45.24725191213572, + "learning_rate": 1.980522837781448e-05, + "loss": 0.1214, + "step": 702 + }, + { + "epoch": 0.912987012987013, + "grad_norm": 7.4152792084445895, + "learning_rate": 1.9804401396093723e-05, + "loss": 0.1239, + "step": 703 + }, + { + "epoch": 0.9142857142857143, + "grad_norm": 18.039197759409518, + "learning_rate": 1.9803572679786696e-05, + "loss": 0.099, + "step": 704 + }, + { + "epoch": 0.9155844155844156, + "grad_norm": 9.245820761727758, + "learning_rate": 1.9802742229040027e-05, + "loss": 0.123, + "step": 705 + }, + { + "epoch": 0.9168831168831169, + "grad_norm": 8.82270385284703, + "learning_rate": 1.9801910044000626e-05, + "loss": 0.1146, + "step": 706 + }, + { + "epoch": 0.9181818181818182, + "grad_norm": 14.18768487196428, + "learning_rate": 1.980107612481573e-05, + "loss": 0.1488, + "step": 707 + }, + { + "epoch": 0.9194805194805195, + "grad_norm": 24.33968002668872, + "learning_rate": 1.980024047163287e-05, + "loss": 0.1534, + "step": 708 + }, + { + "epoch": 0.9207792207792208, + "grad_norm": 9.779540715451327, + "learning_rate": 1.9799403084599893e-05, + "loss": 0.1541, + "step": 709 + }, + { + "epoch": 0.922077922077922, + "grad_norm": 47.96745790209999, + "learning_rate": 1.9798563963864946e-05, + "loss": 0.1498, + "step": 710 + }, + { + "epoch": 0.9233766233766234, + "grad_norm": 25.8674967143919, + "learning_rate": 1.9797723109576493e-05, + "loss": 0.1113, + "step": 711 + }, + { + "epoch": 0.9246753246753247, + "grad_norm": 92.88409548824418, + "learning_rate": 1.979688052188329e-05, + "loss": 0.1595, + "step": 712 + }, + { + "epoch": 0.925974025974026, + "grad_norm": 14.328057823872879, + "learning_rate": 1.9796036200934403e-05, + "loss": 0.1266, + "step": 713 + }, + { + "epoch": 0.9272727272727272, + "grad_norm": 50.18349689776277, + "learning_rate": 1.979519014687922e-05, + "loss": 0.139, + "step": 714 + }, + { + "epoch": 0.9285714285714286, + "grad_norm": 14.842846617127817, + "learning_rate": 1.9794342359867412e-05, + "loss": 0.153, + "step": 715 + }, + { + "epoch": 0.9298701298701298, + "grad_norm": 9.68520678411129, + "learning_rate": 1.979349284004898e-05, + "loss": 0.1317, + "step": 716 + }, + { + "epoch": 0.9311688311688312, + "grad_norm": 49.56055008870536, + "learning_rate": 1.9792641587574212e-05, + "loss": 0.099, + "step": 717 + }, + { + "epoch": 0.9324675324675324, + "grad_norm": 16.817958966096395, + "learning_rate": 1.9791788602593716e-05, + "loss": 0.1153, + "step": 718 + }, + { + "epoch": 0.9337662337662338, + "grad_norm": 31.366517033056958, + "learning_rate": 1.97909338852584e-05, + "loss": 0.1543, + "step": 719 + }, + { + "epoch": 0.935064935064935, + "grad_norm": 12.367176201184943, + "learning_rate": 1.9790077435719477e-05, + "loss": 0.0986, + "step": 720 + }, + { + "epoch": 0.9363636363636364, + "grad_norm": 39.144691184246774, + "learning_rate": 1.978921925412847e-05, + "loss": 0.1498, + "step": 721 + }, + { + "epoch": 0.9376623376623376, + "grad_norm": 10.639051957843272, + "learning_rate": 1.9788359340637213e-05, + "loss": 0.174, + "step": 722 + }, + { + "epoch": 0.938961038961039, + "grad_norm": 14.038720173445704, + "learning_rate": 1.9787497695397836e-05, + "loss": 0.1057, + "step": 723 + }, + { + "epoch": 0.9402597402597402, + "grad_norm": 14.448276974776858, + "learning_rate": 1.9786634318562784e-05, + "loss": 0.1426, + "step": 724 + }, + { + "epoch": 0.9415584415584416, + "grad_norm": 38.889330782342995, + "learning_rate": 1.97857692102848e-05, + "loss": 0.17, + "step": 725 + }, + { + "epoch": 0.9428571428571428, + "grad_norm": 48.55242764374121, + "learning_rate": 1.9784902370716942e-05, + "loss": 0.1389, + "step": 726 + }, + { + "epoch": 0.9441558441558442, + "grad_norm": 47.49228934979205, + "learning_rate": 1.978403380001257e-05, + "loss": 0.1567, + "step": 727 + }, + { + "epoch": 0.9454545454545454, + "grad_norm": 54.244919847339915, + "learning_rate": 1.9783163498325352e-05, + "loss": 0.1476, + "step": 728 + }, + { + "epoch": 0.9467532467532468, + "grad_norm": 40.32714779625412, + "learning_rate": 1.9782291465809256e-05, + "loss": 0.1531, + "step": 729 + }, + { + "epoch": 0.948051948051948, + "grad_norm": 19.885948591202204, + "learning_rate": 1.9781417702618566e-05, + "loss": 0.1165, + "step": 730 + }, + { + "epoch": 0.9493506493506494, + "grad_norm": 22.874943328545495, + "learning_rate": 1.9780542208907866e-05, + "loss": 0.1465, + "step": 731 + }, + { + "epoch": 0.9506493506493506, + "grad_norm": 20.441189952350694, + "learning_rate": 1.977966498483205e-05, + "loss": 0.1425, + "step": 732 + }, + { + "epoch": 0.951948051948052, + "grad_norm": 33.36000671545076, + "learning_rate": 1.9778786030546308e-05, + "loss": 0.1226, + "step": 733 + }, + { + "epoch": 0.9532467532467532, + "grad_norm": 24.186762565841427, + "learning_rate": 1.9777905346206153e-05, + "loss": 0.1588, + "step": 734 + }, + { + "epoch": 0.9545454545454546, + "grad_norm": 12.266639088131573, + "learning_rate": 1.9777022931967395e-05, + "loss": 0.1652, + "step": 735 + }, + { + "epoch": 0.9558441558441558, + "grad_norm": 17.200775825084268, + "learning_rate": 1.9776138787986137e-05, + "loss": 0.1787, + "step": 736 + }, + { + "epoch": 0.9571428571428572, + "grad_norm": 42.060131721258664, + "learning_rate": 1.9775252914418818e-05, + "loss": 0.1059, + "step": 737 + }, + { + "epoch": 0.9584415584415584, + "grad_norm": 15.1621895298883, + "learning_rate": 1.9774365311422155e-05, + "loss": 0.1121, + "step": 738 + }, + { + "epoch": 0.9597402597402598, + "grad_norm": 14.063350567610449, + "learning_rate": 1.9773475979153185e-05, + "loss": 0.1356, + "step": 739 + }, + { + "epoch": 0.961038961038961, + "grad_norm": 32.54485091844219, + "learning_rate": 1.9772584917769245e-05, + "loss": 0.1327, + "step": 740 + }, + { + "epoch": 0.9623376623376624, + "grad_norm": 38.76163163407237, + "learning_rate": 1.977169212742799e-05, + "loss": 0.1296, + "step": 741 + }, + { + "epoch": 0.9636363636363636, + "grad_norm": 25.342569956784324, + "learning_rate": 1.9770797608287362e-05, + "loss": 0.1784, + "step": 742 + }, + { + "epoch": 0.964935064935065, + "grad_norm": 27.575121710140895, + "learning_rate": 1.9769901360505623e-05, + "loss": 0.1633, + "step": 743 + }, + { + "epoch": 0.9662337662337662, + "grad_norm": 40.47301224177095, + "learning_rate": 1.9769003384241333e-05, + "loss": 0.1341, + "step": 744 + }, + { + "epoch": 0.9675324675324676, + "grad_norm": 25.34222192605872, + "learning_rate": 1.9768103679653365e-05, + "loss": 0.1167, + "step": 745 + }, + { + "epoch": 0.9688311688311688, + "grad_norm": 20.93577913596902, + "learning_rate": 1.9767202246900893e-05, + "loss": 0.0989, + "step": 746 + }, + { + "epoch": 0.9701298701298702, + "grad_norm": 8.718909297864654, + "learning_rate": 1.9766299086143398e-05, + "loss": 0.1132, + "step": 747 + }, + { + "epoch": 0.9714285714285714, + "grad_norm": 6.928309624848111, + "learning_rate": 1.9765394197540666e-05, + "loss": 0.1214, + "step": 748 + }, + { + "epoch": 0.9727272727272728, + "grad_norm": 13.022751823690044, + "learning_rate": 1.9764487581252787e-05, + "loss": 0.132, + "step": 749 + }, + { + "epoch": 0.974025974025974, + "grad_norm": 20.58053618526608, + "learning_rate": 1.976357923744016e-05, + "loss": 0.1227, + "step": 750 + }, + { + "epoch": 0.9753246753246754, + "grad_norm": 32.45530016658457, + "learning_rate": 1.9762669166263492e-05, + "loss": 0.1526, + "step": 751 + }, + { + "epoch": 0.9766233766233766, + "grad_norm": 18.864536741027884, + "learning_rate": 1.9761757367883787e-05, + "loss": 0.1473, + "step": 752 + }, + { + "epoch": 0.977922077922078, + "grad_norm": 32.39919604753285, + "learning_rate": 1.9760843842462364e-05, + "loss": 0.1154, + "step": 753 + }, + { + "epoch": 0.9792207792207792, + "grad_norm": 18.028962381228727, + "learning_rate": 1.9759928590160838e-05, + "loss": 0.1028, + "step": 754 + }, + { + "epoch": 0.9805194805194806, + "grad_norm": 28.08692205939195, + "learning_rate": 1.9759011611141138e-05, + "loss": 0.152, + "step": 755 + }, + { + "epoch": 0.9818181818181818, + "grad_norm": 17.051538683239983, + "learning_rate": 1.9758092905565493e-05, + "loss": 0.1187, + "step": 756 + }, + { + "epoch": 0.9831168831168832, + "grad_norm": 24.197145482209784, + "learning_rate": 1.9757172473596444e-05, + "loss": 0.1483, + "step": 757 + }, + { + "epoch": 0.9844155844155844, + "grad_norm": 18.790491154081874, + "learning_rate": 1.975625031539683e-05, + "loss": 0.1414, + "step": 758 + }, + { + "epoch": 0.9857142857142858, + "grad_norm": 12.76885629361089, + "learning_rate": 1.9755326431129797e-05, + "loss": 0.1211, + "step": 759 + }, + { + "epoch": 0.987012987012987, + "grad_norm": 8.915887385057887, + "learning_rate": 1.9754400820958798e-05, + "loss": 0.1777, + "step": 760 + }, + { + "epoch": 0.9883116883116884, + "grad_norm": 31.51148648880449, + "learning_rate": 1.9753473485047596e-05, + "loss": 0.1281, + "step": 761 + }, + { + "epoch": 0.9896103896103896, + "grad_norm": 28.997363448457072, + "learning_rate": 1.975254442356025e-05, + "loss": 0.1524, + "step": 762 + }, + { + "epoch": 0.990909090909091, + "grad_norm": 22.706813410929325, + "learning_rate": 1.975161363666113e-05, + "loss": 0.1564, + "step": 763 + }, + { + "epoch": 0.9922077922077922, + "grad_norm": 46.56055697339971, + "learning_rate": 1.975068112451491e-05, + "loss": 0.1558, + "step": 764 + }, + { + "epoch": 0.9935064935064936, + "grad_norm": 18.86458453454055, + "learning_rate": 1.9749746887286566e-05, + "loss": 0.142, + "step": 765 + }, + { + "epoch": 0.9948051948051948, + "grad_norm": 29.593625938922816, + "learning_rate": 1.9748810925141386e-05, + "loss": 0.1314, + "step": 766 + }, + { + "epoch": 0.9961038961038962, + "grad_norm": 11.263502840436226, + "learning_rate": 1.974787323824496e-05, + "loss": 0.0896, + "step": 767 + }, + { + "epoch": 0.9974025974025974, + "grad_norm": 23.41374366441126, + "learning_rate": 1.974693382676318e-05, + "loss": 0.1219, + "step": 768 + }, + { + "epoch": 0.9987012987012988, + "grad_norm": 9.726567886707256, + "learning_rate": 1.9745992690862256e-05, + "loss": 0.1014, + "step": 769 + }, + { + "epoch": 1.0, + "grad_norm": 11.985812195843046, + "learning_rate": 1.9745049830708676e-05, + "loss": 0.1032, + "step": 770 + }, + { + "epoch": 1.0, + "eval_accuracy": 0.86692869174621, + "eval_f1": 0.8359764550582488, + "eval_loss": 0.19289281964302063, + "eval_precision": 0.8435435346504739, + "eval_recall": 0.8589457122970267, + "eval_runtime": 14.546, + "eval_samples_per_second": 122.439, + "eval_steps_per_second": 0.962, + "step": 770 + }, + { + "epoch": 1.0012987012987014, + "grad_norm": 16.227886640561348, + "learning_rate": 1.9744105246469264e-05, + "loss": 0.1553, + "step": 771 + }, + { + "epoch": 1.0025974025974025, + "grad_norm": 15.780219730903134, + "learning_rate": 1.9743158938311128e-05, + "loss": 0.1303, + "step": 772 + }, + { + "epoch": 1.0038961038961038, + "grad_norm": 48.065510686741106, + "learning_rate": 1.9742210906401686e-05, + "loss": 0.1548, + "step": 773 + }, + { + "epoch": 1.0051948051948052, + "grad_norm": 13.913697859516816, + "learning_rate": 1.974126115090867e-05, + "loss": 0.1551, + "step": 774 + }, + { + "epoch": 1.0064935064935066, + "grad_norm": 43.82171636023464, + "learning_rate": 1.974030967200011e-05, + "loss": 0.1265, + "step": 775 + }, + { + "epoch": 1.0077922077922077, + "grad_norm": 19.74152259550873, + "learning_rate": 1.973935646984433e-05, + "loss": 0.107, + "step": 776 + }, + { + "epoch": 1.009090909090909, + "grad_norm": 33.0326257000995, + "learning_rate": 1.973840154460998e-05, + "loss": 0.1999, + "step": 777 + }, + { + "epoch": 1.0103896103896104, + "grad_norm": 40.01885781954956, + "learning_rate": 1.9737444896466006e-05, + "loss": 0.1635, + "step": 778 + }, + { + "epoch": 1.0116883116883117, + "grad_norm": 17.180280792284776, + "learning_rate": 1.9736486525581648e-05, + "loss": 0.1791, + "step": 779 + }, + { + "epoch": 1.0129870129870129, + "grad_norm": 55.84234222351696, + "learning_rate": 1.973552643212647e-05, + "loss": 0.166, + "step": 780 + }, + { + "epoch": 1.0142857142857142, + "grad_norm": 13.226052278394492, + "learning_rate": 1.9734564616270324e-05, + "loss": 0.1279, + "step": 781 + }, + { + "epoch": 1.0155844155844156, + "grad_norm": 12.675528355727025, + "learning_rate": 1.9733601078183377e-05, + "loss": 0.1124, + "step": 782 + }, + { + "epoch": 1.016883116883117, + "grad_norm": 37.954555544356595, + "learning_rate": 1.9732635818036093e-05, + "loss": 0.1094, + "step": 783 + }, + { + "epoch": 1.018181818181818, + "grad_norm": 12.938420059715503, + "learning_rate": 1.973166883599925e-05, + "loss": 0.1536, + "step": 784 + }, + { + "epoch": 1.0194805194805194, + "grad_norm": 38.65081960529948, + "learning_rate": 1.9730700132243924e-05, + "loss": 0.1376, + "step": 785 + }, + { + "epoch": 1.0207792207792208, + "grad_norm": 23.628272139081222, + "learning_rate": 1.9729729706941495e-05, + "loss": 0.1306, + "step": 786 + }, + { + "epoch": 1.0220779220779221, + "grad_norm": 51.77971220957892, + "learning_rate": 1.9728757560263654e-05, + "loss": 0.1291, + "step": 787 + }, + { + "epoch": 1.0233766233766233, + "grad_norm": 48.95386791171294, + "learning_rate": 1.9727783692382388e-05, + "loss": 0.1396, + "step": 788 + }, + { + "epoch": 1.0246753246753246, + "grad_norm": 62.665732845645785, + "learning_rate": 1.9726808103469998e-05, + "loss": 0.1589, + "step": 789 + }, + { + "epoch": 1.025974025974026, + "grad_norm": 25.071885684664952, + "learning_rate": 1.972583079369908e-05, + "loss": 0.1014, + "step": 790 + }, + { + "epoch": 1.0272727272727273, + "grad_norm": 25.81797829122347, + "learning_rate": 1.9724851763242538e-05, + "loss": 0.1909, + "step": 791 + }, + { + "epoch": 1.0285714285714285, + "grad_norm": 32.664118179604394, + "learning_rate": 1.972387101227359e-05, + "loss": 0.1105, + "step": 792 + }, + { + "epoch": 1.0298701298701298, + "grad_norm": 17.710522095694703, + "learning_rate": 1.9722888540965738e-05, + "loss": 0.1252, + "step": 793 + }, + { + "epoch": 1.0311688311688312, + "grad_norm": 31.393021920491922, + "learning_rate": 1.9721904349492803e-05, + "loss": 0.1299, + "step": 794 + }, + { + "epoch": 1.0324675324675325, + "grad_norm": 8.008771736124826, + "learning_rate": 1.972091843802891e-05, + "loss": 0.1104, + "step": 795 + }, + { + "epoch": 1.0337662337662337, + "grad_norm": 43.64439964053457, + "learning_rate": 1.971993080674849e-05, + "loss": 0.1328, + "step": 796 + }, + { + "epoch": 1.035064935064935, + "grad_norm": 26.109565671581763, + "learning_rate": 1.9718941455826268e-05, + "loss": 0.1556, + "step": 797 + }, + { + "epoch": 1.0363636363636364, + "grad_norm": 79.74701696155712, + "learning_rate": 1.9717950385437277e-05, + "loss": 0.1426, + "step": 798 + }, + { + "epoch": 1.0376623376623377, + "grad_norm": 14.502221476878418, + "learning_rate": 1.971695759575686e-05, + "loss": 0.1399, + "step": 799 + }, + { + "epoch": 1.0389610389610389, + "grad_norm": 65.0701680970009, + "learning_rate": 1.9715963086960662e-05, + "loss": 0.1031, + "step": 800 + }, + { + "epoch": 1.0402597402597402, + "grad_norm": 43.52382924917024, + "learning_rate": 1.971496685922462e-05, + "loss": 0.1417, + "step": 801 + }, + { + "epoch": 1.0415584415584416, + "grad_norm": 97.80727550857769, + "learning_rate": 1.9713968912725005e-05, + "loss": 0.1508, + "step": 802 + }, + { + "epoch": 1.042857142857143, + "grad_norm": 13.783299858204145, + "learning_rate": 1.9712969247638357e-05, + "loss": 0.1218, + "step": 803 + }, + { + "epoch": 1.044155844155844, + "grad_norm": 70.49628835609026, + "learning_rate": 1.9711967864141542e-05, + "loss": 0.144, + "step": 804 + }, + { + "epoch": 1.0454545454545454, + "grad_norm": 37.91274173467588, + "learning_rate": 1.9710964762411723e-05, + "loss": 0.1298, + "step": 805 + }, + { + "epoch": 1.0467532467532468, + "grad_norm": 54.660661002539314, + "learning_rate": 1.9709959942626366e-05, + "loss": 0.1184, + "step": 806 + }, + { + "epoch": 1.0480519480519481, + "grad_norm": 68.31207636595494, + "learning_rate": 1.9708953404963246e-05, + "loss": 0.1299, + "step": 807 + }, + { + "epoch": 1.0493506493506493, + "grad_norm": 57.409100406885855, + "learning_rate": 1.9707945149600433e-05, + "loss": 0.1087, + "step": 808 + }, + { + "epoch": 1.0506493506493506, + "grad_norm": 64.90064961374273, + "learning_rate": 1.9706935176716316e-05, + "loss": 0.1466, + "step": 809 + }, + { + "epoch": 1.051948051948052, + "grad_norm": 21.208390446911963, + "learning_rate": 1.9705923486489572e-05, + "loss": 0.0824, + "step": 810 + }, + { + "epoch": 1.0532467532467533, + "grad_norm": 104.5582727053057, + "learning_rate": 1.970491007909919e-05, + "loss": 0.1692, + "step": 811 + }, + { + "epoch": 1.0545454545454545, + "grad_norm": 27.27290933299614, + "learning_rate": 1.9703894954724462e-05, + "loss": 0.1153, + "step": 812 + }, + { + "epoch": 1.0558441558441558, + "grad_norm": 99.56339557345237, + "learning_rate": 1.970287811354498e-05, + "loss": 0.1476, + "step": 813 + }, + { + "epoch": 1.0571428571428572, + "grad_norm": 7.913137173192939, + "learning_rate": 1.9701859555740647e-05, + "loss": 0.0928, + "step": 814 + }, + { + "epoch": 1.0584415584415585, + "grad_norm": 99.78655925488378, + "learning_rate": 1.970083928149166e-05, + "loss": 0.1599, + "step": 815 + }, + { + "epoch": 1.0597402597402596, + "grad_norm": 27.924022109906392, + "learning_rate": 1.9699817290978532e-05, + "loss": 0.0937, + "step": 816 + }, + { + "epoch": 1.061038961038961, + "grad_norm": 83.66618761639704, + "learning_rate": 1.9698793584382067e-05, + "loss": 0.1613, + "step": 817 + }, + { + "epoch": 1.0623376623376624, + "grad_norm": 25.93280254658599, + "learning_rate": 1.969776816188338e-05, + "loss": 0.1365, + "step": 818 + }, + { + "epoch": 1.0636363636363637, + "grad_norm": 39.6858078553203, + "learning_rate": 1.9696741023663884e-05, + "loss": 0.1134, + "step": 819 + }, + { + "epoch": 1.0649350649350648, + "grad_norm": 13.257754887788789, + "learning_rate": 1.9695712169905307e-05, + "loss": 0.1189, + "step": 820 + }, + { + "epoch": 1.0662337662337662, + "grad_norm": 16.72241143182184, + "learning_rate": 1.9694681600789673e-05, + "loss": 0.1163, + "step": 821 + }, + { + "epoch": 1.0675324675324676, + "grad_norm": 23.43349753951216, + "learning_rate": 1.96936493164993e-05, + "loss": 0.1233, + "step": 822 + }, + { + "epoch": 1.068831168831169, + "grad_norm": 6.931953942367784, + "learning_rate": 1.9692615317216823e-05, + "loss": 0.0995, + "step": 823 + }, + { + "epoch": 1.07012987012987, + "grad_norm": 11.332849414764427, + "learning_rate": 1.969157960312518e-05, + "loss": 0.1081, + "step": 824 + }, + { + "epoch": 1.0714285714285714, + "grad_norm": 12.530131807383082, + "learning_rate": 1.969054217440761e-05, + "loss": 0.0879, + "step": 825 + }, + { + "epoch": 1.0727272727272728, + "grad_norm": 11.89753713336959, + "learning_rate": 1.968950303124765e-05, + "loss": 0.1326, + "step": 826 + }, + { + "epoch": 1.074025974025974, + "grad_norm": 15.867478376786535, + "learning_rate": 1.968846217382914e-05, + "loss": 0.1159, + "step": 827 + }, + { + "epoch": 1.0753246753246752, + "grad_norm": 11.866543176191175, + "learning_rate": 1.9687419602336234e-05, + "loss": 0.1, + "step": 828 + }, + { + "epoch": 1.0766233766233766, + "grad_norm": 32.00870900754855, + "learning_rate": 1.9686375316953383e-05, + "loss": 0.1316, + "step": 829 + }, + { + "epoch": 1.077922077922078, + "grad_norm": 8.130769841202223, + "learning_rate": 1.9685329317865338e-05, + "loss": 0.0963, + "step": 830 + }, + { + "epoch": 1.0792207792207793, + "grad_norm": 36.99886118372149, + "learning_rate": 1.968428160525716e-05, + "loss": 0.1276, + "step": 831 + }, + { + "epoch": 1.0805194805194804, + "grad_norm": 7.301863858398021, + "learning_rate": 1.96832321793142e-05, + "loss": 0.1012, + "step": 832 + }, + { + "epoch": 1.0818181818181818, + "grad_norm": 25.642891780918934, + "learning_rate": 1.9682181040222136e-05, + "loss": 0.1045, + "step": 833 + }, + { + "epoch": 1.0831168831168831, + "grad_norm": 6.209266427474272, + "learning_rate": 1.9681128188166923e-05, + "loss": 0.0934, + "step": 834 + }, + { + "epoch": 1.0844155844155845, + "grad_norm": 18.939784947526682, + "learning_rate": 1.9680073623334833e-05, + "loss": 0.0984, + "step": 835 + }, + { + "epoch": 1.0857142857142856, + "grad_norm": 6.540626166029594, + "learning_rate": 1.9679017345912446e-05, + "loss": 0.0949, + "step": 836 + }, + { + "epoch": 1.087012987012987, + "grad_norm": 9.333624800947732, + "learning_rate": 1.967795935608663e-05, + "loss": 0.1476, + "step": 837 + }, + { + "epoch": 1.0883116883116883, + "grad_norm": 14.373453308879766, + "learning_rate": 1.9676899654044563e-05, + "loss": 0.1, + "step": 838 + }, + { + "epoch": 1.0896103896103897, + "grad_norm": 15.058063062884141, + "learning_rate": 1.9675838239973734e-05, + "loss": 0.1417, + "step": 839 + }, + { + "epoch": 1.0909090909090908, + "grad_norm": 7.787092400109172, + "learning_rate": 1.967477511406192e-05, + "loss": 0.1266, + "step": 840 + }, + { + "epoch": 1.0922077922077922, + "grad_norm": 16.785905764967087, + "learning_rate": 1.9673710276497207e-05, + "loss": 0.0953, + "step": 841 + }, + { + "epoch": 1.0935064935064935, + "grad_norm": 43.91376597106909, + "learning_rate": 1.9672643727467995e-05, + "loss": 0.1108, + "step": 842 + }, + { + "epoch": 1.094805194805195, + "grad_norm": 14.184725723183277, + "learning_rate": 1.967157546716297e-05, + "loss": 0.1776, + "step": 843 + }, + { + "epoch": 1.096103896103896, + "grad_norm": 15.057936114629833, + "learning_rate": 1.967050549577113e-05, + "loss": 0.1187, + "step": 844 + }, + { + "epoch": 1.0974025974025974, + "grad_norm": 46.31241231327463, + "learning_rate": 1.9669433813481774e-05, + "loss": 0.1039, + "step": 845 + }, + { + "epoch": 1.0987012987012987, + "grad_norm": 12.584371336854534, + "learning_rate": 1.9668360420484496e-05, + "loss": 0.1183, + "step": 846 + }, + { + "epoch": 1.1, + "grad_norm": 45.98767368103188, + "learning_rate": 1.966728531696921e-05, + "loss": 0.1044, + "step": 847 + }, + { + "epoch": 1.1012987012987012, + "grad_norm": 36.94520421099805, + "learning_rate": 1.9666208503126115e-05, + "loss": 0.1404, + "step": 848 + }, + { + "epoch": 1.1025974025974026, + "grad_norm": 86.07997755014077, + "learning_rate": 1.966512997914572e-05, + "loss": 0.1526, + "step": 849 + }, + { + "epoch": 1.103896103896104, + "grad_norm": 14.21524457933759, + "learning_rate": 1.9664049745218846e-05, + "loss": 0.094, + "step": 850 + }, + { + "epoch": 1.1051948051948053, + "grad_norm": 66.37218249037929, + "learning_rate": 1.9662967801536593e-05, + "loss": 0.1483, + "step": 851 + }, + { + "epoch": 1.1064935064935064, + "grad_norm": 33.90526190920279, + "learning_rate": 1.9661884148290392e-05, + "loss": 0.0888, + "step": 852 + }, + { + "epoch": 1.1077922077922078, + "grad_norm": 57.232316085444424, + "learning_rate": 1.9660798785671947e-05, + "loss": 0.1324, + "step": 853 + }, + { + "epoch": 1.1090909090909091, + "grad_norm": 49.388099261107, + "learning_rate": 1.9659711713873294e-05, + "loss": 0.0878, + "step": 854 + }, + { + "epoch": 1.1103896103896105, + "grad_norm": 39.9515545888027, + "learning_rate": 1.9658622933086747e-05, + "loss": 0.1344, + "step": 855 + }, + { + "epoch": 1.1116883116883116, + "grad_norm": 36.12322369690276, + "learning_rate": 1.9657532443504934e-05, + "loss": 0.2006, + "step": 856 + }, + { + "epoch": 1.112987012987013, + "grad_norm": 26.78891017491385, + "learning_rate": 1.9656440245320786e-05, + "loss": 0.11, + "step": 857 + }, + { + "epoch": 1.1142857142857143, + "grad_norm": 56.62717112211357, + "learning_rate": 1.965534633872753e-05, + "loss": 0.121, + "step": 858 + }, + { + "epoch": 1.1155844155844157, + "grad_norm": 22.609997059510295, + "learning_rate": 1.9654250723918706e-05, + "loss": 0.0741, + "step": 859 + }, + { + "epoch": 1.1168831168831168, + "grad_norm": 42.49670378966959, + "learning_rate": 1.965315340108814e-05, + "loss": 0.1059, + "step": 860 + }, + { + "epoch": 1.1181818181818182, + "grad_norm": 25.62984766104483, + "learning_rate": 1.9652054370429977e-05, + "loss": 0.1234, + "step": 861 + }, + { + "epoch": 1.1194805194805195, + "grad_norm": 35.64254446918691, + "learning_rate": 1.9650953632138652e-05, + "loss": 0.1095, + "step": 862 + }, + { + "epoch": 1.1207792207792209, + "grad_norm": 41.00584649789639, + "learning_rate": 1.964985118640891e-05, + "loss": 0.1232, + "step": 863 + }, + { + "epoch": 1.122077922077922, + "grad_norm": 29.908862440649834, + "learning_rate": 1.964874703343579e-05, + "loss": 0.1255, + "step": 864 + }, + { + "epoch": 1.1233766233766234, + "grad_norm": 42.47083994972064, + "learning_rate": 1.9647641173414645e-05, + "loss": 0.0929, + "step": 865 + }, + { + "epoch": 1.1246753246753247, + "grad_norm": 9.245760732804978, + "learning_rate": 1.9646533606541116e-05, + "loss": 0.1158, + "step": 866 + }, + { + "epoch": 1.1259740259740258, + "grad_norm": 49.642810008798605, + "learning_rate": 1.9645424333011155e-05, + "loss": 0.1426, + "step": 867 + }, + { + "epoch": 1.1272727272727272, + "grad_norm": 11.827707877170777, + "learning_rate": 1.9644313353021015e-05, + "loss": 0.0992, + "step": 868 + }, + { + "epoch": 1.1285714285714286, + "grad_norm": 43.1991264193531, + "learning_rate": 1.964320066676725e-05, + "loss": 0.0947, + "step": 869 + }, + { + "epoch": 1.12987012987013, + "grad_norm": 9.81846590705676, + "learning_rate": 1.9642086274446713e-05, + "loss": 0.0793, + "step": 870 + }, + { + "epoch": 1.1311688311688313, + "grad_norm": 33.23055564573471, + "learning_rate": 1.964097017625656e-05, + "loss": 0.1354, + "step": 871 + }, + { + "epoch": 1.1324675324675324, + "grad_norm": 16.70393343620691, + "learning_rate": 1.9639852372394258e-05, + "loss": 0.1084, + "step": 872 + }, + { + "epoch": 1.1337662337662338, + "grad_norm": 13.132049463645384, + "learning_rate": 1.9638732863057557e-05, + "loss": 0.1231, + "step": 873 + }, + { + "epoch": 1.135064935064935, + "grad_norm": 13.17997496101103, + "learning_rate": 1.9637611648444528e-05, + "loss": 0.1172, + "step": 874 + }, + { + "epoch": 1.1363636363636362, + "grad_norm": 12.934845435461575, + "learning_rate": 1.9636488728753532e-05, + "loss": 0.1219, + "step": 875 + }, + { + "epoch": 1.1376623376623376, + "grad_norm": 26.093664097262693, + "learning_rate": 1.9635364104183236e-05, + "loss": 0.1202, + "step": 876 + }, + { + "epoch": 1.138961038961039, + "grad_norm": 20.302454308917767, + "learning_rate": 1.963423777493261e-05, + "loss": 0.1273, + "step": 877 + }, + { + "epoch": 1.1402597402597403, + "grad_norm": 6.433720596656636, + "learning_rate": 1.9633109741200915e-05, + "loss": 0.1128, + "step": 878 + }, + { + "epoch": 1.1415584415584417, + "grad_norm": 36.7521817571928, + "learning_rate": 1.963198000318773e-05, + "loss": 0.1632, + "step": 879 + }, + { + "epoch": 1.1428571428571428, + "grad_norm": 34.56249125817019, + "learning_rate": 1.9630848561092923e-05, + "loss": 0.0824, + "step": 880 + }, + { + "epoch": 1.1441558441558441, + "grad_norm": 28.6533655269267, + "learning_rate": 1.962971541511667e-05, + "loss": 0.1201, + "step": 881 + }, + { + "epoch": 1.1454545454545455, + "grad_norm": 19.76114605683019, + "learning_rate": 1.962858056545945e-05, + "loss": 0.1279, + "step": 882 + }, + { + "epoch": 1.1467532467532466, + "grad_norm": 47.66802764828531, + "learning_rate": 1.9627444012322027e-05, + "loss": 0.1194, + "step": 883 + }, + { + "epoch": 1.148051948051948, + "grad_norm": 13.077589069386875, + "learning_rate": 1.962630575590549e-05, + "loss": 0.0981, + "step": 884 + }, + { + "epoch": 1.1493506493506493, + "grad_norm": 57.540055221509334, + "learning_rate": 1.9625165796411213e-05, + "loss": 0.1621, + "step": 885 + }, + { + "epoch": 1.1506493506493507, + "grad_norm": 10.47453342170882, + "learning_rate": 1.9624024134040883e-05, + "loss": 0.1321, + "step": 886 + }, + { + "epoch": 1.151948051948052, + "grad_norm": 79.47717866127486, + "learning_rate": 1.9622880768996476e-05, + "loss": 0.1621, + "step": 887 + }, + { + "epoch": 1.1532467532467532, + "grad_norm": 29.058645849190324, + "learning_rate": 1.962173570148028e-05, + "loss": 0.0995, + "step": 888 + }, + { + "epoch": 1.1545454545454545, + "grad_norm": 49.546558854631876, + "learning_rate": 1.9620588931694873e-05, + "loss": 0.1182, + "step": 889 + }, + { + "epoch": 1.155844155844156, + "grad_norm": 11.295982360124958, + "learning_rate": 1.9619440459843145e-05, + "loss": 0.1216, + "step": 890 + }, + { + "epoch": 1.157142857142857, + "grad_norm": 35.734700894372395, + "learning_rate": 1.961829028612828e-05, + "loss": 0.1024, + "step": 891 + }, + { + "epoch": 1.1584415584415584, + "grad_norm": 33.137656307768104, + "learning_rate": 1.961713841075377e-05, + "loss": 0.1452, + "step": 892 + }, + { + "epoch": 1.1597402597402597, + "grad_norm": 21.73949522826576, + "learning_rate": 1.96159848339234e-05, + "loss": 0.1263, + "step": 893 + }, + { + "epoch": 1.161038961038961, + "grad_norm": 25.230618779435762, + "learning_rate": 1.9614829555841258e-05, + "loss": 0.1327, + "step": 894 + }, + { + "epoch": 1.1623376623376624, + "grad_norm": 17.53111209595618, + "learning_rate": 1.9613672576711745e-05, + "loss": 0.1445, + "step": 895 + }, + { + "epoch": 1.1636363636363636, + "grad_norm": 16.54761503084966, + "learning_rate": 1.961251389673954e-05, + "loss": 0.0863, + "step": 896 + }, + { + "epoch": 1.164935064935065, + "grad_norm": 13.221867900321751, + "learning_rate": 1.9611353516129645e-05, + "loss": 0.13, + "step": 897 + }, + { + "epoch": 1.1662337662337663, + "grad_norm": 18.34788151993121, + "learning_rate": 1.9610191435087343e-05, + "loss": 0.125, + "step": 898 + }, + { + "epoch": 1.1675324675324674, + "grad_norm": 47.97309978124735, + "learning_rate": 1.960902765381824e-05, + "loss": 0.1093, + "step": 899 + }, + { + "epoch": 1.1688311688311688, + "grad_norm": 32.883666706694996, + "learning_rate": 1.9607862172528226e-05, + "loss": 0.0901, + "step": 900 + }, + { + "epoch": 1.1701298701298701, + "grad_norm": 52.54759200093756, + "learning_rate": 1.9606694991423496e-05, + "loss": 0.1034, + "step": 901 + }, + { + "epoch": 1.1714285714285715, + "grad_norm": 26.690000191020367, + "learning_rate": 1.960552611071055e-05, + "loss": 0.1225, + "step": 902 + }, + { + "epoch": 1.1727272727272728, + "grad_norm": 69.20101320621741, + "learning_rate": 1.9604355530596182e-05, + "loss": 0.1277, + "step": 903 + }, + { + "epoch": 1.174025974025974, + "grad_norm": 36.22392273963283, + "learning_rate": 1.9603183251287488e-05, + "loss": 0.1422, + "step": 904 + }, + { + "epoch": 1.1753246753246753, + "grad_norm": 56.57578937368004, + "learning_rate": 1.9602009272991876e-05, + "loss": 0.1016, + "step": 905 + }, + { + "epoch": 1.1766233766233767, + "grad_norm": 17.981917623944636, + "learning_rate": 1.9600833595917037e-05, + "loss": 0.1427, + "step": 906 + }, + { + "epoch": 1.1779220779220778, + "grad_norm": 61.076151023322346, + "learning_rate": 1.959965622027097e-05, + "loss": 0.1407, + "step": 907 + }, + { + "epoch": 1.1792207792207792, + "grad_norm": 6.934550577672918, + "learning_rate": 1.9598477146261983e-05, + "loss": 0.1327, + "step": 908 + }, + { + "epoch": 1.1805194805194805, + "grad_norm": 47.651630990135345, + "learning_rate": 1.959729637409867e-05, + "loss": 0.1016, + "step": 909 + }, + { + "epoch": 1.1818181818181819, + "grad_norm": 14.535810410700133, + "learning_rate": 1.9596113903989932e-05, + "loss": 0.1606, + "step": 910 + }, + { + "epoch": 1.1831168831168832, + "grad_norm": 12.039234030823627, + "learning_rate": 1.9594929736144978e-05, + "loss": 0.1085, + "step": 911 + }, + { + "epoch": 1.1844155844155844, + "grad_norm": 35.12969169085708, + "learning_rate": 1.9593743870773298e-05, + "loss": 0.1014, + "step": 912 + }, + { + "epoch": 1.1857142857142857, + "grad_norm": 27.384964177400118, + "learning_rate": 1.9592556308084706e-05, + "loss": 0.1191, + "step": 913 + }, + { + "epoch": 1.187012987012987, + "grad_norm": 10.210169012754424, + "learning_rate": 1.9591367048289297e-05, + "loss": 0.1315, + "step": 914 + }, + { + "epoch": 1.1883116883116882, + "grad_norm": 24.53860595866022, + "learning_rate": 1.959017609159748e-05, + "loss": 0.1176, + "step": 915 + }, + { + "epoch": 1.1896103896103896, + "grad_norm": 11.192335863533366, + "learning_rate": 1.958898343821995e-05, + "loss": 0.1322, + "step": 916 + }, + { + "epoch": 1.190909090909091, + "grad_norm": 26.880355304350832, + "learning_rate": 1.9587789088367717e-05, + "loss": 0.1071, + "step": 917 + }, + { + "epoch": 1.1922077922077923, + "grad_norm": 17.675371233402437, + "learning_rate": 1.958659304225208e-05, + "loss": 0.1567, + "step": 918 + }, + { + "epoch": 1.1935064935064936, + "grad_norm": 10.354681139419124, + "learning_rate": 1.9585395300084647e-05, + "loss": 0.0859, + "step": 919 + }, + { + "epoch": 1.1948051948051948, + "grad_norm": 11.547519806228944, + "learning_rate": 1.958419586207732e-05, + "loss": 0.0879, + "step": 920 + }, + { + "epoch": 1.1961038961038961, + "grad_norm": 16.200208751680602, + "learning_rate": 1.95829947284423e-05, + "loss": 0.121, + "step": 921 + }, + { + "epoch": 1.1974025974025975, + "grad_norm": 6.4673821189776, + "learning_rate": 1.958179189939209e-05, + "loss": 0.0976, + "step": 922 + }, + { + "epoch": 1.1987012987012986, + "grad_norm": 19.800298617741955, + "learning_rate": 1.95805873751395e-05, + "loss": 0.1202, + "step": 923 + }, + { + "epoch": 1.2, + "grad_norm": 11.215510962952697, + "learning_rate": 1.9579381155897622e-05, + "loss": 0.1248, + "step": 924 + }, + { + "epoch": 1.2012987012987013, + "grad_norm": 18.191252993731403, + "learning_rate": 1.957817324187987e-05, + "loss": 0.1167, + "step": 925 + }, + { + "epoch": 1.2025974025974027, + "grad_norm": 24.072720116228155, + "learning_rate": 1.9576963633299946e-05, + "loss": 0.095, + "step": 926 + }, + { + "epoch": 1.203896103896104, + "grad_norm": 12.987047282587685, + "learning_rate": 1.957575233037185e-05, + "loss": 0.1111, + "step": 927 + }, + { + "epoch": 1.2051948051948052, + "grad_norm": 28.055987176030293, + "learning_rate": 1.9574539333309884e-05, + "loss": 0.1409, + "step": 928 + }, + { + "epoch": 1.2064935064935065, + "grad_norm": 6.279854187496498, + "learning_rate": 1.957332464232865e-05, + "loss": 0.0828, + "step": 929 + }, + { + "epoch": 1.2077922077922079, + "grad_norm": 18.824640508226615, + "learning_rate": 1.957210825764305e-05, + "loss": 0.1239, + "step": 930 + }, + { + "epoch": 1.209090909090909, + "grad_norm": 6.44927568007333, + "learning_rate": 1.957089017946829e-05, + "loss": 0.0641, + "step": 931 + }, + { + "epoch": 1.2103896103896103, + "grad_norm": 17.13862875895041, + "learning_rate": 1.9569670408019865e-05, + "loss": 0.108, + "step": 932 + }, + { + "epoch": 1.2116883116883117, + "grad_norm": 9.0137610415835, + "learning_rate": 1.9568448943513583e-05, + "loss": 0.1154, + "step": 933 + }, + { + "epoch": 1.212987012987013, + "grad_norm": 21.758300921989534, + "learning_rate": 1.956722578616554e-05, + "loss": 0.107, + "step": 934 + }, + { + "epoch": 1.2142857142857142, + "grad_norm": 13.645423997662675, + "learning_rate": 1.9566000936192132e-05, + "loss": 0.0916, + "step": 935 + }, + { + "epoch": 1.2155844155844155, + "grad_norm": 13.657041142629756, + "learning_rate": 1.9564774393810068e-05, + "loss": 0.1264, + "step": 936 + }, + { + "epoch": 1.216883116883117, + "grad_norm": 11.748335938928006, + "learning_rate": 1.9563546159236342e-05, + "loss": 0.1132, + "step": 937 + }, + { + "epoch": 1.2181818181818183, + "grad_norm": 28.89419935057352, + "learning_rate": 1.9562316232688253e-05, + "loss": 0.1129, + "step": 938 + }, + { + "epoch": 1.2194805194805194, + "grad_norm": 15.497513594068266, + "learning_rate": 1.9561084614383396e-05, + "loss": 0.0921, + "step": 939 + }, + { + "epoch": 1.2207792207792207, + "grad_norm": 9.90637203782535, + "learning_rate": 1.955985130453967e-05, + "loss": 0.1019, + "step": 940 + }, + { + "epoch": 1.222077922077922, + "grad_norm": 8.706814087243991, + "learning_rate": 1.9558616303375272e-05, + "loss": 0.0961, + "step": 941 + }, + { + "epoch": 1.2233766233766235, + "grad_norm": 9.43609271329282, + "learning_rate": 1.9557379611108694e-05, + "loss": 0.1223, + "step": 942 + }, + { + "epoch": 1.2246753246753246, + "grad_norm": 10.767594930232084, + "learning_rate": 1.9556141227958734e-05, + "loss": 0.145, + "step": 943 + }, + { + "epoch": 1.225974025974026, + "grad_norm": 32.16877020117008, + "learning_rate": 1.955490115414448e-05, + "loss": 0.1147, + "step": 944 + }, + { + "epoch": 1.2272727272727273, + "grad_norm": 25.152903434005204, + "learning_rate": 1.9553659389885335e-05, + "loss": 0.1192, + "step": 945 + }, + { + "epoch": 1.2285714285714286, + "grad_norm": 5.283669008782753, + "learning_rate": 1.955241593540098e-05, + "loss": 0.1093, + "step": 946 + }, + { + "epoch": 1.2298701298701298, + "grad_norm": 9.753187931457916, + "learning_rate": 1.9551170790911415e-05, + "loss": 0.1197, + "step": 947 + }, + { + "epoch": 1.2311688311688311, + "grad_norm": 25.281897989453565, + "learning_rate": 1.9549923956636924e-05, + "loss": 0.106, + "step": 948 + }, + { + "epoch": 1.2324675324675325, + "grad_norm": 21.07972867893375, + "learning_rate": 1.9548675432798098e-05, + "loss": 0.0974, + "step": 949 + }, + { + "epoch": 1.2337662337662338, + "grad_norm": 12.583704649974251, + "learning_rate": 1.9547425219615826e-05, + "loss": 0.0896, + "step": 950 + }, + { + "epoch": 1.235064935064935, + "grad_norm": 16.000474248533862, + "learning_rate": 1.954617331731129e-05, + "loss": 0.1258, + "step": 951 + }, + { + "epoch": 1.2363636363636363, + "grad_norm": 11.633628589376181, + "learning_rate": 1.954491972610598e-05, + "loss": 0.1078, + "step": 952 + }, + { + "epoch": 1.2376623376623377, + "grad_norm": 18.702812082675084, + "learning_rate": 1.954366444622168e-05, + "loss": 0.1008, + "step": 953 + }, + { + "epoch": 1.238961038961039, + "grad_norm": 10.666558856026562, + "learning_rate": 1.9542407477880468e-05, + "loss": 0.1033, + "step": 954 + }, + { + "epoch": 1.2402597402597402, + "grad_norm": 11.498786655952285, + "learning_rate": 1.954114882130473e-05, + "loss": 0.1019, + "step": 955 + }, + { + "epoch": 1.2415584415584415, + "grad_norm": 30.04034397512009, + "learning_rate": 1.953988847671715e-05, + "loss": 0.1189, + "step": 956 + }, + { + "epoch": 1.2428571428571429, + "grad_norm": 16.97832378500618, + "learning_rate": 1.9538626444340698e-05, + "loss": 0.1304, + "step": 957 + }, + { + "epoch": 1.2441558441558442, + "grad_norm": 55.047317852642514, + "learning_rate": 1.953736272439866e-05, + "loss": 0.1606, + "step": 958 + }, + { + "epoch": 1.2454545454545454, + "grad_norm": 11.5487443735563, + "learning_rate": 1.953609731711461e-05, + "loss": 0.0882, + "step": 959 + }, + { + "epoch": 1.2467532467532467, + "grad_norm": 60.2787740597022, + "learning_rate": 1.953483022271242e-05, + "loss": 0.1114, + "step": 960 + }, + { + "epoch": 1.248051948051948, + "grad_norm": 11.026431948555063, + "learning_rate": 1.9533561441416264e-05, + "loss": 0.1384, + "step": 961 + }, + { + "epoch": 1.2493506493506494, + "grad_norm": 46.25080014033935, + "learning_rate": 1.9532290973450618e-05, + "loss": 0.1138, + "step": 962 + }, + { + "epoch": 1.2506493506493506, + "grad_norm": 15.669782022067288, + "learning_rate": 1.9531018819040246e-05, + "loss": 0.1158, + "step": 963 + }, + { + "epoch": 1.251948051948052, + "grad_norm": 20.60683956740148, + "learning_rate": 1.9529744978410218e-05, + "loss": 0.1142, + "step": 964 + }, + { + "epoch": 1.2532467532467533, + "grad_norm": 11.671685116343081, + "learning_rate": 1.9528469451785907e-05, + "loss": 0.1071, + "step": 965 + }, + { + "epoch": 1.2545454545454544, + "grad_norm": 18.222944856072303, + "learning_rate": 1.9527192239392968e-05, + "loss": 0.1069, + "step": 966 + }, + { + "epoch": 1.2558441558441558, + "grad_norm": 12.600081028161416, + "learning_rate": 1.952591334145737e-05, + "loss": 0.1128, + "step": 967 + }, + { + "epoch": 1.2571428571428571, + "grad_norm": 7.531503290879485, + "learning_rate": 1.9524632758205378e-05, + "loss": 0.091, + "step": 968 + }, + { + "epoch": 1.2584415584415585, + "grad_norm": 17.42926138408269, + "learning_rate": 1.9523350489863545e-05, + "loss": 0.1179, + "step": 969 + }, + { + "epoch": 1.2597402597402598, + "grad_norm": 17.98654779217015, + "learning_rate": 1.952206653665873e-05, + "loss": 0.1534, + "step": 970 + }, + { + "epoch": 1.261038961038961, + "grad_norm": 19.55752263924698, + "learning_rate": 1.9520780898818092e-05, + "loss": 0.1046, + "step": 971 + }, + { + "epoch": 1.2623376623376623, + "grad_norm": 44.68773937888494, + "learning_rate": 1.9519493576569085e-05, + "loss": 0.1331, + "step": 972 + }, + { + "epoch": 1.2636363636363637, + "grad_norm": 51.306363315996585, + "learning_rate": 1.9518204570139456e-05, + "loss": 0.1277, + "step": 973 + }, + { + "epoch": 1.2649350649350648, + "grad_norm": 19.18412579893739, + "learning_rate": 1.951691387975726e-05, + "loss": 0.0827, + "step": 974 + }, + { + "epoch": 1.2662337662337662, + "grad_norm": 24.37709297355283, + "learning_rate": 1.951562150565084e-05, + "loss": 0.1182, + "step": 975 + }, + { + "epoch": 1.2675324675324675, + "grad_norm": 24.542887310930055, + "learning_rate": 1.951432744804885e-05, + "loss": 0.1174, + "step": 976 + }, + { + "epoch": 1.2688311688311689, + "grad_norm": 18.288093264705836, + "learning_rate": 1.9513031707180224e-05, + "loss": 0.0884, + "step": 977 + }, + { + "epoch": 1.2701298701298702, + "grad_norm": 9.59503511932832, + "learning_rate": 1.9511734283274212e-05, + "loss": 0.1074, + "step": 978 + }, + { + "epoch": 1.2714285714285714, + "grad_norm": 11.558664316829088, + "learning_rate": 1.9510435176560346e-05, + "loss": 0.0861, + "step": 979 + }, + { + "epoch": 1.2727272727272727, + "grad_norm": 8.44315629575817, + "learning_rate": 1.9509134387268463e-05, + "loss": 0.0953, + "step": 980 + }, + { + "epoch": 1.274025974025974, + "grad_norm": 18.461167189910473, + "learning_rate": 1.9507831915628706e-05, + "loss": 0.0776, + "step": 981 + }, + { + "epoch": 1.2753246753246752, + "grad_norm": 21.07193146491766, + "learning_rate": 1.9506527761871497e-05, + "loss": 0.1114, + "step": 982 + }, + { + "epoch": 1.2766233766233765, + "grad_norm": 8.138005175465272, + "learning_rate": 1.950522192622757e-05, + "loss": 0.0818, + "step": 983 + }, + { + "epoch": 1.277922077922078, + "grad_norm": 6.682746898039056, + "learning_rate": 1.9503914408927954e-05, + "loss": 0.0905, + "step": 984 + }, + { + "epoch": 1.2792207792207793, + "grad_norm": 12.712105645282618, + "learning_rate": 1.9502605210203974e-05, + "loss": 0.1171, + "step": 985 + }, + { + "epoch": 1.2805194805194806, + "grad_norm": 8.316712822447771, + "learning_rate": 1.9501294330287246e-05, + "loss": 0.1063, + "step": 986 + }, + { + "epoch": 1.2818181818181817, + "grad_norm": 9.722104218576167, + "learning_rate": 1.9499981769409697e-05, + "loss": 0.0658, + "step": 987 + }, + { + "epoch": 1.283116883116883, + "grad_norm": 23.306391241446622, + "learning_rate": 1.949866752780354e-05, + "loss": 0.1141, + "step": 988 + }, + { + "epoch": 1.2844155844155845, + "grad_norm": 23.654133931127447, + "learning_rate": 1.9497351605701294e-05, + "loss": 0.1241, + "step": 989 + }, + { + "epoch": 1.2857142857142856, + "grad_norm": 10.951503033347755, + "learning_rate": 1.9496034003335766e-05, + "loss": 0.1328, + "step": 990 + }, + { + "epoch": 1.287012987012987, + "grad_norm": 4.927434211664772, + "learning_rate": 1.9494714720940065e-05, + "loss": 0.0803, + "step": 991 + }, + { + "epoch": 1.2883116883116883, + "grad_norm": 21.937696462874435, + "learning_rate": 1.94933937587476e-05, + "loss": 0.1218, + "step": 992 + }, + { + "epoch": 1.2896103896103897, + "grad_norm": 15.53830776733688, + "learning_rate": 1.9492071116992072e-05, + "loss": 0.0737, + "step": 993 + }, + { + "epoch": 1.290909090909091, + "grad_norm": 6.906561044716053, + "learning_rate": 1.9490746795907487e-05, + "loss": 0.073, + "step": 994 + }, + { + "epoch": 1.2922077922077921, + "grad_norm": 13.038938466695013, + "learning_rate": 1.9489420795728136e-05, + "loss": 0.143, + "step": 995 + }, + { + "epoch": 1.2935064935064935, + "grad_norm": 11.818667656541981, + "learning_rate": 1.948809311668862e-05, + "loss": 0.1253, + "step": 996 + }, + { + "epoch": 1.2948051948051948, + "grad_norm": 25.644777953213037, + "learning_rate": 1.9486763759023825e-05, + "loss": 0.0958, + "step": 997 + }, + { + "epoch": 1.296103896103896, + "grad_norm": 17.254623628098738, + "learning_rate": 1.9485432722968942e-05, + "loss": 0.0757, + "step": 998 + }, + { + "epoch": 1.2974025974025973, + "grad_norm": 18.32815103822196, + "learning_rate": 1.9484100008759458e-05, + "loss": 0.0972, + "step": 999 + }, + { + "epoch": 1.2987012987012987, + "grad_norm": 28.74775654917307, + "learning_rate": 1.9482765616631156e-05, + "loss": 0.1089, + "step": 1000 + }, + { + "epoch": 1.3, + "grad_norm": 12.034498722662727, + "learning_rate": 1.9481429546820112e-05, + "loss": 0.1428, + "step": 1001 + }, + { + "epoch": 1.3012987012987014, + "grad_norm": 21.17845274796235, + "learning_rate": 1.9480091799562706e-05, + "loss": 0.0996, + "step": 1002 + }, + { + "epoch": 1.3025974025974025, + "grad_norm": 12.092309150411323, + "learning_rate": 1.9478752375095607e-05, + "loss": 0.1505, + "step": 1003 + }, + { + "epoch": 1.3038961038961039, + "grad_norm": 35.959650910859985, + "learning_rate": 1.9477411273655794e-05, + "loss": 0.0878, + "step": 1004 + }, + { + "epoch": 1.3051948051948052, + "grad_norm": 32.864539288658115, + "learning_rate": 1.9476068495480522e-05, + "loss": 0.109, + "step": 1005 + }, + { + "epoch": 1.3064935064935064, + "grad_norm": 16.918030529697422, + "learning_rate": 1.947472404080736e-05, + "loss": 0.1315, + "step": 1006 + }, + { + "epoch": 1.3077922077922077, + "grad_norm": 29.28732642713567, + "learning_rate": 1.947337790987417e-05, + "loss": 0.0858, + "step": 1007 + }, + { + "epoch": 1.309090909090909, + "grad_norm": 9.81541524945221, + "learning_rate": 1.9472030102919102e-05, + "loss": 0.0789, + "step": 1008 + }, + { + "epoch": 1.3103896103896104, + "grad_norm": 18.748594451460463, + "learning_rate": 1.9470680620180612e-05, + "loss": 0.1226, + "step": 1009 + }, + { + "epoch": 1.3116883116883118, + "grad_norm": 16.3784655029746, + "learning_rate": 1.9469329461897454e-05, + "loss": 0.0904, + "step": 1010 + }, + { + "epoch": 1.312987012987013, + "grad_norm": 42.758664107038776, + "learning_rate": 1.9467976628308666e-05, + "loss": 0.1415, + "step": 1011 + }, + { + "epoch": 1.3142857142857143, + "grad_norm": 8.514123378209964, + "learning_rate": 1.946662211965359e-05, + "loss": 0.0963, + "step": 1012 + }, + { + "epoch": 1.3155844155844156, + "grad_norm": 32.963092997942475, + "learning_rate": 1.9465265936171875e-05, + "loss": 0.1165, + "step": 1013 + }, + { + "epoch": 1.3168831168831168, + "grad_norm": 16.93755620799438, + "learning_rate": 1.9463908078103445e-05, + "loss": 0.0989, + "step": 1014 + }, + { + "epoch": 1.3181818181818181, + "grad_norm": 28.52960973437555, + "learning_rate": 1.9462548545688532e-05, + "loss": 0.1398, + "step": 1015 + }, + { + "epoch": 1.3194805194805195, + "grad_norm": 28.288392436514425, + "learning_rate": 1.9461187339167673e-05, + "loss": 0.1181, + "step": 1016 + }, + { + "epoch": 1.3207792207792208, + "grad_norm": 14.87516613253982, + "learning_rate": 1.945982445878168e-05, + "loss": 0.09, + "step": 1017 + }, + { + "epoch": 1.3220779220779222, + "grad_norm": 20.24424912400601, + "learning_rate": 1.9458459904771678e-05, + "loss": 0.1469, + "step": 1018 + }, + { + "epoch": 1.3233766233766233, + "grad_norm": 39.93100760161077, + "learning_rate": 1.9457093677379083e-05, + "loss": 0.0817, + "step": 1019 + }, + { + "epoch": 1.3246753246753247, + "grad_norm": 43.34119025519825, + "learning_rate": 1.9455725776845604e-05, + "loss": 0.1487, + "step": 1020 + }, + { + "epoch": 1.325974025974026, + "grad_norm": 34.757504627032006, + "learning_rate": 1.9454356203413252e-05, + "loss": 0.1126, + "step": 1021 + }, + { + "epoch": 1.3272727272727272, + "grad_norm": 34.188324525452096, + "learning_rate": 1.945298495732433e-05, + "loss": 0.1184, + "step": 1022 + }, + { + "epoch": 1.3285714285714285, + "grad_norm": 29.88443814634753, + "learning_rate": 1.9451612038821436e-05, + "loss": 0.1116, + "step": 1023 + }, + { + "epoch": 1.3298701298701299, + "grad_norm": 12.266860006537707, + "learning_rate": 1.9450237448147463e-05, + "loss": 0.126, + "step": 1024 + }, + { + "epoch": 1.3311688311688312, + "grad_norm": 69.29228636479529, + "learning_rate": 1.9448861185545612e-05, + "loss": 0.1022, + "step": 1025 + }, + { + "epoch": 1.3324675324675326, + "grad_norm": 21.199587721587317, + "learning_rate": 1.944748325125936e-05, + "loss": 0.1107, + "step": 1026 + }, + { + "epoch": 1.3337662337662337, + "grad_norm": 70.76013683359018, + "learning_rate": 1.9446103645532494e-05, + "loss": 0.1343, + "step": 1027 + }, + { + "epoch": 1.335064935064935, + "grad_norm": 28.002446230891454, + "learning_rate": 1.9444722368609095e-05, + "loss": 0.1428, + "step": 1028 + }, + { + "epoch": 1.3363636363636364, + "grad_norm": 41.62306494516917, + "learning_rate": 1.9443339420733537e-05, + "loss": 0.1101, + "step": 1029 + }, + { + "epoch": 1.3376623376623376, + "grad_norm": 10.972201041656279, + "learning_rate": 1.9441954802150486e-05, + "loss": 0.071, + "step": 1030 + }, + { + "epoch": 1.338961038961039, + "grad_norm": 49.739762364943154, + "learning_rate": 1.9440568513104906e-05, + "loss": 0.1388, + "step": 1031 + }, + { + "epoch": 1.3402597402597403, + "grad_norm": 14.620806928433652, + "learning_rate": 1.9439180553842064e-05, + "loss": 0.0846, + "step": 1032 + }, + { + "epoch": 1.3415584415584416, + "grad_norm": 56.57021218157827, + "learning_rate": 1.9437790924607516e-05, + "loss": 0.1539, + "step": 1033 + }, + { + "epoch": 1.342857142857143, + "grad_norm": 15.118165954662341, + "learning_rate": 1.943639962564711e-05, + "loss": 0.101, + "step": 1034 + }, + { + "epoch": 1.344155844155844, + "grad_norm": 51.02939236273399, + "learning_rate": 1.9435006657206998e-05, + "loss": 0.1363, + "step": 1035 + }, + { + "epoch": 1.3454545454545455, + "grad_norm": 10.285539703088427, + "learning_rate": 1.943361201953362e-05, + "loss": 0.0857, + "step": 1036 + }, + { + "epoch": 1.3467532467532468, + "grad_norm": 35.29633627148749, + "learning_rate": 1.943221571287371e-05, + "loss": 0.1051, + "step": 1037 + }, + { + "epoch": 1.348051948051948, + "grad_norm": 21.910223437978114, + "learning_rate": 1.943081773747431e-05, + "loss": 0.0976, + "step": 1038 + }, + { + "epoch": 1.3493506493506493, + "grad_norm": 20.843486145382172, + "learning_rate": 1.9429418093582744e-05, + "loss": 0.1415, + "step": 1039 + }, + { + "epoch": 1.3506493506493507, + "grad_norm": 42.28221015985843, + "learning_rate": 1.942801678144664e-05, + "loss": 0.0779, + "step": 1040 + }, + { + "epoch": 1.351948051948052, + "grad_norm": 19.672727017131834, + "learning_rate": 1.9426613801313907e-05, + "loss": 0.0926, + "step": 1041 + }, + { + "epoch": 1.3532467532467534, + "grad_norm": 47.15123328824099, + "learning_rate": 1.9425209153432767e-05, + "loss": 0.0907, + "step": 1042 + }, + { + "epoch": 1.3545454545454545, + "grad_norm": 10.35234916741717, + "learning_rate": 1.942380283805173e-05, + "loss": 0.1024, + "step": 1043 + }, + { + "epoch": 1.3558441558441559, + "grad_norm": 71.12075172978987, + "learning_rate": 1.9422394855419592e-05, + "loss": 0.1205, + "step": 1044 + }, + { + "epoch": 1.3571428571428572, + "grad_norm": 14.156471961619323, + "learning_rate": 1.9420985205785463e-05, + "loss": 0.1277, + "step": 1045 + }, + { + "epoch": 1.3584415584415583, + "grad_norm": 67.96659151532458, + "learning_rate": 1.941957388939873e-05, + "loss": 0.103, + "step": 1046 + }, + { + "epoch": 1.3597402597402597, + "grad_norm": 5.031311562073995, + "learning_rate": 1.941816090650908e-05, + "loss": 0.1229, + "step": 1047 + }, + { + "epoch": 1.361038961038961, + "grad_norm": 50.774257302271, + "learning_rate": 1.9416746257366505e-05, + "loss": 0.1251, + "step": 1048 + }, + { + "epoch": 1.3623376623376624, + "grad_norm": 24.695029052076976, + "learning_rate": 1.9415329942221276e-05, + "loss": 0.0995, + "step": 1049 + }, + { + "epoch": 1.3636363636363638, + "grad_norm": 43.41460356403345, + "learning_rate": 1.941391196132397e-05, + "loss": 0.0886, + "step": 1050 + }, + { + "epoch": 1.364935064935065, + "grad_norm": 25.882820227110745, + "learning_rate": 1.9412492314925453e-05, + "loss": 0.0926, + "step": 1051 + }, + { + "epoch": 1.3662337662337662, + "grad_norm": 24.900683490805868, + "learning_rate": 1.9411071003276888e-05, + "loss": 0.0866, + "step": 1052 + }, + { + "epoch": 1.3675324675324676, + "grad_norm": 40.90461528702595, + "learning_rate": 1.9409648026629733e-05, + "loss": 0.162, + "step": 1053 + }, + { + "epoch": 1.3688311688311687, + "grad_norm": 15.158384059580197, + "learning_rate": 1.940822338523574e-05, + "loss": 0.075, + "step": 1054 + }, + { + "epoch": 1.37012987012987, + "grad_norm": 47.65492575391997, + "learning_rate": 1.9406797079346954e-05, + "loss": 0.08, + "step": 1055 + }, + { + "epoch": 1.3714285714285714, + "grad_norm": 8.84499435229221, + "learning_rate": 1.9405369109215718e-05, + "loss": 0.1036, + "step": 1056 + }, + { + "epoch": 1.3727272727272728, + "grad_norm": 50.76393097424129, + "learning_rate": 1.9403939475094664e-05, + "loss": 0.1044, + "step": 1057 + }, + { + "epoch": 1.3740259740259742, + "grad_norm": 23.318015871439552, + "learning_rate": 1.9402508177236726e-05, + "loss": 0.084, + "step": 1058 + }, + { + "epoch": 1.3753246753246753, + "grad_norm": 58.13452079492644, + "learning_rate": 1.9401075215895123e-05, + "loss": 0.131, + "step": 1059 + }, + { + "epoch": 1.3766233766233766, + "grad_norm": 23.851075213510118, + "learning_rate": 1.9399640591323376e-05, + "loss": 0.0651, + "step": 1060 + }, + { + "epoch": 1.377922077922078, + "grad_norm": 46.37629495247977, + "learning_rate": 1.93982043037753e-05, + "loss": 0.1074, + "step": 1061 + }, + { + "epoch": 1.3792207792207791, + "grad_norm": 35.58593315802314, + "learning_rate": 1.9396766353505e-05, + "loss": 0.0941, + "step": 1062 + }, + { + "epoch": 1.3805194805194805, + "grad_norm": 50.80483322835675, + "learning_rate": 1.9395326740766872e-05, + "loss": 0.1109, + "step": 1063 + }, + { + "epoch": 1.3818181818181818, + "grad_norm": 45.60850639138541, + "learning_rate": 1.9393885465815615e-05, + "loss": 0.0977, + "step": 1064 + }, + { + "epoch": 1.3831168831168832, + "grad_norm": 51.02534855280732, + "learning_rate": 1.939244252890622e-05, + "loss": 0.1059, + "step": 1065 + }, + { + "epoch": 1.3844155844155845, + "grad_norm": 61.464644483736926, + "learning_rate": 1.9390997930293968e-05, + "loss": 0.0966, + "step": 1066 + }, + { + "epoch": 1.3857142857142857, + "grad_norm": 30.854076996931735, + "learning_rate": 1.9389551670234438e-05, + "loss": 0.1011, + "step": 1067 + }, + { + "epoch": 1.387012987012987, + "grad_norm": 80.04543790931409, + "learning_rate": 1.9388103748983496e-05, + "loss": 0.1274, + "step": 1068 + }, + { + "epoch": 1.3883116883116884, + "grad_norm": 14.343851463503217, + "learning_rate": 1.9386654166797317e-05, + "loss": 0.0668, + "step": 1069 + }, + { + "epoch": 1.3896103896103895, + "grad_norm": 74.70345046654593, + "learning_rate": 1.9385202923932345e-05, + "loss": 0.0885, + "step": 1070 + }, + { + "epoch": 1.3909090909090909, + "grad_norm": 9.31518108819737, + "learning_rate": 1.9383750020645346e-05, + "loss": 0.0911, + "step": 1071 + }, + { + "epoch": 1.3922077922077922, + "grad_norm": 51.00094925138993, + "learning_rate": 1.9382295457193364e-05, + "loss": 0.0959, + "step": 1072 + }, + { + "epoch": 1.3935064935064936, + "grad_norm": 17.991758458522792, + "learning_rate": 1.9380839233833734e-05, + "loss": 0.0876, + "step": 1073 + }, + { + "epoch": 1.394805194805195, + "grad_norm": 40.82727009855829, + "learning_rate": 1.937938135082409e-05, + "loss": 0.1106, + "step": 1074 + }, + { + "epoch": 1.396103896103896, + "grad_norm": 34.882491089463414, + "learning_rate": 1.9377921808422364e-05, + "loss": 0.1198, + "step": 1075 + }, + { + "epoch": 1.3974025974025974, + "grad_norm": 37.55844480893763, + "learning_rate": 1.9376460606886774e-05, + "loss": 0.1042, + "step": 1076 + }, + { + "epoch": 1.3987012987012988, + "grad_norm": 41.3878836740139, + "learning_rate": 1.9374997746475834e-05, + "loss": 0.0857, + "step": 1077 + }, + { + "epoch": 1.4, + "grad_norm": 11.667967612361611, + "learning_rate": 1.9373533227448357e-05, + "loss": 0.0866, + "step": 1078 + }, + { + "epoch": 1.4012987012987013, + "grad_norm": 63.93487881422403, + "learning_rate": 1.937206705006344e-05, + "loss": 0.0914, + "step": 1079 + }, + { + "epoch": 1.4025974025974026, + "grad_norm": 9.258146750619503, + "learning_rate": 1.9370599214580474e-05, + "loss": 0.1123, + "step": 1080 + }, + { + "epoch": 1.403896103896104, + "grad_norm": 47.09261644311387, + "learning_rate": 1.9369129721259156e-05, + "loss": 0.1125, + "step": 1081 + }, + { + "epoch": 1.4051948051948053, + "grad_norm": 13.003218930314107, + "learning_rate": 1.9367658570359463e-05, + "loss": 0.1043, + "step": 1082 + }, + { + "epoch": 1.4064935064935065, + "grad_norm": 18.23828564353552, + "learning_rate": 1.936618576214167e-05, + "loss": 0.1396, + "step": 1083 + }, + { + "epoch": 1.4077922077922078, + "grad_norm": 31.207806625715115, + "learning_rate": 1.9364711296866343e-05, + "loss": 0.1325, + "step": 1084 + }, + { + "epoch": 1.4090909090909092, + "grad_norm": 57.822147107897145, + "learning_rate": 1.9363235174794346e-05, + "loss": 0.1899, + "step": 1085 + }, + { + "epoch": 1.4103896103896103, + "grad_norm": 15.15205145974466, + "learning_rate": 1.9361757396186834e-05, + "loss": 0.1103, + "step": 1086 + }, + { + "epoch": 1.4116883116883117, + "grad_norm": 46.26083955652628, + "learning_rate": 1.936027796130525e-05, + "loss": 0.079, + "step": 1087 + }, + { + "epoch": 1.412987012987013, + "grad_norm": 30.052349555747, + "learning_rate": 1.935879687041134e-05, + "loss": 0.1007, + "step": 1088 + }, + { + "epoch": 1.4142857142857144, + "grad_norm": 57.70373663194927, + "learning_rate": 1.935731412376713e-05, + "loss": 0.0974, + "step": 1089 + }, + { + "epoch": 1.4155844155844157, + "grad_norm": 18.451874383566572, + "learning_rate": 1.9355829721634952e-05, + "loss": 0.0862, + "step": 1090 + }, + { + "epoch": 1.4168831168831169, + "grad_norm": 45.518417689067164, + "learning_rate": 1.9354343664277427e-05, + "loss": 0.0742, + "step": 1091 + }, + { + "epoch": 1.4181818181818182, + "grad_norm": 33.99911689146825, + "learning_rate": 1.9352855951957462e-05, + "loss": 0.0857, + "step": 1092 + }, + { + "epoch": 1.4194805194805196, + "grad_norm": 46.8941969669433, + "learning_rate": 1.9351366584938262e-05, + "loss": 0.1077, + "step": 1093 + }, + { + "epoch": 1.4207792207792207, + "grad_norm": 39.49266825695811, + "learning_rate": 1.934987556348333e-05, + "loss": 0.0997, + "step": 1094 + }, + { + "epoch": 1.422077922077922, + "grad_norm": 20.67107952005338, + "learning_rate": 1.9348382887856446e-05, + "loss": 0.0963, + "step": 1095 + }, + { + "epoch": 1.4233766233766234, + "grad_norm": 65.37341908976738, + "learning_rate": 1.9346888558321706e-05, + "loss": 0.1091, + "step": 1096 + }, + { + "epoch": 1.4246753246753245, + "grad_norm": 21.376189308974997, + "learning_rate": 1.9345392575143473e-05, + "loss": 0.099, + "step": 1097 + }, + { + "epoch": 1.425974025974026, + "grad_norm": 50.14306437321173, + "learning_rate": 1.9343894938586424e-05, + "loss": 0.0805, + "step": 1098 + }, + { + "epoch": 1.4272727272727272, + "grad_norm": 23.664139079285583, + "learning_rate": 1.9342395648915515e-05, + "loss": 0.1082, + "step": 1099 + }, + { + "epoch": 1.4285714285714286, + "grad_norm": 52.21483983815003, + "learning_rate": 1.9340894706396e-05, + "loss": 0.089, + "step": 1100 + }, + { + "epoch": 1.42987012987013, + "grad_norm": 51.04784614722638, + "learning_rate": 1.9339392111293425e-05, + "loss": 0.0854, + "step": 1101 + }, + { + "epoch": 1.431168831168831, + "grad_norm": 53.39758768692599, + "learning_rate": 1.933788786387363e-05, + "loss": 0.1008, + "step": 1102 + }, + { + "epoch": 1.4324675324675324, + "grad_norm": 63.223995476853354, + "learning_rate": 1.9336381964402743e-05, + "loss": 0.1268, + "step": 1103 + }, + { + "epoch": 1.4337662337662338, + "grad_norm": 50.61087537436197, + "learning_rate": 1.9334874413147183e-05, + "loss": 0.0796, + "step": 1104 + }, + { + "epoch": 1.435064935064935, + "grad_norm": 54.83640031564857, + "learning_rate": 1.9333365210373668e-05, + "loss": 0.0864, + "step": 1105 + }, + { + "epoch": 1.4363636363636363, + "grad_norm": 28.687124484521814, + "learning_rate": 1.933185435634921e-05, + "loss": 0.0944, + "step": 1106 + }, + { + "epoch": 1.4376623376623376, + "grad_norm": 44.453485734170414, + "learning_rate": 1.9330341851341098e-05, + "loss": 0.1012, + "step": 1107 + }, + { + "epoch": 1.438961038961039, + "grad_norm": 13.668463280988693, + "learning_rate": 1.9328827695616925e-05, + "loss": 0.0619, + "step": 1108 + }, + { + "epoch": 1.4402597402597404, + "grad_norm": 59.23577085306595, + "learning_rate": 1.9327311889444582e-05, + "loss": 0.0869, + "step": 1109 + }, + { + "epoch": 1.4415584415584415, + "grad_norm": 18.392545999161126, + "learning_rate": 1.9325794433092236e-05, + "loss": 0.091, + "step": 1110 + }, + { + "epoch": 1.4428571428571428, + "grad_norm": 44.10037687852132, + "learning_rate": 1.932427532682836e-05, + "loss": 0.0924, + "step": 1111 + }, + { + "epoch": 1.4441558441558442, + "grad_norm": 40.035946259067266, + "learning_rate": 1.9322754570921713e-05, + "loss": 0.1088, + "step": 1112 + }, + { + "epoch": 1.4454545454545453, + "grad_norm": 33.99570341551419, + "learning_rate": 1.9321232165641336e-05, + "loss": 0.0704, + "step": 1113 + }, + { + "epoch": 1.4467532467532467, + "grad_norm": 39.30443496822773, + "learning_rate": 1.931970811125658e-05, + "loss": 0.106, + "step": 1114 + }, + { + "epoch": 1.448051948051948, + "grad_norm": 10.83301030165558, + "learning_rate": 1.931818240803708e-05, + "loss": 0.0745, + "step": 1115 + }, + { + "epoch": 1.4493506493506494, + "grad_norm": 53.73620350359476, + "learning_rate": 1.9316655056252764e-05, + "loss": 0.1542, + "step": 1116 + }, + { + "epoch": 1.4506493506493507, + "grad_norm": 37.453929248340735, + "learning_rate": 1.9315126056173838e-05, + "loss": 0.0931, + "step": 1117 + }, + { + "epoch": 1.4519480519480519, + "grad_norm": 49.600536137150684, + "learning_rate": 1.9313595408070825e-05, + "loss": 0.07, + "step": 1118 + }, + { + "epoch": 1.4532467532467532, + "grad_norm": 45.47135504300238, + "learning_rate": 1.9312063112214518e-05, + "loss": 0.1123, + "step": 1119 + }, + { + "epoch": 1.4545454545454546, + "grad_norm": 21.190639884012214, + "learning_rate": 1.9310529168876008e-05, + "loss": 0.1126, + "step": 1120 + }, + { + "epoch": 1.4558441558441557, + "grad_norm": 44.50080996715427, + "learning_rate": 1.9308993578326688e-05, + "loss": 0.1098, + "step": 1121 + }, + { + "epoch": 1.457142857142857, + "grad_norm": 12.223796877435062, + "learning_rate": 1.9307456340838225e-05, + "loss": 0.0982, + "step": 1122 + }, + { + "epoch": 1.4584415584415584, + "grad_norm": 44.77005165537213, + "learning_rate": 1.930591745668259e-05, + "loss": 0.1158, + "step": 1123 + }, + { + "epoch": 1.4597402597402598, + "grad_norm": 13.985482578073892, + "learning_rate": 1.9304376926132044e-05, + "loss": 0.0765, + "step": 1124 + }, + { + "epoch": 1.4610389610389611, + "grad_norm": 18.90410529275322, + "learning_rate": 1.9302834749459126e-05, + "loss": 0.1113, + "step": 1125 + }, + { + "epoch": 1.4623376623376623, + "grad_norm": 25.361692330375906, + "learning_rate": 1.9301290926936686e-05, + "loss": 0.0753, + "step": 1126 + }, + { + "epoch": 1.4636363636363636, + "grad_norm": 12.461072349983334, + "learning_rate": 1.929974545883785e-05, + "loss": 0.0442, + "step": 1127 + }, + { + "epoch": 1.464935064935065, + "grad_norm": 42.13815252294741, + "learning_rate": 1.9298198345436044e-05, + "loss": 0.0841, + "step": 1128 + }, + { + "epoch": 1.4662337662337661, + "grad_norm": 11.96407481186497, + "learning_rate": 1.929664958700498e-05, + "loss": 0.0811, + "step": 1129 + }, + { + "epoch": 1.4675324675324675, + "grad_norm": 28.099850677532977, + "learning_rate": 1.9295099183818667e-05, + "loss": 0.0858, + "step": 1130 + }, + { + "epoch": 1.4688311688311688, + "grad_norm": 21.299369807146903, + "learning_rate": 1.9293547136151394e-05, + "loss": 0.1105, + "step": 1131 + }, + { + "epoch": 1.4701298701298702, + "grad_norm": 26.700052949029157, + "learning_rate": 1.9291993444277758e-05, + "loss": 0.0942, + "step": 1132 + }, + { + "epoch": 1.4714285714285715, + "grad_norm": 21.590424386935243, + "learning_rate": 1.929043810847263e-05, + "loss": 0.0991, + "step": 1133 + }, + { + "epoch": 1.4727272727272727, + "grad_norm": 20.613249820773643, + "learning_rate": 1.9288881129011177e-05, + "loss": 0.0815, + "step": 1134 + }, + { + "epoch": 1.474025974025974, + "grad_norm": 32.64270803905538, + "learning_rate": 1.9287322506168862e-05, + "loss": 0.1105, + "step": 1135 + }, + { + "epoch": 1.4753246753246754, + "grad_norm": 30.606506675025628, + "learning_rate": 1.9285762240221434e-05, + "loss": 0.0712, + "step": 1136 + }, + { + "epoch": 1.4766233766233765, + "grad_norm": 12.122935191352726, + "learning_rate": 1.9284200331444935e-05, + "loss": 0.0661, + "step": 1137 + }, + { + "epoch": 1.4779220779220779, + "grad_norm": 12.538507840697362, + "learning_rate": 1.9282636780115697e-05, + "loss": 0.1335, + "step": 1138 + }, + { + "epoch": 1.4792207792207792, + "grad_norm": 17.925970753236967, + "learning_rate": 1.928107158651034e-05, + "loss": 0.1317, + "step": 1139 + }, + { + "epoch": 1.4805194805194806, + "grad_norm": 26.492946580158964, + "learning_rate": 1.927950475090578e-05, + "loss": 0.1102, + "step": 1140 + }, + { + "epoch": 1.481818181818182, + "grad_norm": 8.269904345502884, + "learning_rate": 1.9277936273579213e-05, + "loss": 0.1167, + "step": 1141 + }, + { + "epoch": 1.483116883116883, + "grad_norm": 34.182269766873176, + "learning_rate": 1.927636615480814e-05, + "loss": 0.1239, + "step": 1142 + }, + { + "epoch": 1.4844155844155844, + "grad_norm": 21.676233279198232, + "learning_rate": 1.9274794394870342e-05, + "loss": 0.1054, + "step": 1143 + }, + { + "epoch": 1.4857142857142858, + "grad_norm": 33.30168469754543, + "learning_rate": 1.9273220994043896e-05, + "loss": 0.0655, + "step": 1144 + }, + { + "epoch": 1.487012987012987, + "grad_norm": 8.481161072712373, + "learning_rate": 1.9271645952607164e-05, + "loss": 0.0949, + "step": 1145 + }, + { + "epoch": 1.4883116883116883, + "grad_norm": 16.731298112093803, + "learning_rate": 1.92700692708388e-05, + "loss": 0.1143, + "step": 1146 + }, + { + "epoch": 1.4896103896103896, + "grad_norm": 10.278492679291363, + "learning_rate": 1.9268490949017753e-05, + "loss": 0.0743, + "step": 1147 + }, + { + "epoch": 1.490909090909091, + "grad_norm": 26.390056700027724, + "learning_rate": 1.9266910987423255e-05, + "loss": 0.0931, + "step": 1148 + }, + { + "epoch": 1.4922077922077923, + "grad_norm": 11.673548303519283, + "learning_rate": 1.9265329386334833e-05, + "loss": 0.1534, + "step": 1149 + }, + { + "epoch": 1.4935064935064934, + "grad_norm": 35.283721381599065, + "learning_rate": 1.9263746146032304e-05, + "loss": 0.1122, + "step": 1150 + }, + { + "epoch": 1.4948051948051948, + "grad_norm": 11.868002484823982, + "learning_rate": 1.926216126679577e-05, + "loss": 0.0981, + "step": 1151 + }, + { + "epoch": 1.4961038961038962, + "grad_norm": 26.13798396577956, + "learning_rate": 1.9260574748905633e-05, + "loss": 0.1096, + "step": 1152 + }, + { + "epoch": 1.4974025974025973, + "grad_norm": 26.076093190270736, + "learning_rate": 1.9258986592642568e-05, + "loss": 0.0855, + "step": 1153 + }, + { + "epoch": 1.4987012987012986, + "grad_norm": 21.613031781215387, + "learning_rate": 1.925739679828756e-05, + "loss": 0.0771, + "step": 1154 + }, + { + "epoch": 1.5, + "grad_norm": 23.548927729609066, + "learning_rate": 1.9255805366121863e-05, + "loss": 0.0755, + "step": 1155 + }, + { + "epoch": 1.5012987012987011, + "grad_norm": 11.350635994569172, + "learning_rate": 1.9254212296427043e-05, + "loss": 0.0947, + "step": 1156 + }, + { + "epoch": 1.5025974025974027, + "grad_norm": 28.966094849572727, + "learning_rate": 1.9252617589484942e-05, + "loss": 0.1009, + "step": 1157 + }, + { + "epoch": 1.5038961038961038, + "grad_norm": 19.249198831792068, + "learning_rate": 1.925102124557769e-05, + "loss": 0.0774, + "step": 1158 + }, + { + "epoch": 1.5051948051948052, + "grad_norm": 12.557945037013042, + "learning_rate": 1.9249423264987714e-05, + "loss": 0.076, + "step": 1159 + }, + { + "epoch": 1.5064935064935066, + "grad_norm": 21.761676112008697, + "learning_rate": 1.9247823647997727e-05, + "loss": 0.0952, + "step": 1160 + }, + { + "epoch": 1.5077922077922077, + "grad_norm": 48.84364316385414, + "learning_rate": 1.924622239489073e-05, + "loss": 0.1066, + "step": 1161 + }, + { + "epoch": 1.509090909090909, + "grad_norm": 9.56249411071979, + "learning_rate": 1.924461950595002e-05, + "loss": 0.0738, + "step": 1162 + }, + { + "epoch": 1.5103896103896104, + "grad_norm": 27.88655623373311, + "learning_rate": 1.924301498145917e-05, + "loss": 0.0923, + "step": 1163 + }, + { + "epoch": 1.5116883116883115, + "grad_norm": 5.0144042943577976, + "learning_rate": 1.924140882170206e-05, + "loss": 0.0944, + "step": 1164 + }, + { + "epoch": 1.512987012987013, + "grad_norm": 24.14630575835831, + "learning_rate": 1.923980102696285e-05, + "loss": 0.136, + "step": 1165 + }, + { + "epoch": 1.5142857142857142, + "grad_norm": 17.56854430710703, + "learning_rate": 1.923819159752598e-05, + "loss": 0.1212, + "step": 1166 + }, + { + "epoch": 1.5155844155844156, + "grad_norm": 20.05494551786254, + "learning_rate": 1.92365805336762e-05, + "loss": 0.1219, + "step": 1167 + }, + { + "epoch": 1.516883116883117, + "grad_norm": 70.97597800587714, + "learning_rate": 1.9234967835698533e-05, + "loss": 0.1333, + "step": 1168 + }, + { + "epoch": 1.518181818181818, + "grad_norm": 18.327974997375332, + "learning_rate": 1.9233353503878297e-05, + "loss": 0.0953, + "step": 1169 + }, + { + "epoch": 1.5194805194805194, + "grad_norm": 75.12627753204687, + "learning_rate": 1.9231737538501093e-05, + "loss": 0.0834, + "step": 1170 + }, + { + "epoch": 1.5207792207792208, + "grad_norm": 9.311207969778543, + "learning_rate": 1.923011993985283e-05, + "loss": 0.0894, + "step": 1171 + }, + { + "epoch": 1.522077922077922, + "grad_norm": 67.75705343551864, + "learning_rate": 1.9228500708219674e-05, + "loss": 0.1148, + "step": 1172 + }, + { + "epoch": 1.5233766233766235, + "grad_norm": 33.43096958652541, + "learning_rate": 1.9226879843888114e-05, + "loss": 0.0983, + "step": 1173 + }, + { + "epoch": 1.5246753246753246, + "grad_norm": 61.15324300094886, + "learning_rate": 1.9225257347144903e-05, + "loss": 0.1025, + "step": 1174 + }, + { + "epoch": 1.525974025974026, + "grad_norm": 47.55125138634662, + "learning_rate": 1.9223633218277094e-05, + "loss": 0.0776, + "step": 1175 + }, + { + "epoch": 1.5272727272727273, + "grad_norm": 58.13159550979053, + "learning_rate": 1.922200745757203e-05, + "loss": 0.1334, + "step": 1176 + }, + { + "epoch": 1.5285714285714285, + "grad_norm": 51.11726429086992, + "learning_rate": 1.9220380065317328e-05, + "loss": 0.0894, + "step": 1177 + }, + { + "epoch": 1.5298701298701298, + "grad_norm": 42.56349143852162, + "learning_rate": 1.921875104180092e-05, + "loss": 0.1432, + "step": 1178 + }, + { + "epoch": 1.5311688311688312, + "grad_norm": 77.68228958652348, + "learning_rate": 1.9217120387310996e-05, + "loss": 0.1169, + "step": 1179 + }, + { + "epoch": 1.5324675324675323, + "grad_norm": 35.786287916646415, + "learning_rate": 1.9215488102136066e-05, + "loss": 0.1044, + "step": 1180 + }, + { + "epoch": 1.5337662337662339, + "grad_norm": 94.07461603541564, + "learning_rate": 1.92138541865649e-05, + "loss": 0.1164, + "step": 1181 + }, + { + "epoch": 1.535064935064935, + "grad_norm": 37.08957897683132, + "learning_rate": 1.9212218640886573e-05, + "loss": 0.083, + "step": 1182 + }, + { + "epoch": 1.5363636363636364, + "grad_norm": 90.70181156808303, + "learning_rate": 1.9210581465390445e-05, + "loss": 0.1725, + "step": 1183 + }, + { + "epoch": 1.5376623376623377, + "grad_norm": 33.84444354579334, + "learning_rate": 1.9208942660366164e-05, + "loss": 0.1006, + "step": 1184 + }, + { + "epoch": 1.5389610389610389, + "grad_norm": 93.84978157148899, + "learning_rate": 1.920730222610367e-05, + "loss": 0.1003, + "step": 1185 + }, + { + "epoch": 1.5402597402597402, + "grad_norm": 19.736813370767486, + "learning_rate": 1.9205660162893176e-05, + "loss": 0.0915, + "step": 1186 + }, + { + "epoch": 1.5415584415584416, + "grad_norm": 97.12153324151461, + "learning_rate": 1.92040164710252e-05, + "loss": 0.1334, + "step": 1187 + }, + { + "epoch": 1.5428571428571427, + "grad_norm": 11.65161608729751, + "learning_rate": 1.9202371150790552e-05, + "loss": 0.1145, + "step": 1188 + }, + { + "epoch": 1.5441558441558443, + "grad_norm": 67.85791289534318, + "learning_rate": 1.9200724202480305e-05, + "loss": 0.128, + "step": 1189 + }, + { + "epoch": 1.5454545454545454, + "grad_norm": 13.250223587650886, + "learning_rate": 1.9199075626385847e-05, + "loss": 0.0869, + "step": 1190 + }, + { + "epoch": 1.5467532467532468, + "grad_norm": 51.68862343180047, + "learning_rate": 1.9197425422798832e-05, + "loss": 0.1014, + "step": 1191 + }, + { + "epoch": 1.5480519480519481, + "grad_norm": 19.753193241984302, + "learning_rate": 1.9195773592011225e-05, + "loss": 0.0765, + "step": 1192 + }, + { + "epoch": 1.5493506493506493, + "grad_norm": 27.83714859969863, + "learning_rate": 1.9194120134315256e-05, + "loss": 0.0922, + "step": 1193 + }, + { + "epoch": 1.5506493506493506, + "grad_norm": 44.875853290840155, + "learning_rate": 1.9192465050003462e-05, + "loss": 0.0846, + "step": 1194 + }, + { + "epoch": 1.551948051948052, + "grad_norm": 10.160905345031974, + "learning_rate": 1.9190808339368658e-05, + "loss": 0.0764, + "step": 1195 + }, + { + "epoch": 1.553246753246753, + "grad_norm": 75.89367062436416, + "learning_rate": 1.9189150002703938e-05, + "loss": 0.127, + "step": 1196 + }, + { + "epoch": 1.5545454545454547, + "grad_norm": 9.556484693043718, + "learning_rate": 1.9187490040302703e-05, + "loss": 0.1082, + "step": 1197 + }, + { + "epoch": 1.5558441558441558, + "grad_norm": 50.316231778494, + "learning_rate": 1.918582845245863e-05, + "loss": 0.1146, + "step": 1198 + }, + { + "epoch": 1.5571428571428572, + "grad_norm": 38.96251614381031, + "learning_rate": 1.9184165239465683e-05, + "loss": 0.0945, + "step": 1199 + }, + { + "epoch": 1.5584415584415585, + "grad_norm": 47.23786282164021, + "learning_rate": 1.9182500401618122e-05, + "loss": 0.1349, + "step": 1200 + }, + { + "epoch": 1.5597402597402596, + "grad_norm": 45.37639805445404, + "learning_rate": 1.9180833939210482e-05, + "loss": 0.1102, + "step": 1201 + }, + { + "epoch": 1.561038961038961, + "grad_norm": 11.547903829072315, + "learning_rate": 1.9179165852537596e-05, + "loss": 0.08, + "step": 1202 + }, + { + "epoch": 1.5623376623376624, + "grad_norm": 50.92299866968745, + "learning_rate": 1.917749614189458e-05, + "loss": 0.0987, + "step": 1203 + }, + { + "epoch": 1.5636363636363635, + "grad_norm": 27.505009662316596, + "learning_rate": 1.9175824807576834e-05, + "loss": 0.1025, + "step": 1204 + }, + { + "epoch": 1.564935064935065, + "grad_norm": 62.959279203992125, + "learning_rate": 1.9174151849880055e-05, + "loss": 0.1208, + "step": 1205 + }, + { + "epoch": 1.5662337662337662, + "grad_norm": 22.08236292001386, + "learning_rate": 1.9172477269100223e-05, + "loss": 0.0674, + "step": 1206 + }, + { + "epoch": 1.5675324675324676, + "grad_norm": 32.172907883908486, + "learning_rate": 1.9170801065533593e-05, + "loss": 0.1201, + "step": 1207 + }, + { + "epoch": 1.568831168831169, + "grad_norm": 40.72392242607336, + "learning_rate": 1.9169123239476725e-05, + "loss": 0.1248, + "step": 1208 + }, + { + "epoch": 1.57012987012987, + "grad_norm": 13.736383480252355, + "learning_rate": 1.916744379122646e-05, + "loss": 0.101, + "step": 1209 + }, + { + "epoch": 1.5714285714285714, + "grad_norm": 52.83809491065626, + "learning_rate": 1.916576272107992e-05, + "loss": 0.091, + "step": 1210 + }, + { + "epoch": 1.5727272727272728, + "grad_norm": 8.107311490250531, + "learning_rate": 1.9164080029334517e-05, + "loss": 0.1018, + "step": 1211 + }, + { + "epoch": 1.5740259740259739, + "grad_norm": 61.798422587200726, + "learning_rate": 1.9162395716287955e-05, + "loss": 0.1345, + "step": 1212 + }, + { + "epoch": 1.5753246753246755, + "grad_norm": 19.45561070074833, + "learning_rate": 1.9160709782238222e-05, + "loss": 0.1169, + "step": 1213 + }, + { + "epoch": 1.5766233766233766, + "grad_norm": 54.559729680700585, + "learning_rate": 1.915902222748359e-05, + "loss": 0.0838, + "step": 1214 + }, + { + "epoch": 1.577922077922078, + "grad_norm": 26.964599763525282, + "learning_rate": 1.9157333052322622e-05, + "loss": 0.1143, + "step": 1215 + }, + { + "epoch": 1.5792207792207793, + "grad_norm": 43.208661308651884, + "learning_rate": 1.9155642257054166e-05, + "loss": 0.0929, + "step": 1216 + }, + { + "epoch": 1.5805194805194804, + "grad_norm": 39.16908565603254, + "learning_rate": 1.915394984197735e-05, + "loss": 0.0984, + "step": 1217 + }, + { + "epoch": 1.5818181818181818, + "grad_norm": 23.13557499873426, + "learning_rate": 1.91522558073916e-05, + "loss": 0.1187, + "step": 1218 + }, + { + "epoch": 1.5831168831168831, + "grad_norm": 72.30963827629151, + "learning_rate": 1.915056015359662e-05, + "loss": 0.1187, + "step": 1219 + }, + { + "epoch": 1.5844155844155843, + "grad_norm": 17.72105036569876, + "learning_rate": 1.9148862880892407e-05, + "loss": 0.1178, + "step": 1220 + }, + { + "epoch": 1.5857142857142859, + "grad_norm": 70.02513894769763, + "learning_rate": 1.9147163989579235e-05, + "loss": 0.1048, + "step": 1221 + }, + { + "epoch": 1.587012987012987, + "grad_norm": 10.859368762901815, + "learning_rate": 1.914546347995768e-05, + "loss": 0.1183, + "step": 1222 + }, + { + "epoch": 1.5883116883116883, + "grad_norm": 51.66707823157871, + "learning_rate": 1.9143761352328587e-05, + "loss": 0.1032, + "step": 1223 + }, + { + "epoch": 1.5896103896103897, + "grad_norm": 37.79349050114669, + "learning_rate": 1.91420576069931e-05, + "loss": 0.0999, + "step": 1224 + }, + { + "epoch": 1.5909090909090908, + "grad_norm": 47.914381894386054, + "learning_rate": 1.914035224425264e-05, + "loss": 0.0598, + "step": 1225 + }, + { + "epoch": 1.5922077922077922, + "grad_norm": 41.16296935747927, + "learning_rate": 1.9138645264408917e-05, + "loss": 0.1271, + "step": 1226 + }, + { + "epoch": 1.5935064935064935, + "grad_norm": 30.260108685045306, + "learning_rate": 1.9136936667763935e-05, + "loss": 0.1388, + "step": 1227 + }, + { + "epoch": 1.5948051948051947, + "grad_norm": 54.82569858071155, + "learning_rate": 1.913522645461997e-05, + "loss": 0.1338, + "step": 1228 + }, + { + "epoch": 1.5961038961038962, + "grad_norm": 22.171023130304075, + "learning_rate": 1.91335146252796e-05, + "loss": 0.1107, + "step": 1229 + }, + { + "epoch": 1.5974025974025974, + "grad_norm": 46.947142085304115, + "learning_rate": 1.9131801180045674e-05, + "loss": 0.1194, + "step": 1230 + }, + { + "epoch": 1.5987012987012987, + "grad_norm": 10.278173521285403, + "learning_rate": 1.913008611922133e-05, + "loss": 0.1048, + "step": 1231 + }, + { + "epoch": 1.6, + "grad_norm": 37.97908275037359, + "learning_rate": 1.9128369443110002e-05, + "loss": 0.1317, + "step": 1232 + }, + { + "epoch": 1.6012987012987012, + "grad_norm": 45.56514257468939, + "learning_rate": 1.9126651152015404e-05, + "loss": 0.0697, + "step": 1233 + }, + { + "epoch": 1.6025974025974026, + "grad_norm": 39.1593569649967, + "learning_rate": 1.912493124624153e-05, + "loss": 0.1176, + "step": 1234 + }, + { + "epoch": 1.603896103896104, + "grad_norm": 77.78905387259711, + "learning_rate": 1.9123209726092663e-05, + "loss": 0.1294, + "step": 1235 + }, + { + "epoch": 1.605194805194805, + "grad_norm": 67.83473363572571, + "learning_rate": 1.9121486591873374e-05, + "loss": 0.1252, + "step": 1236 + }, + { + "epoch": 1.6064935064935066, + "grad_norm": 52.15729364603715, + "learning_rate": 1.9119761843888523e-05, + "loss": 0.1064, + "step": 1237 + }, + { + "epoch": 1.6077922077922078, + "grad_norm": 34.95913601597636, + "learning_rate": 1.911803548244325e-05, + "loss": 0.1499, + "step": 1238 + }, + { + "epoch": 1.6090909090909091, + "grad_norm": 75.23004793119945, + "learning_rate": 1.911630750784298e-05, + "loss": 0.1298, + "step": 1239 + }, + { + "epoch": 1.6103896103896105, + "grad_norm": 43.36943926665162, + "learning_rate": 1.9114577920393416e-05, + "loss": 0.1139, + "step": 1240 + }, + { + "epoch": 1.6116883116883116, + "grad_norm": 47.728903853979794, + "learning_rate": 1.911284672040057e-05, + "loss": 0.0858, + "step": 1241 + }, + { + "epoch": 1.612987012987013, + "grad_norm": 13.423509210271318, + "learning_rate": 1.9111113908170714e-05, + "loss": 0.0869, + "step": 1242 + }, + { + "epoch": 1.6142857142857143, + "grad_norm": 55.08398429198415, + "learning_rate": 1.9109379484010418e-05, + "loss": 0.1085, + "step": 1243 + }, + { + "epoch": 1.6155844155844155, + "grad_norm": 16.509421983363453, + "learning_rate": 1.9107643448226536e-05, + "loss": 0.1062, + "step": 1244 + }, + { + "epoch": 1.616883116883117, + "grad_norm": 28.610893907672136, + "learning_rate": 1.9105905801126208e-05, + "loss": 0.0848, + "step": 1245 + }, + { + "epoch": 1.6181818181818182, + "grad_norm": 19.849438959052634, + "learning_rate": 1.9104166543016857e-05, + "loss": 0.1232, + "step": 1246 + }, + { + "epoch": 1.6194805194805195, + "grad_norm": 23.664182854009066, + "learning_rate": 1.9102425674206185e-05, + "loss": 0.0941, + "step": 1247 + }, + { + "epoch": 1.6207792207792209, + "grad_norm": 47.996696295648306, + "learning_rate": 1.910068319500219e-05, + "loss": 0.0683, + "step": 1248 + }, + { + "epoch": 1.622077922077922, + "grad_norm": 28.68327780730416, + "learning_rate": 1.909893910571315e-05, + "loss": 0.1161, + "step": 1249 + }, + { + "epoch": 1.6233766233766234, + "grad_norm": 33.501164397193946, + "learning_rate": 1.9097193406647624e-05, + "loss": 0.0672, + "step": 1250 + }, + { + "epoch": 1.6246753246753247, + "grad_norm": 14.642584135704464, + "learning_rate": 1.9095446098114465e-05, + "loss": 0.1058, + "step": 1251 + }, + { + "epoch": 1.6259740259740258, + "grad_norm": 32.378991829235076, + "learning_rate": 1.90936971804228e-05, + "loss": 0.1192, + "step": 1252 + }, + { + "epoch": 1.6272727272727274, + "grad_norm": 34.45645154640012, + "learning_rate": 1.909194665388205e-05, + "loss": 0.0665, + "step": 1253 + }, + { + "epoch": 1.6285714285714286, + "grad_norm": 10.15648904532918, + "learning_rate": 1.9090194518801916e-05, + "loss": 0.1075, + "step": 1254 + }, + { + "epoch": 1.62987012987013, + "grad_norm": 34.96998482013051, + "learning_rate": 1.9088440775492386e-05, + "loss": 0.0855, + "step": 1255 + }, + { + "epoch": 1.6311688311688313, + "grad_norm": 32.91752820698403, + "learning_rate": 1.9086685424263723e-05, + "loss": 0.1187, + "step": 1256 + }, + { + "epoch": 1.6324675324675324, + "grad_norm": 40.000528673710704, + "learning_rate": 1.908492846542649e-05, + "loss": 0.0912, + "step": 1257 + }, + { + "epoch": 1.6337662337662338, + "grad_norm": 49.18012310983659, + "learning_rate": 1.908316989929152e-05, + "loss": 0.0829, + "step": 1258 + }, + { + "epoch": 1.635064935064935, + "grad_norm": 47.200861740549584, + "learning_rate": 1.9081409726169946e-05, + "loss": 0.1281, + "step": 1259 + }, + { + "epoch": 1.6363636363636362, + "grad_norm": 36.43532803369977, + "learning_rate": 1.9079647946373173e-05, + "loss": 0.0666, + "step": 1260 + }, + { + "epoch": 1.6376623376623378, + "grad_norm": 32.554823334169626, + "learning_rate": 1.9077884560212885e-05, + "loss": 0.0705, + "step": 1261 + }, + { + "epoch": 1.638961038961039, + "grad_norm": 33.841636251305836, + "learning_rate": 1.9076119568001074e-05, + "loss": 0.0799, + "step": 1262 + }, + { + "epoch": 1.6402597402597403, + "grad_norm": 9.370966293889616, + "learning_rate": 1.9074352970049985e-05, + "loss": 0.0718, + "step": 1263 + }, + { + "epoch": 1.6415584415584417, + "grad_norm": 47.17548973195319, + "learning_rate": 1.9072584766672174e-05, + "loss": 0.1117, + "step": 1264 + }, + { + "epoch": 1.6428571428571428, + "grad_norm": 6.647645775481107, + "learning_rate": 1.9070814958180467e-05, + "loss": 0.0882, + "step": 1265 + }, + { + "epoch": 1.6441558441558441, + "grad_norm": 39.05333680639166, + "learning_rate": 1.9069043544887978e-05, + "loss": 0.0844, + "step": 1266 + }, + { + "epoch": 1.6454545454545455, + "grad_norm": 15.500514510506953, + "learning_rate": 1.9067270527108102e-05, + "loss": 0.0924, + "step": 1267 + }, + { + "epoch": 1.6467532467532466, + "grad_norm": 18.28881179503815, + "learning_rate": 1.9065495905154517e-05, + "loss": 0.0816, + "step": 1268 + }, + { + "epoch": 1.6480519480519482, + "grad_norm": 9.848990420397799, + "learning_rate": 1.9063719679341193e-05, + "loss": 0.0967, + "step": 1269 + }, + { + "epoch": 1.6493506493506493, + "grad_norm": 22.680849365135025, + "learning_rate": 1.9061941849982377e-05, + "loss": 0.1057, + "step": 1270 + }, + { + "epoch": 1.6506493506493507, + "grad_norm": 21.414921305004288, + "learning_rate": 1.9060162417392602e-05, + "loss": 0.1081, + "step": 1271 + }, + { + "epoch": 1.651948051948052, + "grad_norm": 20.897539003162926, + "learning_rate": 1.905838138188668e-05, + "loss": 0.0888, + "step": 1272 + }, + { + "epoch": 1.6532467532467532, + "grad_norm": 21.7723769082935, + "learning_rate": 1.905659874377971e-05, + "loss": 0.1219, + "step": 1273 + }, + { + "epoch": 1.6545454545454545, + "grad_norm": 5.844389673055337, + "learning_rate": 1.9054814503387077e-05, + "loss": 0.0964, + "step": 1274 + }, + { + "epoch": 1.655844155844156, + "grad_norm": 5.915123198373267, + "learning_rate": 1.905302866102445e-05, + "loss": 0.0914, + "step": 1275 + }, + { + "epoch": 1.657142857142857, + "grad_norm": 13.422603517538173, + "learning_rate": 1.9051241217007778e-05, + "loss": 0.1138, + "step": 1276 + }, + { + "epoch": 1.6584415584415586, + "grad_norm": 6.504865087751287, + "learning_rate": 1.904945217165329e-05, + "loss": 0.1031, + "step": 1277 + }, + { + "epoch": 1.6597402597402597, + "grad_norm": 29.15099562035742, + "learning_rate": 1.90476615252775e-05, + "loss": 0.092, + "step": 1278 + }, + { + "epoch": 1.661038961038961, + "grad_norm": 16.667715217735104, + "learning_rate": 1.9045869278197216e-05, + "loss": 0.0757, + "step": 1279 + }, + { + "epoch": 1.6623376623376624, + "grad_norm": 26.457857427796437, + "learning_rate": 1.904407543072952e-05, + "loss": 0.0705, + "step": 1280 + }, + { + "epoch": 1.6636363636363636, + "grad_norm": 4.045535817349237, + "learning_rate": 1.904227998319177e-05, + "loss": 0.0796, + "step": 1281 + }, + { + "epoch": 1.664935064935065, + "grad_norm": 28.269783105202325, + "learning_rate": 1.9040482935901624e-05, + "loss": 0.1051, + "step": 1282 + }, + { + "epoch": 1.6662337662337663, + "grad_norm": 10.094825730907534, + "learning_rate": 1.9038684289177008e-05, + "loss": 0.1154, + "step": 1283 + }, + { + "epoch": 1.6675324675324674, + "grad_norm": 21.316419359970634, + "learning_rate": 1.9036884043336137e-05, + "loss": 0.084, + "step": 1284 + }, + { + "epoch": 1.6688311688311688, + "grad_norm": 5.5004697449694495, + "learning_rate": 1.9035082198697515e-05, + "loss": 0.0988, + "step": 1285 + }, + { + "epoch": 1.6701298701298701, + "grad_norm": 22.85204459955462, + "learning_rate": 1.9033278755579917e-05, + "loss": 0.0857, + "step": 1286 + }, + { + "epoch": 1.6714285714285713, + "grad_norm": 10.937896892290654, + "learning_rate": 1.9031473714302405e-05, + "loss": 0.0842, + "step": 1287 + }, + { + "epoch": 1.6727272727272728, + "grad_norm": 28.247935066392397, + "learning_rate": 1.9029667075184333e-05, + "loss": 0.1207, + "step": 1288 + }, + { + "epoch": 1.674025974025974, + "grad_norm": 27.71881973787152, + "learning_rate": 1.9027858838545324e-05, + "loss": 0.1267, + "step": 1289 + }, + { + "epoch": 1.6753246753246753, + "grad_norm": 8.649361930902911, + "learning_rate": 1.9026049004705293e-05, + "loss": 0.0722, + "step": 1290 + }, + { + "epoch": 1.6766233766233767, + "grad_norm": 18.141074122116393, + "learning_rate": 1.902423757398443e-05, + "loss": 0.1064, + "step": 1291 + }, + { + "epoch": 1.6779220779220778, + "grad_norm": 11.439601898906188, + "learning_rate": 1.9022424546703214e-05, + "loss": 0.0922, + "step": 1292 + }, + { + "epoch": 1.6792207792207792, + "grad_norm": 26.711960028967937, + "learning_rate": 1.902060992318241e-05, + "loss": 0.0766, + "step": 1293 + }, + { + "epoch": 1.6805194805194805, + "grad_norm": 16.635922558449202, + "learning_rate": 1.901879370374305e-05, + "loss": 0.066, + "step": 1294 + }, + { + "epoch": 1.6818181818181817, + "grad_norm": 6.189648310759835, + "learning_rate": 1.9016975888706463e-05, + "loss": 0.0917, + "step": 1295 + }, + { + "epoch": 1.6831168831168832, + "grad_norm": 13.686733143896339, + "learning_rate": 1.901515647839425e-05, + "loss": 0.0922, + "step": 1296 + }, + { + "epoch": 1.6844155844155844, + "grad_norm": 31.67811839639656, + "learning_rate": 1.901333547312831e-05, + "loss": 0.0706, + "step": 1297 + }, + { + "epoch": 1.6857142857142857, + "grad_norm": 18.785164409544663, + "learning_rate": 1.9011512873230808e-05, + "loss": 0.1034, + "step": 1298 + }, + { + "epoch": 1.687012987012987, + "grad_norm": 8.794090699868258, + "learning_rate": 1.900968867902419e-05, + "loss": 0.0831, + "step": 1299 + }, + { + "epoch": 1.6883116883116882, + "grad_norm": 7.723609966077085, + "learning_rate": 1.9007862890831202e-05, + "loss": 0.1058, + "step": 1300 + }, + { + "epoch": 1.6896103896103896, + "grad_norm": 12.447418405099315, + "learning_rate": 1.9006035508974857e-05, + "loss": 0.0825, + "step": 1301 + }, + { + "epoch": 1.690909090909091, + "grad_norm": 15.529260150391979, + "learning_rate": 1.9004206533778452e-05, + "loss": 0.0829, + "step": 1302 + }, + { + "epoch": 1.692207792207792, + "grad_norm": 26.338868468954807, + "learning_rate": 1.900237596556557e-05, + "loss": 0.0907, + "step": 1303 + }, + { + "epoch": 1.6935064935064936, + "grad_norm": 20.732407898090656, + "learning_rate": 1.9000543804660073e-05, + "loss": 0.0913, + "step": 1304 + }, + { + "epoch": 1.6948051948051948, + "grad_norm": 39.464958552754155, + "learning_rate": 1.8998710051386103e-05, + "loss": 0.0774, + "step": 1305 + }, + { + "epoch": 1.6961038961038961, + "grad_norm": 7.845982332797729, + "learning_rate": 1.8996874706068088e-05, + "loss": 0.102, + "step": 1306 + }, + { + "epoch": 1.6974025974025975, + "grad_norm": 23.79055573235757, + "learning_rate": 1.899503776903074e-05, + "loss": 0.0861, + "step": 1307 + }, + { + "epoch": 1.6987012987012986, + "grad_norm": 20.41078474549511, + "learning_rate": 1.8993199240599035e-05, + "loss": 0.1263, + "step": 1308 + }, + { + "epoch": 1.7, + "grad_norm": 29.21853711864863, + "learning_rate": 1.8991359121098258e-05, + "loss": 0.1041, + "step": 1309 + }, + { + "epoch": 1.7012987012987013, + "grad_norm": 22.115826125476843, + "learning_rate": 1.8989517410853956e-05, + "loss": 0.1428, + "step": 1310 + }, + { + "epoch": 1.7025974025974024, + "grad_norm": 23.274723330068873, + "learning_rate": 1.898767411019196e-05, + "loss": 0.1094, + "step": 1311 + }, + { + "epoch": 1.703896103896104, + "grad_norm": 31.888602291562204, + "learning_rate": 1.898582921943839e-05, + "loss": 0.0849, + "step": 1312 + }, + { + "epoch": 1.7051948051948052, + "grad_norm": 31.38035437096442, + "learning_rate": 1.898398273891964e-05, + "loss": 0.1301, + "step": 1313 + }, + { + "epoch": 1.7064935064935065, + "grad_norm": 8.838556694591144, + "learning_rate": 1.898213466896239e-05, + "loss": 0.0898, + "step": 1314 + }, + { + "epoch": 1.7077922077922079, + "grad_norm": 9.858042034421453, + "learning_rate": 1.8980285009893596e-05, + "loss": 0.0704, + "step": 1315 + }, + { + "epoch": 1.709090909090909, + "grad_norm": 11.53553992086423, + "learning_rate": 1.8978433762040497e-05, + "loss": 0.1096, + "step": 1316 + }, + { + "epoch": 1.7103896103896103, + "grad_norm": 7.851333570193037, + "learning_rate": 1.897658092573062e-05, + "loss": 0.0779, + "step": 1317 + }, + { + "epoch": 1.7116883116883117, + "grad_norm": 20.317563541252984, + "learning_rate": 1.897472650129176e-05, + "loss": 0.0956, + "step": 1318 + }, + { + "epoch": 1.7129870129870128, + "grad_norm": 22.502655155446305, + "learning_rate": 1.897287048905201e-05, + "loss": 0.0871, + "step": 1319 + }, + { + "epoch": 1.7142857142857144, + "grad_norm": 12.50166887484553, + "learning_rate": 1.897101288933972e-05, + "loss": 0.0898, + "step": 1320 + }, + { + "epoch": 1.7155844155844155, + "grad_norm": 30.841245370101404, + "learning_rate": 1.896915370248355e-05, + "loss": 0.0938, + "step": 1321 + }, + { + "epoch": 1.716883116883117, + "grad_norm": 4.4800023263410225, + "learning_rate": 1.896729292881241e-05, + "loss": 0.0613, + "step": 1322 + }, + { + "epoch": 1.7181818181818183, + "grad_norm": 12.183085250285298, + "learning_rate": 1.8965430568655524e-05, + "loss": 0.0742, + "step": 1323 + }, + { + "epoch": 1.7194805194805194, + "grad_norm": 19.870085467041363, + "learning_rate": 1.8963566622342367e-05, + "loss": 0.1405, + "step": 1324 + }, + { + "epoch": 1.7207792207792207, + "grad_norm": 15.782600836172392, + "learning_rate": 1.8961701090202708e-05, + "loss": 0.1207, + "step": 1325 + }, + { + "epoch": 1.722077922077922, + "grad_norm": 26.977503262192037, + "learning_rate": 1.89598339725666e-05, + "loss": 0.0581, + "step": 1326 + }, + { + "epoch": 1.7233766233766232, + "grad_norm": 6.820692357487086, + "learning_rate": 1.8957965269764366e-05, + "loss": 0.0812, + "step": 1327 + }, + { + "epoch": 1.7246753246753248, + "grad_norm": 12.471791512988402, + "learning_rate": 1.895609498212662e-05, + "loss": 0.1148, + "step": 1328 + }, + { + "epoch": 1.725974025974026, + "grad_norm": 30.396216505538415, + "learning_rate": 1.8954223109984252e-05, + "loss": 0.1053, + "step": 1329 + }, + { + "epoch": 1.7272727272727273, + "grad_norm": 6.68680743850638, + "learning_rate": 1.895234965366843e-05, + "loss": 0.0903, + "step": 1330 + }, + { + "epoch": 1.7285714285714286, + "grad_norm": 46.15353484386625, + "learning_rate": 1.8950474613510605e-05, + "loss": 0.1311, + "step": 1331 + }, + { + "epoch": 1.7298701298701298, + "grad_norm": 14.904035189761615, + "learning_rate": 1.8948597989842504e-05, + "loss": 0.078, + "step": 1332 + }, + { + "epoch": 1.7311688311688311, + "grad_norm": 31.612508151720707, + "learning_rate": 1.894671978299614e-05, + "loss": 0.109, + "step": 1333 + }, + { + "epoch": 1.7324675324675325, + "grad_norm": 12.013847298764787, + "learning_rate": 1.894483999330381e-05, + "loss": 0.0766, + "step": 1334 + }, + { + "epoch": 1.7337662337662336, + "grad_norm": 39.890685829830204, + "learning_rate": 1.8942958621098072e-05, + "loss": 0.0825, + "step": 1335 + }, + { + "epoch": 1.7350649350649352, + "grad_norm": 21.950261828313675, + "learning_rate": 1.894107566671179e-05, + "loss": 0.0897, + "step": 1336 + }, + { + "epoch": 1.7363636363636363, + "grad_norm": 34.072627211347914, + "learning_rate": 1.8939191130478085e-05, + "loss": 0.0918, + "step": 1337 + }, + { + "epoch": 1.7376623376623377, + "grad_norm": 48.48457908325718, + "learning_rate": 1.8937305012730373e-05, + "loss": 0.1162, + "step": 1338 + }, + { + "epoch": 1.738961038961039, + "grad_norm": 55.32776379650705, + "learning_rate": 1.8935417313802344e-05, + "loss": 0.1057, + "step": 1339 + }, + { + "epoch": 1.7402597402597402, + "grad_norm": 31.71260327870839, + "learning_rate": 1.893352803402796e-05, + "loss": 0.0948, + "step": 1340 + }, + { + "epoch": 1.7415584415584415, + "grad_norm": 46.856691248805255, + "learning_rate": 1.8931637173741482e-05, + "loss": 0.1392, + "step": 1341 + }, + { + "epoch": 1.7428571428571429, + "grad_norm": 14.672466025801842, + "learning_rate": 1.8929744733277435e-05, + "loss": 0.0708, + "step": 1342 + }, + { + "epoch": 1.744155844155844, + "grad_norm": 23.334184412702246, + "learning_rate": 1.8927850712970626e-05, + "loss": 0.0904, + "step": 1343 + }, + { + "epoch": 1.7454545454545456, + "grad_norm": 28.742036086360056, + "learning_rate": 1.8925955113156146e-05, + "loss": 0.0813, + "step": 1344 + }, + { + "epoch": 1.7467532467532467, + "grad_norm": 30.85562096587275, + "learning_rate": 1.892405793416936e-05, + "loss": 0.0899, + "step": 1345 + }, + { + "epoch": 1.748051948051948, + "grad_norm": 22.257185035398084, + "learning_rate": 1.8922159176345918e-05, + "loss": 0.1011, + "step": 1346 + }, + { + "epoch": 1.7493506493506494, + "grad_norm": 11.832558978465732, + "learning_rate": 1.8920258840021745e-05, + "loss": 0.0745, + "step": 1347 + }, + { + "epoch": 1.7506493506493506, + "grad_norm": 35.13443260563388, + "learning_rate": 1.8918356925533052e-05, + "loss": 0.0848, + "step": 1348 + }, + { + "epoch": 1.751948051948052, + "grad_norm": 19.48698754334598, + "learning_rate": 1.8916453433216314e-05, + "loss": 0.0577, + "step": 1349 + }, + { + "epoch": 1.7532467532467533, + "grad_norm": 7.657915968552217, + "learning_rate": 1.8914548363408306e-05, + "loss": 0.1271, + "step": 1350 + }, + { + "epoch": 1.7545454545454544, + "grad_norm": 15.455440928138172, + "learning_rate": 1.8912641716446062e-05, + "loss": 0.0691, + "step": 1351 + }, + { + "epoch": 1.755844155844156, + "grad_norm": 4.66305316164773, + "learning_rate": 1.8910733492666907e-05, + "loss": 0.0694, + "step": 1352 + }, + { + "epoch": 1.7571428571428571, + "grad_norm": 29.36065031551231, + "learning_rate": 1.890882369240845e-05, + "loss": 0.0852, + "step": 1353 + }, + { + "epoch": 1.7584415584415585, + "grad_norm": 13.853151080584704, + "learning_rate": 1.890691231600856e-05, + "loss": 0.0638, + "step": 1354 + }, + { + "epoch": 1.7597402597402598, + "grad_norm": 12.209947143056777, + "learning_rate": 1.8904999363805404e-05, + "loss": 0.1071, + "step": 1355 + }, + { + "epoch": 1.761038961038961, + "grad_norm": 9.106915206230772, + "learning_rate": 1.890308483613742e-05, + "loss": 0.0977, + "step": 1356 + }, + { + "epoch": 1.7623376623376623, + "grad_norm": 23.74240569211131, + "learning_rate": 1.8901168733343315e-05, + "loss": 0.0866, + "step": 1357 + }, + { + "epoch": 1.7636363636363637, + "grad_norm": 35.09692731515073, + "learning_rate": 1.8899251055762096e-05, + "loss": 0.0861, + "step": 1358 + }, + { + "epoch": 1.7649350649350648, + "grad_norm": 6.368304994126352, + "learning_rate": 1.8897331803733027e-05, + "loss": 0.0666, + "step": 1359 + }, + { + "epoch": 1.7662337662337664, + "grad_norm": 25.067559570917915, + "learning_rate": 1.889541097759567e-05, + "loss": 0.0877, + "step": 1360 + }, + { + "epoch": 1.7675324675324675, + "grad_norm": 4.924432484131214, + "learning_rate": 1.8893488577689848e-05, + "loss": 0.0577, + "step": 1361 + }, + { + "epoch": 1.7688311688311689, + "grad_norm": 15.898823131574703, + "learning_rate": 1.8891564604355676e-05, + "loss": 0.0786, + "step": 1362 + }, + { + "epoch": 1.7701298701298702, + "grad_norm": 7.992326713459176, + "learning_rate": 1.8889639057933533e-05, + "loss": 0.0975, + "step": 1363 + }, + { + "epoch": 1.7714285714285714, + "grad_norm": 18.772527099451974, + "learning_rate": 1.8887711938764097e-05, + "loss": 0.1014, + "step": 1364 + }, + { + "epoch": 1.7727272727272727, + "grad_norm": 21.622581089583697, + "learning_rate": 1.8885783247188305e-05, + "loss": 0.0764, + "step": 1365 + }, + { + "epoch": 1.774025974025974, + "grad_norm": 10.589381753623668, + "learning_rate": 1.8883852983547383e-05, + "loss": 0.1109, + "step": 1366 + }, + { + "epoch": 1.7753246753246752, + "grad_norm": 12.583245975879464, + "learning_rate": 1.8881921148182827e-05, + "loss": 0.0795, + "step": 1367 + }, + { + "epoch": 1.7766233766233768, + "grad_norm": 7.473381994561002, + "learning_rate": 1.887998774143642e-05, + "loss": 0.0894, + "step": 1368 + }, + { + "epoch": 1.777922077922078, + "grad_norm": 17.83613958648671, + "learning_rate": 1.887805276365021e-05, + "loss": 0.0808, + "step": 1369 + }, + { + "epoch": 1.7792207792207793, + "grad_norm": 28.110884888278775, + "learning_rate": 1.8876116215166546e-05, + "loss": 0.1079, + "step": 1370 + }, + { + "epoch": 1.7805194805194806, + "grad_norm": 9.6435064231997, + "learning_rate": 1.887417809632803e-05, + "loss": 0.0633, + "step": 1371 + }, + { + "epoch": 1.7818181818181817, + "grad_norm": 12.497458202162248, + "learning_rate": 1.8872238407477555e-05, + "loss": 0.1158, + "step": 1372 + }, + { + "epoch": 1.783116883116883, + "grad_norm": 28.97426312159463, + "learning_rate": 1.887029714895829e-05, + "loss": 0.1055, + "step": 1373 + }, + { + "epoch": 1.7844155844155845, + "grad_norm": 23.32508016997953, + "learning_rate": 1.886835432111368e-05, + "loss": 0.1307, + "step": 1374 + }, + { + "epoch": 1.7857142857142856, + "grad_norm": 11.738269402195819, + "learning_rate": 1.886640992428745e-05, + "loss": 0.0607, + "step": 1375 + }, + { + "epoch": 1.7870129870129872, + "grad_norm": 15.742117664471724, + "learning_rate": 1.8864463958823598e-05, + "loss": 0.0624, + "step": 1376 + }, + { + "epoch": 1.7883116883116883, + "grad_norm": 9.314910524533428, + "learning_rate": 1.8862516425066403e-05, + "loss": 0.1184, + "step": 1377 + }, + { + "epoch": 1.7896103896103897, + "grad_norm": 5.7468019057576925, + "learning_rate": 1.886056732336042e-05, + "loss": 0.0765, + "step": 1378 + }, + { + "epoch": 1.790909090909091, + "grad_norm": 7.051446686383138, + "learning_rate": 1.885861665405049e-05, + "loss": 0.0891, + "step": 1379 + }, + { + "epoch": 1.7922077922077921, + "grad_norm": 16.76705011534224, + "learning_rate": 1.8856664417481716e-05, + "loss": 0.0815, + "step": 1380 + }, + { + "epoch": 1.7935064935064935, + "grad_norm": 4.948242901299116, + "learning_rate": 1.8854710613999487e-05, + "loss": 0.0488, + "step": 1381 + }, + { + "epoch": 1.7948051948051948, + "grad_norm": 29.95133390168027, + "learning_rate": 1.885275524394947e-05, + "loss": 0.0937, + "step": 1382 + }, + { + "epoch": 1.796103896103896, + "grad_norm": 13.940538395331254, + "learning_rate": 1.8850798307677607e-05, + "loss": 0.0727, + "step": 1383 + }, + { + "epoch": 1.7974025974025976, + "grad_norm": 15.003680630481782, + "learning_rate": 1.8848839805530117e-05, + "loss": 0.0869, + "step": 1384 + }, + { + "epoch": 1.7987012987012987, + "grad_norm": 6.951120448951751, + "learning_rate": 1.8846879737853497e-05, + "loss": 0.0687, + "step": 1385 + }, + { + "epoch": 1.8, + "grad_norm": 24.26065727050374, + "learning_rate": 1.884491810499452e-05, + "loss": 0.0924, + "step": 1386 + }, + { + "epoch": 1.8012987012987014, + "grad_norm": 12.532655243715746, + "learning_rate": 1.8842954907300236e-05, + "loss": 0.0758, + "step": 1387 + }, + { + "epoch": 1.8025974025974025, + "grad_norm": 4.488622566160861, + "learning_rate": 1.8840990145117978e-05, + "loss": 0.0481, + "step": 1388 + }, + { + "epoch": 1.8038961038961039, + "grad_norm": 7.197936002551289, + "learning_rate": 1.883902381879534e-05, + "loss": 0.0632, + "step": 1389 + }, + { + "epoch": 1.8051948051948052, + "grad_norm": 12.80257442144627, + "learning_rate": 1.8837055928680205e-05, + "loss": 0.0968, + "step": 1390 + }, + { + "epoch": 1.8064935064935064, + "grad_norm": 11.109869901152178, + "learning_rate": 1.883508647512074e-05, + "loss": 0.0794, + "step": 1391 + }, + { + "epoch": 1.807792207792208, + "grad_norm": 8.139956321195733, + "learning_rate": 1.8833115458465367e-05, + "loss": 0.0781, + "step": 1392 + }, + { + "epoch": 1.809090909090909, + "grad_norm": 7.653372030893311, + "learning_rate": 1.8831142879062805e-05, + "loss": 0.1204, + "step": 1393 + }, + { + "epoch": 1.8103896103896104, + "grad_norm": 15.427450979750926, + "learning_rate": 1.882916873726204e-05, + "loss": 0.0504, + "step": 1394 + }, + { + "epoch": 1.8116883116883118, + "grad_norm": 5.709243944312181, + "learning_rate": 1.8827193033412325e-05, + "loss": 0.0739, + "step": 1395 + }, + { + "epoch": 1.812987012987013, + "grad_norm": 9.546708648955306, + "learning_rate": 1.8825215767863215e-05, + "loss": 0.0696, + "step": 1396 + }, + { + "epoch": 1.8142857142857143, + "grad_norm": 10.45982316879138, + "learning_rate": 1.8823236940964515e-05, + "loss": 0.0985, + "step": 1397 + }, + { + "epoch": 1.8155844155844156, + "grad_norm": 19.340699083121688, + "learning_rate": 1.8821256553066326e-05, + "loss": 0.1191, + "step": 1398 + }, + { + "epoch": 1.8168831168831168, + "grad_norm": 17.059975812428526, + "learning_rate": 1.881927460451901e-05, + "loss": 0.0775, + "step": 1399 + }, + { + "epoch": 1.8181818181818183, + "grad_norm": 10.360086747697283, + "learning_rate": 1.881729109567321e-05, + "loss": 0.1195, + "step": 1400 + }, + { + "epoch": 1.8194805194805195, + "grad_norm": 7.9523267079067805, + "learning_rate": 1.8815306026879854e-05, + "loss": 0.067, + "step": 1401 + }, + { + "epoch": 1.8207792207792208, + "grad_norm": 9.672152123363244, + "learning_rate": 1.8813319398490137e-05, + "loss": 0.0888, + "step": 1402 + }, + { + "epoch": 1.8220779220779222, + "grad_norm": 15.325250876506152, + "learning_rate": 1.881133121085553e-05, + "loss": 0.0973, + "step": 1403 + }, + { + "epoch": 1.8233766233766233, + "grad_norm": 6.444780712113342, + "learning_rate": 1.8809341464327776e-05, + "loss": 0.0612, + "step": 1404 + }, + { + "epoch": 1.8246753246753247, + "grad_norm": 22.49140163273471, + "learning_rate": 1.880735015925891e-05, + "loss": 0.1013, + "step": 1405 + }, + { + "epoch": 1.825974025974026, + "grad_norm": 20.34150422003752, + "learning_rate": 1.8805357296001224e-05, + "loss": 0.1134, + "step": 1406 + }, + { + "epoch": 1.8272727272727272, + "grad_norm": 13.156587256470308, + "learning_rate": 1.8803362874907293e-05, + "loss": 0.0627, + "step": 1407 + }, + { + "epoch": 1.8285714285714287, + "grad_norm": 9.957914089496358, + "learning_rate": 1.8801366896329976e-05, + "loss": 0.068, + "step": 1408 + }, + { + "epoch": 1.8298701298701299, + "grad_norm": 8.197401488401475, + "learning_rate": 1.8799369360622394e-05, + "loss": 0.0977, + "step": 1409 + }, + { + "epoch": 1.8311688311688312, + "grad_norm": 10.84939236679237, + "learning_rate": 1.8797370268137948e-05, + "loss": 0.0631, + "step": 1410 + }, + { + "epoch": 1.8324675324675326, + "grad_norm": 13.628329052649935, + "learning_rate": 1.879536961923032e-05, + "loss": 0.0757, + "step": 1411 + }, + { + "epoch": 1.8337662337662337, + "grad_norm": 5.661960410066587, + "learning_rate": 1.8793367414253462e-05, + "loss": 0.0816, + "step": 1412 + }, + { + "epoch": 1.835064935064935, + "grad_norm": 10.406322910664281, + "learning_rate": 1.87913636535616e-05, + "loss": 0.0782, + "step": 1413 + }, + { + "epoch": 1.8363636363636364, + "grad_norm": 5.763562438242753, + "learning_rate": 1.8789358337509238e-05, + "loss": 0.0824, + "step": 1414 + }, + { + "epoch": 1.8376623376623376, + "grad_norm": 26.707652156664466, + "learning_rate": 1.8787351466451156e-05, + "loss": 0.0804, + "step": 1415 + }, + { + "epoch": 1.838961038961039, + "grad_norm": 14.986438750324735, + "learning_rate": 1.8785343040742412e-05, + "loss": 0.1089, + "step": 1416 + }, + { + "epoch": 1.8402597402597403, + "grad_norm": 46.60565168266797, + "learning_rate": 1.8783333060738328e-05, + "loss": 0.0928, + "step": 1417 + }, + { + "epoch": 1.8415584415584414, + "grad_norm": 5.96185884231025, + "learning_rate": 1.878132152679451e-05, + "loss": 0.0714, + "step": 1418 + }, + { + "epoch": 1.842857142857143, + "grad_norm": 34.087255955379696, + "learning_rate": 1.8779308439266838e-05, + "loss": 0.0814, + "step": 1419 + }, + { + "epoch": 1.844155844155844, + "grad_norm": 17.766057639562124, + "learning_rate": 1.8777293798511466e-05, + "loss": 0.1005, + "step": 1420 + }, + { + "epoch": 1.8454545454545455, + "grad_norm": 16.263153636547187, + "learning_rate": 1.8775277604884822e-05, + "loss": 0.0705, + "step": 1421 + }, + { + "epoch": 1.8467532467532468, + "grad_norm": 38.92123650630713, + "learning_rate": 1.8773259858743608e-05, + "loss": 0.0805, + "step": 1422 + }, + { + "epoch": 1.848051948051948, + "grad_norm": 10.303279998782157, + "learning_rate": 1.8771240560444805e-05, + "loss": 0.1086, + "step": 1423 + }, + { + "epoch": 1.8493506493506493, + "grad_norm": 40.44025210097619, + "learning_rate": 1.876921971034566e-05, + "loss": 0.1007, + "step": 1424 + }, + { + "epoch": 1.8506493506493507, + "grad_norm": 24.268176704209303, + "learning_rate": 1.8767197308803707e-05, + "loss": 0.102, + "step": 1425 + }, + { + "epoch": 1.8519480519480518, + "grad_norm": 49.73065871374784, + "learning_rate": 1.8765173356176745e-05, + "loss": 0.0783, + "step": 1426 + }, + { + "epoch": 1.8532467532467534, + "grad_norm": 34.621978411885486, + "learning_rate": 1.8763147852822846e-05, + "loss": 0.0964, + "step": 1427 + }, + { + "epoch": 1.8545454545454545, + "grad_norm": 46.10264817535833, + "learning_rate": 1.8761120799100366e-05, + "loss": 0.0873, + "step": 1428 + }, + { + "epoch": 1.8558441558441559, + "grad_norm": 40.05202845009524, + "learning_rate": 1.8759092195367925e-05, + "loss": 0.1124, + "step": 1429 + }, + { + "epoch": 1.8571428571428572, + "grad_norm": 32.31569259908495, + "learning_rate": 1.8757062041984424e-05, + "loss": 0.1266, + "step": 1430 + }, + { + "epoch": 1.8584415584415583, + "grad_norm": 40.915398796276605, + "learning_rate": 1.8755030339309037e-05, + "loss": 0.0835, + "step": 1431 + }, + { + "epoch": 1.8597402597402597, + "grad_norm": 17.07622155994077, + "learning_rate": 1.875299708770121e-05, + "loss": 0.0843, + "step": 1432 + }, + { + "epoch": 1.861038961038961, + "grad_norm": 43.67304204870263, + "learning_rate": 1.8750962287520658e-05, + "loss": 0.0827, + "step": 1433 + }, + { + "epoch": 1.8623376623376622, + "grad_norm": 8.664670395108397, + "learning_rate": 1.874892593912739e-05, + "loss": 0.1074, + "step": 1434 + }, + { + "epoch": 1.8636363636363638, + "grad_norm": 47.12025932917653, + "learning_rate": 1.8746888042881662e-05, + "loss": 0.0771, + "step": 1435 + }, + { + "epoch": 1.864935064935065, + "grad_norm": 6.012182674981127, + "learning_rate": 1.8744848599144027e-05, + "loss": 0.0895, + "step": 1436 + }, + { + "epoch": 1.8662337662337662, + "grad_norm": 35.26800914075537, + "learning_rate": 1.8742807608275293e-05, + "loss": 0.0859, + "step": 1437 + }, + { + "epoch": 1.8675324675324676, + "grad_norm": 13.631377724811129, + "learning_rate": 1.8740765070636557e-05, + "loss": 0.0949, + "step": 1438 + }, + { + "epoch": 1.8688311688311687, + "grad_norm": 31.426207938229904, + "learning_rate": 1.8738720986589176e-05, + "loss": 0.0989, + "step": 1439 + }, + { + "epoch": 1.87012987012987, + "grad_norm": 12.511489973561597, + "learning_rate": 1.8736675356494793e-05, + "loss": 0.0997, + "step": 1440 + }, + { + "epoch": 1.8714285714285714, + "grad_norm": 13.163067846966571, + "learning_rate": 1.8734628180715318e-05, + "loss": 0.0623, + "step": 1441 + }, + { + "epoch": 1.8727272727272726, + "grad_norm": 19.806751553797287, + "learning_rate": 1.8732579459612935e-05, + "loss": 0.0841, + "step": 1442 + }, + { + "epoch": 1.8740259740259742, + "grad_norm": 16.59872543571445, + "learning_rate": 1.8730529193550104e-05, + "loss": 0.1028, + "step": 1443 + }, + { + "epoch": 1.8753246753246753, + "grad_norm": 8.462853061124404, + "learning_rate": 1.8728477382889557e-05, + "loss": 0.0851, + "step": 1444 + }, + { + "epoch": 1.8766233766233766, + "grad_norm": 32.02937423008012, + "learning_rate": 1.872642402799429e-05, + "loss": 0.1451, + "step": 1445 + }, + { + "epoch": 1.877922077922078, + "grad_norm": 29.525927747398047, + "learning_rate": 1.872436912922759e-05, + "loss": 0.0771, + "step": 1446 + }, + { + "epoch": 1.8792207792207791, + "grad_norm": 14.558801390709398, + "learning_rate": 1.8722312686953005e-05, + "loss": 0.1102, + "step": 1447 + }, + { + "epoch": 1.8805194805194805, + "grad_norm": 35.787700800623014, + "learning_rate": 1.8720254701534356e-05, + "loss": 0.0931, + "step": 1448 + }, + { + "epoch": 1.8818181818181818, + "grad_norm": 6.522565126614798, + "learning_rate": 1.8718195173335744e-05, + "loss": 0.0705, + "step": 1449 + }, + { + "epoch": 1.883116883116883, + "grad_norm": 20.66010970978529, + "learning_rate": 1.871613410272154e-05, + "loss": 0.0953, + "step": 1450 + }, + { + "epoch": 1.8844155844155845, + "grad_norm": 10.534170835787963, + "learning_rate": 1.8714071490056382e-05, + "loss": 0.0945, + "step": 1451 + }, + { + "epoch": 1.8857142857142857, + "grad_norm": 13.076041570047815, + "learning_rate": 1.871200733570519e-05, + "loss": 0.0782, + "step": 1452 + }, + { + "epoch": 1.887012987012987, + "grad_norm": 9.1089694765872, + "learning_rate": 1.8709941640033153e-05, + "loss": 0.0941, + "step": 1453 + }, + { + "epoch": 1.8883116883116884, + "grad_norm": 9.397130602436608, + "learning_rate": 1.8707874403405726e-05, + "loss": 0.0679, + "step": 1454 + }, + { + "epoch": 1.8896103896103895, + "grad_norm": 8.09918531279857, + "learning_rate": 1.8705805626188646e-05, + "loss": 0.0817, + "step": 1455 + }, + { + "epoch": 1.8909090909090909, + "grad_norm": 17.908011828978015, + "learning_rate": 1.870373530874792e-05, + "loss": 0.0707, + "step": 1456 + }, + { + "epoch": 1.8922077922077922, + "grad_norm": 18.183627860253537, + "learning_rate": 1.870166345144983e-05, + "loss": 0.0912, + "step": 1457 + }, + { + "epoch": 1.8935064935064934, + "grad_norm": 7.988061111419561, + "learning_rate": 1.8699590054660922e-05, + "loss": 0.087, + "step": 1458 + }, + { + "epoch": 1.894805194805195, + "grad_norm": 11.391705780997913, + "learning_rate": 1.8697515118748022e-05, + "loss": 0.0837, + "step": 1459 + }, + { + "epoch": 1.896103896103896, + "grad_norm": 10.988806739585417, + "learning_rate": 1.8695438644078227e-05, + "loss": 0.0868, + "step": 1460 + }, + { + "epoch": 1.8974025974025974, + "grad_norm": 5.261983652143334, + "learning_rate": 1.86933606310189e-05, + "loss": 0.0888, + "step": 1461 + }, + { + "epoch": 1.8987012987012988, + "grad_norm": 20.441496265340383, + "learning_rate": 1.8691281079937684e-05, + "loss": 0.1097, + "step": 1462 + }, + { + "epoch": 1.9, + "grad_norm": 8.762333518255904, + "learning_rate": 1.8689199991202493e-05, + "loss": 0.0691, + "step": 1463 + }, + { + "epoch": 1.9012987012987013, + "grad_norm": 18.40356110319363, + "learning_rate": 1.8687117365181514e-05, + "loss": 0.0719, + "step": 1464 + }, + { + "epoch": 1.9025974025974026, + "grad_norm": 12.274887493753537, + "learning_rate": 1.8685033202243196e-05, + "loss": 0.0726, + "step": 1465 + }, + { + "epoch": 1.9038961038961038, + "grad_norm": 8.85805706995349, + "learning_rate": 1.8682947502756273e-05, + "loss": 0.0732, + "step": 1466 + }, + { + "epoch": 1.9051948051948053, + "grad_norm": 24.699922543496925, + "learning_rate": 1.8680860267089742e-05, + "loss": 0.0663, + "step": 1467 + }, + { + "epoch": 1.9064935064935065, + "grad_norm": 15.821035278520396, + "learning_rate": 1.8678771495612874e-05, + "loss": 0.1217, + "step": 1468 + }, + { + "epoch": 1.9077922077922078, + "grad_norm": 13.266556457368077, + "learning_rate": 1.8676681188695222e-05, + "loss": 0.0867, + "step": 1469 + }, + { + "epoch": 1.9090909090909092, + "grad_norm": 7.328839230662229, + "learning_rate": 1.867458934670659e-05, + "loss": 0.1231, + "step": 1470 + }, + { + "epoch": 1.9103896103896103, + "grad_norm": 30.40750232979771, + "learning_rate": 1.8672495970017067e-05, + "loss": 0.0882, + "step": 1471 + }, + { + "epoch": 1.9116883116883117, + "grad_norm": 21.395197368324435, + "learning_rate": 1.867040105899702e-05, + "loss": 0.0803, + "step": 1472 + }, + { + "epoch": 1.912987012987013, + "grad_norm": 18.83773463976758, + "learning_rate": 1.8668304614017067e-05, + "loss": 0.0954, + "step": 1473 + }, + { + "epoch": 1.9142857142857141, + "grad_norm": 28.48814792299126, + "learning_rate": 1.8666206635448116e-05, + "loss": 0.1365, + "step": 1474 + }, + { + "epoch": 1.9155844155844157, + "grad_norm": 15.57114579409072, + "learning_rate": 1.8664107123661337e-05, + "loss": 0.0837, + "step": 1475 + }, + { + "epoch": 1.9168831168831169, + "grad_norm": 15.50438809899305, + "learning_rate": 1.8662006079028176e-05, + "loss": 0.102, + "step": 1476 + }, + { + "epoch": 1.9181818181818182, + "grad_norm": 21.548559441476193, + "learning_rate": 1.8659903501920346e-05, + "loss": 0.0556, + "step": 1477 + }, + { + "epoch": 1.9194805194805196, + "grad_norm": 7.9947492840155245, + "learning_rate": 1.8657799392709838e-05, + "loss": 0.0837, + "step": 1478 + }, + { + "epoch": 1.9207792207792207, + "grad_norm": 25.638008046210786, + "learning_rate": 1.8655693751768902e-05, + "loss": 0.0788, + "step": 1479 + }, + { + "epoch": 1.922077922077922, + "grad_norm": 9.566635638148023, + "learning_rate": 1.865358657947007e-05, + "loss": 0.0865, + "step": 1480 + }, + { + "epoch": 1.9233766233766234, + "grad_norm": 16.252179623152397, + "learning_rate": 1.865147787618614e-05, + "loss": 0.0802, + "step": 1481 + }, + { + "epoch": 1.9246753246753245, + "grad_norm": 8.106825593308077, + "learning_rate": 1.8649367642290187e-05, + "loss": 0.0895, + "step": 1482 + }, + { + "epoch": 1.9259740259740261, + "grad_norm": 13.880225325419051, + "learning_rate": 1.8647255878155544e-05, + "loss": 0.1135, + "step": 1483 + }, + { + "epoch": 1.9272727272727272, + "grad_norm": 26.534271346693437, + "learning_rate": 1.8645142584155825e-05, + "loss": 0.0836, + "step": 1484 + }, + { + "epoch": 1.9285714285714286, + "grad_norm": 39.040041279425246, + "learning_rate": 1.8643027760664916e-05, + "loss": 0.0656, + "step": 1485 + }, + { + "epoch": 1.92987012987013, + "grad_norm": 13.880776091139214, + "learning_rate": 1.8640911408056963e-05, + "loss": 0.0851, + "step": 1486 + }, + { + "epoch": 1.931168831168831, + "grad_norm": 14.015353705189543, + "learning_rate": 1.8638793526706397e-05, + "loss": 0.0965, + "step": 1487 + }, + { + "epoch": 1.9324675324675324, + "grad_norm": 39.60297070267386, + "learning_rate": 1.8636674116987904e-05, + "loss": 0.0876, + "step": 1488 + }, + { + "epoch": 1.9337662337662338, + "grad_norm": 9.006977351887254, + "learning_rate": 1.8634553179276455e-05, + "loss": 0.0892, + "step": 1489 + }, + { + "epoch": 1.935064935064935, + "grad_norm": 70.42412689409267, + "learning_rate": 1.8632430713947283e-05, + "loss": 0.0885, + "step": 1490 + }, + { + "epoch": 1.9363636363636365, + "grad_norm": 15.907520421919669, + "learning_rate": 1.8630306721375887e-05, + "loss": 0.1169, + "step": 1491 + }, + { + "epoch": 1.9376623376623376, + "grad_norm": 50.83319634946522, + "learning_rate": 1.8628181201938045e-05, + "loss": 0.1182, + "step": 1492 + }, + { + "epoch": 1.938961038961039, + "grad_norm": 13.98302675250485, + "learning_rate": 1.8626054156009807e-05, + "loss": 0.1312, + "step": 1493 + }, + { + "epoch": 1.9402597402597404, + "grad_norm": 21.64779989936077, + "learning_rate": 1.8623925583967483e-05, + "loss": 0.1021, + "step": 1494 + }, + { + "epoch": 1.9415584415584415, + "grad_norm": 33.4448162591572, + "learning_rate": 1.8621795486187664e-05, + "loss": 0.0726, + "step": 1495 + }, + { + "epoch": 1.9428571428571428, + "grad_norm": 48.121191640798635, + "learning_rate": 1.8619663863047196e-05, + "loss": 0.0915, + "step": 1496 + }, + { + "epoch": 1.9441558441558442, + "grad_norm": 12.354494172462525, + "learning_rate": 1.861753071492321e-05, + "loss": 0.0914, + "step": 1497 + }, + { + "epoch": 1.9454545454545453, + "grad_norm": 18.618924199913096, + "learning_rate": 1.8615396042193093e-05, + "loss": 0.088, + "step": 1498 + }, + { + "epoch": 1.946753246753247, + "grad_norm": 28.90227802384556, + "learning_rate": 1.8613259845234525e-05, + "loss": 0.0679, + "step": 1499 + }, + { + "epoch": 1.948051948051948, + "grad_norm": 6.820476598118537, + "learning_rate": 1.8611122124425425e-05, + "loss": 0.1003, + "step": 1500 + }, + { + "epoch": 1.9493506493506494, + "grad_norm": 18.056179710390815, + "learning_rate": 1.8608982880144006e-05, + "loss": 0.0941, + "step": 1501 + }, + { + "epoch": 1.9506493506493507, + "grad_norm": 20.49102530146298, + "learning_rate": 1.8606842112768736e-05, + "loss": 0.0706, + "step": 1502 + }, + { + "epoch": 1.9519480519480519, + "grad_norm": 21.649441556107686, + "learning_rate": 1.8604699822678357e-05, + "loss": 0.0923, + "step": 1503 + }, + { + "epoch": 1.9532467532467532, + "grad_norm": 12.765806808841402, + "learning_rate": 1.8602556010251887e-05, + "loss": 0.0651, + "step": 1504 + }, + { + "epoch": 1.9545454545454546, + "grad_norm": 7.91068748402224, + "learning_rate": 1.8600410675868605e-05, + "loss": 0.092, + "step": 1505 + }, + { + "epoch": 1.9558441558441557, + "grad_norm": 26.14169814571844, + "learning_rate": 1.859826381990806e-05, + "loss": 0.1162, + "step": 1506 + }, + { + "epoch": 1.9571428571428573, + "grad_norm": 8.50277849044967, + "learning_rate": 1.8596115442750072e-05, + "loss": 0.1056, + "step": 1507 + }, + { + "epoch": 1.9584415584415584, + "grad_norm": 25.393750724114, + "learning_rate": 1.859396554477473e-05, + "loss": 0.0706, + "step": 1508 + }, + { + "epoch": 1.9597402597402598, + "grad_norm": 9.79625495971635, + "learning_rate": 1.8591814126362397e-05, + "loss": 0.0975, + "step": 1509 + }, + { + "epoch": 1.9610389610389611, + "grad_norm": 28.262661827754453, + "learning_rate": 1.858966118789369e-05, + "loss": 0.0762, + "step": 1510 + }, + { + "epoch": 1.9623376623376623, + "grad_norm": 26.498523005035874, + "learning_rate": 1.8587506729749514e-05, + "loss": 0.1602, + "step": 1511 + }, + { + "epoch": 1.9636363636363636, + "grad_norm": 13.104292754451608, + "learning_rate": 1.858535075231103e-05, + "loss": 0.0984, + "step": 1512 + }, + { + "epoch": 1.964935064935065, + "grad_norm": 28.3238670528527, + "learning_rate": 1.858319325595967e-05, + "loss": 0.1176, + "step": 1513 + }, + { + "epoch": 1.9662337662337661, + "grad_norm": 10.34137287762845, + "learning_rate": 1.8581034241077144e-05, + "loss": 0.1157, + "step": 1514 + }, + { + "epoch": 1.9675324675324677, + "grad_norm": 16.805499836160916, + "learning_rate": 1.8578873708045417e-05, + "loss": 0.0898, + "step": 1515 + }, + { + "epoch": 1.9688311688311688, + "grad_norm": 7.439873051624129, + "learning_rate": 1.857671165724673e-05, + "loss": 0.0861, + "step": 1516 + }, + { + "epoch": 1.9701298701298702, + "grad_norm": 20.6604661501403, + "learning_rate": 1.8574548089063588e-05, + "loss": 0.1288, + "step": 1517 + }, + { + "epoch": 1.9714285714285715, + "grad_norm": 25.45803377766219, + "learning_rate": 1.8572383003878773e-05, + "loss": 0.0902, + "step": 1518 + }, + { + "epoch": 1.9727272727272727, + "grad_norm": 29.48528312713208, + "learning_rate": 1.8570216402075326e-05, + "loss": 0.0514, + "step": 1519 + }, + { + "epoch": 1.974025974025974, + "grad_norm": 36.099662494791296, + "learning_rate": 1.856804828403656e-05, + "loss": 0.0958, + "step": 1520 + }, + { + "epoch": 1.9753246753246754, + "grad_norm": 28.49502008813064, + "learning_rate": 1.8565878650146062e-05, + "loss": 0.0464, + "step": 1521 + }, + { + "epoch": 1.9766233766233765, + "grad_norm": 27.574993345315434, + "learning_rate": 1.8563707500787677e-05, + "loss": 0.0809, + "step": 1522 + }, + { + "epoch": 1.977922077922078, + "grad_norm": 32.459343105045015, + "learning_rate": 1.8561534836345524e-05, + "loss": 0.0819, + "step": 1523 + }, + { + "epoch": 1.9792207792207792, + "grad_norm": 28.47818946833476, + "learning_rate": 1.8559360657203988e-05, + "loss": 0.0883, + "step": 1524 + }, + { + "epoch": 1.9805194805194806, + "grad_norm": 14.923105749619763, + "learning_rate": 1.8557184963747727e-05, + "loss": 0.076, + "step": 1525 + }, + { + "epoch": 1.981818181818182, + "grad_norm": 23.34303949268518, + "learning_rate": 1.8555007756361655e-05, + "loss": 0.0876, + "step": 1526 + }, + { + "epoch": 1.983116883116883, + "grad_norm": 10.432618400789273, + "learning_rate": 1.8552829035430967e-05, + "loss": 0.0849, + "step": 1527 + }, + { + "epoch": 1.9844155844155844, + "grad_norm": 3.9476023075814677, + "learning_rate": 1.8550648801341123e-05, + "loss": 0.0633, + "step": 1528 + }, + { + "epoch": 1.9857142857142858, + "grad_norm": 9.525461099591546, + "learning_rate": 1.8548467054477842e-05, + "loss": 0.0403, + "step": 1529 + }, + { + "epoch": 1.987012987012987, + "grad_norm": 16.73328805972854, + "learning_rate": 1.854628379522712e-05, + "loss": 0.1171, + "step": 1530 + }, + { + "epoch": 1.9883116883116885, + "grad_norm": 15.385241212616041, + "learning_rate": 1.8544099023975218e-05, + "loss": 0.1148, + "step": 1531 + }, + { + "epoch": 1.9896103896103896, + "grad_norm": 8.86758183278733, + "learning_rate": 1.8541912741108664e-05, + "loss": 0.0828, + "step": 1532 + }, + { + "epoch": 1.990909090909091, + "grad_norm": 9.710400593774068, + "learning_rate": 1.853972494701425e-05, + "loss": 0.0937, + "step": 1533 + }, + { + "epoch": 1.9922077922077923, + "grad_norm": 13.461324144892338, + "learning_rate": 1.8537535642079044e-05, + "loss": 0.0808, + "step": 1534 + }, + { + "epoch": 1.9935064935064934, + "grad_norm": 11.361819435392004, + "learning_rate": 1.8535344826690366e-05, + "loss": 0.0837, + "step": 1535 + }, + { + "epoch": 1.9948051948051948, + "grad_norm": 6.0310913188904705, + "learning_rate": 1.853315250123583e-05, + "loss": 0.0757, + "step": 1536 + }, + { + "epoch": 1.9961038961038962, + "grad_norm": 12.903304029911205, + "learning_rate": 1.8530958666103282e-05, + "loss": 0.0946, + "step": 1537 + }, + { + "epoch": 1.9974025974025973, + "grad_norm": 5.977284823236082, + "learning_rate": 1.8528763321680863e-05, + "loss": 0.0909, + "step": 1538 + }, + { + "epoch": 1.9987012987012989, + "grad_norm": 12.18057136598155, + "learning_rate": 1.852656646835697e-05, + "loss": 0.0898, + "step": 1539 + }, + { + "epoch": 2.0, + "grad_norm": 11.053591969153766, + "learning_rate": 1.852436810652027e-05, + "loss": 0.0746, + "step": 1540 + }, + { + "epoch": 2.0, + "eval_accuracy": 0.9152161706906232, + "eval_f1": 0.8929712338768159, + "eval_loss": 0.12412014603614807, + "eval_precision": 0.8943718085328975, + "eval_recall": 0.9145490130559654, + "eval_runtime": 13.1459, + "eval_samples_per_second": 135.479, + "eval_steps_per_second": 1.065, + "step": 1540 + }, + { + "epoch": 2.001298701298701, + "grad_norm": 7.515851259123599, + "learning_rate": 1.8522168236559693e-05, + "loss": 0.0715, + "step": 1541 + }, + { + "epoch": 2.0025974025974027, + "grad_norm": 4.960923855713699, + "learning_rate": 1.851996685886444e-05, + "loss": 0.0292, + "step": 1542 + }, + { + "epoch": 2.003896103896104, + "grad_norm": 6.553160765122292, + "learning_rate": 1.8517763973823977e-05, + "loss": 0.0857, + "step": 1543 + }, + { + "epoch": 2.005194805194805, + "grad_norm": 7.499241212663543, + "learning_rate": 1.8515559581828033e-05, + "loss": 0.0742, + "step": 1544 + }, + { + "epoch": 2.0064935064935066, + "grad_norm": 9.582561962855058, + "learning_rate": 1.851335368326661e-05, + "loss": 0.0773, + "step": 1545 + }, + { + "epoch": 2.0077922077922077, + "grad_norm": 9.119726682860604, + "learning_rate": 1.851114627852997e-05, + "loss": 0.0953, + "step": 1546 + }, + { + "epoch": 2.0090909090909093, + "grad_norm": 12.334115554206047, + "learning_rate": 1.8508937368008656e-05, + "loss": 0.0758, + "step": 1547 + }, + { + "epoch": 2.0103896103896104, + "grad_norm": 12.492886047587662, + "learning_rate": 1.8506726952093454e-05, + "loss": 0.0429, + "step": 1548 + }, + { + "epoch": 2.0116883116883115, + "grad_norm": 24.473965305598934, + "learning_rate": 1.850451503117543e-05, + "loss": 0.0708, + "step": 1549 + }, + { + "epoch": 2.012987012987013, + "grad_norm": 22.12390000276389, + "learning_rate": 1.850230160564592e-05, + "loss": 0.0565, + "step": 1550 + }, + { + "epoch": 2.0142857142857142, + "grad_norm": 26.900940465263762, + "learning_rate": 1.8500086675896528e-05, + "loss": 0.0866, + "step": 1551 + }, + { + "epoch": 2.0155844155844154, + "grad_norm": 22.86591904484376, + "learning_rate": 1.8497870242319103e-05, + "loss": 0.0815, + "step": 1552 + }, + { + "epoch": 2.016883116883117, + "grad_norm": 26.615177454839795, + "learning_rate": 1.8495652305305784e-05, + "loss": 0.1012, + "step": 1553 + }, + { + "epoch": 2.018181818181818, + "grad_norm": 4.57012860087959, + "learning_rate": 1.8493432865248962e-05, + "loss": 0.0654, + "step": 1554 + }, + { + "epoch": 2.0194805194805197, + "grad_norm": 12.612114066969387, + "learning_rate": 1.84912119225413e-05, + "loss": 0.0643, + "step": 1555 + }, + { + "epoch": 2.020779220779221, + "grad_norm": 19.30850063048671, + "learning_rate": 1.8488989477575724e-05, + "loss": 0.0555, + "step": 1556 + }, + { + "epoch": 2.022077922077922, + "grad_norm": 10.389872910812135, + "learning_rate": 1.8486765530745432e-05, + "loss": 0.0713, + "step": 1557 + }, + { + "epoch": 2.0233766233766235, + "grad_norm": 8.141110431409922, + "learning_rate": 1.8484540082443877e-05, + "loss": 0.0644, + "step": 1558 + }, + { + "epoch": 2.0246753246753246, + "grad_norm": 11.894441640463816, + "learning_rate": 1.8482313133064783e-05, + "loss": 0.0444, + "step": 1559 + }, + { + "epoch": 2.0259740259740258, + "grad_norm": 19.467866074818076, + "learning_rate": 1.848008468300215e-05, + "loss": 0.062, + "step": 1560 + }, + { + "epoch": 2.0272727272727273, + "grad_norm": 41.43629414277739, + "learning_rate": 1.847785473265022e-05, + "loss": 0.0706, + "step": 1561 + }, + { + "epoch": 2.0285714285714285, + "grad_norm": 7.331338531990216, + "learning_rate": 1.847562328240352e-05, + "loss": 0.0706, + "step": 1562 + }, + { + "epoch": 2.02987012987013, + "grad_norm": 16.012602398705475, + "learning_rate": 1.847339033265684e-05, + "loss": 0.0497, + "step": 1563 + }, + { + "epoch": 2.031168831168831, + "grad_norm": 13.508566498074776, + "learning_rate": 1.8471155883805223e-05, + "loss": 0.0393, + "step": 1564 + }, + { + "epoch": 2.0324675324675323, + "grad_norm": 17.731281146777604, + "learning_rate": 1.8468919936243993e-05, + "loss": 0.0797, + "step": 1565 + }, + { + "epoch": 2.033766233766234, + "grad_norm": 15.972622202340988, + "learning_rate": 1.8466682490368727e-05, + "loss": 0.067, + "step": 1566 + }, + { + "epoch": 2.035064935064935, + "grad_norm": 16.758145138558667, + "learning_rate": 1.846444354657528e-05, + "loss": 0.0643, + "step": 1567 + }, + { + "epoch": 2.036363636363636, + "grad_norm": 13.154026774146782, + "learning_rate": 1.8462203105259753e-05, + "loss": 0.0956, + "step": 1568 + }, + { + "epoch": 2.0376623376623377, + "grad_norm": 17.86422696412163, + "learning_rate": 1.845996116681853e-05, + "loss": 0.0419, + "step": 1569 + }, + { + "epoch": 2.038961038961039, + "grad_norm": 6.927771051055488, + "learning_rate": 1.8457717731648253e-05, + "loss": 0.0605, + "step": 1570 + }, + { + "epoch": 2.0402597402597404, + "grad_norm": 6.172998400850031, + "learning_rate": 1.8455472800145825e-05, + "loss": 0.0643, + "step": 1571 + }, + { + "epoch": 2.0415584415584416, + "grad_norm": 20.099610374476487, + "learning_rate": 1.845322637270842e-05, + "loss": 0.0691, + "step": 1572 + }, + { + "epoch": 2.0428571428571427, + "grad_norm": 23.29424471594762, + "learning_rate": 1.8450978449733475e-05, + "loss": 0.0871, + "step": 1573 + }, + { + "epoch": 2.0441558441558443, + "grad_norm": 23.529270950651618, + "learning_rate": 1.8448729031618687e-05, + "loss": 0.0656, + "step": 1574 + }, + { + "epoch": 2.0454545454545454, + "grad_norm": 15.752781228780327, + "learning_rate": 1.8446478118762022e-05, + "loss": 0.0853, + "step": 1575 + }, + { + "epoch": 2.0467532467532465, + "grad_norm": 48.095544321604386, + "learning_rate": 1.8444225711561713e-05, + "loss": 0.0838, + "step": 1576 + }, + { + "epoch": 2.048051948051948, + "grad_norm": 7.742108987032386, + "learning_rate": 1.844197181041625e-05, + "loss": 0.0573, + "step": 1577 + }, + { + "epoch": 2.0493506493506493, + "grad_norm": 74.13957973688461, + "learning_rate": 1.843971641572439e-05, + "loss": 0.1, + "step": 1578 + }, + { + "epoch": 2.050649350649351, + "grad_norm": 6.854681611520208, + "learning_rate": 1.8437459527885158e-05, + "loss": 0.0457, + "step": 1579 + }, + { + "epoch": 2.051948051948052, + "grad_norm": 61.50536906401384, + "learning_rate": 1.8435201147297846e-05, + "loss": 0.1089, + "step": 1580 + }, + { + "epoch": 2.053246753246753, + "grad_norm": 40.95691246638915, + "learning_rate": 1.8432941274361995e-05, + "loss": 0.0379, + "step": 1581 + }, + { + "epoch": 2.0545454545454547, + "grad_norm": 72.11346600052715, + "learning_rate": 1.8430679909477428e-05, + "loss": 0.1289, + "step": 1582 + }, + { + "epoch": 2.055844155844156, + "grad_norm": 35.247979654285835, + "learning_rate": 1.8428417053044216e-05, + "loss": 0.0771, + "step": 1583 + }, + { + "epoch": 2.057142857142857, + "grad_norm": 38.27192729743101, + "learning_rate": 1.842615270546271e-05, + "loss": 0.0623, + "step": 1584 + }, + { + "epoch": 2.0584415584415585, + "grad_norm": 53.63387090349023, + "learning_rate": 1.842388686713351e-05, + "loss": 0.0886, + "step": 1585 + }, + { + "epoch": 2.0597402597402596, + "grad_norm": 21.968636515722853, + "learning_rate": 1.8421619538457488e-05, + "loss": 0.0675, + "step": 1586 + }, + { + "epoch": 2.0610389610389612, + "grad_norm": 52.861542611849174, + "learning_rate": 1.8419350719835777e-05, + "loss": 0.0629, + "step": 1587 + }, + { + "epoch": 2.0623376623376624, + "grad_norm": 6.27349012527968, + "learning_rate": 1.8417080411669778e-05, + "loss": 0.0395, + "step": 1588 + }, + { + "epoch": 2.0636363636363635, + "grad_norm": 42.95147893483459, + "learning_rate": 1.841480861436115e-05, + "loss": 0.1066, + "step": 1589 + }, + { + "epoch": 2.064935064935065, + "grad_norm": 37.46237887813078, + "learning_rate": 1.8412535328311813e-05, + "loss": 0.1119, + "step": 1590 + }, + { + "epoch": 2.066233766233766, + "grad_norm": 44.53042862867565, + "learning_rate": 1.841026055392396e-05, + "loss": 0.0618, + "step": 1591 + }, + { + "epoch": 2.0675324675324673, + "grad_norm": 39.49139675046865, + "learning_rate": 1.8407984291600044e-05, + "loss": 0.0704, + "step": 1592 + }, + { + "epoch": 2.068831168831169, + "grad_norm": 44.42540497586482, + "learning_rate": 1.8405706541742773e-05, + "loss": 0.0556, + "step": 1593 + }, + { + "epoch": 2.07012987012987, + "grad_norm": 27.680227745808676, + "learning_rate": 1.840342730475513e-05, + "loss": 0.0543, + "step": 1594 + }, + { + "epoch": 2.0714285714285716, + "grad_norm": 41.77173353994607, + "learning_rate": 1.840114658104035e-05, + "loss": 0.0735, + "step": 1595 + }, + { + "epoch": 2.0727272727272728, + "grad_norm": 20.272003763855793, + "learning_rate": 1.8398864371001945e-05, + "loss": 0.0798, + "step": 1596 + }, + { + "epoch": 2.074025974025974, + "grad_norm": 44.07685740853572, + "learning_rate": 1.8396580675043674e-05, + "loss": 0.0387, + "step": 1597 + }, + { + "epoch": 2.0753246753246755, + "grad_norm": 3.0500370029912154, + "learning_rate": 1.8394295493569572e-05, + "loss": 0.0602, + "step": 1598 + }, + { + "epoch": 2.0766233766233766, + "grad_norm": 33.476318180867096, + "learning_rate": 1.8392008826983926e-05, + "loss": 0.0928, + "step": 1599 + }, + { + "epoch": 2.0779220779220777, + "grad_norm": 8.030426255315165, + "learning_rate": 1.8389720675691297e-05, + "loss": 0.0376, + "step": 1600 + }, + { + "epoch": 2.0792207792207793, + "grad_norm": 19.565178790602314, + "learning_rate": 1.8387431040096494e-05, + "loss": 0.0919, + "step": 1601 + }, + { + "epoch": 2.0805194805194804, + "grad_norm": 21.012483450347204, + "learning_rate": 1.8385139920604607e-05, + "loss": 0.0723, + "step": 1602 + }, + { + "epoch": 2.081818181818182, + "grad_norm": 44.75352311492634, + "learning_rate": 1.8382847317620972e-05, + "loss": 0.0808, + "step": 1603 + }, + { + "epoch": 2.083116883116883, + "grad_norm": 9.196811484075337, + "learning_rate": 1.83805532315512e-05, + "loss": 0.0699, + "step": 1604 + }, + { + "epoch": 2.0844155844155843, + "grad_norm": 42.60537479344422, + "learning_rate": 1.8378257662801156e-05, + "loss": 0.095, + "step": 1605 + }, + { + "epoch": 2.085714285714286, + "grad_norm": 8.002578477984033, + "learning_rate": 1.8375960611776965e-05, + "loss": 0.0705, + "step": 1606 + }, + { + "epoch": 2.087012987012987, + "grad_norm": 56.09451016024491, + "learning_rate": 1.8373662078885028e-05, + "loss": 0.0635, + "step": 1607 + }, + { + "epoch": 2.088311688311688, + "grad_norm": 9.908069537381365, + "learning_rate": 1.8371362064531996e-05, + "loss": 0.0539, + "step": 1608 + }, + { + "epoch": 2.0896103896103897, + "grad_norm": 41.920939161240675, + "learning_rate": 1.8369060569124782e-05, + "loss": 0.0703, + "step": 1609 + }, + { + "epoch": 2.090909090909091, + "grad_norm": 8.95191386784082, + "learning_rate": 1.8366757593070573e-05, + "loss": 0.0733, + "step": 1610 + }, + { + "epoch": 2.0922077922077924, + "grad_norm": 30.398161949669376, + "learning_rate": 1.83644531367768e-05, + "loss": 0.0828, + "step": 1611 + }, + { + "epoch": 2.0935064935064935, + "grad_norm": 7.281469973994752, + "learning_rate": 1.8362147200651172e-05, + "loss": 0.0599, + "step": 1612 + }, + { + "epoch": 2.0948051948051947, + "grad_norm": 31.618103453029534, + "learning_rate": 1.8359839785101652e-05, + "loss": 0.0829, + "step": 1613 + }, + { + "epoch": 2.0961038961038962, + "grad_norm": 10.208945239424285, + "learning_rate": 1.8357530890536465e-05, + "loss": 0.1047, + "step": 1614 + }, + { + "epoch": 2.0974025974025974, + "grad_norm": 39.070634041163245, + "learning_rate": 1.8355220517364094e-05, + "loss": 0.0674, + "step": 1615 + }, + { + "epoch": 2.0987012987012985, + "grad_norm": 6.512084741480914, + "learning_rate": 1.8352908665993296e-05, + "loss": 0.0979, + "step": 1616 + }, + { + "epoch": 2.1, + "grad_norm": 40.93378035868502, + "learning_rate": 1.8350595336833077e-05, + "loss": 0.0601, + "step": 1617 + }, + { + "epoch": 2.101298701298701, + "grad_norm": 8.887543604379816, + "learning_rate": 1.8348280530292712e-05, + "loss": 0.066, + "step": 1618 + }, + { + "epoch": 2.102597402597403, + "grad_norm": 29.35127822998388, + "learning_rate": 1.8345964246781736e-05, + "loss": 0.0983, + "step": 1619 + }, + { + "epoch": 2.103896103896104, + "grad_norm": 30.51570051435816, + "learning_rate": 1.834364648670994e-05, + "loss": 0.0999, + "step": 1620 + }, + { + "epoch": 2.105194805194805, + "grad_norm": 46.858656881811946, + "learning_rate": 1.834132725048738e-05, + "loss": 0.0453, + "step": 1621 + }, + { + "epoch": 2.1064935064935066, + "grad_norm": 23.939774166143046, + "learning_rate": 1.8339006538524373e-05, + "loss": 0.0676, + "step": 1622 + }, + { + "epoch": 2.1077922077922078, + "grad_norm": 33.32797858744367, + "learning_rate": 1.8336684351231498e-05, + "loss": 0.0817, + "step": 1623 + }, + { + "epoch": 2.109090909090909, + "grad_norm": 22.545965474268183, + "learning_rate": 1.83343606890196e-05, + "loss": 0.1002, + "step": 1624 + }, + { + "epoch": 2.1103896103896105, + "grad_norm": 5.208297221837994, + "learning_rate": 1.8332035552299775e-05, + "loss": 0.0888, + "step": 1625 + }, + { + "epoch": 2.1116883116883116, + "grad_norm": 32.44885754934205, + "learning_rate": 1.8329708941483382e-05, + "loss": 0.0731, + "step": 1626 + }, + { + "epoch": 2.112987012987013, + "grad_norm": 10.184412739458217, + "learning_rate": 1.8327380856982046e-05, + "loss": 0.0455, + "step": 1627 + }, + { + "epoch": 2.1142857142857143, + "grad_norm": 26.54688276263037, + "learning_rate": 1.8325051299207647e-05, + "loss": 0.0714, + "step": 1628 + }, + { + "epoch": 2.1155844155844155, + "grad_norm": 21.889898785170164, + "learning_rate": 1.8322720268572333e-05, + "loss": 0.0729, + "step": 1629 + }, + { + "epoch": 2.116883116883117, + "grad_norm": 20.83280858835745, + "learning_rate": 1.8320387765488502e-05, + "loss": 0.0628, + "step": 1630 + }, + { + "epoch": 2.118181818181818, + "grad_norm": 37.524836750471664, + "learning_rate": 1.831805379036882e-05, + "loss": 0.044, + "step": 1631 + }, + { + "epoch": 2.1194805194805193, + "grad_norm": 19.476305373062363, + "learning_rate": 1.8315718343626214e-05, + "loss": 0.0612, + "step": 1632 + }, + { + "epoch": 2.120779220779221, + "grad_norm": 39.79603282327637, + "learning_rate": 1.831338142567387e-05, + "loss": 0.1278, + "step": 1633 + }, + { + "epoch": 2.122077922077922, + "grad_norm": 8.275824023007468, + "learning_rate": 1.831104303692523e-05, + "loss": 0.0642, + "step": 1634 + }, + { + "epoch": 2.1233766233766236, + "grad_norm": 49.00828304955095, + "learning_rate": 1.8308703177794006e-05, + "loss": 0.0823, + "step": 1635 + }, + { + "epoch": 2.1246753246753247, + "grad_norm": 8.76007100748535, + "learning_rate": 1.8306361848694156e-05, + "loss": 0.057, + "step": 1636 + }, + { + "epoch": 2.125974025974026, + "grad_norm": 31.317111586439424, + "learning_rate": 1.8304019050039913e-05, + "loss": 0.0613, + "step": 1637 + }, + { + "epoch": 2.1272727272727274, + "grad_norm": 25.929190868399544, + "learning_rate": 1.8301674782245755e-05, + "loss": 0.0726, + "step": 1638 + }, + { + "epoch": 2.1285714285714286, + "grad_norm": 31.19208465737036, + "learning_rate": 1.8299329045726435e-05, + "loss": 0.0849, + "step": 1639 + }, + { + "epoch": 2.1298701298701297, + "grad_norm": 30.41795789160996, + "learning_rate": 1.829698184089695e-05, + "loss": 0.0539, + "step": 1640 + }, + { + "epoch": 2.1311688311688313, + "grad_norm": 21.91641302677479, + "learning_rate": 1.829463316817258e-05, + "loss": 0.0683, + "step": 1641 + }, + { + "epoch": 2.1324675324675324, + "grad_norm": 41.37637714837891, + "learning_rate": 1.829228302796884e-05, + "loss": 0.0838, + "step": 1642 + }, + { + "epoch": 2.1337662337662335, + "grad_norm": 5.817840036380283, + "learning_rate": 1.8289931420701513e-05, + "loss": 0.0762, + "step": 1643 + }, + { + "epoch": 2.135064935064935, + "grad_norm": 63.70540907626774, + "learning_rate": 1.8287578346786646e-05, + "loss": 0.0766, + "step": 1644 + }, + { + "epoch": 2.1363636363636362, + "grad_norm": 6.590394044489766, + "learning_rate": 1.8285223806640547e-05, + "loss": 0.0721, + "step": 1645 + }, + { + "epoch": 2.137662337662338, + "grad_norm": 69.72650270873311, + "learning_rate": 1.8282867800679774e-05, + "loss": 0.0859, + "step": 1646 + }, + { + "epoch": 2.138961038961039, + "grad_norm": 13.283690896675424, + "learning_rate": 1.828051032932115e-05, + "loss": 0.0854, + "step": 1647 + }, + { + "epoch": 2.14025974025974, + "grad_norm": 51.5738917078316, + "learning_rate": 1.8278151392981762e-05, + "loss": 0.1061, + "step": 1648 + }, + { + "epoch": 2.1415584415584417, + "grad_norm": 31.99164079422246, + "learning_rate": 1.8275790992078945e-05, + "loss": 0.0744, + "step": 1649 + }, + { + "epoch": 2.142857142857143, + "grad_norm": 37.80630655566226, + "learning_rate": 1.82734291270303e-05, + "loss": 0.0627, + "step": 1650 + }, + { + "epoch": 2.1441558441558444, + "grad_norm": 44.05039742645402, + "learning_rate": 1.8271065798253688e-05, + "loss": 0.0457, + "step": 1651 + }, + { + "epoch": 2.1454545454545455, + "grad_norm": 20.538813345642037, + "learning_rate": 1.8268701006167226e-05, + "loss": 0.0675, + "step": 1652 + }, + { + "epoch": 2.1467532467532466, + "grad_norm": 61.992927191063515, + "learning_rate": 1.8266334751189294e-05, + "loss": 0.1002, + "step": 1653 + }, + { + "epoch": 2.148051948051948, + "grad_norm": 28.583892093356248, + "learning_rate": 1.826396703373852e-05, + "loss": 0.05, + "step": 1654 + }, + { + "epoch": 2.1493506493506493, + "grad_norm": 46.64215580633927, + "learning_rate": 1.826159785423381e-05, + "loss": 0.0556, + "step": 1655 + }, + { + "epoch": 2.1506493506493505, + "grad_norm": 15.736288927054042, + "learning_rate": 1.825922721309431e-05, + "loss": 0.0598, + "step": 1656 + }, + { + "epoch": 2.151948051948052, + "grad_norm": 32.85642420727583, + "learning_rate": 1.825685511073943e-05, + "loss": 0.075, + "step": 1657 + }, + { + "epoch": 2.153246753246753, + "grad_norm": 7.324546820675322, + "learning_rate": 1.8254481547588844e-05, + "loss": 0.0466, + "step": 1658 + }, + { + "epoch": 2.1545454545454543, + "grad_norm": 34.75980063646631, + "learning_rate": 1.8252106524062477e-05, + "loss": 0.0722, + "step": 1659 + }, + { + "epoch": 2.155844155844156, + "grad_norm": 7.202923673407684, + "learning_rate": 1.8249730040580523e-05, + "loss": 0.0551, + "step": 1660 + }, + { + "epoch": 2.157142857142857, + "grad_norm": 13.340961059208517, + "learning_rate": 1.824735209756342e-05, + "loss": 0.0482, + "step": 1661 + }, + { + "epoch": 2.1584415584415586, + "grad_norm": 8.445292581785958, + "learning_rate": 1.8244972695431874e-05, + "loss": 0.0306, + "step": 1662 + }, + { + "epoch": 2.1597402597402597, + "grad_norm": 8.056587032863758, + "learning_rate": 1.824259183460685e-05, + "loss": 0.0524, + "step": 1663 + }, + { + "epoch": 2.161038961038961, + "grad_norm": 18.0478013437261, + "learning_rate": 1.8240209515509568e-05, + "loss": 0.042, + "step": 1664 + }, + { + "epoch": 2.1623376623376624, + "grad_norm": 8.253230182990501, + "learning_rate": 1.82378257385615e-05, + "loss": 0.0748, + "step": 1665 + }, + { + "epoch": 2.1636363636363636, + "grad_norm": 26.9387717868515, + "learning_rate": 1.8235440504184386e-05, + "loss": 0.0495, + "step": 1666 + }, + { + "epoch": 2.164935064935065, + "grad_norm": 32.21242934659266, + "learning_rate": 1.823305381280022e-05, + "loss": 0.0956, + "step": 1667 + }, + { + "epoch": 2.1662337662337663, + "grad_norm": 5.114200528115811, + "learning_rate": 1.8230665664831253e-05, + "loss": 0.0692, + "step": 1668 + }, + { + "epoch": 2.1675324675324674, + "grad_norm": 25.682770221591326, + "learning_rate": 1.822827606069999e-05, + "loss": 0.0489, + "step": 1669 + }, + { + "epoch": 2.168831168831169, + "grad_norm": 10.247306061621389, + "learning_rate": 1.82258850008292e-05, + "loss": 0.0489, + "step": 1670 + }, + { + "epoch": 2.17012987012987, + "grad_norm": 15.398187346236444, + "learning_rate": 1.8223492485641916e-05, + "loss": 0.0803, + "step": 1671 + }, + { + "epoch": 2.1714285714285713, + "grad_norm": 8.826942014088996, + "learning_rate": 1.8221098515561405e-05, + "loss": 0.08, + "step": 1672 + }, + { + "epoch": 2.172727272727273, + "grad_norm": 4.174351145433927, + "learning_rate": 1.8218703091011214e-05, + "loss": 0.054, + "step": 1673 + }, + { + "epoch": 2.174025974025974, + "grad_norm": 5.352749964396102, + "learning_rate": 1.821630621241514e-05, + "loss": 0.0458, + "step": 1674 + }, + { + "epoch": 2.175324675324675, + "grad_norm": 5.7405557827731455, + "learning_rate": 1.8213907880197236e-05, + "loss": 0.0933, + "step": 1675 + }, + { + "epoch": 2.1766233766233767, + "grad_norm": 18.82731506280825, + "learning_rate": 1.821150809478181e-05, + "loss": 0.0844, + "step": 1676 + }, + { + "epoch": 2.177922077922078, + "grad_norm": 31.570690595365413, + "learning_rate": 1.8209106856593433e-05, + "loss": 0.0803, + "step": 1677 + }, + { + "epoch": 2.1792207792207794, + "grad_norm": 16.782920269208205, + "learning_rate": 1.820670416605693e-05, + "loss": 0.0678, + "step": 1678 + }, + { + "epoch": 2.1805194805194805, + "grad_norm": 9.25979761215202, + "learning_rate": 1.820430002359738e-05, + "loss": 0.0389, + "step": 1679 + }, + { + "epoch": 2.1818181818181817, + "grad_norm": 13.802282559382446, + "learning_rate": 1.8201894429640125e-05, + "loss": 0.0608, + "step": 1680 + }, + { + "epoch": 2.1831168831168832, + "grad_norm": 7.146954404156662, + "learning_rate": 1.8199487384610758e-05, + "loss": 0.0916, + "step": 1681 + }, + { + "epoch": 2.1844155844155844, + "grad_norm": 6.260815534328283, + "learning_rate": 1.8197078888935137e-05, + "loss": 0.0421, + "step": 1682 + }, + { + "epoch": 2.185714285714286, + "grad_norm": 11.08275861276572, + "learning_rate": 1.819466894303936e-05, + "loss": 0.0315, + "step": 1683 + }, + { + "epoch": 2.187012987012987, + "grad_norm": 12.101289107965702, + "learning_rate": 1.8192257547349805e-05, + "loss": 0.0774, + "step": 1684 + }, + { + "epoch": 2.188311688311688, + "grad_norm": 6.454094568846809, + "learning_rate": 1.8189844702293086e-05, + "loss": 0.105, + "step": 1685 + }, + { + "epoch": 2.18961038961039, + "grad_norm": 13.16866297033139, + "learning_rate": 1.8187430408296088e-05, + "loss": 0.0557, + "step": 1686 + }, + { + "epoch": 2.190909090909091, + "grad_norm": 9.720563283879747, + "learning_rate": 1.8185014665785936e-05, + "loss": 0.0691, + "step": 1687 + }, + { + "epoch": 2.192207792207792, + "grad_norm": 15.890634981928196, + "learning_rate": 1.818259747519003e-05, + "loss": 0.0835, + "step": 1688 + }, + { + "epoch": 2.1935064935064936, + "grad_norm": 9.855028192608817, + "learning_rate": 1.8180178836936012e-05, + "loss": 0.0768, + "step": 1689 + }, + { + "epoch": 2.1948051948051948, + "grad_norm": 12.769386160579511, + "learning_rate": 1.8177758751451787e-05, + "loss": 0.0579, + "step": 1690 + }, + { + "epoch": 2.196103896103896, + "grad_norm": 4.694685730366148, + "learning_rate": 1.817533721916552e-05, + "loss": 0.0388, + "step": 1691 + }, + { + "epoch": 2.1974025974025975, + "grad_norm": 7.830283846195556, + "learning_rate": 1.8172914240505615e-05, + "loss": 0.074, + "step": 1692 + }, + { + "epoch": 2.1987012987012986, + "grad_norm": 9.067849954741057, + "learning_rate": 1.8170489815900754e-05, + "loss": 0.0793, + "step": 1693 + }, + { + "epoch": 2.2, + "grad_norm": 4.164640285946264, + "learning_rate": 1.8168063945779857e-05, + "loss": 0.0629, + "step": 1694 + }, + { + "epoch": 2.2012987012987013, + "grad_norm": 8.859051118420885, + "learning_rate": 1.816563663057211e-05, + "loss": 0.0771, + "step": 1695 + }, + { + "epoch": 2.2025974025974024, + "grad_norm": 11.819373035853602, + "learning_rate": 1.816320787070695e-05, + "loss": 0.0439, + "step": 1696 + }, + { + "epoch": 2.203896103896104, + "grad_norm": 4.3607840178726605, + "learning_rate": 1.8160777666614077e-05, + "loss": 0.076, + "step": 1697 + }, + { + "epoch": 2.205194805194805, + "grad_norm": 11.740179517520762, + "learning_rate": 1.815834601872343e-05, + "loss": 0.0716, + "step": 1698 + }, + { + "epoch": 2.2064935064935067, + "grad_norm": 3.5952723599313408, + "learning_rate": 1.8155912927465224e-05, + "loss": 0.0853, + "step": 1699 + }, + { + "epoch": 2.207792207792208, + "grad_norm": 23.095998761915645, + "learning_rate": 1.8153478393269914e-05, + "loss": 0.0427, + "step": 1700 + }, + { + "epoch": 2.209090909090909, + "grad_norm": 18.153560633570695, + "learning_rate": 1.8151042416568216e-05, + "loss": 0.0561, + "step": 1701 + }, + { + "epoch": 2.2103896103896106, + "grad_norm": 10.538383143501314, + "learning_rate": 1.8148604997791105e-05, + "loss": 0.0601, + "step": 1702 + }, + { + "epoch": 2.2116883116883117, + "grad_norm": 10.835109041920976, + "learning_rate": 1.81461661373698e-05, + "loss": 0.0382, + "step": 1703 + }, + { + "epoch": 2.212987012987013, + "grad_norm": 18.02444126851898, + "learning_rate": 1.814372583573579e-05, + "loss": 0.0668, + "step": 1704 + }, + { + "epoch": 2.2142857142857144, + "grad_norm": 20.03665236480753, + "learning_rate": 1.8141284093320806e-05, + "loss": 0.0664, + "step": 1705 + }, + { + "epoch": 2.2155844155844155, + "grad_norm": 12.593211278369886, + "learning_rate": 1.8138840910556836e-05, + "loss": 0.0917, + "step": 1706 + }, + { + "epoch": 2.2168831168831167, + "grad_norm": 6.539486959955853, + "learning_rate": 1.8136396287876135e-05, + "loss": 0.0704, + "step": 1707 + }, + { + "epoch": 2.2181818181818183, + "grad_norm": 12.910786786131085, + "learning_rate": 1.8133950225711193e-05, + "loss": 0.0987, + "step": 1708 + }, + { + "epoch": 2.2194805194805194, + "grad_norm": 9.226126146209344, + "learning_rate": 1.8131502724494777e-05, + "loss": 0.0549, + "step": 1709 + }, + { + "epoch": 2.220779220779221, + "grad_norm": 17.44399111613745, + "learning_rate": 1.8129053784659888e-05, + "loss": 0.0592, + "step": 1710 + }, + { + "epoch": 2.222077922077922, + "grad_norm": 11.94705384002674, + "learning_rate": 1.8126603406639788e-05, + "loss": 0.0572, + "step": 1711 + }, + { + "epoch": 2.2233766233766232, + "grad_norm": 16.339330140119635, + "learning_rate": 1.8124151590868003e-05, + "loss": 0.0709, + "step": 1712 + }, + { + "epoch": 2.224675324675325, + "grad_norm": 13.362214422513395, + "learning_rate": 1.812169833777831e-05, + "loss": 0.0576, + "step": 1713 + }, + { + "epoch": 2.225974025974026, + "grad_norm": 11.52141366782068, + "learning_rate": 1.811924364780472e-05, + "loss": 0.0842, + "step": 1714 + }, + { + "epoch": 2.227272727272727, + "grad_norm": 6.128694914874965, + "learning_rate": 1.8116787521381532e-05, + "loss": 0.0503, + "step": 1715 + }, + { + "epoch": 2.2285714285714286, + "grad_norm": 5.842593392480932, + "learning_rate": 1.8114329958943272e-05, + "loss": 0.0497, + "step": 1716 + }, + { + "epoch": 2.22987012987013, + "grad_norm": 5.715992764281769, + "learning_rate": 1.8111870960924733e-05, + "loss": 0.0509, + "step": 1717 + }, + { + "epoch": 2.2311688311688314, + "grad_norm": 9.270065757074736, + "learning_rate": 1.8109410527760958e-05, + "loss": 0.058, + "step": 1718 + }, + { + "epoch": 2.2324675324675325, + "grad_norm": 9.71498682692558, + "learning_rate": 1.8106948659887247e-05, + "loss": 0.0551, + "step": 1719 + }, + { + "epoch": 2.2337662337662336, + "grad_norm": 8.816858504465067, + "learning_rate": 1.8104485357739146e-05, + "loss": 0.0809, + "step": 1720 + }, + { + "epoch": 2.235064935064935, + "grad_norm": 19.93964575975691, + "learning_rate": 1.810202062175246e-05, + "loss": 0.0346, + "step": 1721 + }, + { + "epoch": 2.2363636363636363, + "grad_norm": 4.854224192302275, + "learning_rate": 1.809955445236326e-05, + "loss": 0.0747, + "step": 1722 + }, + { + "epoch": 2.2376623376623375, + "grad_norm": 14.054454334645746, + "learning_rate": 1.8097086850007847e-05, + "loss": 0.0443, + "step": 1723 + }, + { + "epoch": 2.238961038961039, + "grad_norm": 4.94439774025682, + "learning_rate": 1.809461781512279e-05, + "loss": 0.0572, + "step": 1724 + }, + { + "epoch": 2.24025974025974, + "grad_norm": 7.090539953225939, + "learning_rate": 1.809214734814491e-05, + "loss": 0.0646, + "step": 1725 + }, + { + "epoch": 2.2415584415584417, + "grad_norm": 4.622312933654094, + "learning_rate": 1.8089675449511277e-05, + "loss": 0.0671, + "step": 1726 + }, + { + "epoch": 2.242857142857143, + "grad_norm": 5.380396209952828, + "learning_rate": 1.808720211965922e-05, + "loss": 0.0661, + "step": 1727 + }, + { + "epoch": 2.244155844155844, + "grad_norm": 8.295428989276664, + "learning_rate": 1.8084727359026316e-05, + "loss": 0.051, + "step": 1728 + }, + { + "epoch": 2.2454545454545456, + "grad_norm": 15.106131881631029, + "learning_rate": 1.8082251168050402e-05, + "loss": 0.0734, + "step": 1729 + }, + { + "epoch": 2.2467532467532467, + "grad_norm": 4.4187312137319505, + "learning_rate": 1.8079773547169557e-05, + "loss": 0.1029, + "step": 1730 + }, + { + "epoch": 2.248051948051948, + "grad_norm": 7.936112254586479, + "learning_rate": 1.8077294496822124e-05, + "loss": 0.0754, + "step": 1731 + }, + { + "epoch": 2.2493506493506494, + "grad_norm": 7.252970531551866, + "learning_rate": 1.8074814017446693e-05, + "loss": 0.0765, + "step": 1732 + }, + { + "epoch": 2.2506493506493506, + "grad_norm": 15.660446948432913, + "learning_rate": 1.8072332109482113e-05, + "loss": 0.0728, + "step": 1733 + }, + { + "epoch": 2.2519480519480517, + "grad_norm": 12.535044763668163, + "learning_rate": 1.806984877336747e-05, + "loss": 0.0886, + "step": 1734 + }, + { + "epoch": 2.2532467532467533, + "grad_norm": 30.1428349654321, + "learning_rate": 1.8067364009542125e-05, + "loss": 0.0789, + "step": 1735 + }, + { + "epoch": 2.2545454545454544, + "grad_norm": 7.702526904801109, + "learning_rate": 1.8064877818445674e-05, + "loss": 0.0816, + "step": 1736 + }, + { + "epoch": 2.255844155844156, + "grad_norm": 32.093265150438945, + "learning_rate": 1.8062390200517975e-05, + "loss": 0.0469, + "step": 1737 + }, + { + "epoch": 2.257142857142857, + "grad_norm": 5.8724242563087365, + "learning_rate": 1.8059901156199132e-05, + "loss": 0.0823, + "step": 1738 + }, + { + "epoch": 2.2584415584415583, + "grad_norm": 18.7754116838312, + "learning_rate": 1.8057410685929505e-05, + "loss": 0.0675, + "step": 1739 + }, + { + "epoch": 2.25974025974026, + "grad_norm": 12.708911362308251, + "learning_rate": 1.8054918790149713e-05, + "loss": 0.0693, + "step": 1740 + }, + { + "epoch": 2.261038961038961, + "grad_norm": 11.24649780588676, + "learning_rate": 1.805242546930061e-05, + "loss": 0.0872, + "step": 1741 + }, + { + "epoch": 2.2623376623376625, + "grad_norm": 22.189735438271267, + "learning_rate": 1.8049930723823315e-05, + "loss": 0.0915, + "step": 1742 + }, + { + "epoch": 2.2636363636363637, + "grad_norm": 12.383186271495587, + "learning_rate": 1.80474345541592e-05, + "loss": 0.0651, + "step": 1743 + }, + { + "epoch": 2.264935064935065, + "grad_norm": 9.255720776627129, + "learning_rate": 1.804493696074988e-05, + "loss": 0.0531, + "step": 1744 + }, + { + "epoch": 2.2662337662337664, + "grad_norm": 12.651597513843669, + "learning_rate": 1.8042437944037234e-05, + "loss": 0.124, + "step": 1745 + }, + { + "epoch": 2.2675324675324675, + "grad_norm": 16.853647309290206, + "learning_rate": 1.8039937504463378e-05, + "loss": 0.0843, + "step": 1746 + }, + { + "epoch": 2.2688311688311686, + "grad_norm": 18.711286604640204, + "learning_rate": 1.8037435642470696e-05, + "loss": 0.0778, + "step": 1747 + }, + { + "epoch": 2.27012987012987, + "grad_norm": 21.54430451398728, + "learning_rate": 1.8034932358501806e-05, + "loss": 0.1153, + "step": 1748 + }, + { + "epoch": 2.2714285714285714, + "grad_norm": 6.037814873331458, + "learning_rate": 1.8032427652999594e-05, + "loss": 0.0593, + "step": 1749 + }, + { + "epoch": 2.2727272727272725, + "grad_norm": 5.949838449695059, + "learning_rate": 1.8029921526407184e-05, + "loss": 0.0693, + "step": 1750 + }, + { + "epoch": 2.274025974025974, + "grad_norm": 7.879969073919631, + "learning_rate": 1.8027413979167968e-05, + "loss": 0.0578, + "step": 1751 + }, + { + "epoch": 2.275324675324675, + "grad_norm": 20.125413007504505, + "learning_rate": 1.802490501172557e-05, + "loss": 0.1041, + "step": 1752 + }, + { + "epoch": 2.2766233766233768, + "grad_norm": 10.188221778015176, + "learning_rate": 1.802239462452388e-05, + "loss": 0.0886, + "step": 1753 + }, + { + "epoch": 2.277922077922078, + "grad_norm": 8.47934081518852, + "learning_rate": 1.8019882818007026e-05, + "loss": 0.062, + "step": 1754 + }, + { + "epoch": 2.279220779220779, + "grad_norm": 5.85016344567073, + "learning_rate": 1.8017369592619398e-05, + "loss": 0.0663, + "step": 1755 + }, + { + "epoch": 2.2805194805194806, + "grad_norm": 5.546736808732752, + "learning_rate": 1.8014854948805636e-05, + "loss": 0.0871, + "step": 1756 + }, + { + "epoch": 2.2818181818181817, + "grad_norm": 10.566912619370882, + "learning_rate": 1.8012338887010633e-05, + "loss": 0.0415, + "step": 1757 + }, + { + "epoch": 2.2831168831168833, + "grad_norm": 4.125100057461264, + "learning_rate": 1.800982140767952e-05, + "loss": 0.062, + "step": 1758 + }, + { + "epoch": 2.2844155844155845, + "grad_norm": 7.196726135569817, + "learning_rate": 1.800730251125769e-05, + "loss": 0.1013, + "step": 1759 + }, + { + "epoch": 2.2857142857142856, + "grad_norm": 13.897699208383985, + "learning_rate": 1.8004782198190783e-05, + "loss": 0.08, + "step": 1760 + }, + { + "epoch": 2.287012987012987, + "grad_norm": 16.49113952705904, + "learning_rate": 1.8002260468924696e-05, + "loss": 0.0306, + "step": 1761 + }, + { + "epoch": 2.2883116883116883, + "grad_norm": 36.6158326682454, + "learning_rate": 1.7999737323905563e-05, + "loss": 0.1016, + "step": 1762 + }, + { + "epoch": 2.2896103896103894, + "grad_norm": 36.67817584813191, + "learning_rate": 1.7997212763579782e-05, + "loss": 0.0433, + "step": 1763 + }, + { + "epoch": 2.290909090909091, + "grad_norm": 35.11956965855619, + "learning_rate": 1.7994686788393995e-05, + "loss": 0.0862, + "step": 1764 + }, + { + "epoch": 2.292207792207792, + "grad_norm": 25.048063839194608, + "learning_rate": 1.7992159398795093e-05, + "loss": 0.0889, + "step": 1765 + }, + { + "epoch": 2.2935064935064933, + "grad_norm": 51.00431865268078, + "learning_rate": 1.7989630595230225e-05, + "loss": 0.075, + "step": 1766 + }, + { + "epoch": 2.294805194805195, + "grad_norm": 33.00306793208659, + "learning_rate": 1.798710037814678e-05, + "loss": 0.0796, + "step": 1767 + }, + { + "epoch": 2.296103896103896, + "grad_norm": 45.51798299565721, + "learning_rate": 1.7984568747992407e-05, + "loss": 0.057, + "step": 1768 + }, + { + "epoch": 2.2974025974025976, + "grad_norm": 38.62672336714045, + "learning_rate": 1.798203570521499e-05, + "loss": 0.0793, + "step": 1769 + }, + { + "epoch": 2.2987012987012987, + "grad_norm": 49.276809863671616, + "learning_rate": 1.7979501250262683e-05, + "loss": 0.1042, + "step": 1770 + }, + { + "epoch": 2.3, + "grad_norm": 53.69290257452379, + "learning_rate": 1.7976965383583876e-05, + "loss": 0.0995, + "step": 1771 + }, + { + "epoch": 2.3012987012987014, + "grad_norm": 42.36362146873065, + "learning_rate": 1.797442810562721e-05, + "loss": 0.0426, + "step": 1772 + }, + { + "epoch": 2.3025974025974025, + "grad_norm": 45.06287550686328, + "learning_rate": 1.7971889416841583e-05, + "loss": 0.0827, + "step": 1773 + }, + { + "epoch": 2.303896103896104, + "grad_norm": 40.35869205238034, + "learning_rate": 1.7969349317676127e-05, + "loss": 0.0582, + "step": 1774 + }, + { + "epoch": 2.3051948051948052, + "grad_norm": 29.79007607433679, + "learning_rate": 1.7966807808580248e-05, + "loss": 0.0513, + "step": 1775 + }, + { + "epoch": 2.3064935064935064, + "grad_norm": 44.22671929581173, + "learning_rate": 1.796426489000358e-05, + "loss": 0.0949, + "step": 1776 + }, + { + "epoch": 2.307792207792208, + "grad_norm": 21.955316510496335, + "learning_rate": 1.7961720562396015e-05, + "loss": 0.0694, + "step": 1777 + }, + { + "epoch": 2.309090909090909, + "grad_norm": 46.04245984122214, + "learning_rate": 1.7959174826207695e-05, + "loss": 0.0715, + "step": 1778 + }, + { + "epoch": 2.31038961038961, + "grad_norm": 22.693977244041562, + "learning_rate": 1.7956627681889008e-05, + "loss": 0.0499, + "step": 1779 + }, + { + "epoch": 2.311688311688312, + "grad_norm": 20.181921818776548, + "learning_rate": 1.7954079129890592e-05, + "loss": 0.0711, + "step": 1780 + }, + { + "epoch": 2.312987012987013, + "grad_norm": 12.436316518937137, + "learning_rate": 1.7951529170663334e-05, + "loss": 0.0526, + "step": 1781 + }, + { + "epoch": 2.314285714285714, + "grad_norm": 19.448979529779415, + "learning_rate": 1.7948977804658373e-05, + "loss": 0.0785, + "step": 1782 + }, + { + "epoch": 2.3155844155844156, + "grad_norm": 22.085707434637612, + "learning_rate": 1.794642503232709e-05, + "loss": 0.0588, + "step": 1783 + }, + { + "epoch": 2.3168831168831168, + "grad_norm": 15.35727284562792, + "learning_rate": 1.7943870854121126e-05, + "loss": 0.0614, + "step": 1784 + }, + { + "epoch": 2.3181818181818183, + "grad_norm": 22.10683448487674, + "learning_rate": 1.794131527049236e-05, + "loss": 0.0386, + "step": 1785 + }, + { + "epoch": 2.3194805194805195, + "grad_norm": 16.887894743599407, + "learning_rate": 1.7938758281892928e-05, + "loss": 0.0821, + "step": 1786 + }, + { + "epoch": 2.3207792207792206, + "grad_norm": 5.458221443428394, + "learning_rate": 1.7936199888775202e-05, + "loss": 0.04, + "step": 1787 + }, + { + "epoch": 2.322077922077922, + "grad_norm": 6.734886311397002, + "learning_rate": 1.7933640091591814e-05, + "loss": 0.0761, + "step": 1788 + }, + { + "epoch": 2.3233766233766233, + "grad_norm": 18.81704089710429, + "learning_rate": 1.7931078890795645e-05, + "loss": 0.0583, + "step": 1789 + }, + { + "epoch": 2.324675324675325, + "grad_norm": 11.518449467325281, + "learning_rate": 1.7928516286839817e-05, + "loss": 0.0534, + "step": 1790 + }, + { + "epoch": 2.325974025974026, + "grad_norm": 11.060543930713145, + "learning_rate": 1.7925952280177705e-05, + "loss": 0.0492, + "step": 1791 + }, + { + "epoch": 2.327272727272727, + "grad_norm": 10.769548450195652, + "learning_rate": 1.7923386871262928e-05, + "loss": 0.0879, + "step": 1792 + }, + { + "epoch": 2.3285714285714287, + "grad_norm": 23.267633740589314, + "learning_rate": 1.7920820060549362e-05, + "loss": 0.0554, + "step": 1793 + }, + { + "epoch": 2.32987012987013, + "grad_norm": 25.859685712850435, + "learning_rate": 1.7918251848491118e-05, + "loss": 0.0544, + "step": 1794 + }, + { + "epoch": 2.331168831168831, + "grad_norm": 46.123236490288036, + "learning_rate": 1.791568223554257e-05, + "loss": 0.0844, + "step": 1795 + }, + { + "epoch": 2.3324675324675326, + "grad_norm": 4.82214704372741, + "learning_rate": 1.791311122215832e-05, + "loss": 0.0378, + "step": 1796 + }, + { + "epoch": 2.3337662337662337, + "grad_norm": 32.58994157079296, + "learning_rate": 1.791053880879324e-05, + "loss": 0.035, + "step": 1797 + }, + { + "epoch": 2.335064935064935, + "grad_norm": 6.288998881181952, + "learning_rate": 1.7907964995902434e-05, + "loss": 0.0357, + "step": 1798 + }, + { + "epoch": 2.3363636363636364, + "grad_norm": 35.456512602163414, + "learning_rate": 1.790538978394126e-05, + "loss": 0.0516, + "step": 1799 + }, + { + "epoch": 2.3376623376623376, + "grad_norm": 9.998720273501664, + "learning_rate": 1.7902813173365326e-05, + "loss": 0.0701, + "step": 1800 + }, + { + "epoch": 2.338961038961039, + "grad_norm": 31.41006070758284, + "learning_rate": 1.7900235164630477e-05, + "loss": 0.0639, + "step": 1801 + }, + { + "epoch": 2.3402597402597403, + "grad_norm": 19.12419630514099, + "learning_rate": 1.7897655758192815e-05, + "loss": 0.0904, + "step": 1802 + }, + { + "epoch": 2.3415584415584414, + "grad_norm": 26.442881294034752, + "learning_rate": 1.7895074954508685e-05, + "loss": 0.0793, + "step": 1803 + }, + { + "epoch": 2.342857142857143, + "grad_norm": 20.110918389729402, + "learning_rate": 1.7892492754034686e-05, + "loss": 0.0745, + "step": 1804 + }, + { + "epoch": 2.344155844155844, + "grad_norm": 15.232939373544829, + "learning_rate": 1.7889909157227652e-05, + "loss": 0.0484, + "step": 1805 + }, + { + "epoch": 2.3454545454545457, + "grad_norm": 34.141974096486855, + "learning_rate": 1.7887324164544674e-05, + "loss": 0.0729, + "step": 1806 + }, + { + "epoch": 2.346753246753247, + "grad_norm": 26.829631332992918, + "learning_rate": 1.7884737776443088e-05, + "loss": 0.0524, + "step": 1807 + }, + { + "epoch": 2.348051948051948, + "grad_norm": 35.964266147569845, + "learning_rate": 1.7882149993380474e-05, + "loss": 0.0953, + "step": 1808 + }, + { + "epoch": 2.3493506493506495, + "grad_norm": 21.662945118366416, + "learning_rate": 1.787956081581466e-05, + "loss": 0.0825, + "step": 1809 + }, + { + "epoch": 2.3506493506493507, + "grad_norm": 36.89443845662454, + "learning_rate": 1.7876970244203722e-05, + "loss": 0.0665, + "step": 1810 + }, + { + "epoch": 2.351948051948052, + "grad_norm": 16.408452023600145, + "learning_rate": 1.7874378279005978e-05, + "loss": 0.0444, + "step": 1811 + }, + { + "epoch": 2.3532467532467534, + "grad_norm": 26.37730608221366, + "learning_rate": 1.7871784920680002e-05, + "loss": 0.0726, + "step": 1812 + }, + { + "epoch": 2.3545454545454545, + "grad_norm": 31.82863926149645, + "learning_rate": 1.7869190169684604e-05, + "loss": 0.0574, + "step": 1813 + }, + { + "epoch": 2.3558441558441556, + "grad_norm": 22.143669021829623, + "learning_rate": 1.7866594026478846e-05, + "loss": 0.0741, + "step": 1814 + }, + { + "epoch": 2.357142857142857, + "grad_norm": 21.861708587640177, + "learning_rate": 1.786399649152204e-05, + "loss": 0.0862, + "step": 1815 + }, + { + "epoch": 2.3584415584415583, + "grad_norm": 24.57701180111012, + "learning_rate": 1.7861397565273736e-05, + "loss": 0.0639, + "step": 1816 + }, + { + "epoch": 2.35974025974026, + "grad_norm": 45.31022686352707, + "learning_rate": 1.7858797248193733e-05, + "loss": 0.0779, + "step": 1817 + }, + { + "epoch": 2.361038961038961, + "grad_norm": 17.03275960638103, + "learning_rate": 1.7856195540742076e-05, + "loss": 0.1007, + "step": 1818 + }, + { + "epoch": 2.362337662337662, + "grad_norm": 18.83367897417229, + "learning_rate": 1.785359244337906e-05, + "loss": 0.0653, + "step": 1819 + }, + { + "epoch": 2.3636363636363638, + "grad_norm": 30.888585866922593, + "learning_rate": 1.785098795656522e-05, + "loss": 0.0852, + "step": 1820 + }, + { + "epoch": 2.364935064935065, + "grad_norm": 9.208209167461105, + "learning_rate": 1.7848382080761343e-05, + "loss": 0.0625, + "step": 1821 + }, + { + "epoch": 2.3662337662337665, + "grad_norm": 44.25687896958553, + "learning_rate": 1.7845774816428458e-05, + "loss": 0.0845, + "step": 1822 + }, + { + "epoch": 2.3675324675324676, + "grad_norm": 17.422374777790015, + "learning_rate": 1.7843166164027835e-05, + "loss": 0.0644, + "step": 1823 + }, + { + "epoch": 2.3688311688311687, + "grad_norm": 32.183446374922866, + "learning_rate": 1.7840556124021e-05, + "loss": 0.0638, + "step": 1824 + }, + { + "epoch": 2.3701298701298703, + "grad_norm": 8.412026719860073, + "learning_rate": 1.7837944696869714e-05, + "loss": 0.0752, + "step": 1825 + }, + { + "epoch": 2.3714285714285714, + "grad_norm": 24.539293450245896, + "learning_rate": 1.7835331883035995e-05, + "loss": 0.0524, + "step": 1826 + }, + { + "epoch": 2.3727272727272726, + "grad_norm": 16.322618196383978, + "learning_rate": 1.7832717682982092e-05, + "loss": 0.0644, + "step": 1827 + }, + { + "epoch": 2.374025974025974, + "grad_norm": 27.18891755062245, + "learning_rate": 1.7830102097170515e-05, + "loss": 0.0661, + "step": 1828 + }, + { + "epoch": 2.3753246753246753, + "grad_norm": 27.742062094372724, + "learning_rate": 1.7827485126064002e-05, + "loss": 0.0457, + "step": 1829 + }, + { + "epoch": 2.3766233766233764, + "grad_norm": 13.261271667703028, + "learning_rate": 1.7824866770125554e-05, + "loss": 0.0838, + "step": 1830 + }, + { + "epoch": 2.377922077922078, + "grad_norm": 65.04023532049875, + "learning_rate": 1.7822247029818405e-05, + "loss": 0.0891, + "step": 1831 + }, + { + "epoch": 2.379220779220779, + "grad_norm": 16.235839959720302, + "learning_rate": 1.7819625905606038e-05, + "loss": 0.0664, + "step": 1832 + }, + { + "epoch": 2.3805194805194807, + "grad_norm": 35.9310935727261, + "learning_rate": 1.7817003397952176e-05, + "loss": 0.0807, + "step": 1833 + }, + { + "epoch": 2.381818181818182, + "grad_norm": 25.959085537406, + "learning_rate": 1.7814379507320795e-05, + "loss": 0.0495, + "step": 1834 + }, + { + "epoch": 2.383116883116883, + "grad_norm": 42.364037913683035, + "learning_rate": 1.781175423417611e-05, + "loss": 0.0516, + "step": 1835 + }, + { + "epoch": 2.3844155844155845, + "grad_norm": 25.307822229670567, + "learning_rate": 1.7809127578982584e-05, + "loss": 0.0606, + "step": 1836 + }, + { + "epoch": 2.3857142857142857, + "grad_norm": 21.85170713771283, + "learning_rate": 1.780649954220492e-05, + "loss": 0.0546, + "step": 1837 + }, + { + "epoch": 2.3870129870129873, + "grad_norm": 27.412038724844667, + "learning_rate": 1.780387012430807e-05, + "loss": 0.0482, + "step": 1838 + }, + { + "epoch": 2.3883116883116884, + "grad_norm": 4.121269731232052, + "learning_rate": 1.7801239325757228e-05, + "loss": 0.0583, + "step": 1839 + }, + { + "epoch": 2.3896103896103895, + "grad_norm": 29.444936951215933, + "learning_rate": 1.7798607147017827e-05, + "loss": 0.0466, + "step": 1840 + }, + { + "epoch": 2.390909090909091, + "grad_norm": 11.93308922918999, + "learning_rate": 1.779597358855556e-05, + "loss": 0.0583, + "step": 1841 + }, + { + "epoch": 2.3922077922077922, + "grad_norm": 16.62058215764619, + "learning_rate": 1.7793338650836344e-05, + "loss": 0.0478, + "step": 1842 + }, + { + "epoch": 2.3935064935064934, + "grad_norm": 11.416045950075167, + "learning_rate": 1.779070233432636e-05, + "loss": 0.0717, + "step": 1843 + }, + { + "epoch": 2.394805194805195, + "grad_norm": 6.807935245915115, + "learning_rate": 1.7788064639492014e-05, + "loss": 0.0697, + "step": 1844 + }, + { + "epoch": 2.396103896103896, + "grad_norm": 9.230240226344671, + "learning_rate": 1.778542556679997e-05, + "loss": 0.0446, + "step": 1845 + }, + { + "epoch": 2.397402597402597, + "grad_norm": 18.815111400705117, + "learning_rate": 1.7782785116717126e-05, + "loss": 0.0557, + "step": 1846 + }, + { + "epoch": 2.398701298701299, + "grad_norm": 10.200109265792909, + "learning_rate": 1.7780143289710632e-05, + "loss": 0.088, + "step": 1847 + }, + { + "epoch": 2.4, + "grad_norm": 13.39431040281573, + "learning_rate": 1.777750008624788e-05, + "loss": 0.0764, + "step": 1848 + }, + { + "epoch": 2.4012987012987015, + "grad_norm": 18.7918717481509, + "learning_rate": 1.7774855506796497e-05, + "loss": 0.1047, + "step": 1849 + }, + { + "epoch": 2.4025974025974026, + "grad_norm": 28.29118144250038, + "learning_rate": 1.777220955182436e-05, + "loss": 0.069, + "step": 1850 + }, + { + "epoch": 2.4038961038961038, + "grad_norm": 8.509883393904763, + "learning_rate": 1.776956222179959e-05, + "loss": 0.059, + "step": 1851 + }, + { + "epoch": 2.4051948051948053, + "grad_norm": 9.301122280403318, + "learning_rate": 1.776691351719056e-05, + "loss": 0.0611, + "step": 1852 + }, + { + "epoch": 2.4064935064935065, + "grad_norm": 14.622718708184866, + "learning_rate": 1.7764263438465864e-05, + "loss": 0.0683, + "step": 1853 + }, + { + "epoch": 2.407792207792208, + "grad_norm": 5.404736574717834, + "learning_rate": 1.7761611986094357e-05, + "loss": 0.0634, + "step": 1854 + }, + { + "epoch": 2.409090909090909, + "grad_norm": 20.757084200601422, + "learning_rate": 1.775895916054513e-05, + "loss": 0.0657, + "step": 1855 + }, + { + "epoch": 2.4103896103896103, + "grad_norm": 4.128638073196671, + "learning_rate": 1.775630496228752e-05, + "loss": 0.0506, + "step": 1856 + }, + { + "epoch": 2.411688311688312, + "grad_norm": 12.824260170390586, + "learning_rate": 1.7753649391791105e-05, + "loss": 0.0671, + "step": 1857 + }, + { + "epoch": 2.412987012987013, + "grad_norm": 4.511239626302904, + "learning_rate": 1.7750992449525703e-05, + "loss": 0.0504, + "step": 1858 + }, + { + "epoch": 2.414285714285714, + "grad_norm": 9.660297000712296, + "learning_rate": 1.7748334135961386e-05, + "loss": 0.0808, + "step": 1859 + }, + { + "epoch": 2.4155844155844157, + "grad_norm": 3.4653353867883108, + "learning_rate": 1.7745674451568454e-05, + "loss": 0.0513, + "step": 1860 + }, + { + "epoch": 2.416883116883117, + "grad_norm": 15.101744731113135, + "learning_rate": 1.7743013396817455e-05, + "loss": 0.0492, + "step": 1861 + }, + { + "epoch": 2.418181818181818, + "grad_norm": 8.976132055219846, + "learning_rate": 1.774035097217919e-05, + "loss": 0.1373, + "step": 1862 + }, + { + "epoch": 2.4194805194805196, + "grad_norm": 9.74984973797226, + "learning_rate": 1.773768717812468e-05, + "loss": 0.0749, + "step": 1863 + }, + { + "epoch": 2.4207792207792207, + "grad_norm": 7.635124933970499, + "learning_rate": 1.773502201512521e-05, + "loss": 0.0731, + "step": 1864 + }, + { + "epoch": 2.4220779220779223, + "grad_norm": 5.045877721389763, + "learning_rate": 1.773235548365229e-05, + "loss": 0.0723, + "step": 1865 + }, + { + "epoch": 2.4233766233766234, + "grad_norm": 9.154532154873316, + "learning_rate": 1.7729687584177694e-05, + "loss": 0.0697, + "step": 1866 + }, + { + "epoch": 2.4246753246753245, + "grad_norm": 8.531225430468817, + "learning_rate": 1.772701831717341e-05, + "loss": 0.0795, + "step": 1867 + }, + { + "epoch": 2.425974025974026, + "grad_norm": 15.157994060941212, + "learning_rate": 1.7724347683111693e-05, + "loss": 0.0667, + "step": 1868 + }, + { + "epoch": 2.4272727272727272, + "grad_norm": 16.82152533944305, + "learning_rate": 1.7721675682465025e-05, + "loss": 0.0814, + "step": 1869 + }, + { + "epoch": 2.4285714285714284, + "grad_norm": 11.080796766253242, + "learning_rate": 1.7719002315706133e-05, + "loss": 0.1002, + "step": 1870 + }, + { + "epoch": 2.42987012987013, + "grad_norm": 22.220475097055747, + "learning_rate": 1.7716327583307988e-05, + "loss": 0.0547, + "step": 1871 + }, + { + "epoch": 2.431168831168831, + "grad_norm": 14.286650316337086, + "learning_rate": 1.7713651485743802e-05, + "loss": 0.0656, + "step": 1872 + }, + { + "epoch": 2.4324675324675322, + "grad_norm": 15.19062740994615, + "learning_rate": 1.771097402348703e-05, + "loss": 0.0522, + "step": 1873 + }, + { + "epoch": 2.433766233766234, + "grad_norm": 32.22608156643629, + "learning_rate": 1.770829519701136e-05, + "loss": 0.0863, + "step": 1874 + }, + { + "epoch": 2.435064935064935, + "grad_norm": 12.537144140688026, + "learning_rate": 1.7705615006790734e-05, + "loss": 0.0589, + "step": 1875 + }, + { + "epoch": 2.4363636363636365, + "grad_norm": 30.82213531392099, + "learning_rate": 1.770293345329932e-05, + "loss": 0.0624, + "step": 1876 + }, + { + "epoch": 2.4376623376623376, + "grad_norm": 12.553357182500484, + "learning_rate": 1.770025053701155e-05, + "loss": 0.0682, + "step": 1877 + }, + { + "epoch": 2.4389610389610388, + "grad_norm": 28.08093438724972, + "learning_rate": 1.7697566258402073e-05, + "loss": 0.0777, + "step": 1878 + }, + { + "epoch": 2.4402597402597404, + "grad_norm": 31.808321832513126, + "learning_rate": 1.7694880617945793e-05, + "loss": 0.0382, + "step": 1879 + }, + { + "epoch": 2.4415584415584415, + "grad_norm": 30.39179053982529, + "learning_rate": 1.7692193616117848e-05, + "loss": 0.0601, + "step": 1880 + }, + { + "epoch": 2.442857142857143, + "grad_norm": 30.746157586417507, + "learning_rate": 1.768950525339362e-05, + "loss": 0.0584, + "step": 1881 + }, + { + "epoch": 2.444155844155844, + "grad_norm": 24.96059744906535, + "learning_rate": 1.768681553024874e-05, + "loss": 0.0473, + "step": 1882 + }, + { + "epoch": 2.4454545454545453, + "grad_norm": 29.042844129505745, + "learning_rate": 1.7684124447159062e-05, + "loss": 0.0398, + "step": 1883 + }, + { + "epoch": 2.446753246753247, + "grad_norm": 8.373349057058634, + "learning_rate": 1.7681432004600694e-05, + "loss": 0.0695, + "step": 1884 + }, + { + "epoch": 2.448051948051948, + "grad_norm": 45.63151178239347, + "learning_rate": 1.767873820304998e-05, + "loss": 0.0785, + "step": 1885 + }, + { + "epoch": 2.449350649350649, + "grad_norm": 7.166290981528156, + "learning_rate": 1.7676043042983503e-05, + "loss": 0.0451, + "step": 1886 + }, + { + "epoch": 2.4506493506493507, + "grad_norm": 47.26659412026907, + "learning_rate": 1.7673346524878096e-05, + "loss": 0.0864, + "step": 1887 + }, + { + "epoch": 2.451948051948052, + "grad_norm": 9.168703115468706, + "learning_rate": 1.7670648649210816e-05, + "loss": 0.102, + "step": 1888 + }, + { + "epoch": 2.453246753246753, + "grad_norm": 37.87431283236583, + "learning_rate": 1.7667949416458973e-05, + "loss": 0.0658, + "step": 1889 + }, + { + "epoch": 2.4545454545454546, + "grad_norm": 15.269572769610619, + "learning_rate": 1.766524882710011e-05, + "loss": 0.0881, + "step": 1890 + }, + { + "epoch": 2.4558441558441557, + "grad_norm": 10.850491509357417, + "learning_rate": 1.766254688161202e-05, + "loss": 0.0654, + "step": 1891 + }, + { + "epoch": 2.4571428571428573, + "grad_norm": 38.95275039376971, + "learning_rate": 1.7659843580472716e-05, + "loss": 0.0802, + "step": 1892 + }, + { + "epoch": 2.4584415584415584, + "grad_norm": 8.015985042838066, + "learning_rate": 1.765713892416048e-05, + "loss": 0.0714, + "step": 1893 + }, + { + "epoch": 2.4597402597402596, + "grad_norm": 46.135014915941326, + "learning_rate": 1.7654432913153802e-05, + "loss": 0.0979, + "step": 1894 + }, + { + "epoch": 2.461038961038961, + "grad_norm": 10.511032566825001, + "learning_rate": 1.765172554793144e-05, + "loss": 0.1345, + "step": 1895 + }, + { + "epoch": 2.4623376623376623, + "grad_norm": 44.32665610897491, + "learning_rate": 1.764901682897237e-05, + "loss": 0.0515, + "step": 1896 + }, + { + "epoch": 2.463636363636364, + "grad_norm": 16.164814272814894, + "learning_rate": 1.764630675675582e-05, + "loss": 0.0406, + "step": 1897 + }, + { + "epoch": 2.464935064935065, + "grad_norm": 34.75945200843531, + "learning_rate": 1.764359533176125e-05, + "loss": 0.0912, + "step": 1898 + }, + { + "epoch": 2.466233766233766, + "grad_norm": 18.50631495202853, + "learning_rate": 1.764088255446837e-05, + "loss": 0.0955, + "step": 1899 + }, + { + "epoch": 2.4675324675324677, + "grad_norm": 13.515465877183454, + "learning_rate": 1.763816842535712e-05, + "loss": 0.0862, + "step": 1900 + }, + { + "epoch": 2.468831168831169, + "grad_norm": 33.85907132844077, + "learning_rate": 1.763545294490767e-05, + "loss": 0.0973, + "step": 1901 + }, + { + "epoch": 2.47012987012987, + "grad_norm": 7.278277276975351, + "learning_rate": 1.763273611360046e-05, + "loss": 0.0513, + "step": 1902 + }, + { + "epoch": 2.4714285714285715, + "grad_norm": 48.995136495660454, + "learning_rate": 1.7630017931916134e-05, + "loss": 0.0566, + "step": 1903 + }, + { + "epoch": 2.4727272727272727, + "grad_norm": 5.297417619379355, + "learning_rate": 1.76272984003356e-05, + "loss": 0.0656, + "step": 1904 + }, + { + "epoch": 2.474025974025974, + "grad_norm": 33.115454770862094, + "learning_rate": 1.7624577519339987e-05, + "loss": 0.0873, + "step": 1905 + }, + { + "epoch": 2.4753246753246754, + "grad_norm": 12.630207493595174, + "learning_rate": 1.7621855289410675e-05, + "loss": 0.0353, + "step": 1906 + }, + { + "epoch": 2.4766233766233765, + "grad_norm": 21.841950710620804, + "learning_rate": 1.761913171102928e-05, + "loss": 0.0613, + "step": 1907 + }, + { + "epoch": 2.477922077922078, + "grad_norm": 35.02065824790259, + "learning_rate": 1.761640678467765e-05, + "loss": 0.1235, + "step": 1908 + }, + { + "epoch": 2.479220779220779, + "grad_norm": 29.000654854585108, + "learning_rate": 1.7613680510837885e-05, + "loss": 0.0614, + "step": 1909 + }, + { + "epoch": 2.4805194805194803, + "grad_norm": 44.56119089175063, + "learning_rate": 1.761095288999231e-05, + "loss": 0.0767, + "step": 1910 + }, + { + "epoch": 2.481818181818182, + "grad_norm": 24.517075403100232, + "learning_rate": 1.760822392262349e-05, + "loss": 0.0667, + "step": 1911 + }, + { + "epoch": 2.483116883116883, + "grad_norm": 45.62309175729297, + "learning_rate": 1.7605493609214238e-05, + "loss": 0.0724, + "step": 1912 + }, + { + "epoch": 2.4844155844155846, + "grad_norm": 13.07308176919026, + "learning_rate": 1.7602761950247595e-05, + "loss": 0.0536, + "step": 1913 + }, + { + "epoch": 2.4857142857142858, + "grad_norm": 30.116995604820012, + "learning_rate": 1.7600028946206845e-05, + "loss": 0.0382, + "step": 1914 + }, + { + "epoch": 2.487012987012987, + "grad_norm": 14.671310237620379, + "learning_rate": 1.759729459757551e-05, + "loss": 0.058, + "step": 1915 + }, + { + "epoch": 2.4883116883116885, + "grad_norm": 19.702696011775284, + "learning_rate": 1.759455890483734e-05, + "loss": 0.0589, + "step": 1916 + }, + { + "epoch": 2.4896103896103896, + "grad_norm": 40.10337661426201, + "learning_rate": 1.7591821868476343e-05, + "loss": 0.0767, + "step": 1917 + }, + { + "epoch": 2.4909090909090907, + "grad_norm": 24.383873757241872, + "learning_rate": 1.758908348897675e-05, + "loss": 0.059, + "step": 1918 + }, + { + "epoch": 2.4922077922077923, + "grad_norm": 48.85200142769423, + "learning_rate": 1.758634376682303e-05, + "loss": 0.0839, + "step": 1919 + }, + { + "epoch": 2.4935064935064934, + "grad_norm": 24.63250820845909, + "learning_rate": 1.7583602702499895e-05, + "loss": 0.0616, + "step": 1920 + }, + { + "epoch": 2.4948051948051946, + "grad_norm": 37.80925599099752, + "learning_rate": 1.7580860296492286e-05, + "loss": 0.0656, + "step": 1921 + }, + { + "epoch": 2.496103896103896, + "grad_norm": 17.01433886959073, + "learning_rate": 1.7578116549285393e-05, + "loss": 0.088, + "step": 1922 + }, + { + "epoch": 2.4974025974025973, + "grad_norm": 12.483332571618368, + "learning_rate": 1.7575371461364635e-05, + "loss": 0.0732, + "step": 1923 + }, + { + "epoch": 2.498701298701299, + "grad_norm": 26.504797706454234, + "learning_rate": 1.7572625033215676e-05, + "loss": 0.0295, + "step": 1924 + }, + { + "epoch": 2.5, + "grad_norm": 29.340078839112934, + "learning_rate": 1.7569877265324404e-05, + "loss": 0.0448, + "step": 1925 + }, + { + "epoch": 2.501298701298701, + "grad_norm": 22.99672349084605, + "learning_rate": 1.7567128158176955e-05, + "loss": 0.0545, + "step": 1926 + }, + { + "epoch": 2.5025974025974027, + "grad_norm": 11.54620376131279, + "learning_rate": 1.7564377712259696e-05, + "loss": 0.0749, + "step": 1927 + }, + { + "epoch": 2.503896103896104, + "grad_norm": 43.2275742640294, + "learning_rate": 1.7561625928059238e-05, + "loss": 0.0594, + "step": 1928 + }, + { + "epoch": 2.5051948051948054, + "grad_norm": 11.614170629288548, + "learning_rate": 1.7558872806062423e-05, + "loss": 0.0392, + "step": 1929 + }, + { + "epoch": 2.5064935064935066, + "grad_norm": 45.97541000268899, + "learning_rate": 1.755611834675633e-05, + "loss": 0.086, + "step": 1930 + }, + { + "epoch": 2.5077922077922077, + "grad_norm": 5.464023449920651, + "learning_rate": 1.7553362550628275e-05, + "loss": 0.0491, + "step": 1931 + }, + { + "epoch": 2.509090909090909, + "grad_norm": 39.8903819683907, + "learning_rate": 1.7550605418165815e-05, + "loss": 0.0766, + "step": 1932 + }, + { + "epoch": 2.5103896103896104, + "grad_norm": 13.48460535584613, + "learning_rate": 1.754784694985673e-05, + "loss": 0.1176, + "step": 1933 + }, + { + "epoch": 2.5116883116883115, + "grad_norm": 18.851936723901126, + "learning_rate": 1.7545087146189057e-05, + "loss": 0.091, + "step": 1934 + }, + { + "epoch": 2.512987012987013, + "grad_norm": 7.183255169117084, + "learning_rate": 1.7542326007651053e-05, + "loss": 0.0464, + "step": 1935 + }, + { + "epoch": 2.5142857142857142, + "grad_norm": 6.982690358591511, + "learning_rate": 1.7539563534731215e-05, + "loss": 0.0719, + "step": 1936 + }, + { + "epoch": 2.5155844155844154, + "grad_norm": 9.267844346448966, + "learning_rate": 1.753679972791828e-05, + "loss": 0.0564, + "step": 1937 + }, + { + "epoch": 2.516883116883117, + "grad_norm": 7.534952885230822, + "learning_rate": 1.7534034587701215e-05, + "loss": 0.0726, + "step": 1938 + }, + { + "epoch": 2.518181818181818, + "grad_norm": 18.563250515740513, + "learning_rate": 1.753126811456923e-05, + "loss": 0.0807, + "step": 1939 + }, + { + "epoch": 2.5194805194805197, + "grad_norm": 12.437169719451294, + "learning_rate": 1.7528500309011768e-05, + "loss": 0.0557, + "step": 1940 + }, + { + "epoch": 2.520779220779221, + "grad_norm": 9.300753516897291, + "learning_rate": 1.7525731171518502e-05, + "loss": 0.0632, + "step": 1941 + }, + { + "epoch": 2.522077922077922, + "grad_norm": 14.82033142309426, + "learning_rate": 1.7522960702579347e-05, + "loss": 0.0564, + "step": 1942 + }, + { + "epoch": 2.5233766233766235, + "grad_norm": 12.44508945789808, + "learning_rate": 1.7520188902684454e-05, + "loss": 0.0562, + "step": 1943 + }, + { + "epoch": 2.5246753246753246, + "grad_norm": 12.881867477715334, + "learning_rate": 1.7517415772324208e-05, + "loss": 0.0628, + "step": 1944 + }, + { + "epoch": 2.525974025974026, + "grad_norm": 29.85671052646669, + "learning_rate": 1.7514641311989223e-05, + "loss": 0.0687, + "step": 1945 + }, + { + "epoch": 2.5272727272727273, + "grad_norm": 7.0301917953957895, + "learning_rate": 1.751186552217036e-05, + "loss": 0.0373, + "step": 1946 + }, + { + "epoch": 2.5285714285714285, + "grad_norm": 22.79715199427616, + "learning_rate": 1.7509088403358708e-05, + "loss": 0.0922, + "step": 1947 + }, + { + "epoch": 2.5298701298701296, + "grad_norm": 29.201467526669756, + "learning_rate": 1.750630995604559e-05, + "loss": 0.0703, + "step": 1948 + }, + { + "epoch": 2.531168831168831, + "grad_norm": 34.91075548434902, + "learning_rate": 1.750353018072257e-05, + "loss": 0.099, + "step": 1949 + }, + { + "epoch": 2.5324675324675323, + "grad_norm": 29.339362326729006, + "learning_rate": 1.750074907788144e-05, + "loss": 0.0579, + "step": 1950 + }, + { + "epoch": 2.533766233766234, + "grad_norm": 15.397545091108823, + "learning_rate": 1.7497966648014236e-05, + "loss": 0.0741, + "step": 1951 + }, + { + "epoch": 2.535064935064935, + "grad_norm": 31.973956973524796, + "learning_rate": 1.749518289161322e-05, + "loss": 0.0626, + "step": 1952 + }, + { + "epoch": 2.536363636363636, + "grad_norm": 6.699581470329517, + "learning_rate": 1.7492397809170887e-05, + "loss": 0.0404, + "step": 1953 + }, + { + "epoch": 2.5376623376623377, + "grad_norm": 25.176525487739355, + "learning_rate": 1.7489611401179977e-05, + "loss": 0.0582, + "step": 1954 + }, + { + "epoch": 2.538961038961039, + "grad_norm": 24.129268797222196, + "learning_rate": 1.7486823668133456e-05, + "loss": 0.0752, + "step": 1955 + }, + { + "epoch": 2.5402597402597404, + "grad_norm": 30.76749835031644, + "learning_rate": 1.7484034610524533e-05, + "loss": 0.0657, + "step": 1956 + }, + { + "epoch": 2.5415584415584416, + "grad_norm": 25.13462598060194, + "learning_rate": 1.7481244228846637e-05, + "loss": 0.0598, + "step": 1957 + }, + { + "epoch": 2.5428571428571427, + "grad_norm": 17.72565771031066, + "learning_rate": 1.7478452523593444e-05, + "loss": 0.0704, + "step": 1958 + }, + { + "epoch": 2.5441558441558443, + "grad_norm": 37.84614580414741, + "learning_rate": 1.7475659495258864e-05, + "loss": 0.0841, + "step": 1959 + }, + { + "epoch": 2.5454545454545454, + "grad_norm": 19.16971958234217, + "learning_rate": 1.747286514433703e-05, + "loss": 0.0522, + "step": 1960 + }, + { + "epoch": 2.546753246753247, + "grad_norm": 32.12640608925153, + "learning_rate": 1.7470069471322325e-05, + "loss": 0.0616, + "step": 1961 + }, + { + "epoch": 2.548051948051948, + "grad_norm": 15.437799187853088, + "learning_rate": 1.7467272476709348e-05, + "loss": 0.0623, + "step": 1962 + }, + { + "epoch": 2.5493506493506493, + "grad_norm": 16.68200378465241, + "learning_rate": 1.7464474160992948e-05, + "loss": 0.067, + "step": 1963 + }, + { + "epoch": 2.5506493506493504, + "grad_norm": 11.450616407071305, + "learning_rate": 1.7461674524668194e-05, + "loss": 0.0712, + "step": 1964 + }, + { + "epoch": 2.551948051948052, + "grad_norm": 18.119037842453917, + "learning_rate": 1.74588735682304e-05, + "loss": 0.0412, + "step": 1965 + }, + { + "epoch": 2.553246753246753, + "grad_norm": 5.620774896530188, + "learning_rate": 1.745607129217511e-05, + "loss": 0.0397, + "step": 1966 + }, + { + "epoch": 2.5545454545454547, + "grad_norm": 19.808486008671817, + "learning_rate": 1.7453267696998094e-05, + "loss": 0.1001, + "step": 1967 + }, + { + "epoch": 2.555844155844156, + "grad_norm": 8.49950937641073, + "learning_rate": 1.745046278319536e-05, + "loss": 0.0803, + "step": 1968 + }, + { + "epoch": 2.557142857142857, + "grad_norm": 37.04243525305229, + "learning_rate": 1.7447656551263166e-05, + "loss": 0.0835, + "step": 1969 + }, + { + "epoch": 2.5584415584415585, + "grad_norm": 9.761632656642849, + "learning_rate": 1.7444849001697976e-05, + "loss": 0.0734, + "step": 1970 + }, + { + "epoch": 2.5597402597402596, + "grad_norm": 25.906776271855847, + "learning_rate": 1.74420401349965e-05, + "loss": 0.0518, + "step": 1971 + }, + { + "epoch": 2.5610389610389612, + "grad_norm": 18.900991632572303, + "learning_rate": 1.7439229951655682e-05, + "loss": 0.0746, + "step": 1972 + }, + { + "epoch": 2.5623376623376624, + "grad_norm": 28.915561753073813, + "learning_rate": 1.74364184521727e-05, + "loss": 0.0695, + "step": 1973 + }, + { + "epoch": 2.5636363636363635, + "grad_norm": 36.707191223984566, + "learning_rate": 1.7433605637044954e-05, + "loss": 0.0551, + "step": 1974 + }, + { + "epoch": 2.564935064935065, + "grad_norm": 38.33392483417342, + "learning_rate": 1.7430791506770096e-05, + "loss": 0.0932, + "step": 1975 + }, + { + "epoch": 2.566233766233766, + "grad_norm": 35.05801743386068, + "learning_rate": 1.742797606184599e-05, + "loss": 0.0424, + "step": 1976 + }, + { + "epoch": 2.5675324675324678, + "grad_norm": 17.088033824019856, + "learning_rate": 1.7425159302770743e-05, + "loss": 0.0629, + "step": 1977 + }, + { + "epoch": 2.568831168831169, + "grad_norm": 59.81263897743532, + "learning_rate": 1.74223412300427e-05, + "loss": 0.0825, + "step": 1978 + }, + { + "epoch": 2.57012987012987, + "grad_norm": 23.872977528565805, + "learning_rate": 1.741952184416043e-05, + "loss": 0.0482, + "step": 1979 + }, + { + "epoch": 2.571428571428571, + "grad_norm": 47.96568982921079, + "learning_rate": 1.7416701145622733e-05, + "loss": 0.0816, + "step": 1980 + }, + { + "epoch": 2.5727272727272728, + "grad_norm": 9.478935453061654, + "learning_rate": 1.7413879134928644e-05, + "loss": 0.0429, + "step": 1981 + }, + { + "epoch": 2.574025974025974, + "grad_norm": 29.91703597830865, + "learning_rate": 1.7411055812577436e-05, + "loss": 0.0838, + "step": 1982 + }, + { + "epoch": 2.5753246753246755, + "grad_norm": 26.175660895790212, + "learning_rate": 1.74082311790686e-05, + "loss": 0.0664, + "step": 1983 + }, + { + "epoch": 2.5766233766233766, + "grad_norm": 17.180890595129675, + "learning_rate": 1.740540523490188e-05, + "loss": 0.0704, + "step": 1984 + }, + { + "epoch": 2.5779220779220777, + "grad_norm": 22.424985670247423, + "learning_rate": 1.7402577980577228e-05, + "loss": 0.08, + "step": 1985 + }, + { + "epoch": 2.5792207792207793, + "grad_norm": 22.303257695448192, + "learning_rate": 1.739974941659485e-05, + "loss": 0.1105, + "step": 1986 + }, + { + "epoch": 2.5805194805194804, + "grad_norm": 23.41880759642185, + "learning_rate": 1.739691954345516e-05, + "loss": 0.0487, + "step": 1987 + }, + { + "epoch": 2.581818181818182, + "grad_norm": 28.02828297683448, + "learning_rate": 1.739408836165883e-05, + "loss": 0.09, + "step": 1988 + }, + { + "epoch": 2.583116883116883, + "grad_norm": 9.720419324120103, + "learning_rate": 1.7391255871706743e-05, + "loss": 0.0459, + "step": 1989 + }, + { + "epoch": 2.5844155844155843, + "grad_norm": 29.598376971170048, + "learning_rate": 1.7388422074100022e-05, + "loss": 0.0643, + "step": 1990 + }, + { + "epoch": 2.585714285714286, + "grad_norm": 8.112037613110262, + "learning_rate": 1.738558696934002e-05, + "loss": 0.0627, + "step": 1991 + }, + { + "epoch": 2.587012987012987, + "grad_norm": 32.48457394963862, + "learning_rate": 1.7382750557928323e-05, + "loss": 0.0726, + "step": 1992 + }, + { + "epoch": 2.5883116883116886, + "grad_norm": 13.189482255367201, + "learning_rate": 1.7379912840366743e-05, + "loss": 0.0452, + "step": 1993 + }, + { + "epoch": 2.5896103896103897, + "grad_norm": 28.253242280886415, + "learning_rate": 1.737707381715733e-05, + "loss": 0.0528, + "step": 1994 + }, + { + "epoch": 2.590909090909091, + "grad_norm": 11.183809907384596, + "learning_rate": 1.737423348880236e-05, + "loss": 0.1006, + "step": 1995 + }, + { + "epoch": 2.592207792207792, + "grad_norm": 12.043678182894508, + "learning_rate": 1.737139185580434e-05, + "loss": 0.0588, + "step": 1996 + }, + { + "epoch": 2.5935064935064935, + "grad_norm": 10.953405068352433, + "learning_rate": 1.736854891866601e-05, + "loss": 0.06, + "step": 1997 + }, + { + "epoch": 2.5948051948051947, + "grad_norm": 11.822669211898019, + "learning_rate": 1.7365704677890344e-05, + "loss": 0.0625, + "step": 1998 + }, + { + "epoch": 2.5961038961038962, + "grad_norm": 4.246381549821179, + "learning_rate": 1.7362859133980538e-05, + "loss": 0.0617, + "step": 1999 + }, + { + "epoch": 2.5974025974025974, + "grad_norm": 12.865638890847858, + "learning_rate": 1.7360012287440023e-05, + "loss": 0.0446, + "step": 2000 + }, + { + "epoch": 2.5987012987012985, + "grad_norm": 18.18843342846506, + "learning_rate": 1.7357164138772466e-05, + "loss": 0.0929, + "step": 2001 + }, + { + "epoch": 2.6, + "grad_norm": 24.269366857366517, + "learning_rate": 1.735431468848175e-05, + "loss": 0.0542, + "step": 2002 + }, + { + "epoch": 2.601298701298701, + "grad_norm": 27.702553302770397, + "learning_rate": 1.7351463937072008e-05, + "loss": 0.0815, + "step": 2003 + }, + { + "epoch": 2.602597402597403, + "grad_norm": 29.995996034948014, + "learning_rate": 1.734861188504758e-05, + "loss": 0.0481, + "step": 2004 + }, + { + "epoch": 2.603896103896104, + "grad_norm": 17.48602152295851, + "learning_rate": 1.7345758532913063e-05, + "loss": 0.0548, + "step": 2005 + }, + { + "epoch": 2.605194805194805, + "grad_norm": 41.63854151500411, + "learning_rate": 1.7342903881173255e-05, + "loss": 0.0533, + "step": 2006 + }, + { + "epoch": 2.6064935064935066, + "grad_norm": 18.60973600423466, + "learning_rate": 1.7340047930333212e-05, + "loss": 0.09, + "step": 2007 + }, + { + "epoch": 2.6077922077922078, + "grad_norm": 32.244771649993865, + "learning_rate": 1.73371906808982e-05, + "loss": 0.0604, + "step": 2008 + }, + { + "epoch": 2.6090909090909093, + "grad_norm": 11.926155498337437, + "learning_rate": 1.733433213337372e-05, + "loss": 0.0817, + "step": 2009 + }, + { + "epoch": 2.6103896103896105, + "grad_norm": 28.36409179440706, + "learning_rate": 1.73314722882655e-05, + "loss": 0.0742, + "step": 2010 + }, + { + "epoch": 2.6116883116883116, + "grad_norm": 5.332892423902022, + "learning_rate": 1.732861114607951e-05, + "loss": 0.0524, + "step": 2011 + }, + { + "epoch": 2.6129870129870127, + "grad_norm": 31.59260221996255, + "learning_rate": 1.7325748707321937e-05, + "loss": 0.0463, + "step": 2012 + }, + { + "epoch": 2.6142857142857143, + "grad_norm": 7.548070287159281, + "learning_rate": 1.7322884972499204e-05, + "loss": 0.0599, + "step": 2013 + }, + { + "epoch": 2.6155844155844155, + "grad_norm": 29.257264274237272, + "learning_rate": 1.7320019942117954e-05, + "loss": 0.053, + "step": 2014 + }, + { + "epoch": 2.616883116883117, + "grad_norm": 15.871613268535341, + "learning_rate": 1.731715361668507e-05, + "loss": 0.0586, + "step": 2015 + }, + { + "epoch": 2.618181818181818, + "grad_norm": 8.871051471988928, + "learning_rate": 1.7314285996707664e-05, + "loss": 0.058, + "step": 2016 + }, + { + "epoch": 2.6194805194805193, + "grad_norm": 37.599034464966316, + "learning_rate": 1.7311417082693063e-05, + "loss": 0.0498, + "step": 2017 + }, + { + "epoch": 2.620779220779221, + "grad_norm": 8.704384807785054, + "learning_rate": 1.7308546875148842e-05, + "loss": 0.0784, + "step": 2018 + }, + { + "epoch": 2.622077922077922, + "grad_norm": 28.282351865520912, + "learning_rate": 1.730567537458279e-05, + "loss": 0.0466, + "step": 2019 + }, + { + "epoch": 2.6233766233766236, + "grad_norm": 26.922371054017493, + "learning_rate": 1.730280258150293e-05, + "loss": 0.0619, + "step": 2020 + }, + { + "epoch": 2.6246753246753247, + "grad_norm": 24.996986866577785, + "learning_rate": 1.729992849641752e-05, + "loss": 0.0474, + "step": 2021 + }, + { + "epoch": 2.625974025974026, + "grad_norm": 46.13880730802449, + "learning_rate": 1.7297053119835035e-05, + "loss": 0.0912, + "step": 2022 + }, + { + "epoch": 2.6272727272727274, + "grad_norm": 21.343886216815292, + "learning_rate": 1.7294176452264185e-05, + "loss": 0.125, + "step": 2023 + }, + { + "epoch": 2.6285714285714286, + "grad_norm": 57.68882024105048, + "learning_rate": 1.7291298494213912e-05, + "loss": 0.0712, + "step": 2024 + }, + { + "epoch": 2.62987012987013, + "grad_norm": 11.945609962707945, + "learning_rate": 1.7288419246193373e-05, + "loss": 0.062, + "step": 2025 + }, + { + "epoch": 2.6311688311688313, + "grad_norm": 42.838826458202, + "learning_rate": 1.7285538708711972e-05, + "loss": 0.0526, + "step": 2026 + }, + { + "epoch": 2.6324675324675324, + "grad_norm": 15.331764360042012, + "learning_rate": 1.7282656882279326e-05, + "loss": 0.0479, + "step": 2027 + }, + { + "epoch": 2.6337662337662335, + "grad_norm": 22.758028813428357, + "learning_rate": 1.7279773767405284e-05, + "loss": 0.0508, + "step": 2028 + }, + { + "epoch": 2.635064935064935, + "grad_norm": 27.683217545385276, + "learning_rate": 1.727688936459993e-05, + "loss": 0.0507, + "step": 2029 + }, + { + "epoch": 2.6363636363636362, + "grad_norm": 12.717350829744532, + "learning_rate": 1.727400367437356e-05, + "loss": 0.0862, + "step": 2030 + }, + { + "epoch": 2.637662337662338, + "grad_norm": 30.9338815515116, + "learning_rate": 1.7271116697236717e-05, + "loss": 0.0878, + "step": 2031 + }, + { + "epoch": 2.638961038961039, + "grad_norm": 5.679610940894336, + "learning_rate": 1.726822843370016e-05, + "loss": 0.1018, + "step": 2032 + }, + { + "epoch": 2.64025974025974, + "grad_norm": 36.84124473006273, + "learning_rate": 1.726533888427488e-05, + "loss": 0.0748, + "step": 2033 + }, + { + "epoch": 2.6415584415584417, + "grad_norm": 17.779307592257037, + "learning_rate": 1.7262448049472088e-05, + "loss": 0.0524, + "step": 2034 + }, + { + "epoch": 2.642857142857143, + "grad_norm": 36.17244256268684, + "learning_rate": 1.725955592980323e-05, + "loss": 0.0958, + "step": 2035 + }, + { + "epoch": 2.6441558441558444, + "grad_norm": 29.394974767596885, + "learning_rate": 1.7256662525779984e-05, + "loss": 0.0528, + "step": 2036 + }, + { + "epoch": 2.6454545454545455, + "grad_norm": 9.498975827238015, + "learning_rate": 1.725376783791424e-05, + "loss": 0.0772, + "step": 2037 + }, + { + "epoch": 2.6467532467532466, + "grad_norm": 52.82097291978553, + "learning_rate": 1.7250871866718127e-05, + "loss": 0.0837, + "step": 2038 + }, + { + "epoch": 2.648051948051948, + "grad_norm": 5.990194403393771, + "learning_rate": 1.7247974612704e-05, + "loss": 0.0606, + "step": 2039 + }, + { + "epoch": 2.6493506493506493, + "grad_norm": 32.61651347061876, + "learning_rate": 1.724507607638443e-05, + "loss": 0.0687, + "step": 2040 + }, + { + "epoch": 2.650649350649351, + "grad_norm": 36.31880982719732, + "learning_rate": 1.7242176258272235e-05, + "loss": 0.0544, + "step": 2041 + }, + { + "epoch": 2.651948051948052, + "grad_norm": 19.18173614021617, + "learning_rate": 1.7239275158880444e-05, + "loss": 0.0516, + "step": 2042 + }, + { + "epoch": 2.653246753246753, + "grad_norm": 35.4111877521288, + "learning_rate": 1.723637277872232e-05, + "loss": 0.0657, + "step": 2043 + }, + { + "epoch": 2.6545454545454543, + "grad_norm": 13.442540155225755, + "learning_rate": 1.7233469118311338e-05, + "loss": 0.0494, + "step": 2044 + }, + { + "epoch": 2.655844155844156, + "grad_norm": 26.518739011108206, + "learning_rate": 1.723056417816122e-05, + "loss": 0.0624, + "step": 2045 + }, + { + "epoch": 2.657142857142857, + "grad_norm": 22.416674447747354, + "learning_rate": 1.7227657958785904e-05, + "loss": 0.0542, + "step": 2046 + }, + { + "epoch": 2.6584415584415586, + "grad_norm": 18.89061939387512, + "learning_rate": 1.7224750460699558e-05, + "loss": 0.0759, + "step": 2047 + }, + { + "epoch": 2.6597402597402597, + "grad_norm": 33.65545139235662, + "learning_rate": 1.722184168441657e-05, + "loss": 0.0696, + "step": 2048 + }, + { + "epoch": 2.661038961038961, + "grad_norm": 32.43027508981188, + "learning_rate": 1.7218931630451563e-05, + "loss": 0.092, + "step": 2049 + }, + { + "epoch": 2.6623376623376624, + "grad_norm": 11.131131130292594, + "learning_rate": 1.7216020299319377e-05, + "loss": 0.059, + "step": 2050 + }, + { + "epoch": 2.6636363636363636, + "grad_norm": 22.666784357461772, + "learning_rate": 1.7213107691535088e-05, + "loss": 0.0436, + "step": 2051 + }, + { + "epoch": 2.664935064935065, + "grad_norm": 5.552035126809306, + "learning_rate": 1.7210193807613983e-05, + "loss": 0.0418, + "step": 2052 + }, + { + "epoch": 2.6662337662337663, + "grad_norm": 18.772204368245365, + "learning_rate": 1.7207278648071593e-05, + "loss": 0.0475, + "step": 2053 + }, + { + "epoch": 2.6675324675324674, + "grad_norm": 6.315294917282179, + "learning_rate": 1.7204362213423656e-05, + "loss": 0.0764, + "step": 2054 + }, + { + "epoch": 2.6688311688311686, + "grad_norm": 20.12205934816386, + "learning_rate": 1.7201444504186158e-05, + "loss": 0.0364, + "step": 2055 + }, + { + "epoch": 2.67012987012987, + "grad_norm": 19.053368124948936, + "learning_rate": 1.7198525520875287e-05, + "loss": 0.0721, + "step": 2056 + }, + { + "epoch": 2.6714285714285713, + "grad_norm": 8.624103679229354, + "learning_rate": 1.7195605264007466e-05, + "loss": 0.0813, + "step": 2057 + }, + { + "epoch": 2.672727272727273, + "grad_norm": 10.347506410220609, + "learning_rate": 1.7192683734099355e-05, + "loss": 0.0804, + "step": 2058 + }, + { + "epoch": 2.674025974025974, + "grad_norm": 30.460463920836418, + "learning_rate": 1.7189760931667815e-05, + "loss": 0.0825, + "step": 2059 + }, + { + "epoch": 2.675324675324675, + "grad_norm": 4.953854935276032, + "learning_rate": 1.7186836857229958e-05, + "loss": 0.0539, + "step": 2060 + }, + { + "epoch": 2.6766233766233767, + "grad_norm": 16.09568039528579, + "learning_rate": 1.71839115113031e-05, + "loss": 0.0504, + "step": 2061 + }, + { + "epoch": 2.677922077922078, + "grad_norm": 28.150303490234556, + "learning_rate": 1.7180984894404796e-05, + "loss": 0.1106, + "step": 2062 + }, + { + "epoch": 2.6792207792207794, + "grad_norm": 25.068534975498032, + "learning_rate": 1.7178057007052823e-05, + "loss": 0.0366, + "step": 2063 + }, + { + "epoch": 2.6805194805194805, + "grad_norm": 32.88520381071973, + "learning_rate": 1.717512784976517e-05, + "loss": 0.0654, + "step": 2064 + }, + { + "epoch": 2.6818181818181817, + "grad_norm": 15.20123721376505, + "learning_rate": 1.7172197423060065e-05, + "loss": 0.0612, + "step": 2065 + }, + { + "epoch": 2.6831168831168832, + "grad_norm": 43.72807991774755, + "learning_rate": 1.7169265727455964e-05, + "loss": 0.1396, + "step": 2066 + }, + { + "epoch": 2.6844155844155844, + "grad_norm": 17.436369628061307, + "learning_rate": 1.7166332763471534e-05, + "loss": 0.0756, + "step": 2067 + }, + { + "epoch": 2.685714285714286, + "grad_norm": 25.51526481789085, + "learning_rate": 1.716339853162567e-05, + "loss": 0.0374, + "step": 2068 + }, + { + "epoch": 2.687012987012987, + "grad_norm": 6.484024919961059, + "learning_rate": 1.71604630324375e-05, + "loss": 0.0645, + "step": 2069 + }, + { + "epoch": 2.688311688311688, + "grad_norm": 10.576878513052483, + "learning_rate": 1.7157526266426364e-05, + "loss": 0.0442, + "step": 2070 + }, + { + "epoch": 2.6896103896103893, + "grad_norm": 7.716848501212515, + "learning_rate": 1.7154588234111837e-05, + "loss": 0.0366, + "step": 2071 + }, + { + "epoch": 2.690909090909091, + "grad_norm": 10.69806727176823, + "learning_rate": 1.715164893601371e-05, + "loss": 0.0864, + "step": 2072 + }, + { + "epoch": 2.692207792207792, + "grad_norm": 17.382716453353357, + "learning_rate": 1.7148708372652002e-05, + "loss": 0.0551, + "step": 2073 + }, + { + "epoch": 2.6935064935064936, + "grad_norm": 7.236828346898619, + "learning_rate": 1.7145766544546956e-05, + "loss": 0.0543, + "step": 2074 + }, + { + "epoch": 2.6948051948051948, + "grad_norm": 9.31674509322659, + "learning_rate": 1.7142823452219036e-05, + "loss": 0.0812, + "step": 2075 + }, + { + "epoch": 2.696103896103896, + "grad_norm": 12.901421123714519, + "learning_rate": 1.7139879096188937e-05, + "loss": 0.0484, + "step": 2076 + }, + { + "epoch": 2.6974025974025975, + "grad_norm": 20.556701794268307, + "learning_rate": 1.7136933476977563e-05, + "loss": 0.0648, + "step": 2077 + }, + { + "epoch": 2.6987012987012986, + "grad_norm": 16.22698981667787, + "learning_rate": 1.713398659510605e-05, + "loss": 0.0518, + "step": 2078 + }, + { + "epoch": 2.7, + "grad_norm": 22.363870315942926, + "learning_rate": 1.7131038451095768e-05, + "loss": 0.0756, + "step": 2079 + }, + { + "epoch": 2.7012987012987013, + "grad_norm": 6.506169007486819, + "learning_rate": 1.7128089045468294e-05, + "loss": 0.0394, + "step": 2080 + }, + { + "epoch": 2.7025974025974024, + "grad_norm": 24.980535491414713, + "learning_rate": 1.7125138378745433e-05, + "loss": 0.0456, + "step": 2081 + }, + { + "epoch": 2.703896103896104, + "grad_norm": 13.927325688358305, + "learning_rate": 1.712218645144922e-05, + "loss": 0.0459, + "step": 2082 + }, + { + "epoch": 2.705194805194805, + "grad_norm": 7.541932502542766, + "learning_rate": 1.71192332641019e-05, + "loss": 0.0668, + "step": 2083 + }, + { + "epoch": 2.7064935064935067, + "grad_norm": 25.805311434866127, + "learning_rate": 1.7116278817225953e-05, + "loss": 0.0681, + "step": 2084 + }, + { + "epoch": 2.707792207792208, + "grad_norm": 9.153236392869868, + "learning_rate": 1.7113323111344077e-05, + "loss": 0.0817, + "step": 2085 + }, + { + "epoch": 2.709090909090909, + "grad_norm": 26.89607232651797, + "learning_rate": 1.7110366146979188e-05, + "loss": 0.0667, + "step": 2086 + }, + { + "epoch": 2.71038961038961, + "grad_norm": 3.8834649373554484, + "learning_rate": 1.710740792465444e-05, + "loss": 0.0746, + "step": 2087 + }, + { + "epoch": 2.7116883116883117, + "grad_norm": 15.395399435867464, + "learning_rate": 1.710444844489319e-05, + "loss": 0.0534, + "step": 2088 + }, + { + "epoch": 2.712987012987013, + "grad_norm": 3.96519669200472, + "learning_rate": 1.710148770821903e-05, + "loss": 0.0448, + "step": 2089 + }, + { + "epoch": 2.7142857142857144, + "grad_norm": 12.038733851422949, + "learning_rate": 1.709852571515577e-05, + "loss": 0.0453, + "step": 2090 + }, + { + "epoch": 2.7155844155844155, + "grad_norm": 5.87213571598351, + "learning_rate": 1.709556246622744e-05, + "loss": 0.0623, + "step": 2091 + }, + { + "epoch": 2.7168831168831167, + "grad_norm": 7.005858351520161, + "learning_rate": 1.7092597961958304e-05, + "loss": 0.0525, + "step": 2092 + }, + { + "epoch": 2.7181818181818183, + "grad_norm": 7.882055793281893, + "learning_rate": 1.7089632202872835e-05, + "loss": 0.0519, + "step": 2093 + }, + { + "epoch": 2.7194805194805194, + "grad_norm": 14.18138834003422, + "learning_rate": 1.708666518949573e-05, + "loss": 0.0621, + "step": 2094 + }, + { + "epoch": 2.720779220779221, + "grad_norm": 8.100402137478005, + "learning_rate": 1.7083696922351913e-05, + "loss": 0.0488, + "step": 2095 + }, + { + "epoch": 2.722077922077922, + "grad_norm": 11.087943075024297, + "learning_rate": 1.708072740196653e-05, + "loss": 0.0509, + "step": 2096 + }, + { + "epoch": 2.7233766233766232, + "grad_norm": 30.010774447242305, + "learning_rate": 1.7077756628864942e-05, + "loss": 0.0505, + "step": 2097 + }, + { + "epoch": 2.724675324675325, + "grad_norm": 4.878259477412357, + "learning_rate": 1.7074784603572735e-05, + "loss": 0.0495, + "step": 2098 + }, + { + "epoch": 2.725974025974026, + "grad_norm": 15.880716013213657, + "learning_rate": 1.707181132661572e-05, + "loss": 0.0581, + "step": 2099 + }, + { + "epoch": 2.7272727272727275, + "grad_norm": 25.704733174372006, + "learning_rate": 1.706883679851992e-05, + "loss": 0.0668, + "step": 2100 + }, + { + "epoch": 2.7285714285714286, + "grad_norm": 20.962164344915713, + "learning_rate": 1.7065861019811598e-05, + "loss": 0.0588, + "step": 2101 + }, + { + "epoch": 2.72987012987013, + "grad_norm": 32.14836349316104, + "learning_rate": 1.7062883991017217e-05, + "loss": 0.0677, + "step": 2102 + }, + { + "epoch": 2.731168831168831, + "grad_norm": 17.58495519632077, + "learning_rate": 1.7059905712663475e-05, + "loss": 0.0447, + "step": 2103 + }, + { + "epoch": 2.7324675324675325, + "grad_norm": 24.63284647143856, + "learning_rate": 1.7056926185277284e-05, + "loss": 0.0489, + "step": 2104 + }, + { + "epoch": 2.7337662337662336, + "grad_norm": 14.850680249860611, + "learning_rate": 1.7053945409385778e-05, + "loss": 0.0818, + "step": 2105 + }, + { + "epoch": 2.735064935064935, + "grad_norm": 14.374410168607335, + "learning_rate": 1.7050963385516314e-05, + "loss": 0.0909, + "step": 2106 + }, + { + "epoch": 2.7363636363636363, + "grad_norm": 3.9730985126701266, + "learning_rate": 1.7047980114196472e-05, + "loss": 0.0602, + "step": 2107 + }, + { + "epoch": 2.7376623376623375, + "grad_norm": 19.769172783532404, + "learning_rate": 1.7044995595954052e-05, + "loss": 0.0502, + "step": 2108 + }, + { + "epoch": 2.738961038961039, + "grad_norm": 7.346296641677996, + "learning_rate": 1.7042009831317062e-05, + "loss": 0.0428, + "step": 2109 + }, + { + "epoch": 2.74025974025974, + "grad_norm": 19.36247338019136, + "learning_rate": 1.703902282081375e-05, + "loss": 0.0357, + "step": 2110 + }, + { + "epoch": 2.7415584415584417, + "grad_norm": 10.174984056612274, + "learning_rate": 1.7036034564972572e-05, + "loss": 0.0491, + "step": 2111 + }, + { + "epoch": 2.742857142857143, + "grad_norm": 12.894667246107247, + "learning_rate": 1.703304506432221e-05, + "loss": 0.0463, + "step": 2112 + }, + { + "epoch": 2.744155844155844, + "grad_norm": 6.364914425216527, + "learning_rate": 1.7030054319391563e-05, + "loss": 0.0566, + "step": 2113 + }, + { + "epoch": 2.7454545454545456, + "grad_norm": 14.919104512059677, + "learning_rate": 1.7027062330709748e-05, + "loss": 0.0516, + "step": 2114 + }, + { + "epoch": 2.7467532467532467, + "grad_norm": 18.633138826137504, + "learning_rate": 1.7024069098806106e-05, + "loss": 0.0399, + "step": 2115 + }, + { + "epoch": 2.7480519480519483, + "grad_norm": 13.535035860491755, + "learning_rate": 1.7021074624210198e-05, + "loss": 0.0676, + "step": 2116 + }, + { + "epoch": 2.7493506493506494, + "grad_norm": 28.307479970302275, + "learning_rate": 1.7018078907451806e-05, + "loss": 0.0362, + "step": 2117 + }, + { + "epoch": 2.7506493506493506, + "grad_norm": 17.917837405241723, + "learning_rate": 1.701508194906093e-05, + "loss": 0.0527, + "step": 2118 + }, + { + "epoch": 2.7519480519480517, + "grad_norm": 8.6408872691462, + "learning_rate": 1.7012083749567786e-05, + "loss": 0.0517, + "step": 2119 + }, + { + "epoch": 2.7532467532467533, + "grad_norm": 15.309805750088747, + "learning_rate": 1.700908430950281e-05, + "loss": 0.0678, + "step": 2120 + }, + { + "epoch": 2.7545454545454544, + "grad_norm": 16.73833424188643, + "learning_rate": 1.7006083629396668e-05, + "loss": 0.065, + "step": 2121 + }, + { + "epoch": 2.755844155844156, + "grad_norm": 15.10931169185886, + "learning_rate": 1.7003081709780235e-05, + "loss": 0.0978, + "step": 2122 + }, + { + "epoch": 2.757142857142857, + "grad_norm": 4.052571843630233, + "learning_rate": 1.7000078551184603e-05, + "loss": 0.0514, + "step": 2123 + }, + { + "epoch": 2.7584415584415583, + "grad_norm": 16.98650334784653, + "learning_rate": 1.6997074154141097e-05, + "loss": 0.0446, + "step": 2124 + }, + { + "epoch": 2.75974025974026, + "grad_norm": 11.692616097930701, + "learning_rate": 1.6994068519181245e-05, + "loss": 0.0655, + "step": 2125 + }, + { + "epoch": 2.761038961038961, + "grad_norm": 9.973370078645406, + "learning_rate": 1.6991061646836806e-05, + "loss": 0.0684, + "step": 2126 + }, + { + "epoch": 2.7623376623376625, + "grad_norm": 20.133489789228683, + "learning_rate": 1.698805353763975e-05, + "loss": 0.0688, + "step": 2127 + }, + { + "epoch": 2.7636363636363637, + "grad_norm": 4.650483731801621, + "learning_rate": 1.698504419212227e-05, + "loss": 0.0599, + "step": 2128 + }, + { + "epoch": 2.764935064935065, + "grad_norm": 29.49383520729832, + "learning_rate": 1.6982033610816774e-05, + "loss": 0.0634, + "step": 2129 + }, + { + "epoch": 2.7662337662337664, + "grad_norm": 11.126871273178708, + "learning_rate": 1.6979021794255902e-05, + "loss": 0.0352, + "step": 2130 + }, + { + "epoch": 2.7675324675324675, + "grad_norm": 8.488684322805808, + "learning_rate": 1.697600874297249e-05, + "loss": 0.0377, + "step": 2131 + }, + { + "epoch": 2.768831168831169, + "grad_norm": 6.254035425327852, + "learning_rate": 1.6972994457499614e-05, + "loss": 0.0684, + "step": 2132 + }, + { + "epoch": 2.77012987012987, + "grad_norm": 4.700527898466244, + "learning_rate": 1.6969978938370548e-05, + "loss": 0.0529, + "step": 2133 + }, + { + "epoch": 2.7714285714285714, + "grad_norm": 6.017651431668933, + "learning_rate": 1.69669621861188e-05, + "loss": 0.0684, + "step": 2134 + }, + { + "epoch": 2.7727272727272725, + "grad_norm": 11.952860210512968, + "learning_rate": 1.6963944201278097e-05, + "loss": 0.0514, + "step": 2135 + }, + { + "epoch": 2.774025974025974, + "grad_norm": 14.218897719601621, + "learning_rate": 1.6960924984382373e-05, + "loss": 0.0504, + "step": 2136 + }, + { + "epoch": 2.775324675324675, + "grad_norm": 22.373478351640028, + "learning_rate": 1.6957904535965783e-05, + "loss": 0.0425, + "step": 2137 + }, + { + "epoch": 2.7766233766233768, + "grad_norm": 26.974796376963983, + "learning_rate": 1.6954882856562705e-05, + "loss": 0.0679, + "step": 2138 + }, + { + "epoch": 2.777922077922078, + "grad_norm": 20.65040196435727, + "learning_rate": 1.6951859946707733e-05, + "loss": 0.0407, + "step": 2139 + }, + { + "epoch": 2.779220779220779, + "grad_norm": 15.213835759131804, + "learning_rate": 1.6948835806935677e-05, + "loss": 0.0882, + "step": 2140 + }, + { + "epoch": 2.7805194805194806, + "grad_norm": 21.085133335133808, + "learning_rate": 1.6945810437781566e-05, + "loss": 0.0668, + "step": 2141 + }, + { + "epoch": 2.7818181818181817, + "grad_norm": 7.197965316161373, + "learning_rate": 1.694278383978064e-05, + "loss": 0.0514, + "step": 2142 + }, + { + "epoch": 2.7831168831168833, + "grad_norm": 24.300740395420284, + "learning_rate": 1.6939756013468372e-05, + "loss": 0.0502, + "step": 2143 + }, + { + "epoch": 2.7844155844155845, + "grad_norm": 8.151251960898133, + "learning_rate": 1.6936726959380434e-05, + "loss": 0.0763, + "step": 2144 + }, + { + "epoch": 2.7857142857142856, + "grad_norm": 11.647906030271802, + "learning_rate": 1.6933696678052726e-05, + "loss": 0.0666, + "step": 2145 + }, + { + "epoch": 2.787012987012987, + "grad_norm": 15.163437731016373, + "learning_rate": 1.6930665170021364e-05, + "loss": 0.0511, + "step": 2146 + }, + { + "epoch": 2.7883116883116883, + "grad_norm": 25.43300859824474, + "learning_rate": 1.6927632435822678e-05, + "loss": 0.0513, + "step": 2147 + }, + { + "epoch": 2.78961038961039, + "grad_norm": 26.83317354429833, + "learning_rate": 1.692459847599322e-05, + "loss": 0.0739, + "step": 2148 + }, + { + "epoch": 2.790909090909091, + "grad_norm": 19.13709783756274, + "learning_rate": 1.692156329106975e-05, + "loss": 0.0526, + "step": 2149 + }, + { + "epoch": 2.792207792207792, + "grad_norm": 7.863177934689875, + "learning_rate": 1.6918526881589254e-05, + "loss": 0.0546, + "step": 2150 + }, + { + "epoch": 2.7935064935064933, + "grad_norm": 37.793091245063316, + "learning_rate": 1.6915489248088934e-05, + "loss": 0.0787, + "step": 2151 + }, + { + "epoch": 2.794805194805195, + "grad_norm": 6.731996594458246, + "learning_rate": 1.6912450391106202e-05, + "loss": 0.0358, + "step": 2152 + }, + { + "epoch": 2.796103896103896, + "grad_norm": 47.50944608754979, + "learning_rate": 1.6909410311178683e-05, + "loss": 0.0709, + "step": 2153 + }, + { + "epoch": 2.7974025974025976, + "grad_norm": 13.533068636545813, + "learning_rate": 1.690636900884424e-05, + "loss": 0.0556, + "step": 2154 + }, + { + "epoch": 2.7987012987012987, + "grad_norm": 49.56268173447333, + "learning_rate": 1.6903326484640928e-05, + "loss": 0.0667, + "step": 2155 + }, + { + "epoch": 2.8, + "grad_norm": 25.11502867907055, + "learning_rate": 1.690028273910703e-05, + "loss": 0.0379, + "step": 2156 + }, + { + "epoch": 2.8012987012987014, + "grad_norm": 37.91741134686253, + "learning_rate": 1.6897237772781046e-05, + "loss": 0.0675, + "step": 2157 + }, + { + "epoch": 2.8025974025974025, + "grad_norm": 41.532663570536414, + "learning_rate": 1.6894191586201677e-05, + "loss": 0.0838, + "step": 2158 + }, + { + "epoch": 2.803896103896104, + "grad_norm": 19.129291230308965, + "learning_rate": 1.689114417990787e-05, + "loss": 0.0231, + "step": 2159 + }, + { + "epoch": 2.8051948051948052, + "grad_norm": 48.147632637974375, + "learning_rate": 1.688809555443875e-05, + "loss": 0.0429, + "step": 2160 + }, + { + "epoch": 2.8064935064935064, + "grad_norm": 20.61810195908609, + "learning_rate": 1.6885045710333693e-05, + "loss": 0.1093, + "step": 2161 + }, + { + "epoch": 2.807792207792208, + "grad_norm": 25.184343137001278, + "learning_rate": 1.6881994648132266e-05, + "loss": 0.0601, + "step": 2162 + }, + { + "epoch": 2.809090909090909, + "grad_norm": 12.09706924895503, + "learning_rate": 1.6878942368374266e-05, + "loss": 0.0792, + "step": 2163 + }, + { + "epoch": 2.8103896103896107, + "grad_norm": 6.114827497933598, + "learning_rate": 1.6875888871599694e-05, + "loss": 0.0399, + "step": 2164 + }, + { + "epoch": 2.811688311688312, + "grad_norm": 5.675699569551319, + "learning_rate": 1.6872834158348776e-05, + "loss": 0.0688, + "step": 2165 + }, + { + "epoch": 2.812987012987013, + "grad_norm": 9.247391684849317, + "learning_rate": 1.6869778229161947e-05, + "loss": 0.1002, + "step": 2166 + }, + { + "epoch": 2.814285714285714, + "grad_norm": 7.860228414283565, + "learning_rate": 1.686672108457986e-05, + "loss": 0.0416, + "step": 2167 + }, + { + "epoch": 2.8155844155844156, + "grad_norm": 12.4261730592762, + "learning_rate": 1.6863662725143383e-05, + "loss": 0.0978, + "step": 2168 + }, + { + "epoch": 2.8168831168831168, + "grad_norm": 23.56732093223641, + "learning_rate": 1.68606031513936e-05, + "loss": 0.043, + "step": 2169 + }, + { + "epoch": 2.8181818181818183, + "grad_norm": 6.2356292349969085, + "learning_rate": 1.6857542363871802e-05, + "loss": 0.0528, + "step": 2170 + }, + { + "epoch": 2.8194805194805195, + "grad_norm": 8.91579745156892, + "learning_rate": 1.685448036311951e-05, + "loss": 0.0501, + "step": 2171 + }, + { + "epoch": 2.8207792207792206, + "grad_norm": 17.33836843720924, + "learning_rate": 1.6851417149678442e-05, + "loss": 0.0534, + "step": 2172 + }, + { + "epoch": 2.822077922077922, + "grad_norm": 24.398325890432936, + "learning_rate": 1.6848352724090545e-05, + "loss": 0.0762, + "step": 2173 + }, + { + "epoch": 2.8233766233766233, + "grad_norm": 12.793897666564266, + "learning_rate": 1.684528708689797e-05, + "loss": 0.0448, + "step": 2174 + }, + { + "epoch": 2.824675324675325, + "grad_norm": 22.03050253070689, + "learning_rate": 1.684222023864309e-05, + "loss": 0.0629, + "step": 2175 + }, + { + "epoch": 2.825974025974026, + "grad_norm": 6.822308496030926, + "learning_rate": 1.6839152179868488e-05, + "loss": 0.0309, + "step": 2176 + }, + { + "epoch": 2.827272727272727, + "grad_norm": 15.07977579324247, + "learning_rate": 1.6836082911116967e-05, + "loss": 0.054, + "step": 2177 + }, + { + "epoch": 2.8285714285714287, + "grad_norm": 11.711333435486665, + "learning_rate": 1.6833012432931532e-05, + "loss": 0.0688, + "step": 2178 + }, + { + "epoch": 2.82987012987013, + "grad_norm": 30.481061503814317, + "learning_rate": 1.682994074585541e-05, + "loss": 0.0669, + "step": 2179 + }, + { + "epoch": 2.8311688311688314, + "grad_norm": 9.329048641222544, + "learning_rate": 1.6826867850432046e-05, + "loss": 0.0844, + "step": 2180 + }, + { + "epoch": 2.8324675324675326, + "grad_norm": 6.431439572851878, + "learning_rate": 1.682379374720509e-05, + "loss": 0.0442, + "step": 2181 + }, + { + "epoch": 2.8337662337662337, + "grad_norm": 15.55584845234338, + "learning_rate": 1.6820718436718413e-05, + "loss": 0.0678, + "step": 2182 + }, + { + "epoch": 2.835064935064935, + "grad_norm": 13.68732837496618, + "learning_rate": 1.6817641919516093e-05, + "loss": 0.0669, + "step": 2183 + }, + { + "epoch": 2.8363636363636364, + "grad_norm": 21.200033027599297, + "learning_rate": 1.6814564196142427e-05, + "loss": 0.0607, + "step": 2184 + }, + { + "epoch": 2.8376623376623376, + "grad_norm": 26.60290701723481, + "learning_rate": 1.6811485267141922e-05, + "loss": 0.0464, + "step": 2185 + }, + { + "epoch": 2.838961038961039, + "grad_norm": 27.996025307621895, + "learning_rate": 1.68084051330593e-05, + "loss": 0.0963, + "step": 2186 + }, + { + "epoch": 2.8402597402597403, + "grad_norm": 21.556722162964746, + "learning_rate": 1.68053237944395e-05, + "loss": 0.0679, + "step": 2187 + }, + { + "epoch": 2.8415584415584414, + "grad_norm": 5.6896524592247015, + "learning_rate": 1.680224125182766e-05, + "loss": 0.0619, + "step": 2188 + }, + { + "epoch": 2.842857142857143, + "grad_norm": 35.8730369667851, + "learning_rate": 1.679915750576915e-05, + "loss": 0.0522, + "step": 2189 + }, + { + "epoch": 2.844155844155844, + "grad_norm": 22.83325184499179, + "learning_rate": 1.6796072556809536e-05, + "loss": 0.1012, + "step": 2190 + }, + { + "epoch": 2.8454545454545457, + "grad_norm": 28.375198054076773, + "learning_rate": 1.6792986405494613e-05, + "loss": 0.0366, + "step": 2191 + }, + { + "epoch": 2.846753246753247, + "grad_norm": 21.051298648114958, + "learning_rate": 1.678989905237037e-05, + "loss": 0.0566, + "step": 2192 + }, + { + "epoch": 2.848051948051948, + "grad_norm": 21.792356785864033, + "learning_rate": 1.678681049798303e-05, + "loss": 0.0632, + "step": 2193 + }, + { + "epoch": 2.849350649350649, + "grad_norm": 27.972226695951438, + "learning_rate": 1.678372074287901e-05, + "loss": 0.0439, + "step": 2194 + }, + { + "epoch": 2.8506493506493507, + "grad_norm": 8.158086154946975, + "learning_rate": 1.678062978760495e-05, + "loss": 0.0425, + "step": 2195 + }, + { + "epoch": 2.851948051948052, + "grad_norm": 36.41258557593457, + "learning_rate": 1.6777537632707694e-05, + "loss": 0.0905, + "step": 2196 + }, + { + "epoch": 2.8532467532467534, + "grad_norm": 9.311899249516372, + "learning_rate": 1.677444427873431e-05, + "loss": 0.0746, + "step": 2197 + }, + { + "epoch": 2.8545454545454545, + "grad_norm": 41.95101987868942, + "learning_rate": 1.677134972623207e-05, + "loss": 0.0864, + "step": 2198 + }, + { + "epoch": 2.8558441558441556, + "grad_norm": 24.47418208074144, + "learning_rate": 1.6768253975748455e-05, + "loss": 0.1071, + "step": 2199 + }, + { + "epoch": 2.857142857142857, + "grad_norm": 47.92228060053897, + "learning_rate": 1.676515702783117e-05, + "loss": 0.0677, + "step": 2200 + }, + { + "epoch": 2.8584415584415583, + "grad_norm": 12.953273027225196, + "learning_rate": 1.676205888302812e-05, + "loss": 0.0611, + "step": 2201 + }, + { + "epoch": 2.85974025974026, + "grad_norm": 4.910505979181398, + "learning_rate": 1.675895954188742e-05, + "loss": 0.0868, + "step": 2202 + }, + { + "epoch": 2.861038961038961, + "grad_norm": 36.32172676121144, + "learning_rate": 1.6755859004957414e-05, + "loss": 0.0582, + "step": 2203 + }, + { + "epoch": 2.862337662337662, + "grad_norm": 12.78695503588204, + "learning_rate": 1.6752757272786645e-05, + "loss": 0.0642, + "step": 2204 + }, + { + "epoch": 2.8636363636363638, + "grad_norm": 40.728319163993646, + "learning_rate": 1.674965434592386e-05, + "loss": 0.0693, + "step": 2205 + }, + { + "epoch": 2.864935064935065, + "grad_norm": 16.414935851922426, + "learning_rate": 1.6746550224918032e-05, + "loss": 0.0627, + "step": 2206 + }, + { + "epoch": 2.8662337662337665, + "grad_norm": 21.254936666884095, + "learning_rate": 1.674344491031834e-05, + "loss": 0.0648, + "step": 2207 + }, + { + "epoch": 2.8675324675324676, + "grad_norm": 35.84619031965191, + "learning_rate": 1.6740338402674175e-05, + "loss": 0.0954, + "step": 2208 + }, + { + "epoch": 2.8688311688311687, + "grad_norm": 14.882995210468634, + "learning_rate": 1.6737230702535132e-05, + "loss": 0.0861, + "step": 2209 + }, + { + "epoch": 2.87012987012987, + "grad_norm": 40.408734981753035, + "learning_rate": 1.6734121810451024e-05, + "loss": 0.0533, + "step": 2210 + }, + { + "epoch": 2.8714285714285714, + "grad_norm": 9.656731914786041, + "learning_rate": 1.673101172697188e-05, + "loss": 0.05, + "step": 2211 + }, + { + "epoch": 2.8727272727272726, + "grad_norm": 25.08015871276543, + "learning_rate": 1.6727900452647925e-05, + "loss": 0.0872, + "step": 2212 + }, + { + "epoch": 2.874025974025974, + "grad_norm": 12.807890311237099, + "learning_rate": 1.672478798802961e-05, + "loss": 0.0571, + "step": 2213 + }, + { + "epoch": 2.8753246753246753, + "grad_norm": 23.482974668919738, + "learning_rate": 1.672167433366758e-05, + "loss": 0.0611, + "step": 2214 + }, + { + "epoch": 2.8766233766233764, + "grad_norm": 22.011481932358983, + "learning_rate": 1.6718559490112712e-05, + "loss": 0.0471, + "step": 2215 + }, + { + "epoch": 2.877922077922078, + "grad_norm": 17.20386818102233, + "learning_rate": 1.6715443457916074e-05, + "loss": 0.0741, + "step": 2216 + }, + { + "epoch": 2.879220779220779, + "grad_norm": 35.053153857455726, + "learning_rate": 1.6712326237628954e-05, + "loss": 0.0905, + "step": 2217 + }, + { + "epoch": 2.8805194805194807, + "grad_norm": 19.55397158516712, + "learning_rate": 1.6709207829802845e-05, + "loss": 0.0622, + "step": 2218 + }, + { + "epoch": 2.881818181818182, + "grad_norm": 34.77907383572348, + "learning_rate": 1.6706088234989457e-05, + "loss": 0.0568, + "step": 2219 + }, + { + "epoch": 2.883116883116883, + "grad_norm": 4.444903507459741, + "learning_rate": 1.6702967453740702e-05, + "loss": 0.0906, + "step": 2220 + }, + { + "epoch": 2.8844155844155845, + "grad_norm": 43.37541091402389, + "learning_rate": 1.669984548660871e-05, + "loss": 0.0891, + "step": 2221 + }, + { + "epoch": 2.8857142857142857, + "grad_norm": 9.079237318562228, + "learning_rate": 1.6696722334145816e-05, + "loss": 0.0652, + "step": 2222 + }, + { + "epoch": 2.8870129870129873, + "grad_norm": 21.57926316807112, + "learning_rate": 1.6693597996904566e-05, + "loss": 0.0623, + "step": 2223 + }, + { + "epoch": 2.8883116883116884, + "grad_norm": 21.58200541175666, + "learning_rate": 1.669047247543771e-05, + "loss": 0.0464, + "step": 2224 + }, + { + "epoch": 2.8896103896103895, + "grad_norm": 7.144148057147307, + "learning_rate": 1.6687345770298215e-05, + "loss": 0.0452, + "step": 2225 + }, + { + "epoch": 2.8909090909090907, + "grad_norm": 32.77216512851378, + "learning_rate": 1.6684217882039262e-05, + "loss": 0.0325, + "step": 2226 + }, + { + "epoch": 2.8922077922077922, + "grad_norm": 7.336404450548681, + "learning_rate": 1.6681088811214224e-05, + "loss": 0.0592, + "step": 2227 + }, + { + "epoch": 2.8935064935064934, + "grad_norm": 12.629384355430172, + "learning_rate": 1.6677958558376696e-05, + "loss": 0.0319, + "step": 2228 + }, + { + "epoch": 2.894805194805195, + "grad_norm": 23.524497997748327, + "learning_rate": 1.6674827124080486e-05, + "loss": 0.0594, + "step": 2229 + }, + { + "epoch": 2.896103896103896, + "grad_norm": 3.459984391926602, + "learning_rate": 1.66716945088796e-05, + "loss": 0.0646, + "step": 2230 + }, + { + "epoch": 2.897402597402597, + "grad_norm": 38.38358499694811, + "learning_rate": 1.6668560713328253e-05, + "loss": 0.0605, + "step": 2231 + }, + { + "epoch": 2.898701298701299, + "grad_norm": 6.543235722068127, + "learning_rate": 1.6665425737980884e-05, + "loss": 0.0537, + "step": 2232 + }, + { + "epoch": 2.9, + "grad_norm": 33.83265614236597, + "learning_rate": 1.666228958339212e-05, + "loss": 0.0515, + "step": 2233 + }, + { + "epoch": 2.9012987012987015, + "grad_norm": 6.4159036157671405, + "learning_rate": 1.665915225011681e-05, + "loss": 0.046, + "step": 2234 + }, + { + "epoch": 2.9025974025974026, + "grad_norm": 16.30274381566198, + "learning_rate": 1.6656013738710017e-05, + "loss": 0.0802, + "step": 2235 + }, + { + "epoch": 2.9038961038961038, + "grad_norm": 20.692985089915116, + "learning_rate": 1.6652874049726987e-05, + "loss": 0.0654, + "step": 2236 + }, + { + "epoch": 2.9051948051948053, + "grad_norm": 16.508351380809692, + "learning_rate": 1.6649733183723206e-05, + "loss": 0.0321, + "step": 2237 + }, + { + "epoch": 2.9064935064935065, + "grad_norm": 13.591707251222175, + "learning_rate": 1.6646591141254348e-05, + "loss": 0.0144, + "step": 2238 + }, + { + "epoch": 2.907792207792208, + "grad_norm": 11.6306872570196, + "learning_rate": 1.6643447922876297e-05, + "loss": 0.0626, + "step": 2239 + }, + { + "epoch": 2.909090909090909, + "grad_norm": 17.25149843647713, + "learning_rate": 1.6640303529145154e-05, + "loss": 0.0933, + "step": 2240 + }, + { + "epoch": 2.9103896103896103, + "grad_norm": 22.012510121505343, + "learning_rate": 1.663715796061722e-05, + "loss": 0.0439, + "step": 2241 + }, + { + "epoch": 2.9116883116883114, + "grad_norm": 8.231611252554526, + "learning_rate": 1.6634011217849e-05, + "loss": 0.0694, + "step": 2242 + }, + { + "epoch": 2.912987012987013, + "grad_norm": 16.750881275564513, + "learning_rate": 1.6630863301397225e-05, + "loss": 0.0674, + "step": 2243 + }, + { + "epoch": 2.914285714285714, + "grad_norm": 6.332892341384294, + "learning_rate": 1.6627714211818815e-05, + "loss": 0.0485, + "step": 2244 + }, + { + "epoch": 2.9155844155844157, + "grad_norm": 5.753826791605336, + "learning_rate": 1.6624563949670903e-05, + "loss": 0.0635, + "step": 2245 + }, + { + "epoch": 2.916883116883117, + "grad_norm": 9.787698902453247, + "learning_rate": 1.6621412515510838e-05, + "loss": 0.0566, + "step": 2246 + }, + { + "epoch": 2.918181818181818, + "grad_norm": 11.357636631293785, + "learning_rate": 1.6618259909896162e-05, + "loss": 0.0464, + "step": 2247 + }, + { + "epoch": 2.9194805194805196, + "grad_norm": 16.880630521204786, + "learning_rate": 1.661510613338463e-05, + "loss": 0.0517, + "step": 2248 + }, + { + "epoch": 2.9207792207792207, + "grad_norm": 11.1452959459255, + "learning_rate": 1.661195118653421e-05, + "loss": 0.0752, + "step": 2249 + }, + { + "epoch": 2.9220779220779223, + "grad_norm": 11.755867908709686, + "learning_rate": 1.660879506990307e-05, + "loss": 0.0344, + "step": 2250 + }, + { + "epoch": 2.9233766233766234, + "grad_norm": 9.183213042204697, + "learning_rate": 1.660563778404959e-05, + "loss": 0.0764, + "step": 2251 + }, + { + "epoch": 2.9246753246753245, + "grad_norm": 14.10242196317308, + "learning_rate": 1.6602479329532352e-05, + "loss": 0.0566, + "step": 2252 + }, + { + "epoch": 2.925974025974026, + "grad_norm": 5.434935518778879, + "learning_rate": 1.6599319706910148e-05, + "loss": 0.0508, + "step": 2253 + }, + { + "epoch": 2.9272727272727272, + "grad_norm": 6.571460554629789, + "learning_rate": 1.6596158916741977e-05, + "loss": 0.0674, + "step": 2254 + }, + { + "epoch": 2.928571428571429, + "grad_norm": 8.061353060268337, + "learning_rate": 1.659299695958704e-05, + "loss": 0.062, + "step": 2255 + }, + { + "epoch": 2.92987012987013, + "grad_norm": 6.173291248647945, + "learning_rate": 1.6589833836004753e-05, + "loss": 0.0676, + "step": 2256 + }, + { + "epoch": 2.931168831168831, + "grad_norm": 13.906095092676853, + "learning_rate": 1.6586669546554725e-05, + "loss": 0.0434, + "step": 2257 + }, + { + "epoch": 2.9324675324675322, + "grad_norm": 4.139825305752846, + "learning_rate": 1.6583504091796785e-05, + "loss": 0.059, + "step": 2258 + }, + { + "epoch": 2.933766233766234, + "grad_norm": 8.206252910777083, + "learning_rate": 1.658033747229096e-05, + "loss": 0.0644, + "step": 2259 + }, + { + "epoch": 2.935064935064935, + "grad_norm": 9.459291741279596, + "learning_rate": 1.657716968859749e-05, + "loss": 0.0344, + "step": 2260 + }, + { + "epoch": 2.9363636363636365, + "grad_norm": 7.509090311851063, + "learning_rate": 1.6574000741276816e-05, + "loss": 0.0316, + "step": 2261 + }, + { + "epoch": 2.9376623376623376, + "grad_norm": 14.996579790941448, + "learning_rate": 1.657083063088958e-05, + "loss": 0.0662, + "step": 2262 + }, + { + "epoch": 2.9389610389610388, + "grad_norm": 12.013976521252493, + "learning_rate": 1.6567659357996637e-05, + "loss": 0.0455, + "step": 2263 + }, + { + "epoch": 2.9402597402597404, + "grad_norm": 16.17075353389033, + "learning_rate": 1.656448692315905e-05, + "loss": 0.0475, + "step": 2264 + }, + { + "epoch": 2.9415584415584415, + "grad_norm": 13.337255116468063, + "learning_rate": 1.656131332693808e-05, + "loss": 0.0578, + "step": 2265 + }, + { + "epoch": 2.942857142857143, + "grad_norm": 14.496241553414302, + "learning_rate": 1.6558138569895196e-05, + "loss": 0.0311, + "step": 2266 + }, + { + "epoch": 2.944155844155844, + "grad_norm": 15.06445504071939, + "learning_rate": 1.6554962652592078e-05, + "loss": 0.0466, + "step": 2267 + }, + { + "epoch": 2.9454545454545453, + "grad_norm": 11.031050805590484, + "learning_rate": 1.65517855755906e-05, + "loss": 0.0467, + "step": 2268 + }, + { + "epoch": 2.946753246753247, + "grad_norm": 14.864468926770147, + "learning_rate": 1.6548607339452853e-05, + "loss": 0.043, + "step": 2269 + }, + { + "epoch": 2.948051948051948, + "grad_norm": 4.175704377058274, + "learning_rate": 1.6545427944741125e-05, + "loss": 0.083, + "step": 2270 + }, + { + "epoch": 2.9493506493506496, + "grad_norm": 21.811907413538766, + "learning_rate": 1.6542247392017914e-05, + "loss": 0.0617, + "step": 2271 + }, + { + "epoch": 2.9506493506493507, + "grad_norm": 4.974416546982046, + "learning_rate": 1.6539065681845917e-05, + "loss": 0.1084, + "step": 2272 + }, + { + "epoch": 2.951948051948052, + "grad_norm": 27.052338318275027, + "learning_rate": 1.6535882814788046e-05, + "loss": 0.072, + "step": 2273 + }, + { + "epoch": 2.953246753246753, + "grad_norm": 7.788334927624452, + "learning_rate": 1.6532698791407402e-05, + "loss": 0.068, + "step": 2274 + }, + { + "epoch": 2.9545454545454546, + "grad_norm": 24.22915336903172, + "learning_rate": 1.6529513612267307e-05, + "loss": 0.0565, + "step": 2275 + }, + { + "epoch": 2.9558441558441557, + "grad_norm": 8.693905138320563, + "learning_rate": 1.6526327277931278e-05, + "loss": 0.0405, + "step": 2276 + }, + { + "epoch": 2.9571428571428573, + "grad_norm": 25.70770106645142, + "learning_rate": 1.652313978896304e-05, + "loss": 0.0626, + "step": 2277 + }, + { + "epoch": 2.9584415584415584, + "grad_norm": 4.770860320230832, + "learning_rate": 1.6519951145926515e-05, + "loss": 0.0314, + "step": 2278 + }, + { + "epoch": 2.9597402597402596, + "grad_norm": 10.594344998295322, + "learning_rate": 1.651676134938584e-05, + "loss": 0.0392, + "step": 2279 + }, + { + "epoch": 2.961038961038961, + "grad_norm": 14.235509310373729, + "learning_rate": 1.6513570399905354e-05, + "loss": 0.0405, + "step": 2280 + }, + { + "epoch": 2.9623376623376623, + "grad_norm": 9.152646428045383, + "learning_rate": 1.6510378298049593e-05, + "loss": 0.0954, + "step": 2281 + }, + { + "epoch": 2.963636363636364, + "grad_norm": 6.949122924377664, + "learning_rate": 1.65071850443833e-05, + "loss": 0.0508, + "step": 2282 + }, + { + "epoch": 2.964935064935065, + "grad_norm": 7.352474162301802, + "learning_rate": 1.6503990639471422e-05, + "loss": 0.0605, + "step": 2283 + }, + { + "epoch": 2.966233766233766, + "grad_norm": 8.368973962128555, + "learning_rate": 1.650079508387911e-05, + "loss": 0.0788, + "step": 2284 + }, + { + "epoch": 2.9675324675324677, + "grad_norm": 15.837553315847726, + "learning_rate": 1.6497598378171724e-05, + "loss": 0.0711, + "step": 2285 + }, + { + "epoch": 2.968831168831169, + "grad_norm": 8.421780272495887, + "learning_rate": 1.649440052291482e-05, + "loss": 0.0311, + "step": 2286 + }, + { + "epoch": 2.9701298701298704, + "grad_norm": 11.384262307436474, + "learning_rate": 1.649120151867416e-05, + "loss": 0.0446, + "step": 2287 + }, + { + "epoch": 2.9714285714285715, + "grad_norm": 14.048990530863016, + "learning_rate": 1.6488001366015708e-05, + "loss": 0.0497, + "step": 2288 + }, + { + "epoch": 2.9727272727272727, + "grad_norm": 12.478719711133294, + "learning_rate": 1.6484800065505627e-05, + "loss": 0.0521, + "step": 2289 + }, + { + "epoch": 2.974025974025974, + "grad_norm": 19.684573819680832, + "learning_rate": 1.64815976177103e-05, + "loss": 0.0377, + "step": 2290 + }, + { + "epoch": 2.9753246753246754, + "grad_norm": 21.726202541606877, + "learning_rate": 1.6478394023196297e-05, + "loss": 0.0414, + "step": 2291 + }, + { + "epoch": 2.9766233766233765, + "grad_norm": 18.53556700834659, + "learning_rate": 1.647518928253039e-05, + "loss": 0.0484, + "step": 2292 + }, + { + "epoch": 2.977922077922078, + "grad_norm": 21.499488338196137, + "learning_rate": 1.6471983396279564e-05, + "loss": 0.0369, + "step": 2293 + }, + { + "epoch": 2.979220779220779, + "grad_norm": 7.361301883121233, + "learning_rate": 1.6468776365010998e-05, + "loss": 0.0929, + "step": 2294 + }, + { + "epoch": 2.9805194805194803, + "grad_norm": 9.366513654781555, + "learning_rate": 1.646556818929208e-05, + "loss": 0.0464, + "step": 2295 + }, + { + "epoch": 2.981818181818182, + "grad_norm": 8.072070384967986, + "learning_rate": 1.64623588696904e-05, + "loss": 0.0662, + "step": 2296 + }, + { + "epoch": 2.983116883116883, + "grad_norm": 14.090378504564816, + "learning_rate": 1.645914840677374e-05, + "loss": 0.0816, + "step": 2297 + }, + { + "epoch": 2.9844155844155846, + "grad_norm": 8.898659480580038, + "learning_rate": 1.6455936801110102e-05, + "loss": 0.0738, + "step": 2298 + }, + { + "epoch": 2.9857142857142858, + "grad_norm": 20.673644566899977, + "learning_rate": 1.645272405326767e-05, + "loss": 0.0979, + "step": 2299 + }, + { + "epoch": 2.987012987012987, + "grad_norm": 5.825667756767621, + "learning_rate": 1.6449510163814854e-05, + "loss": 0.033, + "step": 2300 + }, + { + "epoch": 2.9883116883116885, + "grad_norm": 16.4079973567308, + "learning_rate": 1.6446295133320243e-05, + "loss": 0.0779, + "step": 2301 + }, + { + "epoch": 2.9896103896103896, + "grad_norm": 8.977977104105936, + "learning_rate": 1.644307896235264e-05, + "loss": 0.0486, + "step": 2302 + }, + { + "epoch": 2.990909090909091, + "grad_norm": 16.38673315796228, + "learning_rate": 1.643986165148104e-05, + "loss": 0.076, + "step": 2303 + }, + { + "epoch": 2.9922077922077923, + "grad_norm": 10.80210735830305, + "learning_rate": 1.643664320127466e-05, + "loss": 0.0537, + "step": 2304 + }, + { + "epoch": 2.9935064935064934, + "grad_norm": 12.787268746471005, + "learning_rate": 1.64334236123029e-05, + "loss": 0.0644, + "step": 2305 + }, + { + "epoch": 2.9948051948051946, + "grad_norm": 13.061689352573383, + "learning_rate": 1.6430202885135362e-05, + "loss": 0.0443, + "step": 2306 + }, + { + "epoch": 2.996103896103896, + "grad_norm": 7.779487068913231, + "learning_rate": 1.642698102034186e-05, + "loss": 0.0703, + "step": 2307 + }, + { + "epoch": 2.9974025974025973, + "grad_norm": 23.88203151255705, + "learning_rate": 1.64237580184924e-05, + "loss": 0.0849, + "step": 2308 + }, + { + "epoch": 2.998701298701299, + "grad_norm": 6.398762572899506, + "learning_rate": 1.6420533880157197e-05, + "loss": 0.0466, + "step": 2309 + }, + { + "epoch": 3.0, + "grad_norm": 18.649591554490165, + "learning_rate": 1.641730860590666e-05, + "loss": 0.0371, + "step": 2310 + }, + { + "epoch": 3.0, + "eval_accuracy": 0.9298147108366086, + "eval_f1": 0.9120219958112061, + "eval_loss": 0.12981492280960083, + "eval_precision": 0.9062616453087283, + "eval_recall": 0.9288620428909042, + "eval_runtime": 13.1506, + "eval_samples_per_second": 135.431, + "eval_steps_per_second": 1.065, + "step": 2310 + }, + { + "epoch": 3.001298701298701, + "grad_norm": 4.183885787241933, + "learning_rate": 1.6414082196311402e-05, + "loss": 0.0251, + "step": 2311 + }, + { + "epoch": 3.0025974025974027, + "grad_norm": 23.664256405267007, + "learning_rate": 1.6410854651942235e-05, + "loss": 0.0592, + "step": 2312 + }, + { + "epoch": 3.003896103896104, + "grad_norm": 6.438722509688945, + "learning_rate": 1.6407625973370175e-05, + "loss": 0.0375, + "step": 2313 + }, + { + "epoch": 3.005194805194805, + "grad_norm": 30.004134978457312, + "learning_rate": 1.6404396161166437e-05, + "loss": 0.042, + "step": 2314 + }, + { + "epoch": 3.0064935064935066, + "grad_norm": 7.387498935780062, + "learning_rate": 1.640116521590244e-05, + "loss": 0.0284, + "step": 2315 + }, + { + "epoch": 3.0077922077922077, + "grad_norm": 24.563398431478454, + "learning_rate": 1.6397933138149794e-05, + "loss": 0.0258, + "step": 2316 + }, + { + "epoch": 3.0090909090909093, + "grad_norm": 4.1190459550836, + "learning_rate": 1.6394699928480316e-05, + "loss": 0.0359, + "step": 2317 + }, + { + "epoch": 3.0103896103896104, + "grad_norm": 17.347404909951468, + "learning_rate": 1.6391465587466028e-05, + "loss": 0.0543, + "step": 2318 + }, + { + "epoch": 3.0116883116883115, + "grad_norm": 10.817422386804052, + "learning_rate": 1.6388230115679143e-05, + "loss": 0.0428, + "step": 2319 + }, + { + "epoch": 3.012987012987013, + "grad_norm": 10.987232257318107, + "learning_rate": 1.638499351369208e-05, + "loss": 0.0304, + "step": 2320 + }, + { + "epoch": 3.0142857142857142, + "grad_norm": 7.863796556357488, + "learning_rate": 1.6381755782077453e-05, + "loss": 0.0476, + "step": 2321 + }, + { + "epoch": 3.0155844155844154, + "grad_norm": 13.110141809288342, + "learning_rate": 1.6378516921408077e-05, + "loss": 0.022, + "step": 2322 + }, + { + "epoch": 3.016883116883117, + "grad_norm": 13.38151607930291, + "learning_rate": 1.6375276932256977e-05, + "loss": 0.0365, + "step": 2323 + }, + { + "epoch": 3.018181818181818, + "grad_norm": 11.91929846700814, + "learning_rate": 1.637203581519736e-05, + "loss": 0.0475, + "step": 2324 + }, + { + "epoch": 3.0194805194805197, + "grad_norm": 12.454400521445722, + "learning_rate": 1.6368793570802644e-05, + "loss": 0.0341, + "step": 2325 + }, + { + "epoch": 3.020779220779221, + "grad_norm": 14.067196312304072, + "learning_rate": 1.636555019964645e-05, + "loss": 0.0266, + "step": 2326 + }, + { + "epoch": 3.022077922077922, + "grad_norm": 11.20262820707295, + "learning_rate": 1.6362305702302587e-05, + "loss": 0.0316, + "step": 2327 + }, + { + "epoch": 3.0233766233766235, + "grad_norm": 11.785484352320973, + "learning_rate": 1.6359060079345064e-05, + "loss": 0.0698, + "step": 2328 + }, + { + "epoch": 3.0246753246753246, + "grad_norm": 7.77313525925236, + "learning_rate": 1.6355813331348104e-05, + "loss": 0.0274, + "step": 2329 + }, + { + "epoch": 3.0259740259740258, + "grad_norm": 6.5383140787685745, + "learning_rate": 1.6352565458886114e-05, + "loss": 0.0296, + "step": 2330 + }, + { + "epoch": 3.0272727272727273, + "grad_norm": 19.746901635497164, + "learning_rate": 1.6349316462533705e-05, + "loss": 0.0408, + "step": 2331 + }, + { + "epoch": 3.0285714285714285, + "grad_norm": 6.102231076767843, + "learning_rate": 1.6346066342865682e-05, + "loss": 0.0292, + "step": 2332 + }, + { + "epoch": 3.02987012987013, + "grad_norm": 2.7324521644038313, + "learning_rate": 1.6342815100457064e-05, + "loss": 0.0353, + "step": 2333 + }, + { + "epoch": 3.031168831168831, + "grad_norm": 15.932116268994184, + "learning_rate": 1.633956273588305e-05, + "loss": 0.0574, + "step": 2334 + }, + { + "epoch": 3.0324675324675323, + "grad_norm": 17.040713918937456, + "learning_rate": 1.633630924971904e-05, + "loss": 0.0215, + "step": 2335 + }, + { + "epoch": 3.033766233766234, + "grad_norm": 9.047861628359234, + "learning_rate": 1.6333054642540656e-05, + "loss": 0.0635, + "step": 2336 + }, + { + "epoch": 3.035064935064935, + "grad_norm": 8.750435195999113, + "learning_rate": 1.6329798914923687e-05, + "loss": 0.0465, + "step": 2337 + }, + { + "epoch": 3.036363636363636, + "grad_norm": 12.355808093485225, + "learning_rate": 1.6326542067444136e-05, + "loss": 0.0206, + "step": 2338 + }, + { + "epoch": 3.0376623376623377, + "grad_norm": 6.645046462347423, + "learning_rate": 1.63232841006782e-05, + "loss": 0.0588, + "step": 2339 + }, + { + "epoch": 3.038961038961039, + "grad_norm": 9.917680417433525, + "learning_rate": 1.632002501520228e-05, + "loss": 0.0238, + "step": 2340 + }, + { + "epoch": 3.0402597402597404, + "grad_norm": 6.454688422305678, + "learning_rate": 1.6316764811592968e-05, + "loss": 0.0354, + "step": 2341 + }, + { + "epoch": 3.0415584415584416, + "grad_norm": 4.786198604972967, + "learning_rate": 1.6313503490427052e-05, + "loss": 0.0324, + "step": 2342 + }, + { + "epoch": 3.0428571428571427, + "grad_norm": 12.942172360154474, + "learning_rate": 1.6310241052281534e-05, + "loss": 0.0446, + "step": 2343 + }, + { + "epoch": 3.0441558441558443, + "grad_norm": 10.728993400709788, + "learning_rate": 1.630697749773359e-05, + "loss": 0.0255, + "step": 2344 + }, + { + "epoch": 3.0454545454545454, + "grad_norm": 13.604476593080081, + "learning_rate": 1.630371282736061e-05, + "loss": 0.0286, + "step": 2345 + }, + { + "epoch": 3.0467532467532465, + "grad_norm": 17.98320254591876, + "learning_rate": 1.630044704174018e-05, + "loss": 0.0602, + "step": 2346 + }, + { + "epoch": 3.048051948051948, + "grad_norm": 11.24653621836461, + "learning_rate": 1.6297180141450073e-05, + "loss": 0.0451, + "step": 2347 + }, + { + "epoch": 3.0493506493506493, + "grad_norm": 29.944758288487407, + "learning_rate": 1.629391212706827e-05, + "loss": 0.0313, + "step": 2348 + }, + { + "epoch": 3.050649350649351, + "grad_norm": 15.71631073576057, + "learning_rate": 1.6290642999172947e-05, + "loss": 0.059, + "step": 2349 + }, + { + "epoch": 3.051948051948052, + "grad_norm": 9.827319034545363, + "learning_rate": 1.628737275834247e-05, + "loss": 0.0219, + "step": 2350 + }, + { + "epoch": 3.053246753246753, + "grad_norm": 13.56448457591199, + "learning_rate": 1.6284101405155413e-05, + "loss": 0.0662, + "step": 2351 + }, + { + "epoch": 3.0545454545454547, + "grad_norm": 7.172769573866076, + "learning_rate": 1.6280828940190537e-05, + "loss": 0.0282, + "step": 2352 + }, + { + "epoch": 3.055844155844156, + "grad_norm": 5.917742001088711, + "learning_rate": 1.6277555364026807e-05, + "loss": 0.0183, + "step": 2353 + }, + { + "epoch": 3.057142857142857, + "grad_norm": 12.921797781246614, + "learning_rate": 1.6274280677243377e-05, + "loss": 0.0771, + "step": 2354 + }, + { + "epoch": 3.0584415584415585, + "grad_norm": 9.718121290158027, + "learning_rate": 1.6271004880419606e-05, + "loss": 0.075, + "step": 2355 + }, + { + "epoch": 3.0597402597402596, + "grad_norm": 15.132567632972679, + "learning_rate": 1.6267727974135043e-05, + "loss": 0.0426, + "step": 2356 + }, + { + "epoch": 3.0610389610389612, + "grad_norm": 17.027020856638394, + "learning_rate": 1.6264449958969436e-05, + "loss": 0.0791, + "step": 2357 + }, + { + "epoch": 3.0623376623376624, + "grad_norm": 18.810105883464935, + "learning_rate": 1.626117083550273e-05, + "loss": 0.0465, + "step": 2358 + }, + { + "epoch": 3.0636363636363635, + "grad_norm": 28.461375228830626, + "learning_rate": 1.6257890604315058e-05, + "loss": 0.0438, + "step": 2359 + }, + { + "epoch": 3.064935064935065, + "grad_norm": 11.60788979033136, + "learning_rate": 1.6254609265986767e-05, + "loss": 0.0317, + "step": 2360 + }, + { + "epoch": 3.066233766233766, + "grad_norm": 20.075696963392286, + "learning_rate": 1.625132682109838e-05, + "loss": 0.0267, + "step": 2361 + }, + { + "epoch": 3.0675324675324673, + "grad_norm": 30.795623412750988, + "learning_rate": 1.6248043270230624e-05, + "loss": 0.0646, + "step": 2362 + }, + { + "epoch": 3.068831168831169, + "grad_norm": 25.04510532208064, + "learning_rate": 1.6244758613964428e-05, + "loss": 0.031, + "step": 2363 + }, + { + "epoch": 3.07012987012987, + "grad_norm": 34.063893173117606, + "learning_rate": 1.6241472852880904e-05, + "loss": 0.0368, + "step": 2364 + }, + { + "epoch": 3.0714285714285716, + "grad_norm": 11.82429088331357, + "learning_rate": 1.6238185987561374e-05, + "loss": 0.0901, + "step": 2365 + }, + { + "epoch": 3.0727272727272728, + "grad_norm": 37.723336209687645, + "learning_rate": 1.6234898018587336e-05, + "loss": 0.0416, + "step": 2366 + }, + { + "epoch": 3.074025974025974, + "grad_norm": 8.557552339505069, + "learning_rate": 1.6231608946540505e-05, + "loss": 0.0589, + "step": 2367 + }, + { + "epoch": 3.0753246753246755, + "grad_norm": 38.84974309235112, + "learning_rate": 1.6228318772002777e-05, + "loss": 0.0605, + "step": 2368 + }, + { + "epoch": 3.0766233766233766, + "grad_norm": 26.360422570922804, + "learning_rate": 1.6225027495556248e-05, + "loss": 0.0346, + "step": 2369 + }, + { + "epoch": 3.0779220779220777, + "grad_norm": 35.125864529729526, + "learning_rate": 1.6221735117783208e-05, + "loss": 0.0682, + "step": 2370 + }, + { + "epoch": 3.0792207792207793, + "grad_norm": 29.22960684647252, + "learning_rate": 1.6218441639266136e-05, + "loss": 0.0699, + "step": 2371 + }, + { + "epoch": 3.0805194805194804, + "grad_norm": 36.374529343189, + "learning_rate": 1.621514706058772e-05, + "loss": 0.0848, + "step": 2372 + }, + { + "epoch": 3.081818181818182, + "grad_norm": 15.24142601216938, + "learning_rate": 1.6211851382330827e-05, + "loss": 0.0489, + "step": 2373 + }, + { + "epoch": 3.083116883116883, + "grad_norm": 11.514705536178962, + "learning_rate": 1.6208554605078528e-05, + "loss": 0.0554, + "step": 2374 + }, + { + "epoch": 3.0844155844155843, + "grad_norm": 25.492343615975695, + "learning_rate": 1.6205256729414086e-05, + "loss": 0.0311, + "step": 2375 + }, + { + "epoch": 3.085714285714286, + "grad_norm": 11.203124569506347, + "learning_rate": 1.620195775592096e-05, + "loss": 0.0381, + "step": 2376 + }, + { + "epoch": 3.087012987012987, + "grad_norm": 27.067359696089188, + "learning_rate": 1.61986576851828e-05, + "loss": 0.045, + "step": 2377 + }, + { + "epoch": 3.088311688311688, + "grad_norm": 20.866761316485558, + "learning_rate": 1.6195356517783455e-05, + "loss": 0.0415, + "step": 2378 + }, + { + "epoch": 3.0896103896103897, + "grad_norm": 18.570901461661396, + "learning_rate": 1.619205425430696e-05, + "loss": 0.0763, + "step": 2379 + }, + { + "epoch": 3.090909090909091, + "grad_norm": 37.97443752658283, + "learning_rate": 1.618875089533755e-05, + "loss": 0.0535, + "step": 2380 + }, + { + "epoch": 3.0922077922077924, + "grad_norm": 9.080263433340294, + "learning_rate": 1.6185446441459652e-05, + "loss": 0.0367, + "step": 2381 + }, + { + "epoch": 3.0935064935064935, + "grad_norm": 37.85340006064281, + "learning_rate": 1.6182140893257885e-05, + "loss": 0.0353, + "step": 2382 + }, + { + "epoch": 3.0948051948051947, + "grad_norm": 7.946106172059584, + "learning_rate": 1.617883425131707e-05, + "loss": 0.0296, + "step": 2383 + }, + { + "epoch": 3.0961038961038962, + "grad_norm": 30.99415265432052, + "learning_rate": 1.617552651622221e-05, + "loss": 0.0397, + "step": 2384 + }, + { + "epoch": 3.0974025974025974, + "grad_norm": 28.07208185162711, + "learning_rate": 1.6172217688558508e-05, + "loss": 0.0749, + "step": 2385 + }, + { + "epoch": 3.0987012987012985, + "grad_norm": 35.114089520201816, + "learning_rate": 1.6168907768911362e-05, + "loss": 0.0403, + "step": 2386 + }, + { + "epoch": 3.1, + "grad_norm": 28.84023500309252, + "learning_rate": 1.6165596757866357e-05, + "loss": 0.0582, + "step": 2387 + }, + { + "epoch": 3.101298701298701, + "grad_norm": 33.41714774816695, + "learning_rate": 1.6162284656009276e-05, + "loss": 0.0491, + "step": 2388 + }, + { + "epoch": 3.102597402597403, + "grad_norm": 28.330638181172347, + "learning_rate": 1.6158971463926088e-05, + "loss": 0.0551, + "step": 2389 + }, + { + "epoch": 3.103896103896104, + "grad_norm": 14.636656024921809, + "learning_rate": 1.6155657182202968e-05, + "loss": 0.043, + "step": 2390 + }, + { + "epoch": 3.105194805194805, + "grad_norm": 33.026970949965516, + "learning_rate": 1.6152341811426267e-05, + "loss": 0.0399, + "step": 2391 + }, + { + "epoch": 3.1064935064935066, + "grad_norm": 5.980591494370592, + "learning_rate": 1.614902535218255e-05, + "loss": 0.0369, + "step": 2392 + }, + { + "epoch": 3.1077922077922078, + "grad_norm": 23.57728226718987, + "learning_rate": 1.6145707805058547e-05, + "loss": 0.028, + "step": 2393 + }, + { + "epoch": 3.109090909090909, + "grad_norm": 27.172123317925664, + "learning_rate": 1.6142389170641207e-05, + "loss": 0.0348, + "step": 2394 + }, + { + "epoch": 3.1103896103896105, + "grad_norm": 18.26218026408267, + "learning_rate": 1.6139069449517656e-05, + "loss": 0.0261, + "step": 2395 + }, + { + "epoch": 3.1116883116883116, + "grad_norm": 26.37478903140657, + "learning_rate": 1.6135748642275214e-05, + "loss": 0.0393, + "step": 2396 + }, + { + "epoch": 3.112987012987013, + "grad_norm": 6.531269925755529, + "learning_rate": 1.61324267495014e-05, + "loss": 0.0629, + "step": 2397 + }, + { + "epoch": 3.1142857142857143, + "grad_norm": 21.44078355752576, + "learning_rate": 1.612910377178392e-05, + "loss": 0.0335, + "step": 2398 + }, + { + "epoch": 3.1155844155844155, + "grad_norm": 10.026186574741587, + "learning_rate": 1.6125779709710668e-05, + "loss": 0.0278, + "step": 2399 + }, + { + "epoch": 3.116883116883117, + "grad_norm": 18.157585434417737, + "learning_rate": 1.6122454563869732e-05, + "loss": 0.0651, + "step": 2400 + }, + { + "epoch": 3.118181818181818, + "grad_norm": 22.88571411669533, + "learning_rate": 1.6119128334849406e-05, + "loss": 0.0516, + "step": 2401 + }, + { + "epoch": 3.1194805194805193, + "grad_norm": 31.95263388659759, + "learning_rate": 1.6115801023238154e-05, + "loss": 0.0712, + "step": 2402 + }, + { + "epoch": 3.120779220779221, + "grad_norm": 7.2801530696266985, + "learning_rate": 1.6112472629624644e-05, + "loss": 0.0651, + "step": 2403 + }, + { + "epoch": 3.122077922077922, + "grad_norm": 10.831142980048915, + "learning_rate": 1.610914315459773e-05, + "loss": 0.0394, + "step": 2404 + }, + { + "epoch": 3.1233766233766236, + "grad_norm": 14.180411067436758, + "learning_rate": 1.6105812598746462e-05, + "loss": 0.036, + "step": 2405 + }, + { + "epoch": 3.1246753246753247, + "grad_norm": 10.449965584685588, + "learning_rate": 1.610248096266008e-05, + "loss": 0.0506, + "step": 2406 + }, + { + "epoch": 3.125974025974026, + "grad_norm": 12.112962188660758, + "learning_rate": 1.609914824692801e-05, + "loss": 0.0515, + "step": 2407 + }, + { + "epoch": 3.1272727272727274, + "grad_norm": 11.252711795316499, + "learning_rate": 1.609581445213988e-05, + "loss": 0.0634, + "step": 2408 + }, + { + "epoch": 3.1285714285714286, + "grad_norm": 13.616673262789357, + "learning_rate": 1.60924795788855e-05, + "loss": 0.0345, + "step": 2409 + }, + { + "epoch": 3.1298701298701297, + "grad_norm": 20.015984932528003, + "learning_rate": 1.6089143627754863e-05, + "loss": 0.0476, + "step": 2410 + }, + { + "epoch": 3.1311688311688313, + "grad_norm": 11.517002658496331, + "learning_rate": 1.608580659933818e-05, + "loss": 0.0304, + "step": 2411 + }, + { + "epoch": 3.1324675324675324, + "grad_norm": 20.419678558903186, + "learning_rate": 1.6082468494225818e-05, + "loss": 0.0896, + "step": 2412 + }, + { + "epoch": 3.1337662337662335, + "grad_norm": 13.188271356194054, + "learning_rate": 1.6079129313008363e-05, + "loss": 0.045, + "step": 2413 + }, + { + "epoch": 3.135064935064935, + "grad_norm": 11.746669841527337, + "learning_rate": 1.6075789056276577e-05, + "loss": 0.0669, + "step": 2414 + }, + { + "epoch": 3.1363636363636362, + "grad_norm": 5.679386549402815, + "learning_rate": 1.6072447724621417e-05, + "loss": 0.0546, + "step": 2415 + }, + { + "epoch": 3.137662337662338, + "grad_norm": 9.592393152872148, + "learning_rate": 1.6069105318634024e-05, + "loss": 0.016, + "step": 2416 + }, + { + "epoch": 3.138961038961039, + "grad_norm": 9.82375743673902, + "learning_rate": 1.6065761838905735e-05, + "loss": 0.0524, + "step": 2417 + }, + { + "epoch": 3.14025974025974, + "grad_norm": 6.237003749561289, + "learning_rate": 1.606241728602808e-05, + "loss": 0.0462, + "step": 2418 + }, + { + "epoch": 3.1415584415584417, + "grad_norm": 17.18133392426163, + "learning_rate": 1.605907166059277e-05, + "loss": 0.0554, + "step": 2419 + }, + { + "epoch": 3.142857142857143, + "grad_norm": 11.331118590997908, + "learning_rate": 1.6055724963191714e-05, + "loss": 0.046, + "step": 2420 + }, + { + "epoch": 3.1441558441558444, + "grad_norm": 20.30770917230055, + "learning_rate": 1.6052377194417e-05, + "loss": 0.0295, + "step": 2421 + }, + { + "epoch": 3.1454545454545455, + "grad_norm": 11.84021231095325, + "learning_rate": 1.6049028354860918e-05, + "loss": 0.0416, + "step": 2422 + }, + { + "epoch": 3.1467532467532466, + "grad_norm": 13.690813467484706, + "learning_rate": 1.604567844511594e-05, + "loss": 0.0417, + "step": 2423 + }, + { + "epoch": 3.148051948051948, + "grad_norm": 7.006701258593577, + "learning_rate": 1.6042327465774728e-05, + "loss": 0.0603, + "step": 2424 + }, + { + "epoch": 3.1493506493506493, + "grad_norm": 12.839098181997985, + "learning_rate": 1.603897541743014e-05, + "loss": 0.0299, + "step": 2425 + }, + { + "epoch": 3.1506493506493505, + "grad_norm": 15.278576500040202, + "learning_rate": 1.6035622300675212e-05, + "loss": 0.0285, + "step": 2426 + }, + { + "epoch": 3.151948051948052, + "grad_norm": 15.546986527994125, + "learning_rate": 1.6032268116103177e-05, + "loss": 0.0502, + "step": 2427 + }, + { + "epoch": 3.153246753246753, + "grad_norm": 10.295778645380494, + "learning_rate": 1.602891286430745e-05, + "loss": 0.0543, + "step": 2428 + }, + { + "epoch": 3.1545454545454543, + "grad_norm": 12.931240988028808, + "learning_rate": 1.6025556545881646e-05, + "loss": 0.0409, + "step": 2429 + }, + { + "epoch": 3.155844155844156, + "grad_norm": 8.797905558038714, + "learning_rate": 1.6022199161419554e-05, + "loss": 0.0264, + "step": 2430 + }, + { + "epoch": 3.157142857142857, + "grad_norm": 22.97109468558496, + "learning_rate": 1.601884071151517e-05, + "loss": 0.0546, + "step": 2431 + }, + { + "epoch": 3.1584415584415586, + "grad_norm": 8.137538343019608, + "learning_rate": 1.6015481196762658e-05, + "loss": 0.0559, + "step": 2432 + }, + { + "epoch": 3.1597402597402597, + "grad_norm": 17.628712844949867, + "learning_rate": 1.601212061775639e-05, + "loss": 0.0365, + "step": 2433 + }, + { + "epoch": 3.161038961038961, + "grad_norm": 14.774867209881197, + "learning_rate": 1.6008758975090905e-05, + "loss": 0.032, + "step": 2434 + }, + { + "epoch": 3.1623376623376624, + "grad_norm": 9.248445502403692, + "learning_rate": 1.600539626936095e-05, + "loss": 0.0268, + "step": 2435 + }, + { + "epoch": 3.1636363636363636, + "grad_norm": 33.42187355145025, + "learning_rate": 1.6002032501161454e-05, + "loss": 0.0596, + "step": 2436 + }, + { + "epoch": 3.164935064935065, + "grad_norm": 9.517775377375859, + "learning_rate": 1.5998667671087527e-05, + "loss": 0.0427, + "step": 2437 + }, + { + "epoch": 3.1662337662337663, + "grad_norm": 34.57182155951531, + "learning_rate": 1.599530177973447e-05, + "loss": 0.0265, + "step": 2438 + }, + { + "epoch": 3.1675324675324674, + "grad_norm": 5.574705360592123, + "learning_rate": 1.5991934827697782e-05, + "loss": 0.0423, + "step": 2439 + }, + { + "epoch": 3.168831168831169, + "grad_norm": 39.83536102915724, + "learning_rate": 1.598856681557313e-05, + "loss": 0.0707, + "step": 2440 + }, + { + "epoch": 3.17012987012987, + "grad_norm": 9.371056351414754, + "learning_rate": 1.5985197743956392e-05, + "loss": 0.0539, + "step": 2441 + }, + { + "epoch": 3.1714285714285713, + "grad_norm": 22.72315384792833, + "learning_rate": 1.5981827613443612e-05, + "loss": 0.0345, + "step": 2442 + }, + { + "epoch": 3.172727272727273, + "grad_norm": 20.08478335531911, + "learning_rate": 1.5978456424631033e-05, + "loss": 0.016, + "step": 2443 + }, + { + "epoch": 3.174025974025974, + "grad_norm": 17.035502858424373, + "learning_rate": 1.5975084178115086e-05, + "loss": 0.0376, + "step": 2444 + }, + { + "epoch": 3.175324675324675, + "grad_norm": 18.67252494426682, + "learning_rate": 1.5971710874492382e-05, + "loss": 0.0611, + "step": 2445 + }, + { + "epoch": 3.1766233766233767, + "grad_norm": 4.237077910322895, + "learning_rate": 1.5968336514359727e-05, + "loss": 0.0334, + "step": 2446 + }, + { + "epoch": 3.177922077922078, + "grad_norm": 25.46510254717697, + "learning_rate": 1.5964961098314105e-05, + "loss": 0.0416, + "step": 2447 + }, + { + "epoch": 3.1792207792207794, + "grad_norm": 3.7845572042452997, + "learning_rate": 1.5961584626952692e-05, + "loss": 0.0288, + "step": 2448 + }, + { + "epoch": 3.1805194805194805, + "grad_norm": 17.859626052173134, + "learning_rate": 1.5958207100872857e-05, + "loss": 0.0474, + "step": 2449 + }, + { + "epoch": 3.1818181818181817, + "grad_norm": 13.855832497097703, + "learning_rate": 1.5954828520672144e-05, + "loss": 0.0159, + "step": 2450 + }, + { + "epoch": 3.1831168831168832, + "grad_norm": 10.221160972812743, + "learning_rate": 1.5951448886948286e-05, + "loss": 0.0447, + "step": 2451 + }, + { + "epoch": 3.1844155844155844, + "grad_norm": 23.71838826394589, + "learning_rate": 1.5948068200299212e-05, + "loss": 0.057, + "step": 2452 + }, + { + "epoch": 3.185714285714286, + "grad_norm": 6.58935696264577, + "learning_rate": 1.5944686461323022e-05, + "loss": 0.0344, + "step": 2453 + }, + { + "epoch": 3.187012987012987, + "grad_norm": 19.213205080093147, + "learning_rate": 1.5941303670618018e-05, + "loss": 0.0321, + "step": 2454 + }, + { + "epoch": 3.188311688311688, + "grad_norm": 4.429362689939051, + "learning_rate": 1.5937919828782674e-05, + "loss": 0.0427, + "step": 2455 + }, + { + "epoch": 3.18961038961039, + "grad_norm": 4.280418359323422, + "learning_rate": 1.593453493641566e-05, + "loss": 0.0395, + "step": 2456 + }, + { + "epoch": 3.190909090909091, + "grad_norm": 21.019687363111576, + "learning_rate": 1.593114899411583e-05, + "loss": 0.0483, + "step": 2457 + }, + { + "epoch": 3.192207792207792, + "grad_norm": 9.953006127641949, + "learning_rate": 1.5927762002482213e-05, + "loss": 0.0552, + "step": 2458 + }, + { + "epoch": 3.1935064935064936, + "grad_norm": 7.445471936858932, + "learning_rate": 1.5924373962114048e-05, + "loss": 0.0233, + "step": 2459 + }, + { + "epoch": 3.1948051948051948, + "grad_norm": 4.381328085140121, + "learning_rate": 1.592098487361073e-05, + "loss": 0.0167, + "step": 2460 + }, + { + "epoch": 3.196103896103896, + "grad_norm": 7.697943456798237, + "learning_rate": 1.591759473757186e-05, + "loss": 0.0397, + "step": 2461 + }, + { + "epoch": 3.1974025974025975, + "grad_norm": 2.7757019918882455, + "learning_rate": 1.5914203554597217e-05, + "loss": 0.0163, + "step": 2462 + }, + { + "epoch": 3.1987012987012986, + "grad_norm": 5.149810974693576, + "learning_rate": 1.5910811325286768e-05, + "loss": 0.0186, + "step": 2463 + }, + { + "epoch": 3.2, + "grad_norm": 8.379362983850728, + "learning_rate": 1.5907418050240657e-05, + "loss": 0.0472, + "step": 2464 + }, + { + "epoch": 3.2012987012987013, + "grad_norm": 8.562591365725591, + "learning_rate": 1.5904023730059227e-05, + "loss": 0.0346, + "step": 2465 + }, + { + "epoch": 3.2025974025974024, + "grad_norm": 6.1146353092598575, + "learning_rate": 1.5900628365342994e-05, + "loss": 0.0284, + "step": 2466 + }, + { + "epoch": 3.203896103896104, + "grad_norm": 5.16968743260755, + "learning_rate": 1.5897231956692664e-05, + "loss": 0.0405, + "step": 2467 + }, + { + "epoch": 3.205194805194805, + "grad_norm": 8.882237687592555, + "learning_rate": 1.589383450470913e-05, + "loss": 0.0209, + "step": 2468 + }, + { + "epoch": 3.2064935064935067, + "grad_norm": 3.5719046419040734, + "learning_rate": 1.589043600999346e-05, + "loss": 0.0251, + "step": 2469 + }, + { + "epoch": 3.207792207792208, + "grad_norm": 5.714509914112394, + "learning_rate": 1.5887036473146916e-05, + "loss": 0.0284, + "step": 2470 + }, + { + "epoch": 3.209090909090909, + "grad_norm": 11.673640125432213, + "learning_rate": 1.5883635894770935e-05, + "loss": 0.0335, + "step": 2471 + }, + { + "epoch": 3.2103896103896106, + "grad_norm": 11.24885573634582, + "learning_rate": 1.588023427546716e-05, + "loss": 0.0339, + "step": 2472 + }, + { + "epoch": 3.2116883116883117, + "grad_norm": 13.639624989811924, + "learning_rate": 1.5876831615837386e-05, + "loss": 0.037, + "step": 2473 + }, + { + "epoch": 3.212987012987013, + "grad_norm": 6.020500655872691, + "learning_rate": 1.5873427916483618e-05, + "loss": 0.0199, + "step": 2474 + }, + { + "epoch": 3.2142857142857144, + "grad_norm": 11.38403364861548, + "learning_rate": 1.5870023178008035e-05, + "loss": 0.0412, + "step": 2475 + }, + { + "epoch": 3.2155844155844155, + "grad_norm": 10.327169319925973, + "learning_rate": 1.5866617401012995e-05, + "loss": 0.0532, + "step": 2476 + }, + { + "epoch": 3.2168831168831167, + "grad_norm": 8.739913462450984, + "learning_rate": 1.5863210586101048e-05, + "loss": 0.031, + "step": 2477 + }, + { + "epoch": 3.2181818181818183, + "grad_norm": 12.476197058829388, + "learning_rate": 1.5859802733874928e-05, + "loss": 0.055, + "step": 2478 + }, + { + "epoch": 3.2194805194805194, + "grad_norm": 11.752321994992526, + "learning_rate": 1.5856393844937544e-05, + "loss": 0.0215, + "step": 2479 + }, + { + "epoch": 3.220779220779221, + "grad_norm": 6.945654994859079, + "learning_rate": 1.5852983919892e-05, + "loss": 0.0394, + "step": 2480 + }, + { + "epoch": 3.222077922077922, + "grad_norm": 11.4184282712036, + "learning_rate": 1.584957295934157e-05, + "loss": 0.0296, + "step": 2481 + }, + { + "epoch": 3.2233766233766232, + "grad_norm": 7.410971830539309, + "learning_rate": 1.5846160963889723e-05, + "loss": 0.023, + "step": 2482 + }, + { + "epoch": 3.224675324675325, + "grad_norm": 6.420031917015504, + "learning_rate": 1.5842747934140103e-05, + "loss": 0.0446, + "step": 2483 + }, + { + "epoch": 3.225974025974026, + "grad_norm": 11.562495571592846, + "learning_rate": 1.5839333870696544e-05, + "loss": 0.0268, + "step": 2484 + }, + { + "epoch": 3.227272727272727, + "grad_norm": 6.132026433733215, + "learning_rate": 1.5835918774163056e-05, + "loss": 0.0121, + "step": 2485 + }, + { + "epoch": 3.2285714285714286, + "grad_norm": 18.55305475895002, + "learning_rate": 1.5832502645143837e-05, + "loss": 0.0548, + "step": 2486 + }, + { + "epoch": 3.22987012987013, + "grad_norm": 3.4874400016752536, + "learning_rate": 1.5829085484243264e-05, + "loss": 0.0235, + "step": 2487 + }, + { + "epoch": 3.2311688311688314, + "grad_norm": 13.622361528227234, + "learning_rate": 1.58256672920659e-05, + "loss": 0.0231, + "step": 2488 + }, + { + "epoch": 3.2324675324675325, + "grad_norm": 11.868255669954364, + "learning_rate": 1.5822248069216487e-05, + "loss": 0.0166, + "step": 2489 + }, + { + "epoch": 3.2337662337662336, + "grad_norm": 13.602156647329144, + "learning_rate": 1.581882781629995e-05, + "loss": 0.0382, + "step": 2490 + }, + { + "epoch": 3.235064935064935, + "grad_norm": 12.799731495445867, + "learning_rate": 1.5815406533921402e-05, + "loss": 0.031, + "step": 2491 + }, + { + "epoch": 3.2363636363636363, + "grad_norm": 14.7707495164333, + "learning_rate": 1.5811984222686127e-05, + "loss": 0.042, + "step": 2492 + }, + { + "epoch": 3.2376623376623375, + "grad_norm": 7.586841148323603, + "learning_rate": 1.5808560883199604e-05, + "loss": 0.0669, + "step": 2493 + }, + { + "epoch": 3.238961038961039, + "grad_norm": 14.4114130182429, + "learning_rate": 1.5805136516067482e-05, + "loss": 0.0289, + "step": 2494 + }, + { + "epoch": 3.24025974025974, + "grad_norm": 4.692415354051082, + "learning_rate": 1.5801711121895598e-05, + "loss": 0.0365, + "step": 2495 + }, + { + "epoch": 3.2415584415584417, + "grad_norm": 6.26516182851298, + "learning_rate": 1.5798284701289972e-05, + "loss": 0.0273, + "step": 2496 + }, + { + "epoch": 3.242857142857143, + "grad_norm": 7.882683672718905, + "learning_rate": 1.57948572548568e-05, + "loss": 0.0378, + "step": 2497 + }, + { + "epoch": 3.244155844155844, + "grad_norm": 4.658393600787512, + "learning_rate": 1.5791428783202465e-05, + "loss": 0.0734, + "step": 2498 + }, + { + "epoch": 3.2454545454545456, + "grad_norm": 9.792516506417202, + "learning_rate": 1.578799928693353e-05, + "loss": 0.0534, + "step": 2499 + }, + { + "epoch": 3.2467532467532467, + "grad_norm": 7.7361018803934405, + "learning_rate": 1.5784568766656734e-05, + "loss": 0.0589, + "step": 2500 + }, + { + "epoch": 3.248051948051948, + "grad_norm": 10.71620029781344, + "learning_rate": 1.5781137222979013e-05, + "loss": 0.0498, + "step": 2501 + }, + { + "epoch": 3.2493506493506494, + "grad_norm": 13.854488540334932, + "learning_rate": 1.5777704656507462e-05, + "loss": 0.0537, + "step": 2502 + }, + { + "epoch": 3.2506493506493506, + "grad_norm": 12.20799076453884, + "learning_rate": 1.5774271067849368e-05, + "loss": 0.0499, + "step": 2503 + }, + { + "epoch": 3.2519480519480517, + "grad_norm": 11.69910009209532, + "learning_rate": 1.5770836457612203e-05, + "loss": 0.037, + "step": 2504 + }, + { + "epoch": 3.2532467532467533, + "grad_norm": 20.182077816478643, + "learning_rate": 1.5767400826403613e-05, + "loss": 0.0409, + "step": 2505 + }, + { + "epoch": 3.2545454545454544, + "grad_norm": 17.740995232998202, + "learning_rate": 1.576396417483143e-05, + "loss": 0.0318, + "step": 2506 + }, + { + "epoch": 3.255844155844156, + "grad_norm": 12.51590271113076, + "learning_rate": 1.576052650350366e-05, + "loss": 0.0746, + "step": 2507 + }, + { + "epoch": 3.257142857142857, + "grad_norm": 18.84401518950666, + "learning_rate": 1.5757087813028494e-05, + "loss": 0.063, + "step": 2508 + }, + { + "epoch": 3.2584415584415583, + "grad_norm": 16.806364053736534, + "learning_rate": 1.57536481040143e-05, + "loss": 0.04, + "step": 2509 + }, + { + "epoch": 3.25974025974026, + "grad_norm": 4.681094592212491, + "learning_rate": 1.575020737706963e-05, + "loss": 0.0301, + "step": 2510 + }, + { + "epoch": 3.261038961038961, + "grad_norm": 21.764573381490706, + "learning_rate": 1.5746765632803218e-05, + "loss": 0.0333, + "step": 2511 + }, + { + "epoch": 3.2623376623376625, + "grad_norm": 6.528040704040589, + "learning_rate": 1.574332287182397e-05, + "loss": 0.0333, + "step": 2512 + }, + { + "epoch": 3.2636363636363637, + "grad_norm": 23.667093390613843, + "learning_rate": 1.5739879094740978e-05, + "loss": 0.0449, + "step": 2513 + }, + { + "epoch": 3.264935064935065, + "grad_norm": 8.991417043479705, + "learning_rate": 1.5736434302163514e-05, + "loss": 0.0322, + "step": 2514 + }, + { + "epoch": 3.2662337662337664, + "grad_norm": 17.3086751280151, + "learning_rate": 1.573298849470102e-05, + "loss": 0.0362, + "step": 2515 + }, + { + "epoch": 3.2675324675324675, + "grad_norm": 13.6338089007882, + "learning_rate": 1.5729541672963134e-05, + "loss": 0.0641, + "step": 2516 + }, + { + "epoch": 3.2688311688311686, + "grad_norm": 14.017705439920533, + "learning_rate": 1.572609383755966e-05, + "loss": 0.0294, + "step": 2517 + }, + { + "epoch": 3.27012987012987, + "grad_norm": 5.6713980769076, + "learning_rate": 1.5722644989100587e-05, + "loss": 0.0361, + "step": 2518 + }, + { + "epoch": 3.2714285714285714, + "grad_norm": 12.036484502054044, + "learning_rate": 1.5719195128196085e-05, + "loss": 0.0391, + "step": 2519 + }, + { + "epoch": 3.2727272727272725, + "grad_norm": 7.5682521059677335, + "learning_rate": 1.5715744255456497e-05, + "loss": 0.0415, + "step": 2520 + }, + { + "epoch": 3.274025974025974, + "grad_norm": 4.0313829986410346, + "learning_rate": 1.571229237149235e-05, + "loss": 0.0213, + "step": 2521 + }, + { + "epoch": 3.275324675324675, + "grad_norm": 17.60818957346189, + "learning_rate": 1.5708839476914347e-05, + "loss": 0.0369, + "step": 2522 + }, + { + "epoch": 3.2766233766233768, + "grad_norm": 8.013693400325433, + "learning_rate": 1.5705385572333375e-05, + "loss": 0.026, + "step": 2523 + }, + { + "epoch": 3.277922077922078, + "grad_norm": 13.135444331023981, + "learning_rate": 1.5701930658360493e-05, + "loss": 0.0461, + "step": 2524 + }, + { + "epoch": 3.279220779220779, + "grad_norm": 5.377913910624734, + "learning_rate": 1.569847473560694e-05, + "loss": 0.0409, + "step": 2525 + }, + { + "epoch": 3.2805194805194806, + "grad_norm": 10.17772898803105, + "learning_rate": 1.5695017804684134e-05, + "loss": 0.0608, + "step": 2526 + }, + { + "epoch": 3.2818181818181817, + "grad_norm": 7.265101225996652, + "learning_rate": 1.5691559866203676e-05, + "loss": 0.0289, + "step": 2527 + }, + { + "epoch": 3.2831168831168833, + "grad_norm": 10.482100652076097, + "learning_rate": 1.5688100920777343e-05, + "loss": 0.0431, + "step": 2528 + }, + { + "epoch": 3.2844155844155845, + "grad_norm": 8.78460449913222, + "learning_rate": 1.568464096901708e-05, + "loss": 0.0395, + "step": 2529 + }, + { + "epoch": 3.2857142857142856, + "grad_norm": 20.051720685107163, + "learning_rate": 1.568118001153503e-05, + "loss": 0.072, + "step": 2530 + }, + { + "epoch": 3.287012987012987, + "grad_norm": 5.67376973760364, + "learning_rate": 1.5677718048943495e-05, + "loss": 0.0153, + "step": 2531 + }, + { + "epoch": 3.2883116883116883, + "grad_norm": 26.101932105035402, + "learning_rate": 1.5674255081854967e-05, + "loss": 0.0427, + "step": 2532 + }, + { + "epoch": 3.2896103896103894, + "grad_norm": 8.816053428683245, + "learning_rate": 1.5670791110882103e-05, + "loss": 0.067, + "step": 2533 + }, + { + "epoch": 3.290909090909091, + "grad_norm": 19.426128328157848, + "learning_rate": 1.5667326136637754e-05, + "loss": 0.0247, + "step": 2534 + }, + { + "epoch": 3.292207792207792, + "grad_norm": 4.340283745551044, + "learning_rate": 1.566386015973494e-05, + "loss": 0.0358, + "step": 2535 + }, + { + "epoch": 3.2935064935064933, + "grad_norm": 10.124313411712867, + "learning_rate": 1.5660393180786854e-05, + "loss": 0.0431, + "step": 2536 + }, + { + "epoch": 3.294805194805195, + "grad_norm": 9.59634787976438, + "learning_rate": 1.5656925200406875e-05, + "loss": 0.0459, + "step": 2537 + }, + { + "epoch": 3.296103896103896, + "grad_norm": 5.303652655222251, + "learning_rate": 1.5653456219208547e-05, + "loss": 0.0136, + "step": 2538 + }, + { + "epoch": 3.2974025974025976, + "grad_norm": 16.064404736960075, + "learning_rate": 1.564998623780561e-05, + "loss": 0.0358, + "step": 2539 + }, + { + "epoch": 3.2987012987012987, + "grad_norm": 11.76104866138784, + "learning_rate": 1.5646515256811968e-05, + "loss": 0.0545, + "step": 2540 + }, + { + "epoch": 3.3, + "grad_norm": 13.97723719306187, + "learning_rate": 1.56430432768417e-05, + "loss": 0.0399, + "step": 2541 + }, + { + "epoch": 3.3012987012987014, + "grad_norm": 2.7057250734191642, + "learning_rate": 1.5639570298509067e-05, + "loss": 0.0274, + "step": 2542 + }, + { + "epoch": 3.3025974025974025, + "grad_norm": 10.507775669565632, + "learning_rate": 1.5636096322428506e-05, + "loss": 0.0287, + "step": 2543 + }, + { + "epoch": 3.303896103896104, + "grad_norm": 8.896332029967775, + "learning_rate": 1.563262134921463e-05, + "loss": 0.0291, + "step": 2544 + }, + { + "epoch": 3.3051948051948052, + "grad_norm": 4.149257502779458, + "learning_rate": 1.5629145379482228e-05, + "loss": 0.0297, + "step": 2545 + }, + { + "epoch": 3.3064935064935064, + "grad_norm": 19.141991246529997, + "learning_rate": 1.562566841384626e-05, + "loss": 0.022, + "step": 2546 + }, + { + "epoch": 3.307792207792208, + "grad_norm": 6.4797922752676556, + "learning_rate": 1.562219045292188e-05, + "loss": 0.0307, + "step": 2547 + }, + { + "epoch": 3.309090909090909, + "grad_norm": 18.686324619232224, + "learning_rate": 1.5618711497324394e-05, + "loss": 0.036, + "step": 2548 + }, + { + "epoch": 3.31038961038961, + "grad_norm": 10.842603345280954, + "learning_rate": 1.5615231547669305e-05, + "loss": 0.0374, + "step": 2549 + }, + { + "epoch": 3.311688311688312, + "grad_norm": 16.388752091719006, + "learning_rate": 1.561175060457228e-05, + "loss": 0.0227, + "step": 2550 + }, + { + "epoch": 3.312987012987013, + "grad_norm": 8.150939177866935, + "learning_rate": 1.5608268668649157e-05, + "loss": 0.0297, + "step": 2551 + }, + { + "epoch": 3.314285714285714, + "grad_norm": 19.255112493472698, + "learning_rate": 1.5604785740515965e-05, + "loss": 0.0419, + "step": 2552 + }, + { + "epoch": 3.3155844155844156, + "grad_norm": 19.251832915531242, + "learning_rate": 1.56013018207889e-05, + "loss": 0.0543, + "step": 2553 + }, + { + "epoch": 3.3168831168831168, + "grad_norm": 15.148130216068058, + "learning_rate": 1.5597816910084334e-05, + "loss": 0.0383, + "step": 2554 + }, + { + "epoch": 3.3181818181818183, + "grad_norm": 30.36150798604851, + "learning_rate": 1.559433100901881e-05, + "loss": 0.0495, + "step": 2555 + }, + { + "epoch": 3.3194805194805195, + "grad_norm": 10.974153371432251, + "learning_rate": 1.5590844118209055e-05, + "loss": 0.0469, + "step": 2556 + }, + { + "epoch": 3.3207792207792206, + "grad_norm": 20.45640609694339, + "learning_rate": 1.5587356238271966e-05, + "loss": 0.0665, + "step": 2557 + }, + { + "epoch": 3.322077922077922, + "grad_norm": 17.729968928615268, + "learning_rate": 1.558386736982462e-05, + "loss": 0.0473, + "step": 2558 + }, + { + "epoch": 3.3233766233766233, + "grad_norm": 10.254024152673814, + "learning_rate": 1.5580377513484252e-05, + "loss": 0.0328, + "step": 2559 + }, + { + "epoch": 3.324675324675325, + "grad_norm": 27.099585282446586, + "learning_rate": 1.5576886669868297e-05, + "loss": 0.056, + "step": 2560 + }, + { + "epoch": 3.325974025974026, + "grad_norm": 10.991155203385059, + "learning_rate": 1.5573394839594344e-05, + "loss": 0.0546, + "step": 2561 + }, + { + "epoch": 3.327272727272727, + "grad_norm": 18.862716447986124, + "learning_rate": 1.556990202328017e-05, + "loss": 0.0332, + "step": 2562 + }, + { + "epoch": 3.3285714285714287, + "grad_norm": 5.00651564731284, + "learning_rate": 1.556640822154372e-05, + "loss": 0.0295, + "step": 2563 + }, + { + "epoch": 3.32987012987013, + "grad_norm": 16.54318859100646, + "learning_rate": 1.5562913435003113e-05, + "loss": 0.0185, + "step": 2564 + }, + { + "epoch": 3.331168831168831, + "grad_norm": 7.962669583062521, + "learning_rate": 1.5559417664276643e-05, + "loss": 0.0352, + "step": 2565 + }, + { + "epoch": 3.3324675324675326, + "grad_norm": 7.331355448143134, + "learning_rate": 1.5555920909982782e-05, + "loss": 0.0855, + "step": 2566 + }, + { + "epoch": 3.3337662337662337, + "grad_norm": 22.785920250763812, + "learning_rate": 1.555242317274017e-05, + "loss": 0.0408, + "step": 2567 + }, + { + "epoch": 3.335064935064935, + "grad_norm": 9.665569453613081, + "learning_rate": 1.554892445316762e-05, + "loss": 0.0372, + "step": 2568 + }, + { + "epoch": 3.3363636363636364, + "grad_norm": 35.72827532061015, + "learning_rate": 1.554542475188413e-05, + "loss": 0.0817, + "step": 2569 + }, + { + "epoch": 3.3376623376623376, + "grad_norm": 5.9175613747270805, + "learning_rate": 1.554192406950886e-05, + "loss": 0.0129, + "step": 2570 + }, + { + "epoch": 3.338961038961039, + "grad_norm": 21.80276518604063, + "learning_rate": 1.5538422406661152e-05, + "loss": 0.0218, + "step": 2571 + }, + { + "epoch": 3.3402597402597403, + "grad_norm": 24.8711261223846, + "learning_rate": 1.553491976396051e-05, + "loss": 0.0248, + "step": 2572 + }, + { + "epoch": 3.3415584415584414, + "grad_norm": 26.088054851959313, + "learning_rate": 1.5531416142026624e-05, + "loss": 0.0478, + "step": 2573 + }, + { + "epoch": 3.342857142857143, + "grad_norm": 15.00724518550546, + "learning_rate": 1.5527911541479345e-05, + "loss": 0.0593, + "step": 2574 + }, + { + "epoch": 3.344155844155844, + "grad_norm": 5.017330620675103, + "learning_rate": 1.5524405962938717e-05, + "loss": 0.0348, + "step": 2575 + }, + { + "epoch": 3.3454545454545457, + "grad_norm": 15.514580149190532, + "learning_rate": 1.5520899407024928e-05, + "loss": 0.0469, + "step": 2576 + }, + { + "epoch": 3.346753246753247, + "grad_norm": 21.208756703396297, + "learning_rate": 1.551739187435837e-05, + "loss": 0.0353, + "step": 2577 + }, + { + "epoch": 3.348051948051948, + "grad_norm": 19.069332859395487, + "learning_rate": 1.5513883365559583e-05, + "loss": 0.0368, + "step": 2578 + }, + { + "epoch": 3.3493506493506495, + "grad_norm": 16.50602546394576, + "learning_rate": 1.551037388124929e-05, + "loss": 0.0364, + "step": 2579 + }, + { + "epoch": 3.3506493506493507, + "grad_norm": 20.379652816000753, + "learning_rate": 1.550686342204839e-05, + "loss": 0.0752, + "step": 2580 + }, + { + "epoch": 3.351948051948052, + "grad_norm": 8.104934299752601, + "learning_rate": 1.5503351988577947e-05, + "loss": 0.0501, + "step": 2581 + }, + { + "epoch": 3.3532467532467534, + "grad_norm": 9.054942840910053, + "learning_rate": 1.54998395814592e-05, + "loss": 0.0557, + "step": 2582 + }, + { + "epoch": 3.3545454545454545, + "grad_norm": 5.0025702316149605, + "learning_rate": 1.5496326201313566e-05, + "loss": 0.0163, + "step": 2583 + }, + { + "epoch": 3.3558441558441556, + "grad_norm": 14.483964755963905, + "learning_rate": 1.5492811848762624e-05, + "loss": 0.0281, + "step": 2584 + }, + { + "epoch": 3.357142857142857, + "grad_norm": 12.589229546461663, + "learning_rate": 1.5489296524428133e-05, + "loss": 0.0175, + "step": 2585 + }, + { + "epoch": 3.3584415584415583, + "grad_norm": 5.419967541703944, + "learning_rate": 1.548578022893202e-05, + "loss": 0.0318, + "step": 2586 + }, + { + "epoch": 3.35974025974026, + "grad_norm": 5.089597255090377, + "learning_rate": 1.5482262962896387e-05, + "loss": 0.0244, + "step": 2587 + }, + { + "epoch": 3.361038961038961, + "grad_norm": 4.1514098723977115, + "learning_rate": 1.5478744726943508e-05, + "loss": 0.0268, + "step": 2588 + }, + { + "epoch": 3.362337662337662, + "grad_norm": 10.558817836484193, + "learning_rate": 1.5475225521695817e-05, + "loss": 0.0255, + "step": 2589 + }, + { + "epoch": 3.3636363636363638, + "grad_norm": 7.71045666721526, + "learning_rate": 1.5471705347775933e-05, + "loss": 0.0498, + "step": 2590 + }, + { + "epoch": 3.364935064935065, + "grad_norm": 13.700212129426902, + "learning_rate": 1.5468184205806646e-05, + "loss": 0.0525, + "step": 2591 + }, + { + "epoch": 3.3662337662337665, + "grad_norm": 5.059075434380803, + "learning_rate": 1.546466209641091e-05, + "loss": 0.0152, + "step": 2592 + }, + { + "epoch": 3.3675324675324676, + "grad_norm": 18.8494329380806, + "learning_rate": 1.546113902021186e-05, + "loss": 0.0344, + "step": 2593 + }, + { + "epoch": 3.3688311688311687, + "grad_norm": 3.5592071277180906, + "learning_rate": 1.5457614977832782e-05, + "loss": 0.0175, + "step": 2594 + }, + { + "epoch": 3.3701298701298703, + "grad_norm": 21.701337930840264, + "learning_rate": 1.5454089969897158e-05, + "loss": 0.0393, + "step": 2595 + }, + { + "epoch": 3.3714285714285714, + "grad_norm": 16.636611239765244, + "learning_rate": 1.5450563997028628e-05, + "loss": 0.0564, + "step": 2596 + }, + { + "epoch": 3.3727272727272726, + "grad_norm": 36.14827183815564, + "learning_rate": 1.5447037059851e-05, + "loss": 0.0759, + "step": 2597 + }, + { + "epoch": 3.374025974025974, + "grad_norm": 15.640856040835457, + "learning_rate": 1.5443509158988254e-05, + "loss": 0.0403, + "step": 2598 + }, + { + "epoch": 3.3753246753246753, + "grad_norm": 22.91179265906459, + "learning_rate": 1.5439980295064557e-05, + "loss": 0.0426, + "step": 2599 + }, + { + "epoch": 3.3766233766233764, + "grad_norm": 35.35401419771444, + "learning_rate": 1.5436450468704215e-05, + "loss": 0.0406, + "step": 2600 + }, + { + "epoch": 3.377922077922078, + "grad_norm": 18.72668081873439, + "learning_rate": 1.5432919680531736e-05, + "loss": 0.0352, + "step": 2601 + }, + { + "epoch": 3.379220779220779, + "grad_norm": 34.18507261520058, + "learning_rate": 1.542938793117178e-05, + "loss": 0.0353, + "step": 2602 + }, + { + "epoch": 3.3805194805194807, + "grad_norm": 9.718840796934924, + "learning_rate": 1.542585522124917e-05, + "loss": 0.0362, + "step": 2603 + }, + { + "epoch": 3.381818181818182, + "grad_norm": 34.556270054729225, + "learning_rate": 1.542232155138893e-05, + "loss": 0.0506, + "step": 2604 + }, + { + "epoch": 3.383116883116883, + "grad_norm": 20.450367249287684, + "learning_rate": 1.5418786922216216e-05, + "loss": 0.0352, + "step": 2605 + }, + { + "epoch": 3.3844155844155845, + "grad_norm": 21.615275995407732, + "learning_rate": 1.5415251334356383e-05, + "loss": 0.0235, + "step": 2606 + }, + { + "epoch": 3.3857142857142857, + "grad_norm": 25.65681389275723, + "learning_rate": 1.541171478843494e-05, + "loss": 0.0512, + "step": 2607 + }, + { + "epoch": 3.3870129870129873, + "grad_norm": 6.951476899506869, + "learning_rate": 1.5408177285077564e-05, + "loss": 0.038, + "step": 2608 + }, + { + "epoch": 3.3883116883116884, + "grad_norm": 36.55158251130923, + "learning_rate": 1.5404638824910116e-05, + "loss": 0.0609, + "step": 2609 + }, + { + "epoch": 3.3896103896103895, + "grad_norm": 6.539377066529639, + "learning_rate": 1.5401099408558615e-05, + "loss": 0.0284, + "step": 2610 + }, + { + "epoch": 3.390909090909091, + "grad_norm": 24.371161200590766, + "learning_rate": 1.539755903664925e-05, + "loss": 0.0534, + "step": 2611 + }, + { + "epoch": 3.3922077922077922, + "grad_norm": 17.023883979352068, + "learning_rate": 1.539401770980838e-05, + "loss": 0.0305, + "step": 2612 + }, + { + "epoch": 3.3935064935064934, + "grad_norm": 9.622075868069851, + "learning_rate": 1.5390475428662534e-05, + "loss": 0.0315, + "step": 2613 + }, + { + "epoch": 3.394805194805195, + "grad_norm": 24.366771479673346, + "learning_rate": 1.5386932193838412e-05, + "loss": 0.0659, + "step": 2614 + }, + { + "epoch": 3.396103896103896, + "grad_norm": 6.9359224652078915, + "learning_rate": 1.5383388005962878e-05, + "loss": 0.0171, + "step": 2615 + }, + { + "epoch": 3.397402597402597, + "grad_norm": 19.82251968205052, + "learning_rate": 1.537984286566297e-05, + "loss": 0.0258, + "step": 2616 + }, + { + "epoch": 3.398701298701299, + "grad_norm": 16.815114403402916, + "learning_rate": 1.537629677356588e-05, + "loss": 0.0366, + "step": 2617 + }, + { + "epoch": 3.4, + "grad_norm": 22.87750680433941, + "learning_rate": 1.537274973029899e-05, + "loss": 0.0539, + "step": 2618 + }, + { + "epoch": 3.4012987012987015, + "grad_norm": 4.224634024584742, + "learning_rate": 1.536920173648984e-05, + "loss": 0.0387, + "step": 2619 + }, + { + "epoch": 3.4025974025974026, + "grad_norm": 6.026308431528538, + "learning_rate": 1.5365652792766134e-05, + "loss": 0.0269, + "step": 2620 + }, + { + "epoch": 3.4038961038961038, + "grad_norm": 7.244988513051976, + "learning_rate": 1.536210289975575e-05, + "loss": 0.0683, + "step": 2621 + }, + { + "epoch": 3.4051948051948053, + "grad_norm": 15.322761178932252, + "learning_rate": 1.5358552058086725e-05, + "loss": 0.0491, + "step": 2622 + }, + { + "epoch": 3.4064935064935065, + "grad_norm": 4.272604945673308, + "learning_rate": 1.5355000268387286e-05, + "loss": 0.0345, + "step": 2623 + }, + { + "epoch": 3.407792207792208, + "grad_norm": 20.071339111670053, + "learning_rate": 1.53514475312858e-05, + "loss": 0.0374, + "step": 2624 + }, + { + "epoch": 3.409090909090909, + "grad_norm": 4.669812997437132, + "learning_rate": 1.534789384741082e-05, + "loss": 0.0329, + "step": 2625 + }, + { + "epoch": 3.4103896103896103, + "grad_norm": 14.710189643637245, + "learning_rate": 1.534433921739105e-05, + "loss": 0.061, + "step": 2626 + }, + { + "epoch": 3.411688311688312, + "grad_norm": 14.382858851411017, + "learning_rate": 1.5340783641855388e-05, + "loss": 0.0652, + "step": 2627 + }, + { + "epoch": 3.412987012987013, + "grad_norm": 6.546907415933409, + "learning_rate": 1.5337227121432876e-05, + "loss": 0.054, + "step": 2628 + }, + { + "epoch": 3.414285714285714, + "grad_norm": 23.6105574234002, + "learning_rate": 1.533366965675273e-05, + "loss": 0.04, + "step": 2629 + }, + { + "epoch": 3.4155844155844157, + "grad_norm": 12.598378158647096, + "learning_rate": 1.5330111248444332e-05, + "loss": 0.0218, + "step": 2630 + }, + { + "epoch": 3.416883116883117, + "grad_norm": 16.93888719157786, + "learning_rate": 1.5326551897137236e-05, + "loss": 0.057, + "step": 2631 + }, + { + "epoch": 3.418181818181818, + "grad_norm": 4.956943620997017, + "learning_rate": 1.532299160346116e-05, + "loss": 0.0286, + "step": 2632 + }, + { + "epoch": 3.4194805194805196, + "grad_norm": 11.009608742684318, + "learning_rate": 1.5319430368045987e-05, + "loss": 0.0382, + "step": 2633 + }, + { + "epoch": 3.4207792207792207, + "grad_norm": 4.285742173883653, + "learning_rate": 1.5315868191521762e-05, + "loss": 0.0373, + "step": 2634 + }, + { + "epoch": 3.4220779220779223, + "grad_norm": 11.545314292486115, + "learning_rate": 1.5312305074518712e-05, + "loss": 0.0838, + "step": 2635 + }, + { + "epoch": 3.4233766233766234, + "grad_norm": 10.56965687891732, + "learning_rate": 1.5308741017667213e-05, + "loss": 0.0416, + "step": 2636 + }, + { + "epoch": 3.4246753246753245, + "grad_norm": 11.061291491203011, + "learning_rate": 1.5305176021597817e-05, + "loss": 0.0157, + "step": 2637 + }, + { + "epoch": 3.425974025974026, + "grad_norm": 12.03947775025989, + "learning_rate": 1.5301610086941244e-05, + "loss": 0.0366, + "step": 2638 + }, + { + "epoch": 3.4272727272727272, + "grad_norm": 5.663230797281292, + "learning_rate": 1.5298043214328368e-05, + "loss": 0.0461, + "step": 2639 + }, + { + "epoch": 3.4285714285714284, + "grad_norm": 4.6527566985309585, + "learning_rate": 1.5294475404390245e-05, + "loss": 0.0333, + "step": 2640 + }, + { + "epoch": 3.42987012987013, + "grad_norm": 6.212778012836857, + "learning_rate": 1.5290906657758085e-05, + "loss": 0.0469, + "step": 2641 + }, + { + "epoch": 3.431168831168831, + "grad_norm": 9.386792069861897, + "learning_rate": 1.528733697506327e-05, + "loss": 0.0307, + "step": 2642 + }, + { + "epoch": 3.4324675324675322, + "grad_norm": 3.084745760722714, + "learning_rate": 1.528376635693734e-05, + "loss": 0.0243, + "step": 2643 + }, + { + "epoch": 3.433766233766234, + "grad_norm": 12.672627167550356, + "learning_rate": 1.5280194804012012e-05, + "loss": 0.0537, + "step": 2644 + }, + { + "epoch": 3.435064935064935, + "grad_norm": 9.831923458636473, + "learning_rate": 1.5276622316919158e-05, + "loss": 0.0495, + "step": 2645 + }, + { + "epoch": 3.4363636363636365, + "grad_norm": 5.913144240646676, + "learning_rate": 1.527304889629082e-05, + "loss": 0.0391, + "step": 2646 + }, + { + "epoch": 3.4376623376623376, + "grad_norm": 4.729083411642438, + "learning_rate": 1.5269474542759206e-05, + "loss": 0.0378, + "step": 2647 + }, + { + "epoch": 3.4389610389610388, + "grad_norm": 10.481346985205976, + "learning_rate": 1.5265899256956687e-05, + "loss": 0.0393, + "step": 2648 + }, + { + "epoch": 3.4402597402597404, + "grad_norm": 8.302508423831624, + "learning_rate": 1.52623230395158e-05, + "loss": 0.0125, + "step": 2649 + }, + { + "epoch": 3.4415584415584415, + "grad_norm": 5.004480476203915, + "learning_rate": 1.525874589106924e-05, + "loss": 0.0501, + "step": 2650 + }, + { + "epoch": 3.442857142857143, + "grad_norm": 7.519436564693977, + "learning_rate": 1.5255167812249882e-05, + "loss": 0.0561, + "step": 2651 + }, + { + "epoch": 3.444155844155844, + "grad_norm": 10.439537247750488, + "learning_rate": 1.5251588803690753e-05, + "loss": 0.0689, + "step": 2652 + }, + { + "epoch": 3.4454545454545453, + "grad_norm": 9.134020955579174, + "learning_rate": 1.5248008866025048e-05, + "loss": 0.0336, + "step": 2653 + }, + { + "epoch": 3.446753246753247, + "grad_norm": 13.285441578728602, + "learning_rate": 1.5244427999886124e-05, + "loss": 0.0577, + "step": 2654 + }, + { + "epoch": 3.448051948051948, + "grad_norm": 7.671226398510718, + "learning_rate": 1.5240846205907512e-05, + "loss": 0.0141, + "step": 2655 + }, + { + "epoch": 3.449350649350649, + "grad_norm": 13.909102404899325, + "learning_rate": 1.5237263484722893e-05, + "loss": 0.0529, + "step": 2656 + }, + { + "epoch": 3.4506493506493507, + "grad_norm": 4.13352772496371, + "learning_rate": 1.5233679836966122e-05, + "loss": 0.0266, + "step": 2657 + }, + { + "epoch": 3.451948051948052, + "grad_norm": 14.661375669113616, + "learning_rate": 1.5230095263271211e-05, + "loss": 0.031, + "step": 2658 + }, + { + "epoch": 3.453246753246753, + "grad_norm": 4.198230664578419, + "learning_rate": 1.5226509764272348e-05, + "loss": 0.0203, + "step": 2659 + }, + { + "epoch": 3.4545454545454546, + "grad_norm": 17.484421308008194, + "learning_rate": 1.5222923340603868e-05, + "loss": 0.0344, + "step": 2660 + }, + { + "epoch": 3.4558441558441557, + "grad_norm": 7.5834359510542635, + "learning_rate": 1.5219335992900282e-05, + "loss": 0.0355, + "step": 2661 + }, + { + "epoch": 3.4571428571428573, + "grad_norm": 15.171796679132663, + "learning_rate": 1.5215747721796259e-05, + "loss": 0.0366, + "step": 2662 + }, + { + "epoch": 3.4584415584415584, + "grad_norm": 15.211663467949982, + "learning_rate": 1.5212158527926635e-05, + "loss": 0.0536, + "step": 2663 + }, + { + "epoch": 3.4597402597402596, + "grad_norm": 16.202283598833116, + "learning_rate": 1.5208568411926406e-05, + "loss": 0.0294, + "step": 2664 + }, + { + "epoch": 3.461038961038961, + "grad_norm": 12.902181105347783, + "learning_rate": 1.520497737443073e-05, + "loss": 0.0383, + "step": 2665 + }, + { + "epoch": 3.4623376623376623, + "grad_norm": 3.6399088363186243, + "learning_rate": 1.5201385416074934e-05, + "loss": 0.0391, + "step": 2666 + }, + { + "epoch": 3.463636363636364, + "grad_norm": 23.981751589373218, + "learning_rate": 1.5197792537494498e-05, + "loss": 0.0239, + "step": 2667 + }, + { + "epoch": 3.464935064935065, + "grad_norm": 4.258406116461223, + "learning_rate": 1.5194198739325078e-05, + "loss": 0.0274, + "step": 2668 + }, + { + "epoch": 3.466233766233766, + "grad_norm": 12.131774353352094, + "learning_rate": 1.5190604022202484e-05, + "loss": 0.0481, + "step": 2669 + }, + { + "epoch": 3.4675324675324677, + "grad_norm": 19.9056635402713, + "learning_rate": 1.518700838676269e-05, + "loss": 0.0324, + "step": 2670 + }, + { + "epoch": 3.468831168831169, + "grad_norm": 9.774249534443738, + "learning_rate": 1.518341183364183e-05, + "loss": 0.0453, + "step": 2671 + }, + { + "epoch": 3.47012987012987, + "grad_norm": 33.16440957381356, + "learning_rate": 1.5179814363476207e-05, + "loss": 0.0352, + "step": 2672 + }, + { + "epoch": 3.4714285714285715, + "grad_norm": 8.515429419055199, + "learning_rate": 1.517621597690228e-05, + "loss": 0.0324, + "step": 2673 + }, + { + "epoch": 3.4727272727272727, + "grad_norm": 12.911686793111928, + "learning_rate": 1.5172616674556673e-05, + "loss": 0.023, + "step": 2674 + }, + { + "epoch": 3.474025974025974, + "grad_norm": 13.00893894393131, + "learning_rate": 1.5169016457076172e-05, + "loss": 0.0316, + "step": 2675 + }, + { + "epoch": 3.4753246753246754, + "grad_norm": 5.2519476001986956, + "learning_rate": 1.5165415325097718e-05, + "loss": 0.0245, + "step": 2676 + }, + { + "epoch": 3.4766233766233765, + "grad_norm": 10.181269522981545, + "learning_rate": 1.5161813279258431e-05, + "loss": 0.0134, + "step": 2677 + }, + { + "epoch": 3.477922077922078, + "grad_norm": 19.71626536551108, + "learning_rate": 1.5158210320195577e-05, + "loss": 0.0397, + "step": 2678 + }, + { + "epoch": 3.479220779220779, + "grad_norm": 12.3445822065009, + "learning_rate": 1.5154606448546587e-05, + "loss": 0.036, + "step": 2679 + }, + { + "epoch": 3.4805194805194803, + "grad_norm": 20.91009787276524, + "learning_rate": 1.5151001664949055e-05, + "loss": 0.0772, + "step": 2680 + }, + { + "epoch": 3.481818181818182, + "grad_norm": 9.283060313748141, + "learning_rate": 1.5147395970040738e-05, + "loss": 0.0794, + "step": 2681 + }, + { + "epoch": 3.483116883116883, + "grad_norm": 10.12214505895812, + "learning_rate": 1.5143789364459552e-05, + "loss": 0.0317, + "step": 2682 + }, + { + "epoch": 3.4844155844155846, + "grad_norm": 19.1141115891684, + "learning_rate": 1.5140181848843574e-05, + "loss": 0.0496, + "step": 2683 + }, + { + "epoch": 3.4857142857142858, + "grad_norm": 19.24717612963016, + "learning_rate": 1.5136573423831043e-05, + "loss": 0.0604, + "step": 2684 + }, + { + "epoch": 3.487012987012987, + "grad_norm": 20.960535775160476, + "learning_rate": 1.5132964090060358e-05, + "loss": 0.0452, + "step": 2685 + }, + { + "epoch": 3.4883116883116885, + "grad_norm": 16.69880147592467, + "learning_rate": 1.5129353848170081e-05, + "loss": 0.0348, + "step": 2686 + }, + { + "epoch": 3.4896103896103896, + "grad_norm": 22.737532891341573, + "learning_rate": 1.5125742698798929e-05, + "loss": 0.0581, + "step": 2687 + }, + { + "epoch": 3.4909090909090907, + "grad_norm": 8.757788584286734, + "learning_rate": 1.5122130642585786e-05, + "loss": 0.0179, + "step": 2688 + }, + { + "epoch": 3.4922077922077923, + "grad_norm": 20.268398869124663, + "learning_rate": 1.5118517680169695e-05, + "loss": 0.0265, + "step": 2689 + }, + { + "epoch": 3.4935064935064934, + "grad_norm": 20.656807338435332, + "learning_rate": 1.5114903812189855e-05, + "loss": 0.0533, + "step": 2690 + }, + { + "epoch": 3.4948051948051946, + "grad_norm": 8.582969612190134, + "learning_rate": 1.5111289039285632e-05, + "loss": 0.0321, + "step": 2691 + }, + { + "epoch": 3.496103896103896, + "grad_norm": 12.40247850360799, + "learning_rate": 1.5107673362096546e-05, + "loss": 0.0341, + "step": 2692 + }, + { + "epoch": 3.4974025974025973, + "grad_norm": 4.465929246192896, + "learning_rate": 1.5104056781262282e-05, + "loss": 0.0199, + "step": 2693 + }, + { + "epoch": 3.498701298701299, + "grad_norm": 16.137517302476464, + "learning_rate": 1.5100439297422677e-05, + "loss": 0.0442, + "step": 2694 + }, + { + "epoch": 3.5, + "grad_norm": 6.124803871645755, + "learning_rate": 1.5096820911217745e-05, + "loss": 0.0419, + "step": 2695 + }, + { + "epoch": 3.501298701298701, + "grad_norm": 21.42179480415258, + "learning_rate": 1.5093201623287631e-05, + "loss": 0.0445, + "step": 2696 + }, + { + "epoch": 3.5025974025974027, + "grad_norm": 13.999271471597705, + "learning_rate": 1.5089581434272667e-05, + "loss": 0.0854, + "step": 2697 + }, + { + "epoch": 3.503896103896104, + "grad_norm": 17.370892200195186, + "learning_rate": 1.5085960344813333e-05, + "loss": 0.0527, + "step": 2698 + }, + { + "epoch": 3.5051948051948054, + "grad_norm": 4.778280534594769, + "learning_rate": 1.5082338355550267e-05, + "loss": 0.0398, + "step": 2699 + }, + { + "epoch": 3.5064935064935066, + "grad_norm": 16.66340529974551, + "learning_rate": 1.507871546712427e-05, + "loss": 0.0625, + "step": 2700 + }, + { + "epoch": 3.5077922077922077, + "grad_norm": 9.177121831861857, + "learning_rate": 1.5075091680176299e-05, + "loss": 0.0447, + "step": 2701 + }, + { + "epoch": 3.509090909090909, + "grad_norm": 16.431395344533158, + "learning_rate": 1.5071466995347471e-05, + "loss": 0.0458, + "step": 2702 + }, + { + "epoch": 3.5103896103896104, + "grad_norm": 12.354349334029688, + "learning_rate": 1.5067841413279063e-05, + "loss": 0.0254, + "step": 2703 + }, + { + "epoch": 3.5116883116883115, + "grad_norm": 5.363512992520653, + "learning_rate": 1.5064214934612507e-05, + "loss": 0.0522, + "step": 2704 + }, + { + "epoch": 3.512987012987013, + "grad_norm": 18.96537210141487, + "learning_rate": 1.5060587559989403e-05, + "loss": 0.0218, + "step": 2705 + }, + { + "epoch": 3.5142857142857142, + "grad_norm": 8.855191112688788, + "learning_rate": 1.5056959290051496e-05, + "loss": 0.0394, + "step": 2706 + }, + { + "epoch": 3.5155844155844154, + "grad_norm": 14.982754812845435, + "learning_rate": 1.5053330125440701e-05, + "loss": 0.0333, + "step": 2707 + }, + { + "epoch": 3.516883116883117, + "grad_norm": 13.796763724488034, + "learning_rate": 1.5049700066799084e-05, + "loss": 0.0591, + "step": 2708 + }, + { + "epoch": 3.518181818181818, + "grad_norm": 26.173091885641917, + "learning_rate": 1.5046069114768873e-05, + "loss": 0.049, + "step": 2709 + }, + { + "epoch": 3.5194805194805197, + "grad_norm": 13.835876420912655, + "learning_rate": 1.5042437269992451e-05, + "loss": 0.011, + "step": 2710 + }, + { + "epoch": 3.520779220779221, + "grad_norm": 20.921258764380276, + "learning_rate": 1.5038804533112368e-05, + "loss": 0.021, + "step": 2711 + }, + { + "epoch": 3.522077922077922, + "grad_norm": 25.334313519663812, + "learning_rate": 1.5035170904771315e-05, + "loss": 0.0537, + "step": 2712 + }, + { + "epoch": 3.5233766233766235, + "grad_norm": 22.226622179011105, + "learning_rate": 1.5031536385612158e-05, + "loss": 0.0548, + "step": 2713 + }, + { + "epoch": 3.5246753246753246, + "grad_norm": 14.955599697113009, + "learning_rate": 1.5027900976277907e-05, + "loss": 0.0491, + "step": 2714 + }, + { + "epoch": 3.525974025974026, + "grad_norm": 4.861421753200918, + "learning_rate": 1.5024264677411738e-05, + "loss": 0.0252, + "step": 2715 + }, + { + "epoch": 3.5272727272727273, + "grad_norm": 20.16616927477854, + "learning_rate": 1.5020627489656984e-05, + "loss": 0.0211, + "step": 2716 + }, + { + "epoch": 3.5285714285714285, + "grad_norm": 6.534755426239281, + "learning_rate": 1.5016989413657127e-05, + "loss": 0.0501, + "step": 2717 + }, + { + "epoch": 3.5298701298701296, + "grad_norm": 9.190383125749065, + "learning_rate": 1.5013350450055821e-05, + "loss": 0.0424, + "step": 2718 + }, + { + "epoch": 3.531168831168831, + "grad_norm": 17.571017842106496, + "learning_rate": 1.5009710599496862e-05, + "loss": 0.0389, + "step": 2719 + }, + { + "epoch": 3.5324675324675323, + "grad_norm": 8.413547605189645, + "learning_rate": 1.5006069862624214e-05, + "loss": 0.0369, + "step": 2720 + }, + { + "epoch": 3.533766233766234, + "grad_norm": 13.000163652387261, + "learning_rate": 1.5002428240081987e-05, + "loss": 0.0334, + "step": 2721 + }, + { + "epoch": 3.535064935064935, + "grad_norm": 9.2412381315696, + "learning_rate": 1.4998785732514459e-05, + "loss": 0.0407, + "step": 2722 + }, + { + "epoch": 3.536363636363636, + "grad_norm": 16.08854408844238, + "learning_rate": 1.4995142340566054e-05, + "loss": 0.089, + "step": 2723 + }, + { + "epoch": 3.5376623376623377, + "grad_norm": 18.12589918334963, + "learning_rate": 1.4991498064881362e-05, + "loss": 0.0294, + "step": 2724 + }, + { + "epoch": 3.538961038961039, + "grad_norm": 18.603822856584372, + "learning_rate": 1.4987852906105125e-05, + "loss": 0.0407, + "step": 2725 + }, + { + "epoch": 3.5402597402597404, + "grad_norm": 13.954393563038423, + "learning_rate": 1.498420686488224e-05, + "loss": 0.0536, + "step": 2726 + }, + { + "epoch": 3.5415584415584416, + "grad_norm": 5.588551459610778, + "learning_rate": 1.4980559941857762e-05, + "loss": 0.06, + "step": 2727 + }, + { + "epoch": 3.5428571428571427, + "grad_norm": 15.101026513841736, + "learning_rate": 1.49769121376769e-05, + "loss": 0.0466, + "step": 2728 + }, + { + "epoch": 3.5441558441558443, + "grad_norm": 8.233624587831475, + "learning_rate": 1.4973263452985023e-05, + "loss": 0.0399, + "step": 2729 + }, + { + "epoch": 3.5454545454545454, + "grad_norm": 15.616364516586707, + "learning_rate": 1.496961388842765e-05, + "loss": 0.0426, + "step": 2730 + }, + { + "epoch": 3.546753246753247, + "grad_norm": 6.969850553380492, + "learning_rate": 1.4965963444650465e-05, + "loss": 0.0711, + "step": 2731 + }, + { + "epoch": 3.548051948051948, + "grad_norm": 3.988320447926815, + "learning_rate": 1.4962312122299294e-05, + "loss": 0.0326, + "step": 2732 + }, + { + "epoch": 3.5493506493506493, + "grad_norm": 13.398331102394202, + "learning_rate": 1.495865992202013e-05, + "loss": 0.0365, + "step": 2733 + }, + { + "epoch": 3.5506493506493504, + "grad_norm": 8.392227017836372, + "learning_rate": 1.4955006844459115e-05, + "loss": 0.086, + "step": 2734 + }, + { + "epoch": 3.551948051948052, + "grad_norm": 14.804259938180836, + "learning_rate": 1.4951352890262549e-05, + "loss": 0.0401, + "step": 2735 + }, + { + "epoch": 3.553246753246753, + "grad_norm": 9.167717968086706, + "learning_rate": 1.4947698060076888e-05, + "loss": 0.0545, + "step": 2736 + }, + { + "epoch": 3.5545454545454547, + "grad_norm": 13.695645120301657, + "learning_rate": 1.494404235454874e-05, + "loss": 0.0582, + "step": 2737 + }, + { + "epoch": 3.555844155844156, + "grad_norm": 7.4998789700682025, + "learning_rate": 1.494038577432487e-05, + "loss": 0.0536, + "step": 2738 + }, + { + "epoch": 3.557142857142857, + "grad_norm": 13.527870657198456, + "learning_rate": 1.4936728320052196e-05, + "loss": 0.039, + "step": 2739 + }, + { + "epoch": 3.5584415584415585, + "grad_norm": 9.351568246537136, + "learning_rate": 1.4933069992377793e-05, + "loss": 0.0293, + "step": 2740 + }, + { + "epoch": 3.5597402597402596, + "grad_norm": 21.828842854354626, + "learning_rate": 1.4929410791948886e-05, + "loss": 0.0831, + "step": 2741 + }, + { + "epoch": 3.5610389610389612, + "grad_norm": 7.662760170181546, + "learning_rate": 1.4925750719412863e-05, + "loss": 0.0501, + "step": 2742 + }, + { + "epoch": 3.5623376623376624, + "grad_norm": 30.568336222490796, + "learning_rate": 1.4922089775417257e-05, + "loss": 0.0376, + "step": 2743 + }, + { + "epoch": 3.5636363636363635, + "grad_norm": 7.09778804172361, + "learning_rate": 1.491842796060976e-05, + "loss": 0.0532, + "step": 2744 + }, + { + "epoch": 3.564935064935065, + "grad_norm": 24.413059941389047, + "learning_rate": 1.4914765275638215e-05, + "loss": 0.0698, + "step": 2745 + }, + { + "epoch": 3.566233766233766, + "grad_norm": 9.705824455753076, + "learning_rate": 1.4911101721150625e-05, + "loss": 0.0348, + "step": 2746 + }, + { + "epoch": 3.5675324675324678, + "grad_norm": 22.243825637246857, + "learning_rate": 1.490743729779514e-05, + "loss": 0.0445, + "step": 2747 + }, + { + "epoch": 3.568831168831169, + "grad_norm": 19.82002207733103, + "learning_rate": 1.4903772006220067e-05, + "loss": 0.0499, + "step": 2748 + }, + { + "epoch": 3.57012987012987, + "grad_norm": 16.2123486043267, + "learning_rate": 1.490010584707387e-05, + "loss": 0.0331, + "step": 2749 + }, + { + "epoch": 3.571428571428571, + "grad_norm": 25.065753376841002, + "learning_rate": 1.4896438821005157e-05, + "loss": 0.0381, + "step": 2750 + }, + { + "epoch": 3.5727272727272728, + "grad_norm": 9.046474560400952, + "learning_rate": 1.4892770928662699e-05, + "loss": 0.0281, + "step": 2751 + }, + { + "epoch": 3.574025974025974, + "grad_norm": 22.458776506533674, + "learning_rate": 1.4889102170695416e-05, + "loss": 0.0524, + "step": 2752 + }, + { + "epoch": 3.5753246753246755, + "grad_norm": 9.03616940628358, + "learning_rate": 1.4885432547752377e-05, + "loss": 0.0571, + "step": 2753 + }, + { + "epoch": 3.5766233766233766, + "grad_norm": 10.691360546134744, + "learning_rate": 1.4881762060482814e-05, + "loss": 0.0245, + "step": 2754 + }, + { + "epoch": 3.5779220779220777, + "grad_norm": 6.90696004466469, + "learning_rate": 1.4878090709536107e-05, + "loss": 0.0378, + "step": 2755 + }, + { + "epoch": 3.5792207792207793, + "grad_norm": 7.958661824394447, + "learning_rate": 1.4874418495561783e-05, + "loss": 0.0442, + "step": 2756 + }, + { + "epoch": 3.5805194805194804, + "grad_norm": 11.059765212572787, + "learning_rate": 1.4870745419209531e-05, + "loss": 0.0295, + "step": 2757 + }, + { + "epoch": 3.581818181818182, + "grad_norm": 9.29301510815726, + "learning_rate": 1.4867071481129186e-05, + "loss": 0.0298, + "step": 2758 + }, + { + "epoch": 3.583116883116883, + "grad_norm": 10.027153451396723, + "learning_rate": 1.4863396681970739e-05, + "loss": 0.0455, + "step": 2759 + }, + { + "epoch": 3.5844155844155843, + "grad_norm": 7.049207992631224, + "learning_rate": 1.4859721022384335e-05, + "loss": 0.0309, + "step": 2760 + }, + { + "epoch": 3.585714285714286, + "grad_norm": 10.274033007361016, + "learning_rate": 1.4856044503020263e-05, + "loss": 0.0421, + "step": 2761 + }, + { + "epoch": 3.587012987012987, + "grad_norm": 10.183674615634013, + "learning_rate": 1.4852367124528975e-05, + "loss": 0.0205, + "step": 2762 + }, + { + "epoch": 3.5883116883116886, + "grad_norm": 7.269348374077022, + "learning_rate": 1.4848688887561068e-05, + "loss": 0.0299, + "step": 2763 + }, + { + "epoch": 3.5896103896103897, + "grad_norm": 9.169714550872248, + "learning_rate": 1.4845009792767289e-05, + "loss": 0.0327, + "step": 2764 + }, + { + "epoch": 3.590909090909091, + "grad_norm": 4.80828980205508, + "learning_rate": 1.4841329840798543e-05, + "loss": 0.0315, + "step": 2765 + }, + { + "epoch": 3.592207792207792, + "grad_norm": 5.880856199411808, + "learning_rate": 1.4837649032305885e-05, + "loss": 0.0502, + "step": 2766 + }, + { + "epoch": 3.5935064935064935, + "grad_norm": 12.567033149510932, + "learning_rate": 1.4833967367940523e-05, + "loss": 0.044, + "step": 2767 + }, + { + "epoch": 3.5948051948051947, + "grad_norm": 11.816219521718839, + "learning_rate": 1.483028484835381e-05, + "loss": 0.0526, + "step": 2768 + }, + { + "epoch": 3.5961038961038962, + "grad_norm": 6.206281081957932, + "learning_rate": 1.482660147419725e-05, + "loss": 0.03, + "step": 2769 + }, + { + "epoch": 3.5974025974025974, + "grad_norm": 8.380178711453398, + "learning_rate": 1.4822917246122512e-05, + "loss": 0.0605, + "step": 2770 + }, + { + "epoch": 3.5987012987012985, + "grad_norm": 5.817712717376197, + "learning_rate": 1.4819232164781396e-05, + "loss": 0.0517, + "step": 2771 + }, + { + "epoch": 3.6, + "grad_norm": 5.503244352296483, + "learning_rate": 1.4815546230825875e-05, + "loss": 0.043, + "step": 2772 + }, + { + "epoch": 3.601298701298701, + "grad_norm": 2.9693263063178565, + "learning_rate": 1.4811859444908053e-05, + "loss": 0.033, + "step": 2773 + }, + { + "epoch": 3.602597402597403, + "grad_norm": 5.912888966473904, + "learning_rate": 1.4808171807680194e-05, + "loss": 0.021, + "step": 2774 + }, + { + "epoch": 3.603896103896104, + "grad_norm": 11.64142968302923, + "learning_rate": 1.4804483319794711e-05, + "loss": 0.0472, + "step": 2775 + }, + { + "epoch": 3.605194805194805, + "grad_norm": 11.82960453728387, + "learning_rate": 1.4800793981904175e-05, + "loss": 0.0437, + "step": 2776 + }, + { + "epoch": 3.6064935064935066, + "grad_norm": 10.798513186594198, + "learning_rate": 1.4797103794661292e-05, + "loss": 0.0544, + "step": 2777 + }, + { + "epoch": 3.6077922077922078, + "grad_norm": 5.330854181192362, + "learning_rate": 1.4793412758718933e-05, + "loss": 0.0257, + "step": 2778 + }, + { + "epoch": 3.6090909090909093, + "grad_norm": 18.413677499679647, + "learning_rate": 1.4789720874730106e-05, + "loss": 0.0567, + "step": 2779 + }, + { + "epoch": 3.6103896103896105, + "grad_norm": 6.3980941302321925, + "learning_rate": 1.4786028143347984e-05, + "loss": 0.0442, + "step": 2780 + }, + { + "epoch": 3.6116883116883116, + "grad_norm": 20.31327690823794, + "learning_rate": 1.4782334565225877e-05, + "loss": 0.0618, + "step": 2781 + }, + { + "epoch": 3.6129870129870127, + "grad_norm": 5.745444415218033, + "learning_rate": 1.4778640141017249e-05, + "loss": 0.0279, + "step": 2782 + }, + { + "epoch": 3.6142857142857143, + "grad_norm": 14.061475665720847, + "learning_rate": 1.4774944871375715e-05, + "loss": 0.0476, + "step": 2783 + }, + { + "epoch": 3.6155844155844155, + "grad_norm": 18.267185391176938, + "learning_rate": 1.4771248756955042e-05, + "loss": 0.0301, + "step": 2784 + }, + { + "epoch": 3.616883116883117, + "grad_norm": 8.787518714734428, + "learning_rate": 1.4767551798409141e-05, + "loss": 0.0125, + "step": 2785 + }, + { + "epoch": 3.618181818181818, + "grad_norm": 20.07606700634818, + "learning_rate": 1.4763853996392075e-05, + "loss": 0.0414, + "step": 2786 + }, + { + "epoch": 3.6194805194805193, + "grad_norm": 7.185599741262751, + "learning_rate": 1.4760155351558055e-05, + "loss": 0.052, + "step": 2787 + }, + { + "epoch": 3.620779220779221, + "grad_norm": 12.35823809261525, + "learning_rate": 1.4756455864561442e-05, + "loss": 0.0606, + "step": 2788 + }, + { + "epoch": 3.622077922077922, + "grad_norm": 16.98044669977127, + "learning_rate": 1.4752755536056751e-05, + "loss": 0.0421, + "step": 2789 + }, + { + "epoch": 3.6233766233766236, + "grad_norm": 14.184541607345025, + "learning_rate": 1.4749054366698632e-05, + "loss": 0.059, + "step": 2790 + }, + { + "epoch": 3.6246753246753247, + "grad_norm": 7.609178529538834, + "learning_rate": 1.4745352357141903e-05, + "loss": 0.0275, + "step": 2791 + }, + { + "epoch": 3.625974025974026, + "grad_norm": 7.029605649298671, + "learning_rate": 1.474164950804151e-05, + "loss": 0.0658, + "step": 2792 + }, + { + "epoch": 3.6272727272727274, + "grad_norm": 6.932977221525121, + "learning_rate": 1.4737945820052566e-05, + "loss": 0.0251, + "step": 2793 + }, + { + "epoch": 3.6285714285714286, + "grad_norm": 4.274090255269545, + "learning_rate": 1.4734241293830321e-05, + "loss": 0.0325, + "step": 2794 + }, + { + "epoch": 3.62987012987013, + "grad_norm": 12.039963129949486, + "learning_rate": 1.4730535930030172e-05, + "loss": 0.0298, + "step": 2795 + }, + { + "epoch": 3.6311688311688313, + "grad_norm": 4.512367950704392, + "learning_rate": 1.4726829729307679e-05, + "loss": 0.0412, + "step": 2796 + }, + { + "epoch": 3.6324675324675324, + "grad_norm": 7.457093915002099, + "learning_rate": 1.4723122692318532e-05, + "loss": 0.0173, + "step": 2797 + }, + { + "epoch": 3.6337662337662335, + "grad_norm": 4.573647679492863, + "learning_rate": 1.4719414819718583e-05, + "loss": 0.0176, + "step": 2798 + }, + { + "epoch": 3.635064935064935, + "grad_norm": 4.202326236375847, + "learning_rate": 1.4715706112163817e-05, + "loss": 0.0374, + "step": 2799 + }, + { + "epoch": 3.6363636363636362, + "grad_norm": 15.117933830853199, + "learning_rate": 1.4711996570310383e-05, + "loss": 0.0394, + "step": 2800 + }, + { + "epoch": 3.637662337662338, + "grad_norm": 4.418412279673898, + "learning_rate": 1.4708286194814565e-05, + "loss": 0.0388, + "step": 2801 + }, + { + "epoch": 3.638961038961039, + "grad_norm": 8.72710425922234, + "learning_rate": 1.4704574986332803e-05, + "loss": 0.0331, + "step": 2802 + }, + { + "epoch": 3.64025974025974, + "grad_norm": 16.17936373393096, + "learning_rate": 1.4700862945521679e-05, + "loss": 0.0463, + "step": 2803 + }, + { + "epoch": 3.6415584415584417, + "grad_norm": 8.886585939489208, + "learning_rate": 1.4697150073037925e-05, + "loss": 0.0349, + "step": 2804 + }, + { + "epoch": 3.642857142857143, + "grad_norm": 23.92603343631102, + "learning_rate": 1.4693436369538416e-05, + "loss": 0.0262, + "step": 2805 + }, + { + "epoch": 3.6441558441558444, + "grad_norm": 7.099757258511428, + "learning_rate": 1.4689721835680183e-05, + "loss": 0.0507, + "step": 2806 + }, + { + "epoch": 3.6454545454545455, + "grad_norm": 24.05260137943423, + "learning_rate": 1.4686006472120396e-05, + "loss": 0.0354, + "step": 2807 + }, + { + "epoch": 3.6467532467532466, + "grad_norm": 9.14795396286759, + "learning_rate": 1.468229027951637e-05, + "loss": 0.0337, + "step": 2808 + }, + { + "epoch": 3.648051948051948, + "grad_norm": 21.509392414950128, + "learning_rate": 1.4678573258525577e-05, + "loss": 0.044, + "step": 2809 + }, + { + "epoch": 3.6493506493506493, + "grad_norm": 16.505497201771465, + "learning_rate": 1.4674855409805622e-05, + "loss": 0.0354, + "step": 2810 + }, + { + "epoch": 3.650649350649351, + "grad_norm": 31.60329847158903, + "learning_rate": 1.467113673401427e-05, + "loss": 0.0573, + "step": 2811 + }, + { + "epoch": 3.651948051948052, + "grad_norm": 12.378816540088065, + "learning_rate": 1.4667417231809422e-05, + "loss": 0.045, + "step": 2812 + }, + { + "epoch": 3.653246753246753, + "grad_norm": 25.341078527779494, + "learning_rate": 1.4663696903849131e-05, + "loss": 0.0539, + "step": 2813 + }, + { + "epoch": 3.6545454545454543, + "grad_norm": 21.016304680734347, + "learning_rate": 1.4659975750791594e-05, + "loss": 0.0339, + "step": 2814 + }, + { + "epoch": 3.655844155844156, + "grad_norm": 18.739477132617747, + "learning_rate": 1.465625377329515e-05, + "loss": 0.0414, + "step": 2815 + }, + { + "epoch": 3.657142857142857, + "grad_norm": 20.6271909076218, + "learning_rate": 1.4652530972018296e-05, + "loss": 0.0257, + "step": 2816 + }, + { + "epoch": 3.6584415584415586, + "grad_norm": 8.184734623405973, + "learning_rate": 1.4648807347619662e-05, + "loss": 0.0457, + "step": 2817 + }, + { + "epoch": 3.6597402597402597, + "grad_norm": 20.973866852173582, + "learning_rate": 1.4645082900758027e-05, + "loss": 0.0305, + "step": 2818 + }, + { + "epoch": 3.661038961038961, + "grad_norm": 23.44148032215228, + "learning_rate": 1.464135763209232e-05, + "loss": 0.0488, + "step": 2819 + }, + { + "epoch": 3.6623376623376624, + "grad_norm": 18.787533890918528, + "learning_rate": 1.4637631542281611e-05, + "loss": 0.0334, + "step": 2820 + }, + { + "epoch": 3.6636363636363636, + "grad_norm": 26.950942062577877, + "learning_rate": 1.463390463198512e-05, + "loss": 0.0477, + "step": 2821 + }, + { + "epoch": 3.664935064935065, + "grad_norm": 8.89203860065401, + "learning_rate": 1.4630176901862205e-05, + "loss": 0.0476, + "step": 2822 + }, + { + "epoch": 3.6662337662337663, + "grad_norm": 13.223067093840891, + "learning_rate": 1.4626448352572371e-05, + "loss": 0.0326, + "step": 2823 + }, + { + "epoch": 3.6675324675324674, + "grad_norm": 20.598387259578896, + "learning_rate": 1.4622718984775274e-05, + "loss": 0.0416, + "step": 2824 + }, + { + "epoch": 3.6688311688311686, + "grad_norm": 20.918972969764653, + "learning_rate": 1.4618988799130707e-05, + "loss": 0.0414, + "step": 2825 + }, + { + "epoch": 3.67012987012987, + "grad_norm": 17.44758433045498, + "learning_rate": 1.4615257796298612e-05, + "loss": 0.0876, + "step": 2826 + }, + { + "epoch": 3.6714285714285713, + "grad_norm": 10.409802038758794, + "learning_rate": 1.4611525976939081e-05, + "loss": 0.0441, + "step": 2827 + }, + { + "epoch": 3.672727272727273, + "grad_norm": 25.576990322131394, + "learning_rate": 1.4607793341712336e-05, + "loss": 0.0884, + "step": 2828 + }, + { + "epoch": 3.674025974025974, + "grad_norm": 16.045404580584588, + "learning_rate": 1.4604059891278754e-05, + "loss": 0.0552, + "step": 2829 + }, + { + "epoch": 3.675324675324675, + "grad_norm": 33.31627415126043, + "learning_rate": 1.4600325626298855e-05, + "loss": 0.0589, + "step": 2830 + }, + { + "epoch": 3.6766233766233767, + "grad_norm": 22.19926110062321, + "learning_rate": 1.45965905474333e-05, + "loss": 0.0409, + "step": 2831 + }, + { + "epoch": 3.677922077922078, + "grad_norm": 19.3222461156973, + "learning_rate": 1.4592854655342901e-05, + "loss": 0.0256, + "step": 2832 + }, + { + "epoch": 3.6792207792207794, + "grad_norm": 33.8491166798349, + "learning_rate": 1.4589117950688603e-05, + "loss": 0.0288, + "step": 2833 + }, + { + "epoch": 3.6805194805194805, + "grad_norm": 6.818085808207528, + "learning_rate": 1.4585380434131504e-05, + "loss": 0.0274, + "step": 2834 + }, + { + "epoch": 3.6818181818181817, + "grad_norm": 28.20630096275035, + "learning_rate": 1.4581642106332843e-05, + "loss": 0.0403, + "step": 2835 + }, + { + "epoch": 3.6831168831168832, + "grad_norm": 22.745545972283182, + "learning_rate": 1.4577902967953995e-05, + "loss": 0.0297, + "step": 2836 + }, + { + "epoch": 3.6844155844155844, + "grad_norm": 30.992165688373362, + "learning_rate": 1.4574163019656493e-05, + "loss": 0.0296, + "step": 2837 + }, + { + "epoch": 3.685714285714286, + "grad_norm": 34.03166953934165, + "learning_rate": 1.4570422262102001e-05, + "loss": 0.0394, + "step": 2838 + }, + { + "epoch": 3.687012987012987, + "grad_norm": 23.43381586604714, + "learning_rate": 1.4566680695952333e-05, + "loss": 0.08, + "step": 2839 + }, + { + "epoch": 3.688311688311688, + "grad_norm": 38.25958282612306, + "learning_rate": 1.4562938321869445e-05, + "loss": 0.0537, + "step": 2840 + }, + { + "epoch": 3.6896103896103893, + "grad_norm": 13.585068294875024, + "learning_rate": 1.455919514051543e-05, + "loss": 0.0729, + "step": 2841 + }, + { + "epoch": 3.690909090909091, + "grad_norm": 30.76028210113323, + "learning_rate": 1.4555451152552529e-05, + "loss": 0.0299, + "step": 2842 + }, + { + "epoch": 3.692207792207792, + "grad_norm": 8.739132989722895, + "learning_rate": 1.455170635864313e-05, + "loss": 0.0339, + "step": 2843 + }, + { + "epoch": 3.6935064935064936, + "grad_norm": 22.55811628803188, + "learning_rate": 1.4547960759449753e-05, + "loss": 0.0521, + "step": 2844 + }, + { + "epoch": 3.6948051948051948, + "grad_norm": 19.241410702847976, + "learning_rate": 1.454421435563507e-05, + "loss": 0.0513, + "step": 2845 + }, + { + "epoch": 3.696103896103896, + "grad_norm": 4.302511469709936, + "learning_rate": 1.454046714786189e-05, + "loss": 0.0346, + "step": 2846 + }, + { + "epoch": 3.6974025974025975, + "grad_norm": 22.1719278745936, + "learning_rate": 1.4536719136793168e-05, + "loss": 0.0253, + "step": 2847 + }, + { + "epoch": 3.6987012987012986, + "grad_norm": 10.200619767293494, + "learning_rate": 1.4532970323091992e-05, + "loss": 0.0395, + "step": 2848 + }, + { + "epoch": 3.7, + "grad_norm": 5.7566736273265064, + "learning_rate": 1.4529220707421607e-05, + "loss": 0.0489, + "step": 2849 + }, + { + "epoch": 3.7012987012987013, + "grad_norm": 20.00733745039456, + "learning_rate": 1.4525470290445392e-05, + "loss": 0.031, + "step": 2850 + }, + { + "epoch": 3.7025974025974024, + "grad_norm": 5.721817484760584, + "learning_rate": 1.4521719072826858e-05, + "loss": 0.0263, + "step": 2851 + }, + { + "epoch": 3.703896103896104, + "grad_norm": 19.636081971868524, + "learning_rate": 1.4517967055229677e-05, + "loss": 0.0431, + "step": 2852 + }, + { + "epoch": 3.705194805194805, + "grad_norm": 5.410752222542494, + "learning_rate": 1.4514214238317648e-05, + "loss": 0.0465, + "step": 2853 + }, + { + "epoch": 3.7064935064935067, + "grad_norm": 5.0602023328823025, + "learning_rate": 1.4510460622754718e-05, + "loss": 0.0486, + "step": 2854 + }, + { + "epoch": 3.707792207792208, + "grad_norm": 9.024376580900755, + "learning_rate": 1.4506706209204972e-05, + "loss": 0.0573, + "step": 2855 + }, + { + "epoch": 3.709090909090909, + "grad_norm": 4.090971819239661, + "learning_rate": 1.4502950998332642e-05, + "loss": 0.0429, + "step": 2856 + }, + { + "epoch": 3.71038961038961, + "grad_norm": 9.07226588492291, + "learning_rate": 1.4499194990802088e-05, + "loss": 0.045, + "step": 2857 + }, + { + "epoch": 3.7116883116883117, + "grad_norm": 8.688282363999807, + "learning_rate": 1.4495438187277827e-05, + "loss": 0.0545, + "step": 2858 + }, + { + "epoch": 3.712987012987013, + "grad_norm": 16.299610784125253, + "learning_rate": 1.4491680588424504e-05, + "loss": 0.0281, + "step": 2859 + }, + { + "epoch": 3.7142857142857144, + "grad_norm": 7.297742048140453, + "learning_rate": 1.4487922194906915e-05, + "loss": 0.0299, + "step": 2860 + }, + { + "epoch": 3.7155844155844155, + "grad_norm": 20.54379991293518, + "learning_rate": 1.4484163007389991e-05, + "loss": 0.0376, + "step": 2861 + }, + { + "epoch": 3.7168831168831167, + "grad_norm": 7.166525934241753, + "learning_rate": 1.44804030265388e-05, + "loss": 0.0371, + "step": 2862 + }, + { + "epoch": 3.7181818181818183, + "grad_norm": 23.370582798379438, + "learning_rate": 1.4476642253018565e-05, + "loss": 0.0416, + "step": 2863 + }, + { + "epoch": 3.7194805194805194, + "grad_norm": 13.74713829872153, + "learning_rate": 1.4472880687494626e-05, + "loss": 0.055, + "step": 2864 + }, + { + "epoch": 3.720779220779221, + "grad_norm": 22.48443460534222, + "learning_rate": 1.4469118330632483e-05, + "loss": 0.0426, + "step": 2865 + }, + { + "epoch": 3.722077922077922, + "grad_norm": 15.147937950682566, + "learning_rate": 1.4465355183097765e-05, + "loss": 0.0358, + "step": 2866 + }, + { + "epoch": 3.7233766233766232, + "grad_norm": 21.63319805282534, + "learning_rate": 1.446159124555625e-05, + "loss": 0.0581, + "step": 2867 + }, + { + "epoch": 3.724675324675325, + "grad_norm": 18.38164617741696, + "learning_rate": 1.4457826518673847e-05, + "loss": 0.0342, + "step": 2868 + }, + { + "epoch": 3.725974025974026, + "grad_norm": 10.269509037585626, + "learning_rate": 1.4454061003116611e-05, + "loss": 0.0281, + "step": 2869 + }, + { + "epoch": 3.7272727272727275, + "grad_norm": 17.323142885997203, + "learning_rate": 1.4450294699550728e-05, + "loss": 0.0248, + "step": 2870 + }, + { + "epoch": 3.7285714285714286, + "grad_norm": 15.925468786016694, + "learning_rate": 1.4446527608642538e-05, + "loss": 0.0633, + "step": 2871 + }, + { + "epoch": 3.72987012987013, + "grad_norm": 24.11270471850016, + "learning_rate": 1.4442759731058502e-05, + "loss": 0.0362, + "step": 2872 + }, + { + "epoch": 3.731168831168831, + "grad_norm": 15.897523036183157, + "learning_rate": 1.4438991067465238e-05, + "loss": 0.0324, + "step": 2873 + }, + { + "epoch": 3.7324675324675325, + "grad_norm": 10.94009912727905, + "learning_rate": 1.4435221618529487e-05, + "loss": 0.0192, + "step": 2874 + }, + { + "epoch": 3.7337662337662336, + "grad_norm": 14.594246604054513, + "learning_rate": 1.4431451384918144e-05, + "loss": 0.0388, + "step": 2875 + }, + { + "epoch": 3.735064935064935, + "grad_norm": 7.487919755704311, + "learning_rate": 1.4427680367298234e-05, + "loss": 0.0287, + "step": 2876 + }, + { + "epoch": 3.7363636363636363, + "grad_norm": 7.746294909885998, + "learning_rate": 1.4423908566336914e-05, + "loss": 0.0321, + "step": 2877 + }, + { + "epoch": 3.7376623376623375, + "grad_norm": 14.092711414933584, + "learning_rate": 1.44201359827015e-05, + "loss": 0.0454, + "step": 2878 + }, + { + "epoch": 3.738961038961039, + "grad_norm": 7.482383540363228, + "learning_rate": 1.4416362617059424e-05, + "loss": 0.0307, + "step": 2879 + }, + { + "epoch": 3.74025974025974, + "grad_norm": 19.75031490920195, + "learning_rate": 1.4412588470078275e-05, + "loss": 0.0226, + "step": 2880 + }, + { + "epoch": 3.7415584415584417, + "grad_norm": 12.988161258347532, + "learning_rate": 1.4408813542425763e-05, + "loss": 0.0285, + "step": 2881 + }, + { + "epoch": 3.742857142857143, + "grad_norm": 14.723310726113137, + "learning_rate": 1.4405037834769753e-05, + "loss": 0.0322, + "step": 2882 + }, + { + "epoch": 3.744155844155844, + "grad_norm": 14.963834380998813, + "learning_rate": 1.4401261347778235e-05, + "loss": 0.0391, + "step": 2883 + }, + { + "epoch": 3.7454545454545456, + "grad_norm": 8.109766655092972, + "learning_rate": 1.4397484082119346e-05, + "loss": 0.0286, + "step": 2884 + }, + { + "epoch": 3.7467532467532467, + "grad_norm": 18.511760206064032, + "learning_rate": 1.4393706038461353e-05, + "loss": 0.0543, + "step": 2885 + }, + { + "epoch": 3.7480519480519483, + "grad_norm": 8.499064326719898, + "learning_rate": 1.4389927217472663e-05, + "loss": 0.039, + "step": 2886 + }, + { + "epoch": 3.7493506493506494, + "grad_norm": 13.918299401297553, + "learning_rate": 1.4386147619821827e-05, + "loss": 0.0567, + "step": 2887 + }, + { + "epoch": 3.7506493506493506, + "grad_norm": 8.879167571649035, + "learning_rate": 1.4382367246177525e-05, + "loss": 0.0662, + "step": 2888 + }, + { + "epoch": 3.7519480519480517, + "grad_norm": 30.117291285371287, + "learning_rate": 1.4378586097208577e-05, + "loss": 0.0448, + "step": 2889 + }, + { + "epoch": 3.7532467532467533, + "grad_norm": 7.323335305029314, + "learning_rate": 1.4374804173583943e-05, + "loss": 0.0361, + "step": 2890 + }, + { + "epoch": 3.7545454545454544, + "grad_norm": 23.431257555932127, + "learning_rate": 1.4371021475972715e-05, + "loss": 0.0349, + "step": 2891 + }, + { + "epoch": 3.755844155844156, + "grad_norm": 14.783452459062282, + "learning_rate": 1.4367238005044124e-05, + "loss": 0.015, + "step": 2892 + }, + { + "epoch": 3.757142857142857, + "grad_norm": 11.87410739406909, + "learning_rate": 1.4363453761467541e-05, + "loss": 0.0245, + "step": 2893 + }, + { + "epoch": 3.7584415584415583, + "grad_norm": 25.697664387524874, + "learning_rate": 1.4359668745912472e-05, + "loss": 0.0153, + "step": 2894 + }, + { + "epoch": 3.75974025974026, + "grad_norm": 5.748078241480792, + "learning_rate": 1.4355882959048553e-05, + "loss": 0.0244, + "step": 2895 + }, + { + "epoch": 3.761038961038961, + "grad_norm": 26.128722572694787, + "learning_rate": 1.435209640154557e-05, + "loss": 0.0426, + "step": 2896 + }, + { + "epoch": 3.7623376623376625, + "grad_norm": 8.414436963442611, + "learning_rate": 1.434830907407343e-05, + "loss": 0.06, + "step": 2897 + }, + { + "epoch": 3.7636363636363637, + "grad_norm": 17.010046120067916, + "learning_rate": 1.434452097730219e-05, + "loss": 0.0699, + "step": 2898 + }, + { + "epoch": 3.764935064935065, + "grad_norm": 20.501429750793385, + "learning_rate": 1.4340732111902033e-05, + "loss": 0.0373, + "step": 2899 + }, + { + "epoch": 3.7662337662337664, + "grad_norm": 4.312933446561273, + "learning_rate": 1.4336942478543283e-05, + "loss": 0.0235, + "step": 2900 + }, + { + "epoch": 3.7675324675324675, + "grad_norm": 27.562024220536497, + "learning_rate": 1.43331520778964e-05, + "loss": 0.0412, + "step": 2901 + }, + { + "epoch": 3.768831168831169, + "grad_norm": 12.391679751086329, + "learning_rate": 1.4329360910631975e-05, + "loss": 0.0711, + "step": 2902 + }, + { + "epoch": 3.77012987012987, + "grad_norm": 10.341289404162206, + "learning_rate": 1.432556897742074e-05, + "loss": 0.0599, + "step": 2903 + }, + { + "epoch": 3.7714285714285714, + "grad_norm": 30.14905008104709, + "learning_rate": 1.4321776278933564e-05, + "loss": 0.0449, + "step": 2904 + }, + { + "epoch": 3.7727272727272725, + "grad_norm": 6.0395722832015, + "learning_rate": 1.4317982815841442e-05, + "loss": 0.0534, + "step": 2905 + }, + { + "epoch": 3.774025974025974, + "grad_norm": 32.404908891830544, + "learning_rate": 1.4314188588815514e-05, + "loss": 0.0249, + "step": 2906 + }, + { + "epoch": 3.775324675324675, + "grad_norm": 17.26759641933844, + "learning_rate": 1.4310393598527049e-05, + "loss": 0.0389, + "step": 2907 + }, + { + "epoch": 3.7766233766233768, + "grad_norm": 22.364786338639465, + "learning_rate": 1.4306597845647453e-05, + "loss": 0.0245, + "step": 2908 + }, + { + "epoch": 3.777922077922078, + "grad_norm": 27.613515481281777, + "learning_rate": 1.430280133084827e-05, + "loss": 0.046, + "step": 2909 + }, + { + "epoch": 3.779220779220779, + "grad_norm": 7.377728732071294, + "learning_rate": 1.4299004054801177e-05, + "loss": 0.0344, + "step": 2910 + }, + { + "epoch": 3.7805194805194806, + "grad_norm": 30.96308977881074, + "learning_rate": 1.429520601817798e-05, + "loss": 0.0504, + "step": 2911 + }, + { + "epoch": 3.7818181818181817, + "grad_norm": 8.2464575429027, + "learning_rate": 1.4291407221650632e-05, + "loss": 0.0337, + "step": 2912 + }, + { + "epoch": 3.7831168831168833, + "grad_norm": 10.288192979469128, + "learning_rate": 1.4287607665891204e-05, + "loss": 0.0534, + "step": 2913 + }, + { + "epoch": 3.7844155844155845, + "grad_norm": 29.831104529536372, + "learning_rate": 1.4283807351571914e-05, + "loss": 0.0756, + "step": 2914 + }, + { + "epoch": 3.7857142857142856, + "grad_norm": 16.092349264161392, + "learning_rate": 1.4280006279365114e-05, + "loss": 0.0433, + "step": 2915 + }, + { + "epoch": 3.787012987012987, + "grad_norm": 26.187191196855473, + "learning_rate": 1.427620444994328e-05, + "loss": 0.0554, + "step": 2916 + }, + { + "epoch": 3.7883116883116883, + "grad_norm": 22.752234459612847, + "learning_rate": 1.4272401863979032e-05, + "loss": 0.0481, + "step": 2917 + }, + { + "epoch": 3.78961038961039, + "grad_norm": 8.323620513713593, + "learning_rate": 1.4268598522145121e-05, + "loss": 0.027, + "step": 2918 + }, + { + "epoch": 3.790909090909091, + "grad_norm": 21.900006196605215, + "learning_rate": 1.4264794425114428e-05, + "loss": 0.034, + "step": 2919 + }, + { + "epoch": 3.792207792207792, + "grad_norm": 7.285917306910062, + "learning_rate": 1.4260989573559974e-05, + "loss": 0.0518, + "step": 2920 + }, + { + "epoch": 3.7935064935064933, + "grad_norm": 7.547016461254018, + "learning_rate": 1.4257183968154906e-05, + "loss": 0.0798, + "step": 2921 + }, + { + "epoch": 3.794805194805195, + "grad_norm": 18.21751837948121, + "learning_rate": 1.4253377609572509e-05, + "loss": 0.0309, + "step": 2922 + }, + { + "epoch": 3.796103896103896, + "grad_norm": 11.782906726275355, + "learning_rate": 1.4249570498486207e-05, + "loss": 0.0226, + "step": 2923 + }, + { + "epoch": 3.7974025974025976, + "grad_norm": 13.230687466044719, + "learning_rate": 1.4245762635569542e-05, + "loss": 0.0394, + "step": 2924 + }, + { + "epoch": 3.7987012987012987, + "grad_norm": 21.562278445793492, + "learning_rate": 1.4241954021496205e-05, + "loss": 0.0653, + "step": 2925 + }, + { + "epoch": 3.8, + "grad_norm": 2.9394355287344567, + "learning_rate": 1.423814465694001e-05, + "loss": 0.0195, + "step": 2926 + }, + { + "epoch": 3.8012987012987014, + "grad_norm": 29.807591670760733, + "learning_rate": 1.4234334542574906e-05, + "loss": 0.0406, + "step": 2927 + }, + { + "epoch": 3.8025974025974025, + "grad_norm": 4.229541869107677, + "learning_rate": 1.4230523679074972e-05, + "loss": 0.0279, + "step": 2928 + }, + { + "epoch": 3.803896103896104, + "grad_norm": 17.89766796476549, + "learning_rate": 1.4226712067114427e-05, + "loss": 0.0413, + "step": 2929 + }, + { + "epoch": 3.8051948051948052, + "grad_norm": 13.639494134036887, + "learning_rate": 1.4222899707367619e-05, + "loss": 0.0314, + "step": 2930 + }, + { + "epoch": 3.8064935064935064, + "grad_norm": 10.076227237890159, + "learning_rate": 1.4219086600509025e-05, + "loss": 0.0248, + "step": 2931 + }, + { + "epoch": 3.807792207792208, + "grad_norm": 18.62879346067206, + "learning_rate": 1.4215272747213258e-05, + "loss": 0.0235, + "step": 2932 + }, + { + "epoch": 3.809090909090909, + "grad_norm": 5.876585090473158, + "learning_rate": 1.4211458148155059e-05, + "loss": 0.037, + "step": 2933 + }, + { + "epoch": 3.8103896103896107, + "grad_norm": 14.833933021531305, + "learning_rate": 1.4207642804009307e-05, + "loss": 0.0317, + "step": 2934 + }, + { + "epoch": 3.811688311688312, + "grad_norm": 18.49601612255392, + "learning_rate": 1.4203826715451008e-05, + "loss": 0.0362, + "step": 2935 + }, + { + "epoch": 3.812987012987013, + "grad_norm": 11.933247350601704, + "learning_rate": 1.4200009883155302e-05, + "loss": 0.0547, + "step": 2936 + }, + { + "epoch": 3.814285714285714, + "grad_norm": 20.646744203540983, + "learning_rate": 1.4196192307797456e-05, + "loss": 0.0626, + "step": 2937 + }, + { + "epoch": 3.8155844155844156, + "grad_norm": 7.5545150452810805, + "learning_rate": 1.4192373990052877e-05, + "loss": 0.0544, + "step": 2938 + }, + { + "epoch": 3.8168831168831168, + "grad_norm": 28.544171180420882, + "learning_rate": 1.4188554930597094e-05, + "loss": 0.0547, + "step": 2939 + }, + { + "epoch": 3.8181818181818183, + "grad_norm": 7.7854081914228255, + "learning_rate": 1.4184735130105776e-05, + "loss": 0.015, + "step": 2940 + }, + { + "epoch": 3.8194805194805195, + "grad_norm": 20.26042811199473, + "learning_rate": 1.4180914589254716e-05, + "loss": 0.057, + "step": 2941 + }, + { + "epoch": 3.8207792207792206, + "grad_norm": 18.752416707783407, + "learning_rate": 1.417709330871984e-05, + "loss": 0.031, + "step": 2942 + }, + { + "epoch": 3.822077922077922, + "grad_norm": 11.894230054090734, + "learning_rate": 1.4173271289177208e-05, + "loss": 0.0456, + "step": 2943 + }, + { + "epoch": 3.8233766233766233, + "grad_norm": 13.534191927205173, + "learning_rate": 1.416944853130301e-05, + "loss": 0.0586, + "step": 2944 + }, + { + "epoch": 3.824675324675325, + "grad_norm": 4.408630490270197, + "learning_rate": 1.4165625035773563e-05, + "loss": 0.0303, + "step": 2945 + }, + { + "epoch": 3.825974025974026, + "grad_norm": 6.609331745260988, + "learning_rate": 1.416180080326531e-05, + "loss": 0.0264, + "step": 2946 + }, + { + "epoch": 3.827272727272727, + "grad_norm": 15.573414709436792, + "learning_rate": 1.4157975834454844e-05, + "loss": 0.0257, + "step": 2947 + }, + { + "epoch": 3.8285714285714287, + "grad_norm": 5.58476975664751, + "learning_rate": 1.4154150130018867e-05, + "loss": 0.0193, + "step": 2948 + }, + { + "epoch": 3.82987012987013, + "grad_norm": 10.417972118627803, + "learning_rate": 1.415032369063422e-05, + "loss": 0.0312, + "step": 2949 + }, + { + "epoch": 3.8311688311688314, + "grad_norm": 8.43058977020254, + "learning_rate": 1.4146496516977872e-05, + "loss": 0.0601, + "step": 2950 + }, + { + "epoch": 3.8324675324675326, + "grad_norm": 7.2699908537995865, + "learning_rate": 1.4142668609726928e-05, + "loss": 0.0317, + "step": 2951 + }, + { + "epoch": 3.8337662337662337, + "grad_norm": 6.334810508976759, + "learning_rate": 1.4138839969558614e-05, + "loss": 0.0704, + "step": 2952 + }, + { + "epoch": 3.835064935064935, + "grad_norm": 5.566656038941535, + "learning_rate": 1.4135010597150291e-05, + "loss": 0.0529, + "step": 2953 + }, + { + "epoch": 3.8363636363636364, + "grad_norm": 8.699830198130527, + "learning_rate": 1.4131180493179448e-05, + "loss": 0.0425, + "step": 2954 + }, + { + "epoch": 3.8376623376623376, + "grad_norm": 13.766715988068642, + "learning_rate": 1.4127349658323709e-05, + "loss": 0.03, + "step": 2955 + }, + { + "epoch": 3.838961038961039, + "grad_norm": 22.32512752657778, + "learning_rate": 1.4123518093260811e-05, + "loss": 0.0561, + "step": 2956 + }, + { + "epoch": 3.8402597402597403, + "grad_norm": 6.159306737125377, + "learning_rate": 1.411968579866864e-05, + "loss": 0.0313, + "step": 2957 + }, + { + "epoch": 3.8415584415584414, + "grad_norm": 11.791217927663919, + "learning_rate": 1.41158527752252e-05, + "loss": 0.0684, + "step": 2958 + }, + { + "epoch": 3.842857142857143, + "grad_norm": 14.495106148781394, + "learning_rate": 1.4112019023608626e-05, + "loss": 0.0499, + "step": 2959 + }, + { + "epoch": 3.844155844155844, + "grad_norm": 10.148723830790262, + "learning_rate": 1.410818454449718e-05, + "loss": 0.0423, + "step": 2960 + }, + { + "epoch": 3.8454545454545457, + "grad_norm": 10.119773732731264, + "learning_rate": 1.4104349338569256e-05, + "loss": 0.0202, + "step": 2961 + }, + { + "epoch": 3.846753246753247, + "grad_norm": 8.325183830116693, + "learning_rate": 1.4100513406503377e-05, + "loss": 0.0239, + "step": 2962 + }, + { + "epoch": 3.848051948051948, + "grad_norm": 11.048321059016622, + "learning_rate": 1.4096676748978193e-05, + "loss": 0.0576, + "step": 2963 + }, + { + "epoch": 3.849350649350649, + "grad_norm": 25.153051730514047, + "learning_rate": 1.4092839366672479e-05, + "loss": 0.0576, + "step": 2964 + }, + { + "epoch": 3.8506493506493507, + "grad_norm": 6.4245684974259, + "learning_rate": 1.4089001260265143e-05, + "loss": 0.0428, + "step": 2965 + }, + { + "epoch": 3.851948051948052, + "grad_norm": 17.32426692246966, + "learning_rate": 1.4085162430435225e-05, + "loss": 0.0259, + "step": 2966 + }, + { + "epoch": 3.8532467532467534, + "grad_norm": 18.841889468802886, + "learning_rate": 1.4081322877861876e-05, + "loss": 0.023, + "step": 2967 + }, + { + "epoch": 3.8545454545454545, + "grad_norm": 5.855935287645721, + "learning_rate": 1.4077482603224394e-05, + "loss": 0.0235, + "step": 2968 + }, + { + "epoch": 3.8558441558441556, + "grad_norm": 24.543354201336605, + "learning_rate": 1.4073641607202198e-05, + "loss": 0.0429, + "step": 2969 + }, + { + "epoch": 3.857142857142857, + "grad_norm": 6.735786256998592, + "learning_rate": 1.4069799890474829e-05, + "loss": 0.05, + "step": 2970 + }, + { + "epoch": 3.8584415584415583, + "grad_norm": 10.690237253572432, + "learning_rate": 1.4065957453721963e-05, + "loss": 0.0281, + "step": 2971 + }, + { + "epoch": 3.85974025974026, + "grad_norm": 23.192728238689387, + "learning_rate": 1.4062114297623402e-05, + "loss": 0.0234, + "step": 2972 + }, + { + "epoch": 3.861038961038961, + "grad_norm": 6.771551724379689, + "learning_rate": 1.4058270422859071e-05, + "loss": 0.0505, + "step": 2973 + }, + { + "epoch": 3.862337662337662, + "grad_norm": 21.883222848534736, + "learning_rate": 1.4054425830109025e-05, + "loss": 0.0321, + "step": 2974 + }, + { + "epoch": 3.8636363636363638, + "grad_norm": 16.004314554410712, + "learning_rate": 1.4050580520053448e-05, + "loss": 0.0351, + "step": 2975 + }, + { + "epoch": 3.864935064935065, + "grad_norm": 16.304295852620044, + "learning_rate": 1.4046734493372646e-05, + "loss": 0.0187, + "step": 2976 + }, + { + "epoch": 3.8662337662337665, + "grad_norm": 22.45719674168719, + "learning_rate": 1.404288775074706e-05, + "loss": 0.0395, + "step": 2977 + }, + { + "epoch": 3.8675324675324676, + "grad_norm": 6.286097322022884, + "learning_rate": 1.4039040292857247e-05, + "loss": 0.0293, + "step": 2978 + }, + { + "epoch": 3.8688311688311687, + "grad_norm": 22.974597361395052, + "learning_rate": 1.4035192120383901e-05, + "loss": 0.0312, + "step": 2979 + }, + { + "epoch": 3.87012987012987, + "grad_norm": 12.4231331945388, + "learning_rate": 1.403134323400783e-05, + "loss": 0.0372, + "step": 2980 + }, + { + "epoch": 3.8714285714285714, + "grad_norm": 19.29886760468947, + "learning_rate": 1.4027493634409983e-05, + "loss": 0.0585, + "step": 2981 + }, + { + "epoch": 3.8727272727272726, + "grad_norm": 24.364393071259435, + "learning_rate": 1.4023643322271425e-05, + "loss": 0.0406, + "step": 2982 + }, + { + "epoch": 3.874025974025974, + "grad_norm": 8.040019330865062, + "learning_rate": 1.401979229827335e-05, + "loss": 0.0289, + "step": 2983 + }, + { + "epoch": 3.8753246753246753, + "grad_norm": 28.141367846787794, + "learning_rate": 1.4015940563097078e-05, + "loss": 0.0409, + "step": 2984 + }, + { + "epoch": 3.8766233766233764, + "grad_norm": 10.289471333330757, + "learning_rate": 1.4012088117424054e-05, + "loss": 0.026, + "step": 2985 + }, + { + "epoch": 3.877922077922078, + "grad_norm": 16.280755583966602, + "learning_rate": 1.4008234961935849e-05, + "loss": 0.0538, + "step": 2986 + }, + { + "epoch": 3.879220779220779, + "grad_norm": 23.642686486611883, + "learning_rate": 1.400438109731416e-05, + "loss": 0.0334, + "step": 2987 + }, + { + "epoch": 3.8805194805194807, + "grad_norm": 5.39992533882519, + "learning_rate": 1.4000526524240814e-05, + "loss": 0.0271, + "step": 2988 + }, + { + "epoch": 3.881818181818182, + "grad_norm": 23.15372484865068, + "learning_rate": 1.3996671243397753e-05, + "loss": 0.0548, + "step": 2989 + }, + { + "epoch": 3.883116883116883, + "grad_norm": 17.333178527464696, + "learning_rate": 1.3992815255467051e-05, + "loss": 0.0433, + "step": 2990 + }, + { + "epoch": 3.8844155844155845, + "grad_norm": 16.39781490105334, + "learning_rate": 1.3988958561130907e-05, + "loss": 0.0509, + "step": 2991 + }, + { + "epoch": 3.8857142857142857, + "grad_norm": 20.267804022148223, + "learning_rate": 1.3985101161071645e-05, + "loss": 0.0354, + "step": 2992 + }, + { + "epoch": 3.8870129870129873, + "grad_norm": 5.593355820699934, + "learning_rate": 1.3981243055971712e-05, + "loss": 0.044, + "step": 2993 + }, + { + "epoch": 3.8883116883116884, + "grad_norm": 23.52162280017055, + "learning_rate": 1.3977384246513681e-05, + "loss": 0.0574, + "step": 2994 + }, + { + "epoch": 3.8896103896103895, + "grad_norm": 11.342777539033229, + "learning_rate": 1.3973524733380247e-05, + "loss": 0.0769, + "step": 2995 + }, + { + "epoch": 3.8909090909090907, + "grad_norm": 14.409157477120884, + "learning_rate": 1.3969664517254233e-05, + "loss": 0.0443, + "step": 2996 + }, + { + "epoch": 3.8922077922077922, + "grad_norm": 6.232576827090496, + "learning_rate": 1.3965803598818588e-05, + "loss": 0.0342, + "step": 2997 + }, + { + "epoch": 3.8935064935064934, + "grad_norm": 9.957602918841122, + "learning_rate": 1.3961941978756379e-05, + "loss": 0.0421, + "step": 2998 + }, + { + "epoch": 3.894805194805195, + "grad_norm": 13.670770921637148, + "learning_rate": 1.3958079657750803e-05, + "loss": 0.0403, + "step": 2999 + }, + { + "epoch": 3.896103896103896, + "grad_norm": 4.838200272610608, + "learning_rate": 1.3954216636485176e-05, + "loss": 0.0442, + "step": 3000 + }, + { + "epoch": 3.897402597402597, + "grad_norm": 19.560495735179124, + "learning_rate": 1.3950352915642941e-05, + "loss": 0.0657, + "step": 3001 + }, + { + "epoch": 3.898701298701299, + "grad_norm": 19.34474295560659, + "learning_rate": 1.3946488495907663e-05, + "loss": 0.0344, + "step": 3002 + }, + { + "epoch": 3.9, + "grad_norm": 17.52395271942666, + "learning_rate": 1.3942623377963039e-05, + "loss": 0.0255, + "step": 3003 + }, + { + "epoch": 3.9012987012987015, + "grad_norm": 4.557545641449821, + "learning_rate": 1.3938757562492873e-05, + "loss": 0.0313, + "step": 3004 + }, + { + "epoch": 3.9025974025974026, + "grad_norm": 22.095699139257647, + "learning_rate": 1.3934891050181108e-05, + "loss": 0.0566, + "step": 3005 + }, + { + "epoch": 3.9038961038961038, + "grad_norm": 13.303127151763494, + "learning_rate": 1.3931023841711797e-05, + "loss": 0.0344, + "step": 3006 + }, + { + "epoch": 3.9051948051948053, + "grad_norm": 10.035347167616694, + "learning_rate": 1.3927155937769132e-05, + "loss": 0.0262, + "step": 3007 + }, + { + "epoch": 3.9064935064935065, + "grad_norm": 17.370701003965127, + "learning_rate": 1.3923287339037413e-05, + "loss": 0.041, + "step": 3008 + }, + { + "epoch": 3.907792207792208, + "grad_norm": 3.612160176131464, + "learning_rate": 1.3919418046201072e-05, + "loss": 0.0134, + "step": 3009 + }, + { + "epoch": 3.909090909090909, + "grad_norm": 7.058928240283739, + "learning_rate": 1.391554805994466e-05, + "loss": 0.037, + "step": 3010 + }, + { + "epoch": 3.9103896103896103, + "grad_norm": 11.728713121885788, + "learning_rate": 1.3911677380952853e-05, + "loss": 0.0299, + "step": 3011 + }, + { + "epoch": 3.9116883116883114, + "grad_norm": 4.202156657416837, + "learning_rate": 1.3907806009910445e-05, + "loss": 0.0202, + "step": 3012 + }, + { + "epoch": 3.912987012987013, + "grad_norm": 7.350269877529516, + "learning_rate": 1.3903933947502357e-05, + "loss": 0.0503, + "step": 3013 + }, + { + "epoch": 3.914285714285714, + "grad_norm": 13.784440910989083, + "learning_rate": 1.3900061194413637e-05, + "loss": 0.0569, + "step": 3014 + }, + { + "epoch": 3.9155844155844157, + "grad_norm": 3.941981040186317, + "learning_rate": 1.3896187751329438e-05, + "loss": 0.0459, + "step": 3015 + }, + { + "epoch": 3.916883116883117, + "grad_norm": 22.710703320497984, + "learning_rate": 1.3892313618935053e-05, + "loss": 0.0288, + "step": 3016 + }, + { + "epoch": 3.918181818181818, + "grad_norm": 7.2716003308519905, + "learning_rate": 1.3888438797915892e-05, + "loss": 0.0359, + "step": 3017 + }, + { + "epoch": 3.9194805194805196, + "grad_norm": 14.638330427525336, + "learning_rate": 1.3884563288957478e-05, + "loss": 0.0439, + "step": 3018 + }, + { + "epoch": 3.9207792207792207, + "grad_norm": 23.878851494165662, + "learning_rate": 1.3880687092745468e-05, + "loss": 0.0336, + "step": 3019 + }, + { + "epoch": 3.9220779220779223, + "grad_norm": 4.892176586722383, + "learning_rate": 1.3876810209965637e-05, + "loss": 0.0245, + "step": 3020 + }, + { + "epoch": 3.9233766233766234, + "grad_norm": 27.578727466192724, + "learning_rate": 1.3872932641303874e-05, + "loss": 0.0253, + "step": 3021 + }, + { + "epoch": 3.9246753246753245, + "grad_norm": 9.196930810383497, + "learning_rate": 1.3869054387446199e-05, + "loss": 0.0389, + "step": 3022 + }, + { + "epoch": 3.925974025974026, + "grad_norm": 9.52156115176582, + "learning_rate": 1.3865175449078748e-05, + "loss": 0.0347, + "step": 3023 + }, + { + "epoch": 3.9272727272727272, + "grad_norm": 14.960350678614358, + "learning_rate": 1.3861295826887782e-05, + "loss": 0.0303, + "step": 3024 + }, + { + "epoch": 3.928571428571429, + "grad_norm": 5.0013309062050135, + "learning_rate": 1.3857415521559677e-05, + "loss": 0.0493, + "step": 3025 + }, + { + "epoch": 3.92987012987013, + "grad_norm": 13.850402404759281, + "learning_rate": 1.3853534533780932e-05, + "loss": 0.0251, + "step": 3026 + }, + { + "epoch": 3.931168831168831, + "grad_norm": 14.861023602498252, + "learning_rate": 1.384965286423817e-05, + "loss": 0.0288, + "step": 3027 + }, + { + "epoch": 3.9324675324675322, + "grad_norm": 13.300608983415845, + "learning_rate": 1.3845770513618138e-05, + "loss": 0.0409, + "step": 3028 + }, + { + "epoch": 3.933766233766234, + "grad_norm": 14.249110447234923, + "learning_rate": 1.3841887482607689e-05, + "loss": 0.041, + "step": 3029 + }, + { + "epoch": 3.935064935064935, + "grad_norm": 7.954997409574959, + "learning_rate": 1.3838003771893813e-05, + "loss": 0.0338, + "step": 3030 + }, + { + "epoch": 3.9363636363636365, + "grad_norm": 14.347684231666847, + "learning_rate": 1.3834119382163609e-05, + "loss": 0.0326, + "step": 3031 + }, + { + "epoch": 3.9376623376623376, + "grad_norm": 26.591143721859012, + "learning_rate": 1.3830234314104298e-05, + "loss": 0.0528, + "step": 3032 + }, + { + "epoch": 3.9389610389610388, + "grad_norm": 16.06924099818245, + "learning_rate": 1.3826348568403228e-05, + "loss": 0.0302, + "step": 3033 + }, + { + "epoch": 3.9402597402597404, + "grad_norm": 14.936134403794197, + "learning_rate": 1.3822462145747858e-05, + "loss": 0.0368, + "step": 3034 + }, + { + "epoch": 3.9415584415584415, + "grad_norm": 11.901477767668837, + "learning_rate": 1.3818575046825774e-05, + "loss": 0.0364, + "step": 3035 + }, + { + "epoch": 3.942857142857143, + "grad_norm": 9.930198707867367, + "learning_rate": 1.3814687272324672e-05, + "loss": 0.0272, + "step": 3036 + }, + { + "epoch": 3.944155844155844, + "grad_norm": 17.08377796670919, + "learning_rate": 1.381079882293238e-05, + "loss": 0.0463, + "step": 3037 + }, + { + "epoch": 3.9454545454545453, + "grad_norm": 3.8341549238599666, + "learning_rate": 1.3806909699336837e-05, + "loss": 0.0202, + "step": 3038 + }, + { + "epoch": 3.946753246753247, + "grad_norm": 16.561544257921323, + "learning_rate": 1.38030199022261e-05, + "loss": 0.0585, + "step": 3039 + }, + { + "epoch": 3.948051948051948, + "grad_norm": 4.0376094291125435, + "learning_rate": 1.3799129432288356e-05, + "loss": 0.0135, + "step": 3040 + }, + { + "epoch": 3.9493506493506496, + "grad_norm": 6.136844514150804, + "learning_rate": 1.3795238290211896e-05, + "loss": 0.0484, + "step": 3041 + }, + { + "epoch": 3.9506493506493507, + "grad_norm": 15.132430329297993, + "learning_rate": 1.3791346476685144e-05, + "loss": 0.0674, + "step": 3042 + }, + { + "epoch": 3.951948051948052, + "grad_norm": 11.019076573795083, + "learning_rate": 1.3787453992396631e-05, + "loss": 0.0799, + "step": 3043 + }, + { + "epoch": 3.953246753246753, + "grad_norm": 3.3622525132882926, + "learning_rate": 1.3783560838035015e-05, + "loss": 0.0421, + "step": 3044 + }, + { + "epoch": 3.9545454545454546, + "grad_norm": 6.280759599348839, + "learning_rate": 1.3779667014289067e-05, + "loss": 0.0529, + "step": 3045 + }, + { + "epoch": 3.9558441558441557, + "grad_norm": 9.669062518136611, + "learning_rate": 1.3775772521847683e-05, + "loss": 0.0557, + "step": 3046 + }, + { + "epoch": 3.9571428571428573, + "grad_norm": 19.50182622765462, + "learning_rate": 1.3771877361399868e-05, + "loss": 0.0442, + "step": 3047 + }, + { + "epoch": 3.9584415584415584, + "grad_norm": 14.726845817618154, + "learning_rate": 1.3767981533634755e-05, + "loss": 0.0402, + "step": 3048 + }, + { + "epoch": 3.9597402597402596, + "grad_norm": 11.440811690696686, + "learning_rate": 1.3764085039241588e-05, + "loss": 0.039, + "step": 3049 + }, + { + "epoch": 3.961038961038961, + "grad_norm": 15.161649873948166, + "learning_rate": 1.3760187878909735e-05, + "loss": 0.0363, + "step": 3050 + }, + { + "epoch": 3.9623376623376623, + "grad_norm": 10.72817726248374, + "learning_rate": 1.3756290053328671e-05, + "loss": 0.0451, + "step": 3051 + }, + { + "epoch": 3.963636363636364, + "grad_norm": 9.51036866828214, + "learning_rate": 1.3752391563188003e-05, + "loss": 0.0214, + "step": 3052 + }, + { + "epoch": 3.964935064935065, + "grad_norm": 15.460956160496698, + "learning_rate": 1.3748492409177448e-05, + "loss": 0.0525, + "step": 3053 + }, + { + "epoch": 3.966233766233766, + "grad_norm": 5.304524528402238, + "learning_rate": 1.3744592591986834e-05, + "loss": 0.0329, + "step": 3054 + }, + { + "epoch": 3.9675324675324677, + "grad_norm": 15.768565040954481, + "learning_rate": 1.3740692112306122e-05, + "loss": 0.0372, + "step": 3055 + }, + { + "epoch": 3.968831168831169, + "grad_norm": 7.705381904416042, + "learning_rate": 1.3736790970825376e-05, + "loss": 0.0344, + "step": 3056 + }, + { + "epoch": 3.9701298701298704, + "grad_norm": 4.764156326853063, + "learning_rate": 1.3732889168234787e-05, + "loss": 0.0553, + "step": 3057 + }, + { + "epoch": 3.9714285714285715, + "grad_norm": 5.877476448988908, + "learning_rate": 1.3728986705224651e-05, + "loss": 0.024, + "step": 3058 + }, + { + "epoch": 3.9727272727272727, + "grad_norm": 6.759154500603701, + "learning_rate": 1.3725083582485397e-05, + "loss": 0.0334, + "step": 3059 + }, + { + "epoch": 3.974025974025974, + "grad_norm": 13.145077865442108, + "learning_rate": 1.3721179800707556e-05, + "loss": 0.0387, + "step": 3060 + }, + { + "epoch": 3.9753246753246754, + "grad_norm": 3.6231874766869536, + "learning_rate": 1.3717275360581788e-05, + "loss": 0.0137, + "step": 3061 + }, + { + "epoch": 3.9766233766233765, + "grad_norm": 6.2311173288276045, + "learning_rate": 1.3713370262798856e-05, + "loss": 0.0443, + "step": 3062 + }, + { + "epoch": 3.977922077922078, + "grad_norm": 11.357606038461798, + "learning_rate": 1.3709464508049648e-05, + "loss": 0.0348, + "step": 3063 + }, + { + "epoch": 3.979220779220779, + "grad_norm": 6.866147201769705, + "learning_rate": 1.3705558097025173e-05, + "loss": 0.0491, + "step": 3064 + }, + { + "epoch": 3.9805194805194803, + "grad_norm": 14.393411110204228, + "learning_rate": 1.3701651030416542e-05, + "loss": 0.0449, + "step": 3065 + }, + { + "epoch": 3.981818181818182, + "grad_norm": 6.157856374851522, + "learning_rate": 1.3697743308914996e-05, + "loss": 0.0494, + "step": 3066 + }, + { + "epoch": 3.983116883116883, + "grad_norm": 6.348445779110279, + "learning_rate": 1.369383493321188e-05, + "loss": 0.0479, + "step": 3067 + }, + { + "epoch": 3.9844155844155846, + "grad_norm": 6.22295735737115, + "learning_rate": 1.3689925903998668e-05, + "loss": 0.0402, + "step": 3068 + }, + { + "epoch": 3.9857142857142858, + "grad_norm": 6.125280186215149, + "learning_rate": 1.3686016221966933e-05, + "loss": 0.0233, + "step": 3069 + }, + { + "epoch": 3.987012987012987, + "grad_norm": 8.608520342359547, + "learning_rate": 1.368210588780838e-05, + "loss": 0.0325, + "step": 3070 + }, + { + "epoch": 3.9883116883116885, + "grad_norm": 8.82673069650894, + "learning_rate": 1.367819490221482e-05, + "loss": 0.0258, + "step": 3071 + }, + { + "epoch": 3.9896103896103896, + "grad_norm": 6.396052662557485, + "learning_rate": 1.3674283265878181e-05, + "loss": 0.0346, + "step": 3072 + }, + { + "epoch": 3.990909090909091, + "grad_norm": 11.233285384360014, + "learning_rate": 1.3670370979490501e-05, + "loss": 0.0401, + "step": 3073 + }, + { + "epoch": 3.9922077922077923, + "grad_norm": 9.602035660876446, + "learning_rate": 1.3666458043743948e-05, + "loss": 0.0314, + "step": 3074 + }, + { + "epoch": 3.9935064935064934, + "grad_norm": 9.598853796498455, + "learning_rate": 1.3662544459330787e-05, + "loss": 0.0295, + "step": 3075 + }, + { + "epoch": 3.9948051948051946, + "grad_norm": 19.54213318198762, + "learning_rate": 1.3658630226943412e-05, + "loss": 0.0121, + "step": 3076 + }, + { + "epoch": 3.996103896103896, + "grad_norm": 13.010776417165603, + "learning_rate": 1.365471534727432e-05, + "loss": 0.0504, + "step": 3077 + }, + { + "epoch": 3.9974025974025973, + "grad_norm": 13.756724888231108, + "learning_rate": 1.3650799821016134e-05, + "loss": 0.0258, + "step": 3078 + }, + { + "epoch": 3.998701298701299, + "grad_norm": 24.36084655659811, + "learning_rate": 1.3646883648861581e-05, + "loss": 0.0425, + "step": 3079 + }, + { + "epoch": 4.0, + "grad_norm": 12.70406144274415, + "learning_rate": 1.3642966831503508e-05, + "loss": 0.0458, + "step": 3080 + }, + { + "epoch": 4.0, + "eval_accuracy": 0.9208309938236946, + "eval_f1": 0.9024250574855844, + "eval_loss": 0.1669396013021469, + "eval_precision": 0.897222697057996, + "eval_recall": 0.9196895103666123, + "eval_runtime": 13.2149, + "eval_samples_per_second": 134.773, + "eval_steps_per_second": 1.059, + "step": 3080 + }, + { + "epoch": 4.001298701298701, + "grad_norm": 14.934543664131173, + "learning_rate": 1.3639049369634878e-05, + "loss": 0.0138, + "step": 3081 + }, + { + "epoch": 4.002597402597402, + "grad_norm": 5.876802900616323, + "learning_rate": 1.3635131263948757e-05, + "loss": 0.0141, + "step": 3082 + }, + { + "epoch": 4.003896103896104, + "grad_norm": 11.156304888267634, + "learning_rate": 1.3631212515138344e-05, + "loss": 0.0532, + "step": 3083 + }, + { + "epoch": 4.005194805194805, + "grad_norm": 16.63466358946404, + "learning_rate": 1.3627293123896932e-05, + "loss": 0.017, + "step": 3084 + }, + { + "epoch": 4.0064935064935066, + "grad_norm": 1.8929580401457629, + "learning_rate": 1.3623373090917942e-05, + "loss": 0.0127, + "step": 3085 + }, + { + "epoch": 4.007792207792208, + "grad_norm": 12.866858645068492, + "learning_rate": 1.3619452416894896e-05, + "loss": 0.0129, + "step": 3086 + }, + { + "epoch": 4.009090909090909, + "grad_norm": 13.246395512830839, + "learning_rate": 1.3615531102521443e-05, + "loss": 0.0222, + "step": 3087 + }, + { + "epoch": 4.01038961038961, + "grad_norm": 25.204627004664506, + "learning_rate": 1.3611609148491333e-05, + "loss": 0.0502, + "step": 3088 + }, + { + "epoch": 4.011688311688312, + "grad_norm": 5.093571828346039, + "learning_rate": 1.3607686555498442e-05, + "loss": 0.0139, + "step": 3089 + }, + { + "epoch": 4.012987012987013, + "grad_norm": 15.612145078654317, + "learning_rate": 1.3603763324236745e-05, + "loss": 0.0267, + "step": 3090 + }, + { + "epoch": 4.014285714285714, + "grad_norm": 4.003182568142123, + "learning_rate": 1.359983945540034e-05, + "loss": 0.0217, + "step": 3091 + }, + { + "epoch": 4.015584415584415, + "grad_norm": 14.775834524004592, + "learning_rate": 1.3595914949683433e-05, + "loss": 0.0449, + "step": 3092 + }, + { + "epoch": 4.0168831168831165, + "grad_norm": 5.748386875817497, + "learning_rate": 1.3591989807780346e-05, + "loss": 0.0383, + "step": 3093 + }, + { + "epoch": 4.0181818181818185, + "grad_norm": 9.246213079123928, + "learning_rate": 1.3588064030385509e-05, + "loss": 0.0146, + "step": 3094 + }, + { + "epoch": 4.01948051948052, + "grad_norm": 10.04401917732273, + "learning_rate": 1.358413761819347e-05, + "loss": 0.0207, + "step": 3095 + }, + { + "epoch": 4.020779220779221, + "grad_norm": 6.272153613713951, + "learning_rate": 1.3580210571898884e-05, + "loss": 0.0116, + "step": 3096 + }, + { + "epoch": 4.022077922077922, + "grad_norm": 8.141601224937553, + "learning_rate": 1.357628289219652e-05, + "loss": 0.0348, + "step": 3097 + }, + { + "epoch": 4.023376623376623, + "grad_norm": 4.502631521192026, + "learning_rate": 1.3572354579781265e-05, + "loss": 0.0234, + "step": 3098 + }, + { + "epoch": 4.024675324675325, + "grad_norm": 7.774497990561703, + "learning_rate": 1.3568425635348108e-05, + "loss": 0.0158, + "step": 3099 + }, + { + "epoch": 4.025974025974026, + "grad_norm": 10.059802107291596, + "learning_rate": 1.3564496059592159e-05, + "loss": 0.0382, + "step": 3100 + }, + { + "epoch": 4.027272727272727, + "grad_norm": 13.092133169791225, + "learning_rate": 1.3560565853208627e-05, + "loss": 0.0261, + "step": 3101 + }, + { + "epoch": 4.0285714285714285, + "grad_norm": 17.838621249307195, + "learning_rate": 1.3556635016892848e-05, + "loss": 0.0183, + "step": 3102 + }, + { + "epoch": 4.02987012987013, + "grad_norm": 9.89043996604508, + "learning_rate": 1.3552703551340259e-05, + "loss": 0.0299, + "step": 3103 + }, + { + "epoch": 4.031168831168831, + "grad_norm": 21.12708445678405, + "learning_rate": 1.3548771457246412e-05, + "loss": 0.0163, + "step": 3104 + }, + { + "epoch": 4.032467532467533, + "grad_norm": 4.630904842897571, + "learning_rate": 1.354483873530697e-05, + "loss": 0.0312, + "step": 3105 + }, + { + "epoch": 4.033766233766234, + "grad_norm": 11.49448875713658, + "learning_rate": 1.3540905386217708e-05, + "loss": 0.016, + "step": 3106 + }, + { + "epoch": 4.035064935064935, + "grad_norm": 16.64013645231536, + "learning_rate": 1.353697141067451e-05, + "loss": 0.017, + "step": 3107 + }, + { + "epoch": 4.036363636363636, + "grad_norm": 12.306039515991936, + "learning_rate": 1.353303680937337e-05, + "loss": 0.0489, + "step": 3108 + }, + { + "epoch": 4.037662337662337, + "grad_norm": 11.654425448159312, + "learning_rate": 1.3529101583010396e-05, + "loss": 0.0317, + "step": 3109 + }, + { + "epoch": 4.038961038961039, + "grad_norm": 11.171821461476393, + "learning_rate": 1.3525165732281802e-05, + "loss": 0.0327, + "step": 3110 + }, + { + "epoch": 4.04025974025974, + "grad_norm": 15.319842322685181, + "learning_rate": 1.352122925788392e-05, + "loss": 0.0565, + "step": 3111 + }, + { + "epoch": 4.041558441558442, + "grad_norm": 19.739280222182547, + "learning_rate": 1.3517292160513184e-05, + "loss": 0.0309, + "step": 3112 + }, + { + "epoch": 4.042857142857143, + "grad_norm": 7.495204393122061, + "learning_rate": 1.3513354440866144e-05, + "loss": 0.0372, + "step": 3113 + }, + { + "epoch": 4.044155844155844, + "grad_norm": 12.32160630651299, + "learning_rate": 1.3509416099639456e-05, + "loss": 0.0217, + "step": 3114 + }, + { + "epoch": 4.045454545454546, + "grad_norm": 5.0794013592883545, + "learning_rate": 1.3505477137529892e-05, + "loss": 0.0452, + "step": 3115 + }, + { + "epoch": 4.046753246753247, + "grad_norm": 10.21174081054195, + "learning_rate": 1.3501537555234323e-05, + "loss": 0.0406, + "step": 3116 + }, + { + "epoch": 4.048051948051948, + "grad_norm": 5.3673580937826495, + "learning_rate": 1.3497597353449743e-05, + "loss": 0.0319, + "step": 3117 + }, + { + "epoch": 4.049350649350649, + "grad_norm": 8.896654990974499, + "learning_rate": 1.3493656532873246e-05, + "loss": 0.0105, + "step": 3118 + }, + { + "epoch": 4.05064935064935, + "grad_norm": 5.596873034232895, + "learning_rate": 1.3489715094202039e-05, + "loss": 0.0325, + "step": 3119 + }, + { + "epoch": 4.0519480519480515, + "grad_norm": 4.895075130472986, + "learning_rate": 1.3485773038133437e-05, + "loss": 0.0152, + "step": 3120 + }, + { + "epoch": 4.0532467532467535, + "grad_norm": 9.3338080332178, + "learning_rate": 1.3481830365364866e-05, + "loss": 0.0174, + "step": 3121 + }, + { + "epoch": 4.054545454545455, + "grad_norm": 5.849318105064706, + "learning_rate": 1.347788707659386e-05, + "loss": 0.0379, + "step": 3122 + }, + { + "epoch": 4.055844155844156, + "grad_norm": 4.564224092997429, + "learning_rate": 1.3473943172518064e-05, + "loss": 0.0414, + "step": 3123 + }, + { + "epoch": 4.057142857142857, + "grad_norm": 9.407191757753763, + "learning_rate": 1.346999865383523e-05, + "loss": 0.0152, + "step": 3124 + }, + { + "epoch": 4.058441558441558, + "grad_norm": 4.682678144939733, + "learning_rate": 1.3466053521243214e-05, + "loss": 0.017, + "step": 3125 + }, + { + "epoch": 4.05974025974026, + "grad_norm": 7.540155823183294, + "learning_rate": 1.3462107775439991e-05, + "loss": 0.0108, + "step": 3126 + }, + { + "epoch": 4.061038961038961, + "grad_norm": 4.199355768282487, + "learning_rate": 1.3458161417123639e-05, + "loss": 0.0218, + "step": 3127 + }, + { + "epoch": 4.062337662337662, + "grad_norm": 10.325761107301847, + "learning_rate": 1.345421444699234e-05, + "loss": 0.0398, + "step": 3128 + }, + { + "epoch": 4.0636363636363635, + "grad_norm": 3.503243861688969, + "learning_rate": 1.345026686574439e-05, + "loss": 0.0152, + "step": 3129 + }, + { + "epoch": 4.064935064935065, + "grad_norm": 6.594052693520304, + "learning_rate": 1.3446318674078194e-05, + "loss": 0.0162, + "step": 3130 + }, + { + "epoch": 4.066233766233767, + "grad_norm": 5.140909880251675, + "learning_rate": 1.3442369872692262e-05, + "loss": 0.0128, + "step": 3131 + }, + { + "epoch": 4.067532467532468, + "grad_norm": 7.35785701745714, + "learning_rate": 1.343842046228521e-05, + "loss": 0.0177, + "step": 3132 + }, + { + "epoch": 4.068831168831169, + "grad_norm": 10.76095733356172, + "learning_rate": 1.343447044355577e-05, + "loss": 0.0226, + "step": 3133 + }, + { + "epoch": 4.07012987012987, + "grad_norm": 3.41206177685542, + "learning_rate": 1.3430519817202768e-05, + "loss": 0.0242, + "step": 3134 + }, + { + "epoch": 4.071428571428571, + "grad_norm": 5.126938777983507, + "learning_rate": 1.3426568583925153e-05, + "loss": 0.0348, + "step": 3135 + }, + { + "epoch": 4.072727272727272, + "grad_norm": 7.496819537680193, + "learning_rate": 1.3422616744421966e-05, + "loss": 0.0425, + "step": 3136 + }, + { + "epoch": 4.074025974025974, + "grad_norm": 7.040083982508461, + "learning_rate": 1.3418664299392373e-05, + "loss": 0.0293, + "step": 3137 + }, + { + "epoch": 4.0753246753246755, + "grad_norm": 5.908481745280185, + "learning_rate": 1.3414711249535626e-05, + "loss": 0.0118, + "step": 3138 + }, + { + "epoch": 4.076623376623377, + "grad_norm": 5.1189843958818875, + "learning_rate": 1.3410757595551104e-05, + "loss": 0.0097, + "step": 3139 + }, + { + "epoch": 4.077922077922078, + "grad_norm": 5.098477971946128, + "learning_rate": 1.3406803338138278e-05, + "loss": 0.0209, + "step": 3140 + }, + { + "epoch": 4.079220779220779, + "grad_norm": 2.9050440887285744, + "learning_rate": 1.3402848477996738e-05, + "loss": 0.0104, + "step": 3141 + }, + { + "epoch": 4.080519480519481, + "grad_norm": 9.34479674030673, + "learning_rate": 1.3398893015826166e-05, + "loss": 0.0308, + "step": 3142 + }, + { + "epoch": 4.081818181818182, + "grad_norm": 5.968880161536896, + "learning_rate": 1.3394936952326369e-05, + "loss": 0.0384, + "step": 3143 + }, + { + "epoch": 4.083116883116883, + "grad_norm": 11.611337978712648, + "learning_rate": 1.3390980288197244e-05, + "loss": 0.0205, + "step": 3144 + }, + { + "epoch": 4.084415584415584, + "grad_norm": 11.999105632672046, + "learning_rate": 1.3387023024138803e-05, + "loss": 0.0085, + "step": 3145 + }, + { + "epoch": 4.085714285714285, + "grad_norm": 13.622750599431372, + "learning_rate": 1.3383065160851158e-05, + "loss": 0.0398, + "step": 3146 + }, + { + "epoch": 4.087012987012987, + "grad_norm": 13.877408438064684, + "learning_rate": 1.3379106699034536e-05, + "loss": 0.023, + "step": 3147 + }, + { + "epoch": 4.088311688311689, + "grad_norm": 5.380388383304574, + "learning_rate": 1.3375147639389261e-05, + "loss": 0.021, + "step": 3148 + }, + { + "epoch": 4.08961038961039, + "grad_norm": 11.765047463386088, + "learning_rate": 1.3371187982615767e-05, + "loss": 0.032, + "step": 3149 + }, + { + "epoch": 4.090909090909091, + "grad_norm": 8.80287506160992, + "learning_rate": 1.3367227729414596e-05, + "loss": 0.0237, + "step": 3150 + }, + { + "epoch": 4.092207792207792, + "grad_norm": 10.510635949870748, + "learning_rate": 1.3363266880486388e-05, + "loss": 0.0078, + "step": 3151 + }, + { + "epoch": 4.093506493506493, + "grad_norm": 8.345188596676174, + "learning_rate": 1.3359305436531895e-05, + "loss": 0.0084, + "step": 3152 + }, + { + "epoch": 4.094805194805195, + "grad_norm": 5.505648273237033, + "learning_rate": 1.3355343398251973e-05, + "loss": 0.022, + "step": 3153 + }, + { + "epoch": 4.096103896103896, + "grad_norm": 7.008013744987074, + "learning_rate": 1.3351380766347581e-05, + "loss": 0.0247, + "step": 3154 + }, + { + "epoch": 4.097402597402597, + "grad_norm": 7.478418087773441, + "learning_rate": 1.3347417541519786e-05, + "loss": 0.008, + "step": 3155 + }, + { + "epoch": 4.0987012987012985, + "grad_norm": 3.6256220567259922, + "learning_rate": 1.334345372446976e-05, + "loss": 0.0147, + "step": 3156 + }, + { + "epoch": 4.1, + "grad_norm": 7.425384504885163, + "learning_rate": 1.3339489315898773e-05, + "loss": 0.0142, + "step": 3157 + }, + { + "epoch": 4.101298701298702, + "grad_norm": 6.008448392751938, + "learning_rate": 1.3335524316508208e-05, + "loss": 0.021, + "step": 3158 + }, + { + "epoch": 4.102597402597403, + "grad_norm": 10.267143287425526, + "learning_rate": 1.3331558726999547e-05, + "loss": 0.0198, + "step": 3159 + }, + { + "epoch": 4.103896103896104, + "grad_norm": 4.998696842448022, + "learning_rate": 1.3327592548074387e-05, + "loss": 0.0306, + "step": 3160 + }, + { + "epoch": 4.105194805194805, + "grad_norm": 8.025874548163305, + "learning_rate": 1.3323625780434406e-05, + "loss": 0.0444, + "step": 3161 + }, + { + "epoch": 4.106493506493506, + "grad_norm": 6.147586693684919, + "learning_rate": 1.3319658424781415e-05, + "loss": 0.0375, + "step": 3162 + }, + { + "epoch": 4.107792207792208, + "grad_norm": 3.6956503602957413, + "learning_rate": 1.3315690481817307e-05, + "loss": 0.0184, + "step": 3163 + }, + { + "epoch": 4.109090909090909, + "grad_norm": 10.172998336920049, + "learning_rate": 1.331172195224409e-05, + "loss": 0.0098, + "step": 3164 + }, + { + "epoch": 4.1103896103896105, + "grad_norm": 4.8436313679256315, + "learning_rate": 1.3307752836763874e-05, + "loss": 0.0393, + "step": 3165 + }, + { + "epoch": 4.111688311688312, + "grad_norm": 10.458385875292851, + "learning_rate": 1.330378313607887e-05, + "loss": 0.0335, + "step": 3166 + }, + { + "epoch": 4.112987012987013, + "grad_norm": 9.251282467587886, + "learning_rate": 1.3299812850891396e-05, + "loss": 0.014, + "step": 3167 + }, + { + "epoch": 4.114285714285714, + "grad_norm": 11.344859729052441, + "learning_rate": 1.3295841981903865e-05, + "loss": 0.0215, + "step": 3168 + }, + { + "epoch": 4.115584415584416, + "grad_norm": 10.328969726536949, + "learning_rate": 1.3291870529818809e-05, + "loss": 0.0252, + "step": 3169 + }, + { + "epoch": 4.116883116883117, + "grad_norm": 3.4114781171841138, + "learning_rate": 1.3287898495338845e-05, + "loss": 0.0082, + "step": 3170 + }, + { + "epoch": 4.118181818181818, + "grad_norm": 14.614398244153604, + "learning_rate": 1.3283925879166711e-05, + "loss": 0.0157, + "step": 3171 + }, + { + "epoch": 4.119480519480519, + "grad_norm": 6.656760844359067, + "learning_rate": 1.3279952682005229e-05, + "loss": 0.0265, + "step": 3172 + }, + { + "epoch": 4.12077922077922, + "grad_norm": 11.377353685556823, + "learning_rate": 1.3275978904557341e-05, + "loss": 0.0157, + "step": 3173 + }, + { + "epoch": 4.1220779220779225, + "grad_norm": 10.754868651539203, + "learning_rate": 1.3272004547526084e-05, + "loss": 0.0166, + "step": 3174 + }, + { + "epoch": 4.123376623376624, + "grad_norm": 4.401545894170428, + "learning_rate": 1.326802961161459e-05, + "loss": 0.0143, + "step": 3175 + }, + { + "epoch": 4.124675324675325, + "grad_norm": 7.517454288674153, + "learning_rate": 1.3264054097526111e-05, + "loss": 0.0056, + "step": 3176 + }, + { + "epoch": 4.125974025974026, + "grad_norm": 7.119653455051146, + "learning_rate": 1.3260078005963988e-05, + "loss": 0.016, + "step": 3177 + }, + { + "epoch": 4.127272727272727, + "grad_norm": 3.2691250476437954, + "learning_rate": 1.3256101337631666e-05, + "loss": 0.0342, + "step": 3178 + }, + { + "epoch": 4.128571428571428, + "grad_norm": 14.27699343360444, + "learning_rate": 1.3252124093232692e-05, + "loss": 0.0089, + "step": 3179 + }, + { + "epoch": 4.12987012987013, + "grad_norm": 6.266152546956543, + "learning_rate": 1.3248146273470725e-05, + "loss": 0.0153, + "step": 3180 + }, + { + "epoch": 4.131168831168831, + "grad_norm": 6.372490124331612, + "learning_rate": 1.3244167879049508e-05, + "loss": 0.0276, + "step": 3181 + }, + { + "epoch": 4.132467532467532, + "grad_norm": 7.039178049618962, + "learning_rate": 1.3240188910672902e-05, + "loss": 0.0288, + "step": 3182 + }, + { + "epoch": 4.1337662337662335, + "grad_norm": 4.554282815946494, + "learning_rate": 1.3236209369044857e-05, + "loss": 0.0212, + "step": 3183 + }, + { + "epoch": 4.135064935064935, + "grad_norm": 5.44059352315613, + "learning_rate": 1.3232229254869435e-05, + "loss": 0.0127, + "step": 3184 + }, + { + "epoch": 4.136363636363637, + "grad_norm": 7.343110938917417, + "learning_rate": 1.322824856885079e-05, + "loss": 0.0583, + "step": 3185 + }, + { + "epoch": 4.137662337662338, + "grad_norm": 4.732037683539564, + "learning_rate": 1.3224267311693186e-05, + "loss": 0.0223, + "step": 3186 + }, + { + "epoch": 4.138961038961039, + "grad_norm": 8.542907211800523, + "learning_rate": 1.3220285484100981e-05, + "loss": 0.028, + "step": 3187 + }, + { + "epoch": 4.14025974025974, + "grad_norm": 7.39181807475459, + "learning_rate": 1.3216303086778636e-05, + "loss": 0.017, + "step": 3188 + }, + { + "epoch": 4.141558441558441, + "grad_norm": 8.251501971995685, + "learning_rate": 1.3212320120430718e-05, + "loss": 0.0247, + "step": 3189 + }, + { + "epoch": 4.142857142857143, + "grad_norm": 14.960629580697896, + "learning_rate": 1.3208336585761883e-05, + "loss": 0.0161, + "step": 3190 + }, + { + "epoch": 4.144155844155844, + "grad_norm": 7.022153644626419, + "learning_rate": 1.3204352483476898e-05, + "loss": 0.0274, + "step": 3191 + }, + { + "epoch": 4.1454545454545455, + "grad_norm": 20.329710081866, + "learning_rate": 1.3200367814280626e-05, + "loss": 0.0147, + "step": 3192 + }, + { + "epoch": 4.146753246753247, + "grad_norm": 13.725302344167705, + "learning_rate": 1.3196382578878032e-05, + "loss": 0.0233, + "step": 3193 + }, + { + "epoch": 4.148051948051948, + "grad_norm": 11.002862160212532, + "learning_rate": 1.3192396777974182e-05, + "loss": 0.0294, + "step": 3194 + }, + { + "epoch": 4.14935064935065, + "grad_norm": 24.215163292526988, + "learning_rate": 1.3188410412274237e-05, + "loss": 0.0214, + "step": 3195 + }, + { + "epoch": 4.150649350649351, + "grad_norm": 3.7255061495869586, + "learning_rate": 1.3184423482483463e-05, + "loss": 0.0254, + "step": 3196 + }, + { + "epoch": 4.151948051948052, + "grad_norm": 25.013504107852683, + "learning_rate": 1.3180435989307227e-05, + "loss": 0.0318, + "step": 3197 + }, + { + "epoch": 4.153246753246753, + "grad_norm": 8.188764396438213, + "learning_rate": 1.317644793345099e-05, + "loss": 0.0105, + "step": 3198 + }, + { + "epoch": 4.154545454545454, + "grad_norm": 12.388974590647367, + "learning_rate": 1.3172459315620315e-05, + "loss": 0.0276, + "step": 3199 + }, + { + "epoch": 4.1558441558441555, + "grad_norm": 12.342141657060095, + "learning_rate": 1.3168470136520863e-05, + "loss": 0.0592, + "step": 3200 + }, + { + "epoch": 4.1571428571428575, + "grad_norm": 3.772808788949958, + "learning_rate": 1.3164480396858401e-05, + "loss": 0.0145, + "step": 3201 + }, + { + "epoch": 4.158441558441559, + "grad_norm": 8.64400261671058, + "learning_rate": 1.316049009733879e-05, + "loss": 0.039, + "step": 3202 + }, + { + "epoch": 4.15974025974026, + "grad_norm": 28.40518823088921, + "learning_rate": 1.3156499238667983e-05, + "loss": 0.0738, + "step": 3203 + }, + { + "epoch": 4.161038961038961, + "grad_norm": 7.805837168202388, + "learning_rate": 1.3152507821552049e-05, + "loss": 0.0294, + "step": 3204 + }, + { + "epoch": 4.162337662337662, + "grad_norm": 35.310864580915116, + "learning_rate": 1.314851584669714e-05, + "loss": 0.0427, + "step": 3205 + }, + { + "epoch": 4.163636363636364, + "grad_norm": 4.430808714903632, + "learning_rate": 1.3144523314809516e-05, + "loss": 0.0194, + "step": 3206 + }, + { + "epoch": 4.164935064935065, + "grad_norm": 23.950920454272573, + "learning_rate": 1.314053022659553e-05, + "loss": 0.0155, + "step": 3207 + }, + { + "epoch": 4.166233766233766, + "grad_norm": 17.73038088477666, + "learning_rate": 1.3136536582761636e-05, + "loss": 0.0149, + "step": 3208 + }, + { + "epoch": 4.167532467532467, + "grad_norm": 9.867061463804871, + "learning_rate": 1.3132542384014389e-05, + "loss": 0.0302, + "step": 3209 + }, + { + "epoch": 4.1688311688311686, + "grad_norm": 22.382115991417198, + "learning_rate": 1.3128547631060434e-05, + "loss": 0.0234, + "step": 3210 + }, + { + "epoch": 4.17012987012987, + "grad_norm": 12.491979765229718, + "learning_rate": 1.312455232460652e-05, + "loss": 0.0328, + "step": 3211 + }, + { + "epoch": 4.171428571428572, + "grad_norm": 15.520313928695028, + "learning_rate": 1.3120556465359502e-05, + "loss": 0.026, + "step": 3212 + }, + { + "epoch": 4.172727272727273, + "grad_norm": 9.216624021523911, + "learning_rate": 1.311656005402631e-05, + "loss": 0.0177, + "step": 3213 + }, + { + "epoch": 4.174025974025974, + "grad_norm": 4.5851545470723085, + "learning_rate": 1.3112563091313998e-05, + "loss": 0.024, + "step": 3214 + }, + { + "epoch": 4.175324675324675, + "grad_norm": 9.929381040672515, + "learning_rate": 1.31085655779297e-05, + "loss": 0.0122, + "step": 3215 + }, + { + "epoch": 4.176623376623376, + "grad_norm": 3.0010838369134722, + "learning_rate": 1.3104567514580648e-05, + "loss": 0.0272, + "step": 3216 + }, + { + "epoch": 4.177922077922078, + "grad_norm": 5.5867560775088965, + "learning_rate": 1.3100568901974185e-05, + "loss": 0.0279, + "step": 3217 + }, + { + "epoch": 4.179220779220779, + "grad_norm": 10.95305444398832, + "learning_rate": 1.3096569740817736e-05, + "loss": 0.0401, + "step": 3218 + }, + { + "epoch": 4.1805194805194805, + "grad_norm": 5.686741057732427, + "learning_rate": 1.309257003181883e-05, + "loss": 0.0182, + "step": 3219 + }, + { + "epoch": 4.181818181818182, + "grad_norm": 8.138434978320493, + "learning_rate": 1.3088569775685094e-05, + "loss": 0.023, + "step": 3220 + }, + { + "epoch": 4.183116883116883, + "grad_norm": 11.026197956078315, + "learning_rate": 1.308456897312425e-05, + "loss": 0.0144, + "step": 3221 + }, + { + "epoch": 4.184415584415585, + "grad_norm": 7.00392638390651, + "learning_rate": 1.3080567624844114e-05, + "loss": 0.0298, + "step": 3222 + }, + { + "epoch": 4.185714285714286, + "grad_norm": 12.547298252517924, + "learning_rate": 1.3076565731552602e-05, + "loss": 0.0221, + "step": 3223 + }, + { + "epoch": 4.187012987012987, + "grad_norm": 9.415947366231654, + "learning_rate": 1.3072563293957725e-05, + "loss": 0.0203, + "step": 3224 + }, + { + "epoch": 4.188311688311688, + "grad_norm": 10.141149005997068, + "learning_rate": 1.3068560312767595e-05, + "loss": 0.0409, + "step": 3225 + }, + { + "epoch": 4.189610389610389, + "grad_norm": 11.634539005953199, + "learning_rate": 1.306455678869041e-05, + "loss": 0.0398, + "step": 3226 + }, + { + "epoch": 4.190909090909091, + "grad_norm": 5.681284596902167, + "learning_rate": 1.3060552722434476e-05, + "loss": 0.0325, + "step": 3227 + }, + { + "epoch": 4.1922077922077925, + "grad_norm": 11.539257646746467, + "learning_rate": 1.3056548114708182e-05, + "loss": 0.0244, + "step": 3228 + }, + { + "epoch": 4.193506493506494, + "grad_norm": 13.292576167365874, + "learning_rate": 1.3052542966220028e-05, + "loss": 0.0297, + "step": 3229 + }, + { + "epoch": 4.194805194805195, + "grad_norm": 11.64870053211232, + "learning_rate": 1.304853727767859e-05, + "loss": 0.0223, + "step": 3230 + }, + { + "epoch": 4.196103896103896, + "grad_norm": 22.48619903367214, + "learning_rate": 1.304453104979256e-05, + "loss": 0.0408, + "step": 3231 + }, + { + "epoch": 4.197402597402597, + "grad_norm": 9.90837524927709, + "learning_rate": 1.3040524283270722e-05, + "loss": 0.0153, + "step": 3232 + }, + { + "epoch": 4.198701298701299, + "grad_norm": 15.973366244554281, + "learning_rate": 1.3036516978821935e-05, + "loss": 0.0134, + "step": 3233 + }, + { + "epoch": 4.2, + "grad_norm": 12.036239770264903, + "learning_rate": 1.3032509137155175e-05, + "loss": 0.0496, + "step": 3234 + }, + { + "epoch": 4.201298701298701, + "grad_norm": 19.941824880842997, + "learning_rate": 1.3028500758979507e-05, + "loss": 0.0315, + "step": 3235 + }, + { + "epoch": 4.202597402597402, + "grad_norm": 10.352238504437413, + "learning_rate": 1.302449184500409e-05, + "loss": 0.0109, + "step": 3236 + }, + { + "epoch": 4.203896103896104, + "grad_norm": 5.299285681701042, + "learning_rate": 1.3020482395938174e-05, + "loss": 0.0109, + "step": 3237 + }, + { + "epoch": 4.205194805194806, + "grad_norm": 12.843599897638187, + "learning_rate": 1.3016472412491113e-05, + "loss": 0.0179, + "step": 3238 + }, + { + "epoch": 4.206493506493507, + "grad_norm": 13.614552640690848, + "learning_rate": 1.3012461895372343e-05, + "loss": 0.0333, + "step": 3239 + }, + { + "epoch": 4.207792207792208, + "grad_norm": 10.400364429245911, + "learning_rate": 1.3008450845291408e-05, + "loss": 0.0163, + "step": 3240 + }, + { + "epoch": 4.209090909090909, + "grad_norm": 16.719562810021014, + "learning_rate": 1.3004439262957931e-05, + "loss": 0.0103, + "step": 3241 + }, + { + "epoch": 4.21038961038961, + "grad_norm": 3.0340624503436278, + "learning_rate": 1.3000427149081647e-05, + "loss": 0.0156, + "step": 3242 + }, + { + "epoch": 4.211688311688311, + "grad_norm": 16.072945506090385, + "learning_rate": 1.299641450437237e-05, + "loss": 0.0216, + "step": 3243 + }, + { + "epoch": 4.212987012987013, + "grad_norm": 6.197600254299947, + "learning_rate": 1.2992401329540014e-05, + "loss": 0.0161, + "step": 3244 + }, + { + "epoch": 4.214285714285714, + "grad_norm": 15.79616351566482, + "learning_rate": 1.2988387625294592e-05, + "loss": 0.0467, + "step": 3245 + }, + { + "epoch": 4.2155844155844155, + "grad_norm": 14.95494124361543, + "learning_rate": 1.2984373392346195e-05, + "loss": 0.0358, + "step": 3246 + }, + { + "epoch": 4.216883116883117, + "grad_norm": 11.842133250135793, + "learning_rate": 1.2980358631405027e-05, + "loss": 0.0194, + "step": 3247 + }, + { + "epoch": 4.218181818181818, + "grad_norm": 11.798315874884294, + "learning_rate": 1.2976343343181371e-05, + "loss": 0.0172, + "step": 3248 + }, + { + "epoch": 4.21948051948052, + "grad_norm": 9.264861118649522, + "learning_rate": 1.297232752838561e-05, + "loss": 0.0503, + "step": 3249 + }, + { + "epoch": 4.220779220779221, + "grad_norm": 5.616209921689897, + "learning_rate": 1.296831118772822e-05, + "loss": 0.0188, + "step": 3250 + }, + { + "epoch": 4.222077922077922, + "grad_norm": 8.059272646819032, + "learning_rate": 1.2964294321919763e-05, + "loss": 0.0225, + "step": 3251 + }, + { + "epoch": 4.223376623376623, + "grad_norm": 3.9091294177918066, + "learning_rate": 1.2960276931670905e-05, + "loss": 0.0301, + "step": 3252 + }, + { + "epoch": 4.224675324675324, + "grad_norm": 5.41607788497412, + "learning_rate": 1.29562590176924e-05, + "loss": 0.0414, + "step": 3253 + }, + { + "epoch": 4.225974025974026, + "grad_norm": 5.971876611047876, + "learning_rate": 1.2952240580695088e-05, + "loss": 0.0182, + "step": 3254 + }, + { + "epoch": 4.2272727272727275, + "grad_norm": 6.659406721634659, + "learning_rate": 1.2948221621389914e-05, + "loss": 0.0362, + "step": 3255 + }, + { + "epoch": 4.228571428571429, + "grad_norm": 7.183321260772134, + "learning_rate": 1.2944202140487905e-05, + "loss": 0.0398, + "step": 3256 + }, + { + "epoch": 4.22987012987013, + "grad_norm": 10.188959536826188, + "learning_rate": 1.294018213870018e-05, + "loss": 0.0336, + "step": 3257 + }, + { + "epoch": 4.231168831168831, + "grad_norm": 5.125446029819703, + "learning_rate": 1.2936161616737965e-05, + "loss": 0.0263, + "step": 3258 + }, + { + "epoch": 4.232467532467532, + "grad_norm": 9.15922533921995, + "learning_rate": 1.2932140575312557e-05, + "loss": 0.0105, + "step": 3259 + }, + { + "epoch": 4.233766233766234, + "grad_norm": 5.53524332321439, + "learning_rate": 1.2928119015135363e-05, + "loss": 0.0281, + "step": 3260 + }, + { + "epoch": 4.235064935064935, + "grad_norm": 2.8029877396621963, + "learning_rate": 1.2924096936917869e-05, + "loss": 0.0058, + "step": 3261 + }, + { + "epoch": 4.236363636363636, + "grad_norm": 14.441697261568393, + "learning_rate": 1.2920074341371658e-05, + "loss": 0.0206, + "step": 3262 + }, + { + "epoch": 4.2376623376623375, + "grad_norm": 3.3391065111411136, + "learning_rate": 1.2916051229208406e-05, + "loss": 0.0257, + "step": 3263 + }, + { + "epoch": 4.238961038961039, + "grad_norm": 8.693077585656033, + "learning_rate": 1.2912027601139875e-05, + "loss": 0.045, + "step": 3264 + }, + { + "epoch": 4.240259740259741, + "grad_norm": 9.09742781105825, + "learning_rate": 1.290800345787793e-05, + "loss": 0.0308, + "step": 3265 + }, + { + "epoch": 4.241558441558442, + "grad_norm": 9.565188528920267, + "learning_rate": 1.2903978800134509e-05, + "loss": 0.0386, + "step": 3266 + }, + { + "epoch": 4.242857142857143, + "grad_norm": 4.504276769791098, + "learning_rate": 1.2899953628621654e-05, + "loss": 0.0449, + "step": 3267 + }, + { + "epoch": 4.244155844155844, + "grad_norm": 11.821352727168863, + "learning_rate": 1.2895927944051503e-05, + "loss": 0.0258, + "step": 3268 + }, + { + "epoch": 4.245454545454545, + "grad_norm": 5.338913263739233, + "learning_rate": 1.2891901747136269e-05, + "loss": 0.0353, + "step": 3269 + }, + { + "epoch": 4.246753246753247, + "grad_norm": 9.068614438720916, + "learning_rate": 1.288787503858826e-05, + "loss": 0.0464, + "step": 3270 + }, + { + "epoch": 4.248051948051948, + "grad_norm": 7.378549857895497, + "learning_rate": 1.2883847819119883e-05, + "loss": 0.0072, + "step": 3271 + }, + { + "epoch": 4.249350649350649, + "grad_norm": 10.020832840811886, + "learning_rate": 1.2879820089443631e-05, + "loss": 0.0223, + "step": 3272 + }, + { + "epoch": 4.250649350649351, + "grad_norm": 7.121996960306325, + "learning_rate": 1.2875791850272086e-05, + "loss": 0.018, + "step": 3273 + }, + { + "epoch": 4.251948051948052, + "grad_norm": 4.570359673117334, + "learning_rate": 1.2871763102317916e-05, + "loss": 0.0157, + "step": 3274 + }, + { + "epoch": 4.253246753246753, + "grad_norm": 3.8900065805375257, + "learning_rate": 1.286773384629389e-05, + "loss": 0.0268, + "step": 3275 + }, + { + "epoch": 4.254545454545455, + "grad_norm": 7.592651818971176, + "learning_rate": 1.2863704082912856e-05, + "loss": 0.0107, + "step": 3276 + }, + { + "epoch": 4.255844155844156, + "grad_norm": 10.682605666400793, + "learning_rate": 1.285967381288776e-05, + "loss": 0.0389, + "step": 3277 + }, + { + "epoch": 4.257142857142857, + "grad_norm": 6.326444245625554, + "learning_rate": 1.2855643036931626e-05, + "loss": 0.0441, + "step": 3278 + }, + { + "epoch": 4.258441558441558, + "grad_norm": 10.915409934709443, + "learning_rate": 1.2851611755757587e-05, + "loss": 0.045, + "step": 3279 + }, + { + "epoch": 4.259740259740259, + "grad_norm": 7.8321358525550115, + "learning_rate": 1.2847579970078846e-05, + "loss": 0.0202, + "step": 3280 + }, + { + "epoch": 4.261038961038961, + "grad_norm": 10.0546954291364, + "learning_rate": 1.2843547680608705e-05, + "loss": 0.0283, + "step": 3281 + }, + { + "epoch": 4.2623376623376625, + "grad_norm": 4.548104874200484, + "learning_rate": 1.2839514888060556e-05, + "loss": 0.0077, + "step": 3282 + }, + { + "epoch": 4.263636363636364, + "grad_norm": 11.902272664328454, + "learning_rate": 1.283548159314787e-05, + "loss": 0.0161, + "step": 3283 + }, + { + "epoch": 4.264935064935065, + "grad_norm": 4.928695419170289, + "learning_rate": 1.2831447796584221e-05, + "loss": 0.0235, + "step": 3284 + }, + { + "epoch": 4.266233766233766, + "grad_norm": 6.9687133703333695, + "learning_rate": 1.2827413499083262e-05, + "loss": 0.0271, + "step": 3285 + }, + { + "epoch": 4.267532467532467, + "grad_norm": 12.180527619425348, + "learning_rate": 1.2823378701358737e-05, + "loss": 0.0179, + "step": 3286 + }, + { + "epoch": 4.268831168831169, + "grad_norm": 8.061536011465249, + "learning_rate": 1.281934340412448e-05, + "loss": 0.0349, + "step": 3287 + }, + { + "epoch": 4.27012987012987, + "grad_norm": 5.201497725197897, + "learning_rate": 1.2815307608094414e-05, + "loss": 0.0126, + "step": 3288 + }, + { + "epoch": 4.271428571428571, + "grad_norm": 8.400000328511128, + "learning_rate": 1.2811271313982543e-05, + "loss": 0.0243, + "step": 3289 + }, + { + "epoch": 4.2727272727272725, + "grad_norm": 3.579006386563467, + "learning_rate": 1.280723452250297e-05, + "loss": 0.0163, + "step": 3290 + }, + { + "epoch": 4.274025974025974, + "grad_norm": 7.993834376253763, + "learning_rate": 1.2803197234369878e-05, + "loss": 0.0227, + "step": 3291 + }, + { + "epoch": 4.275324675324676, + "grad_norm": 4.959277983248826, + "learning_rate": 1.2799159450297542e-05, + "loss": 0.028, + "step": 3292 + }, + { + "epoch": 4.276623376623377, + "grad_norm": 3.5950167823278414, + "learning_rate": 1.2795121171000324e-05, + "loss": 0.0115, + "step": 3293 + }, + { + "epoch": 4.277922077922078, + "grad_norm": 4.247611394285891, + "learning_rate": 1.2791082397192671e-05, + "loss": 0.0126, + "step": 3294 + }, + { + "epoch": 4.279220779220779, + "grad_norm": 7.1978489511697195, + "learning_rate": 1.2787043129589118e-05, + "loss": 0.0327, + "step": 3295 + }, + { + "epoch": 4.28051948051948, + "grad_norm": 6.077200008669098, + "learning_rate": 1.2783003368904295e-05, + "loss": 0.0143, + "step": 3296 + }, + { + "epoch": 4.281818181818182, + "grad_norm": 5.782399920654351, + "learning_rate": 1.2778963115852906e-05, + "loss": 0.0315, + "step": 3297 + }, + { + "epoch": 4.283116883116883, + "grad_norm": 3.4393076284644453, + "learning_rate": 1.277492237114975e-05, + "loss": 0.0182, + "step": 3298 + }, + { + "epoch": 4.2844155844155845, + "grad_norm": 12.110852003701227, + "learning_rate": 1.2770881135509716e-05, + "loss": 0.0184, + "step": 3299 + }, + { + "epoch": 4.285714285714286, + "grad_norm": 9.935616718546736, + "learning_rate": 1.2766839409647774e-05, + "loss": 0.0247, + "step": 3300 + }, + { + "epoch": 4.287012987012987, + "grad_norm": 14.236636955299137, + "learning_rate": 1.2762797194278984e-05, + "loss": 0.0202, + "step": 3301 + }, + { + "epoch": 4.288311688311689, + "grad_norm": 10.6045050067263, + "learning_rate": 1.2758754490118488e-05, + "loss": 0.0339, + "step": 3302 + }, + { + "epoch": 4.28961038961039, + "grad_norm": 15.389075439954121, + "learning_rate": 1.275471129788152e-05, + "loss": 0.023, + "step": 3303 + }, + { + "epoch": 4.290909090909091, + "grad_norm": 14.363066538628605, + "learning_rate": 1.2750667618283397e-05, + "loss": 0.0125, + "step": 3304 + }, + { + "epoch": 4.292207792207792, + "grad_norm": 2.843485718846649, + "learning_rate": 1.2746623452039526e-05, + "loss": 0.0155, + "step": 3305 + }, + { + "epoch": 4.293506493506493, + "grad_norm": 14.798946045016276, + "learning_rate": 1.2742578799865393e-05, + "loss": 0.0128, + "step": 3306 + }, + { + "epoch": 4.294805194805194, + "grad_norm": 10.058329388502193, + "learning_rate": 1.2738533662476578e-05, + "loss": 0.017, + "step": 3307 + }, + { + "epoch": 4.296103896103896, + "grad_norm": 7.3622353052238765, + "learning_rate": 1.2734488040588742e-05, + "loss": 0.0182, + "step": 3308 + }, + { + "epoch": 4.297402597402598, + "grad_norm": 14.06016780174021, + "learning_rate": 1.2730441934917631e-05, + "loss": 0.0311, + "step": 3309 + }, + { + "epoch": 4.298701298701299, + "grad_norm": 9.030914233651949, + "learning_rate": 1.2726395346179086e-05, + "loss": 0.0258, + "step": 3310 + }, + { + "epoch": 4.3, + "grad_norm": 14.676179894741063, + "learning_rate": 1.2722348275089016e-05, + "loss": 0.0166, + "step": 3311 + }, + { + "epoch": 4.301298701298701, + "grad_norm": 18.977089609379668, + "learning_rate": 1.2718300722363431e-05, + "loss": 0.0198, + "step": 3312 + }, + { + "epoch": 4.302597402597403, + "grad_norm": 7.881088439572385, + "learning_rate": 1.2714252688718418e-05, + "loss": 0.0065, + "step": 3313 + }, + { + "epoch": 4.303896103896104, + "grad_norm": 19.85090938694214, + "learning_rate": 1.2710204174870155e-05, + "loss": 0.0234, + "step": 3314 + }, + { + "epoch": 4.305194805194805, + "grad_norm": 13.718036331703516, + "learning_rate": 1.2706155181534897e-05, + "loss": 0.0504, + "step": 3315 + }, + { + "epoch": 4.306493506493506, + "grad_norm": 18.81746131401524, + "learning_rate": 1.2702105709428993e-05, + "loss": 0.0169, + "step": 3316 + }, + { + "epoch": 4.3077922077922075, + "grad_norm": 22.77989121499448, + "learning_rate": 1.269805575926887e-05, + "loss": 0.01, + "step": 3317 + }, + { + "epoch": 4.309090909090909, + "grad_norm": 8.21287163550533, + "learning_rate": 1.2694005331771041e-05, + "loss": 0.0098, + "step": 3318 + }, + { + "epoch": 4.310389610389611, + "grad_norm": 17.33331332367833, + "learning_rate": 1.2689954427652107e-05, + "loss": 0.0504, + "step": 3319 + }, + { + "epoch": 4.311688311688312, + "grad_norm": 19.99733431388382, + "learning_rate": 1.2685903047628747e-05, + "loss": 0.0156, + "step": 3320 + }, + { + "epoch": 4.312987012987013, + "grad_norm": 14.76814821415546, + "learning_rate": 1.268185119241773e-05, + "loss": 0.0366, + "step": 3321 + }, + { + "epoch": 4.314285714285714, + "grad_norm": 25.432945666334277, + "learning_rate": 1.2677798862735907e-05, + "loss": 0.0269, + "step": 3322 + }, + { + "epoch": 4.315584415584415, + "grad_norm": 4.374371201497526, + "learning_rate": 1.2673746059300209e-05, + "loss": 0.0251, + "step": 3323 + }, + { + "epoch": 4.316883116883117, + "grad_norm": 22.718871736122214, + "learning_rate": 1.2669692782827657e-05, + "loss": 0.0097, + "step": 3324 + }, + { + "epoch": 4.318181818181818, + "grad_norm": 15.768644843783362, + "learning_rate": 1.2665639034035355e-05, + "loss": 0.0334, + "step": 3325 + }, + { + "epoch": 4.3194805194805195, + "grad_norm": 12.892409479557392, + "learning_rate": 1.2661584813640485e-05, + "loss": 0.0692, + "step": 3326 + }, + { + "epoch": 4.320779220779221, + "grad_norm": 23.50978146886155, + "learning_rate": 1.265753012236032e-05, + "loss": 0.0191, + "step": 3327 + }, + { + "epoch": 4.322077922077922, + "grad_norm": 7.52674603034745, + "learning_rate": 1.265347496091221e-05, + "loss": 0.022, + "step": 3328 + }, + { + "epoch": 4.323376623376624, + "grad_norm": 20.132907505129907, + "learning_rate": 1.2649419330013589e-05, + "loss": 0.0323, + "step": 3329 + }, + { + "epoch": 4.324675324675325, + "grad_norm": 22.039995327465668, + "learning_rate": 1.2645363230381982e-05, + "loss": 0.0452, + "step": 3330 + }, + { + "epoch": 4.325974025974026, + "grad_norm": 9.095795908215255, + "learning_rate": 1.2641306662734985e-05, + "loss": 0.0249, + "step": 3331 + }, + { + "epoch": 4.327272727272727, + "grad_norm": 27.323071444715847, + "learning_rate": 1.2637249627790284e-05, + "loss": 0.0254, + "step": 3332 + }, + { + "epoch": 4.328571428571428, + "grad_norm": 8.866791828120874, + "learning_rate": 1.2633192126265646e-05, + "loss": 0.0135, + "step": 3333 + }, + { + "epoch": 4.32987012987013, + "grad_norm": 16.315372711661226, + "learning_rate": 1.2629134158878919e-05, + "loss": 0.0221, + "step": 3334 + }, + { + "epoch": 4.3311688311688314, + "grad_norm": 24.423750894080655, + "learning_rate": 1.2625075726348042e-05, + "loss": 0.0208, + "step": 3335 + }, + { + "epoch": 4.332467532467533, + "grad_norm": 6.095660947844862, + "learning_rate": 1.2621016829391022e-05, + "loss": 0.0159, + "step": 3336 + }, + { + "epoch": 4.333766233766234, + "grad_norm": 24.52680353730199, + "learning_rate": 1.2616957468725958e-05, + "loss": 0.0441, + "step": 3337 + }, + { + "epoch": 4.335064935064935, + "grad_norm": 15.502121832676567, + "learning_rate": 1.2612897645071032e-05, + "loss": 0.0359, + "step": 3338 + }, + { + "epoch": 4.336363636363636, + "grad_norm": 15.50275090366949, + "learning_rate": 1.2608837359144497e-05, + "loss": 0.0319, + "step": 3339 + }, + { + "epoch": 4.337662337662338, + "grad_norm": 26.57712227248337, + "learning_rate": 1.2604776611664701e-05, + "loss": 0.0292, + "step": 3340 + }, + { + "epoch": 4.338961038961039, + "grad_norm": 6.786275772172715, + "learning_rate": 1.2600715403350066e-05, + "loss": 0.0083, + "step": 3341 + }, + { + "epoch": 4.34025974025974, + "grad_norm": 28.358559970921704, + "learning_rate": 1.25966537349191e-05, + "loss": 0.0509, + "step": 3342 + }, + { + "epoch": 4.341558441558441, + "grad_norm": 13.626083271822353, + "learning_rate": 1.2592591607090383e-05, + "loss": 0.0172, + "step": 3343 + }, + { + "epoch": 4.3428571428571425, + "grad_norm": 22.224326931562825, + "learning_rate": 1.2588529020582593e-05, + "loss": 0.0183, + "step": 3344 + }, + { + "epoch": 4.3441558441558445, + "grad_norm": 21.921244276916763, + "learning_rate": 1.258446597611447e-05, + "loss": 0.0131, + "step": 3345 + }, + { + "epoch": 4.345454545454546, + "grad_norm": 7.497403908645142, + "learning_rate": 1.2580402474404854e-05, + "loss": 0.0322, + "step": 3346 + }, + { + "epoch": 4.346753246753247, + "grad_norm": 17.538926367517778, + "learning_rate": 1.2576338516172646e-05, + "loss": 0.04, + "step": 3347 + }, + { + "epoch": 4.348051948051948, + "grad_norm": 12.374825825542464, + "learning_rate": 1.2572274102136846e-05, + "loss": 0.0166, + "step": 3348 + }, + { + "epoch": 4.349350649350649, + "grad_norm": 10.495322340643883, + "learning_rate": 1.256820923301652e-05, + "loss": 0.0186, + "step": 3349 + }, + { + "epoch": 4.35064935064935, + "grad_norm": 17.397021001952062, + "learning_rate": 1.2564143909530828e-05, + "loss": 0.0194, + "step": 3350 + }, + { + "epoch": 4.351948051948052, + "grad_norm": 3.831603648039545, + "learning_rate": 1.2560078132399e-05, + "loss": 0.0349, + "step": 3351 + }, + { + "epoch": 4.353246753246753, + "grad_norm": 14.41794115447292, + "learning_rate": 1.2556011902340348e-05, + "loss": 0.0468, + "step": 3352 + }, + { + "epoch": 4.3545454545454545, + "grad_norm": 6.809069300716716, + "learning_rate": 1.2551945220074271e-05, + "loss": 0.0107, + "step": 3353 + }, + { + "epoch": 4.355844155844156, + "grad_norm": 2.8391977705263045, + "learning_rate": 1.2547878086320235e-05, + "loss": 0.023, + "step": 3354 + }, + { + "epoch": 4.357142857142857, + "grad_norm": 10.940364910049874, + "learning_rate": 1.2543810501797802e-05, + "loss": 0.0167, + "step": 3355 + }, + { + "epoch": 4.358441558441559, + "grad_norm": 9.373580818404172, + "learning_rate": 1.25397424672266e-05, + "loss": 0.0165, + "step": 3356 + }, + { + "epoch": 4.35974025974026, + "grad_norm": 10.073764343783576, + "learning_rate": 1.253567398332635e-05, + "loss": 0.0106, + "step": 3357 + }, + { + "epoch": 4.361038961038961, + "grad_norm": 12.712282856153271, + "learning_rate": 1.2531605050816831e-05, + "loss": 0.0313, + "step": 3358 + }, + { + "epoch": 4.362337662337662, + "grad_norm": 1.8016483832823802, + "learning_rate": 1.2527535670417927e-05, + "loss": 0.0102, + "step": 3359 + }, + { + "epoch": 4.363636363636363, + "grad_norm": 14.36029977715041, + "learning_rate": 1.2523465842849584e-05, + "loss": 0.0119, + "step": 3360 + }, + { + "epoch": 4.364935064935065, + "grad_norm": 5.11312010813834, + "learning_rate": 1.2519395568831839e-05, + "loss": 0.0303, + "step": 3361 + }, + { + "epoch": 4.3662337662337665, + "grad_norm": 6.654649488192243, + "learning_rate": 1.2515324849084789e-05, + "loss": 0.0216, + "step": 3362 + }, + { + "epoch": 4.367532467532468, + "grad_norm": 15.977370289243629, + "learning_rate": 1.2511253684328634e-05, + "loss": 0.0446, + "step": 3363 + }, + { + "epoch": 4.368831168831169, + "grad_norm": 3.9152513008190226, + "learning_rate": 1.2507182075283633e-05, + "loss": 0.0145, + "step": 3364 + }, + { + "epoch": 4.37012987012987, + "grad_norm": 9.207434943730325, + "learning_rate": 1.2503110022670135e-05, + "loss": 0.0116, + "step": 3365 + }, + { + "epoch": 4.371428571428572, + "grad_norm": 7.127731521376642, + "learning_rate": 1.2499037527208566e-05, + "loss": 0.0131, + "step": 3366 + }, + { + "epoch": 4.372727272727273, + "grad_norm": 10.141925085676323, + "learning_rate": 1.2494964589619424e-05, + "loss": 0.0397, + "step": 3367 + }, + { + "epoch": 4.374025974025974, + "grad_norm": 4.607265570850028, + "learning_rate": 1.2490891210623297e-05, + "loss": 0.008, + "step": 3368 + }, + { + "epoch": 4.375324675324675, + "grad_norm": 11.486093637236005, + "learning_rate": 1.2486817390940835e-05, + "loss": 0.0506, + "step": 3369 + }, + { + "epoch": 4.376623376623376, + "grad_norm": 5.9276419824900275, + "learning_rate": 1.2482743131292781e-05, + "loss": 0.0346, + "step": 3370 + }, + { + "epoch": 4.3779220779220775, + "grad_norm": 2.6007463680792386, + "learning_rate": 1.2478668432399942e-05, + "loss": 0.0084, + "step": 3371 + }, + { + "epoch": 4.37922077922078, + "grad_norm": 5.997286975918016, + "learning_rate": 1.247459329498322e-05, + "loss": 0.0145, + "step": 3372 + }, + { + "epoch": 4.380519480519481, + "grad_norm": 5.677700598456514, + "learning_rate": 1.2470517719763576e-05, + "loss": 0.0245, + "step": 3373 + }, + { + "epoch": 4.381818181818182, + "grad_norm": 7.6022436139767935, + "learning_rate": 1.2466441707462065e-05, + "loss": 0.0321, + "step": 3374 + }, + { + "epoch": 4.383116883116883, + "grad_norm": 3.427207910393584, + "learning_rate": 1.2462365258799803e-05, + "loss": 0.0154, + "step": 3375 + }, + { + "epoch": 4.384415584415584, + "grad_norm": 2.894550638463261, + "learning_rate": 1.2458288374498e-05, + "loss": 0.0139, + "step": 3376 + }, + { + "epoch": 4.385714285714286, + "grad_norm": 2.844569204055512, + "learning_rate": 1.2454211055277931e-05, + "loss": 0.0169, + "step": 3377 + }, + { + "epoch": 4.387012987012987, + "grad_norm": 4.7152177308609895, + "learning_rate": 1.2450133301860951e-05, + "loss": 0.0073, + "step": 3378 + }, + { + "epoch": 4.388311688311688, + "grad_norm": 3.269304131931076, + "learning_rate": 1.2446055114968498e-05, + "loss": 0.0129, + "step": 3379 + }, + { + "epoch": 4.3896103896103895, + "grad_norm": 6.490410334058398, + "learning_rate": 1.2441976495322072e-05, + "loss": 0.0199, + "step": 3380 + }, + { + "epoch": 4.390909090909091, + "grad_norm": 4.806468261030809, + "learning_rate": 1.2437897443643267e-05, + "loss": 0.0132, + "step": 3381 + }, + { + "epoch": 4.392207792207792, + "grad_norm": 5.645599132293577, + "learning_rate": 1.243381796065374e-05, + "loss": 0.0137, + "step": 3382 + }, + { + "epoch": 4.393506493506494, + "grad_norm": 5.223532681939227, + "learning_rate": 1.2429738047075235e-05, + "loss": 0.0117, + "step": 3383 + }, + { + "epoch": 4.394805194805195, + "grad_norm": 11.44150140529227, + "learning_rate": 1.242565770362956e-05, + "loss": 0.011, + "step": 3384 + }, + { + "epoch": 4.396103896103896, + "grad_norm": 7.712909430435109, + "learning_rate": 1.2421576931038613e-05, + "loss": 0.0113, + "step": 3385 + }, + { + "epoch": 4.397402597402597, + "grad_norm": 14.570877428069284, + "learning_rate": 1.2417495730024355e-05, + "loss": 0.0427, + "step": 3386 + }, + { + "epoch": 4.398701298701298, + "grad_norm": 7.576054228896691, + "learning_rate": 1.2413414101308834e-05, + "loss": 0.0209, + "step": 3387 + }, + { + "epoch": 4.4, + "grad_norm": 11.029200741247031, + "learning_rate": 1.2409332045614162e-05, + "loss": 0.0348, + "step": 3388 + }, + { + "epoch": 4.4012987012987015, + "grad_norm": 8.583067068366253, + "learning_rate": 1.2405249563662539e-05, + "loss": 0.0313, + "step": 3389 + }, + { + "epoch": 4.402597402597403, + "grad_norm": 7.1861470220039925, + "learning_rate": 1.2401166656176229e-05, + "loss": 0.0199, + "step": 3390 + }, + { + "epoch": 4.403896103896104, + "grad_norm": 8.959390279788376, + "learning_rate": 1.239708332387758e-05, + "loss": 0.0132, + "step": 3391 + }, + { + "epoch": 4.405194805194805, + "grad_norm": 4.737884479026286, + "learning_rate": 1.239299956748901e-05, + "loss": 0.0159, + "step": 3392 + }, + { + "epoch": 4.406493506493507, + "grad_norm": 7.841505354099999, + "learning_rate": 1.2388915387733013e-05, + "loss": 0.0254, + "step": 3393 + }, + { + "epoch": 4.407792207792208, + "grad_norm": 6.768821246849492, + "learning_rate": 1.238483078533216e-05, + "loss": 0.0495, + "step": 3394 + }, + { + "epoch": 4.409090909090909, + "grad_norm": 8.974329124246177, + "learning_rate": 1.2380745761009094e-05, + "loss": 0.0164, + "step": 3395 + }, + { + "epoch": 4.41038961038961, + "grad_norm": 12.86925592348931, + "learning_rate": 1.2376660315486535e-05, + "loss": 0.0231, + "step": 3396 + }, + { + "epoch": 4.411688311688311, + "grad_norm": 4.388492897385481, + "learning_rate": 1.2372574449487281e-05, + "loss": 0.0126, + "step": 3397 + }, + { + "epoch": 4.4129870129870135, + "grad_norm": 16.968114264556363, + "learning_rate": 1.2368488163734188e-05, + "loss": 0.044, + "step": 3398 + }, + { + "epoch": 4.414285714285715, + "grad_norm": 7.8056519777510305, + "learning_rate": 1.2364401458950212e-05, + "loss": 0.0098, + "step": 3399 + }, + { + "epoch": 4.415584415584416, + "grad_norm": 11.833928642985905, + "learning_rate": 1.236031433585836e-05, + "loss": 0.0153, + "step": 3400 + }, + { + "epoch": 4.416883116883117, + "grad_norm": 12.930851151268225, + "learning_rate": 1.2356226795181724e-05, + "loss": 0.0174, + "step": 3401 + }, + { + "epoch": 4.418181818181818, + "grad_norm": 10.105051965683685, + "learning_rate": 1.235213883764347e-05, + "loss": 0.015, + "step": 3402 + }, + { + "epoch": 4.419480519480519, + "grad_norm": 9.37882417353873, + "learning_rate": 1.2348050463966835e-05, + "loss": 0.0256, + "step": 3403 + }, + { + "epoch": 4.420779220779221, + "grad_norm": 12.651455499897388, + "learning_rate": 1.2343961674875133e-05, + "loss": 0.0264, + "step": 3404 + }, + { + "epoch": 4.422077922077922, + "grad_norm": 12.097361718459634, + "learning_rate": 1.2339872471091745e-05, + "loss": 0.0154, + "step": 3405 + }, + { + "epoch": 4.423376623376623, + "grad_norm": 8.233364505975869, + "learning_rate": 1.233578285334013e-05, + "loss": 0.0303, + "step": 3406 + }, + { + "epoch": 4.4246753246753245, + "grad_norm": 4.621372973720773, + "learning_rate": 1.2331692822343823e-05, + "loss": 0.0243, + "step": 3407 + }, + { + "epoch": 4.425974025974026, + "grad_norm": 12.820516578589581, + "learning_rate": 1.2327602378826427e-05, + "loss": 0.0433, + "step": 3408 + }, + { + "epoch": 4.427272727272728, + "grad_norm": 9.417150046095772, + "learning_rate": 1.232351152351162e-05, + "loss": 0.0183, + "step": 3409 + }, + { + "epoch": 4.428571428571429, + "grad_norm": 3.41508945886792, + "learning_rate": 1.2319420257123149e-05, + "loss": 0.0176, + "step": 3410 + }, + { + "epoch": 4.42987012987013, + "grad_norm": 12.905515497452166, + "learning_rate": 1.2315328580384841e-05, + "loss": 0.0133, + "step": 3411 + }, + { + "epoch": 4.431168831168831, + "grad_norm": 4.858137651372793, + "learning_rate": 1.2311236494020593e-05, + "loss": 0.0072, + "step": 3412 + }, + { + "epoch": 4.432467532467532, + "grad_norm": 3.747977134584351, + "learning_rate": 1.2307143998754372e-05, + "loss": 0.0337, + "step": 3413 + }, + { + "epoch": 4.433766233766233, + "grad_norm": 13.628897933662477, + "learning_rate": 1.2303051095310214e-05, + "loss": 0.0157, + "step": 3414 + }, + { + "epoch": 4.435064935064935, + "grad_norm": 11.357191291403787, + "learning_rate": 1.2298957784412241e-05, + "loss": 0.0334, + "step": 3415 + }, + { + "epoch": 4.4363636363636365, + "grad_norm": 10.69684493785464, + "learning_rate": 1.2294864066784631e-05, + "loss": 0.0115, + "step": 3416 + }, + { + "epoch": 4.437662337662338, + "grad_norm": 12.04860815328898, + "learning_rate": 1.2290769943151645e-05, + "loss": 0.0446, + "step": 3417 + }, + { + "epoch": 4.438961038961039, + "grad_norm": 5.21934630659172, + "learning_rate": 1.2286675414237613e-05, + "loss": 0.031, + "step": 3418 + }, + { + "epoch": 4.44025974025974, + "grad_norm": 12.574891853023603, + "learning_rate": 1.2282580480766928e-05, + "loss": 0.0171, + "step": 3419 + }, + { + "epoch": 4.441558441558442, + "grad_norm": 9.322147942411446, + "learning_rate": 1.2278485143464068e-05, + "loss": 0.0115, + "step": 3420 + }, + { + "epoch": 4.442857142857143, + "grad_norm": 1.7042081950763428, + "learning_rate": 1.2274389403053574e-05, + "loss": 0.0114, + "step": 3421 + }, + { + "epoch": 4.444155844155844, + "grad_norm": 14.608946746244484, + "learning_rate": 1.2270293260260066e-05, + "loss": 0.0197, + "step": 3422 + }, + { + "epoch": 4.445454545454545, + "grad_norm": 5.836021369352794, + "learning_rate": 1.2266196715808223e-05, + "loss": 0.0107, + "step": 3423 + }, + { + "epoch": 4.4467532467532465, + "grad_norm": 13.985795677299917, + "learning_rate": 1.2262099770422808e-05, + "loss": 0.0317, + "step": 3424 + }, + { + "epoch": 4.448051948051948, + "grad_norm": 5.068722787370574, + "learning_rate": 1.2258002424828645e-05, + "loss": 0.0201, + "step": 3425 + }, + { + "epoch": 4.44935064935065, + "grad_norm": 11.355553701466103, + "learning_rate": 1.2253904679750637e-05, + "loss": 0.0433, + "step": 3426 + }, + { + "epoch": 4.450649350649351, + "grad_norm": 4.45026096707175, + "learning_rate": 1.2249806535913747e-05, + "loss": 0.0099, + "step": 3427 + }, + { + "epoch": 4.451948051948052, + "grad_norm": 2.6820859658234424, + "learning_rate": 1.2245707994043021e-05, + "loss": 0.0265, + "step": 3428 + }, + { + "epoch": 4.453246753246753, + "grad_norm": 8.95682393046263, + "learning_rate": 1.2241609054863565e-05, + "loss": 0.0134, + "step": 3429 + }, + { + "epoch": 4.454545454545454, + "grad_norm": 6.53139146877564, + "learning_rate": 1.223750971910057e-05, + "loss": 0.0071, + "step": 3430 + }, + { + "epoch": 4.455844155844156, + "grad_norm": 8.572853015425352, + "learning_rate": 1.223340998747927e-05, + "loss": 0.0089, + "step": 3431 + }, + { + "epoch": 4.457142857142857, + "grad_norm": 7.946820541499452, + "learning_rate": 1.2229309860724999e-05, + "loss": 0.0404, + "step": 3432 + }, + { + "epoch": 4.458441558441558, + "grad_norm": 7.099946664479982, + "learning_rate": 1.2225209339563144e-05, + "loss": 0.0611, + "step": 3433 + }, + { + "epoch": 4.45974025974026, + "grad_norm": 10.878323198437304, + "learning_rate": 1.2221108424719166e-05, + "loss": 0.0311, + "step": 3434 + }, + { + "epoch": 4.461038961038961, + "grad_norm": 11.390154978403562, + "learning_rate": 1.2217007116918598e-05, + "loss": 0.0377, + "step": 3435 + }, + { + "epoch": 4.462337662337663, + "grad_norm": 11.647717244546845, + "learning_rate": 1.2212905416887033e-05, + "loss": 0.0244, + "step": 3436 + }, + { + "epoch": 4.463636363636364, + "grad_norm": 7.884399427557095, + "learning_rate": 1.2208803325350146e-05, + "loss": 0.0238, + "step": 3437 + }, + { + "epoch": 4.464935064935065, + "grad_norm": 13.86756051862557, + "learning_rate": 1.2204700843033675e-05, + "loss": 0.007, + "step": 3438 + }, + { + "epoch": 4.466233766233766, + "grad_norm": 4.945316592094499, + "learning_rate": 1.2200597970663426e-05, + "loss": 0.0398, + "step": 3439 + }, + { + "epoch": 4.467532467532467, + "grad_norm": 12.298468044693204, + "learning_rate": 1.2196494708965274e-05, + "loss": 0.0084, + "step": 3440 + }, + { + "epoch": 4.468831168831169, + "grad_norm": 9.475885134400475, + "learning_rate": 1.2192391058665169e-05, + "loss": 0.0126, + "step": 3441 + }, + { + "epoch": 4.47012987012987, + "grad_norm": 7.433176126469104, + "learning_rate": 1.218828702048912e-05, + "loss": 0.0134, + "step": 3442 + }, + { + "epoch": 4.4714285714285715, + "grad_norm": 11.183768625883687, + "learning_rate": 1.2184182595163216e-05, + "loss": 0.0337, + "step": 3443 + }, + { + "epoch": 4.472727272727273, + "grad_norm": 4.63585323194941, + "learning_rate": 1.2180077783413601e-05, + "loss": 0.019, + "step": 3444 + }, + { + "epoch": 4.474025974025974, + "grad_norm": 9.331094650234622, + "learning_rate": 1.2175972585966505e-05, + "loss": 0.0213, + "step": 3445 + }, + { + "epoch": 4.475324675324675, + "grad_norm": 8.195815198965605, + "learning_rate": 1.2171867003548208e-05, + "loss": 0.0187, + "step": 3446 + }, + { + "epoch": 4.476623376623377, + "grad_norm": 7.976348195800345, + "learning_rate": 1.2167761036885067e-05, + "loss": 0.0105, + "step": 3447 + }, + { + "epoch": 4.477922077922078, + "grad_norm": 15.822012269123308, + "learning_rate": 1.2163654686703508e-05, + "loss": 0.0226, + "step": 3448 + }, + { + "epoch": 4.479220779220779, + "grad_norm": 6.611983064157571, + "learning_rate": 1.215954795373002e-05, + "loss": 0.0145, + "step": 3449 + }, + { + "epoch": 4.48051948051948, + "grad_norm": 9.121651780725184, + "learning_rate": 1.2155440838691165e-05, + "loss": 0.0053, + "step": 3450 + }, + { + "epoch": 4.4818181818181815, + "grad_norm": 3.6039456489986104, + "learning_rate": 1.2151333342313572e-05, + "loss": 0.0078, + "step": 3451 + }, + { + "epoch": 4.4831168831168835, + "grad_norm": 8.723316256408207, + "learning_rate": 1.2147225465323932e-05, + "loss": 0.0288, + "step": 3452 + }, + { + "epoch": 4.484415584415585, + "grad_norm": 4.462454782994781, + "learning_rate": 1.2143117208449008e-05, + "loss": 0.0238, + "step": 3453 + }, + { + "epoch": 4.485714285714286, + "grad_norm": 5.800153596076456, + "learning_rate": 1.2139008572415633e-05, + "loss": 0.0261, + "step": 3454 + }, + { + "epoch": 4.487012987012987, + "grad_norm": 9.577325742661499, + "learning_rate": 1.21348995579507e-05, + "loss": 0.0151, + "step": 3455 + }, + { + "epoch": 4.488311688311688, + "grad_norm": 7.168944609484856, + "learning_rate": 1.2130790165781172e-05, + "loss": 0.0152, + "step": 3456 + }, + { + "epoch": 4.489610389610389, + "grad_norm": 14.086317785246584, + "learning_rate": 1.2126680396634081e-05, + "loss": 0.0225, + "step": 3457 + }, + { + "epoch": 4.490909090909091, + "grad_norm": 6.8021060386544185, + "learning_rate": 1.2122570251236522e-05, + "loss": 0.0194, + "step": 3458 + }, + { + "epoch": 4.492207792207792, + "grad_norm": 6.694975432805661, + "learning_rate": 1.211845973031566e-05, + "loss": 0.0223, + "step": 3459 + }, + { + "epoch": 4.4935064935064934, + "grad_norm": 13.601636375590491, + "learning_rate": 1.2114348834598722e-05, + "loss": 0.0182, + "step": 3460 + }, + { + "epoch": 4.494805194805195, + "grad_norm": 5.150890733020374, + "learning_rate": 1.2110237564813008e-05, + "loss": 0.0113, + "step": 3461 + }, + { + "epoch": 4.496103896103896, + "grad_norm": 7.5437637833663755, + "learning_rate": 1.2106125921685877e-05, + "loss": 0.0366, + "step": 3462 + }, + { + "epoch": 4.497402597402598, + "grad_norm": 13.737778573995493, + "learning_rate": 1.2102013905944755e-05, + "loss": 0.0072, + "step": 3463 + }, + { + "epoch": 4.498701298701299, + "grad_norm": 6.6016676701655745, + "learning_rate": 1.2097901518317147e-05, + "loss": 0.0171, + "step": 3464 + }, + { + "epoch": 4.5, + "grad_norm": 15.49150264901704, + "learning_rate": 1.2093788759530603e-05, + "loss": 0.028, + "step": 3465 + }, + { + "epoch": 4.501298701298701, + "grad_norm": 9.893491984843687, + "learning_rate": 1.2089675630312755e-05, + "loss": 0.0173, + "step": 3466 + }, + { + "epoch": 4.502597402597402, + "grad_norm": 5.076383219636045, + "learning_rate": 1.2085562131391284e-05, + "loss": 0.0153, + "step": 3467 + }, + { + "epoch": 4.503896103896103, + "grad_norm": 10.982115252449939, + "learning_rate": 1.2081448263493958e-05, + "loss": 0.0662, + "step": 3468 + }, + { + "epoch": 4.505194805194805, + "grad_norm": 4.344072710602639, + "learning_rate": 1.2077334027348597e-05, + "loss": 0.0131, + "step": 3469 + }, + { + "epoch": 4.5064935064935066, + "grad_norm": 9.757898411103055, + "learning_rate": 1.2073219423683082e-05, + "loss": 0.0161, + "step": 3470 + }, + { + "epoch": 4.507792207792208, + "grad_norm": 13.733387242979793, + "learning_rate": 1.2069104453225374e-05, + "loss": 0.0583, + "step": 3471 + }, + { + "epoch": 4.509090909090909, + "grad_norm": 4.605293915452005, + "learning_rate": 1.2064989116703483e-05, + "loss": 0.0362, + "step": 3472 + }, + { + "epoch": 4.510389610389611, + "grad_norm": 8.202252627552157, + "learning_rate": 1.2060873414845492e-05, + "loss": 0.0309, + "step": 3473 + }, + { + "epoch": 4.511688311688312, + "grad_norm": 10.923391950420905, + "learning_rate": 1.2056757348379549e-05, + "loss": 0.0306, + "step": 3474 + }, + { + "epoch": 4.512987012987013, + "grad_norm": 1.919986716959262, + "learning_rate": 1.2052640918033863e-05, + "loss": 0.0065, + "step": 3475 + }, + { + "epoch": 4.514285714285714, + "grad_norm": 3.6711771749487507, + "learning_rate": 1.2048524124536712e-05, + "loss": 0.0363, + "step": 3476 + }, + { + "epoch": 4.515584415584415, + "grad_norm": 6.391635308384745, + "learning_rate": 1.2044406968616433e-05, + "loss": 0.0464, + "step": 3477 + }, + { + "epoch": 4.5168831168831165, + "grad_norm": 8.16999799601075, + "learning_rate": 1.2040289451001433e-05, + "loss": 0.0267, + "step": 3478 + }, + { + "epoch": 4.5181818181818185, + "grad_norm": 5.252693582040066, + "learning_rate": 1.2036171572420174e-05, + "loss": 0.0312, + "step": 3479 + }, + { + "epoch": 4.51948051948052, + "grad_norm": 7.882672479495417, + "learning_rate": 1.2032053333601195e-05, + "loss": 0.0187, + "step": 3480 + }, + { + "epoch": 4.520779220779221, + "grad_norm": 5.83323551796789, + "learning_rate": 1.2027934735273086e-05, + "loss": 0.0321, + "step": 3481 + }, + { + "epoch": 4.522077922077922, + "grad_norm": 6.6153152942366615, + "learning_rate": 1.2023815778164508e-05, + "loss": 0.0107, + "step": 3482 + }, + { + "epoch": 4.523376623376623, + "grad_norm": 3.6953982435105917, + "learning_rate": 1.201969646300418e-05, + "loss": 0.005, + "step": 3483 + }, + { + "epoch": 4.524675324675325, + "grad_norm": 6.790455256252264, + "learning_rate": 1.2015576790520893e-05, + "loss": 0.0246, + "step": 3484 + }, + { + "epoch": 4.525974025974026, + "grad_norm": 4.343071817279417, + "learning_rate": 1.201145676144349e-05, + "loss": 0.0187, + "step": 3485 + }, + { + "epoch": 4.527272727272727, + "grad_norm": 7.947042339282201, + "learning_rate": 1.200733637650089e-05, + "loss": 0.0239, + "step": 3486 + }, + { + "epoch": 4.5285714285714285, + "grad_norm": 5.802992148569939, + "learning_rate": 1.2003215636422064e-05, + "loss": 0.0252, + "step": 3487 + }, + { + "epoch": 4.52987012987013, + "grad_norm": 5.725103236957299, + "learning_rate": 1.1999094541936047e-05, + "loss": 0.0225, + "step": 3488 + }, + { + "epoch": 4.531168831168831, + "grad_norm": 5.4936908367784465, + "learning_rate": 1.1994973093771945e-05, + "loss": 0.0367, + "step": 3489 + }, + { + "epoch": 4.532467532467533, + "grad_norm": 7.171300718194679, + "learning_rate": 1.1990851292658916e-05, + "loss": 0.0187, + "step": 3490 + }, + { + "epoch": 4.533766233766234, + "grad_norm": 10.015957708042563, + "learning_rate": 1.198672913932619e-05, + "loss": 0.0139, + "step": 3491 + }, + { + "epoch": 4.535064935064935, + "grad_norm": 4.726424562882882, + "learning_rate": 1.198260663450305e-05, + "loss": 0.0268, + "step": 3492 + }, + { + "epoch": 4.536363636363636, + "grad_norm": 7.067059208812518, + "learning_rate": 1.1978483778918855e-05, + "loss": 0.026, + "step": 3493 + }, + { + "epoch": 4.537662337662337, + "grad_norm": 5.605292982475464, + "learning_rate": 1.1974360573303006e-05, + "loss": 0.0288, + "step": 3494 + }, + { + "epoch": 4.538961038961039, + "grad_norm": 4.409873192640926, + "learning_rate": 1.1970237018384983e-05, + "loss": 0.0126, + "step": 3495 + }, + { + "epoch": 4.54025974025974, + "grad_norm": 12.370143532855066, + "learning_rate": 1.196611311489432e-05, + "loss": 0.0129, + "step": 3496 + }, + { + "epoch": 4.541558441558442, + "grad_norm": 4.1456133019161285, + "learning_rate": 1.1961988863560615e-05, + "loss": 0.0251, + "step": 3497 + }, + { + "epoch": 4.542857142857143, + "grad_norm": 5.403432006337554, + "learning_rate": 1.1957864265113528e-05, + "loss": 0.0283, + "step": 3498 + }, + { + "epoch": 4.544155844155844, + "grad_norm": 10.124031454268865, + "learning_rate": 1.1953739320282778e-05, + "loss": 0.0293, + "step": 3499 + }, + { + "epoch": 4.545454545454545, + "grad_norm": 6.465100210748105, + "learning_rate": 1.194961402979815e-05, + "loss": 0.0507, + "step": 3500 + }, + { + "epoch": 4.546753246753247, + "grad_norm": 12.344824312330937, + "learning_rate": 1.1945488394389479e-05, + "loss": 0.0138, + "step": 3501 + }, + { + "epoch": 4.548051948051948, + "grad_norm": 7.542807372861196, + "learning_rate": 1.194136241478668e-05, + "loss": 0.0136, + "step": 3502 + }, + { + "epoch": 4.549350649350649, + "grad_norm": 9.979026170925529, + "learning_rate": 1.1937236091719706e-05, + "loss": 0.0132, + "step": 3503 + }, + { + "epoch": 4.55064935064935, + "grad_norm": 2.2902172825203735, + "learning_rate": 1.1933109425918592e-05, + "loss": 0.0185, + "step": 3504 + }, + { + "epoch": 4.551948051948052, + "grad_norm": 8.56299638629528, + "learning_rate": 1.192898241811342e-05, + "loss": 0.012, + "step": 3505 + }, + { + "epoch": 4.5532467532467535, + "grad_norm": 3.7888813315733203, + "learning_rate": 1.1924855069034336e-05, + "loss": 0.0335, + "step": 3506 + }, + { + "epoch": 4.554545454545455, + "grad_norm": 3.5634427236246724, + "learning_rate": 1.1920727379411547e-05, + "loss": 0.0182, + "step": 3507 + }, + { + "epoch": 4.555844155844156, + "grad_norm": 3.4225830817687637, + "learning_rate": 1.1916599349975326e-05, + "loss": 0.0154, + "step": 3508 + }, + { + "epoch": 4.557142857142857, + "grad_norm": 4.14068690219181, + "learning_rate": 1.191247098145599e-05, + "loss": 0.0195, + "step": 3509 + }, + { + "epoch": 4.558441558441558, + "grad_norm": 3.366020118867262, + "learning_rate": 1.1908342274583937e-05, + "loss": 0.0329, + "step": 3510 + }, + { + "epoch": 4.55974025974026, + "grad_norm": 10.870915242793219, + "learning_rate": 1.1904213230089608e-05, + "loss": 0.0095, + "step": 3511 + }, + { + "epoch": 4.561038961038961, + "grad_norm": 5.233775331354727, + "learning_rate": 1.1900083848703511e-05, + "loss": 0.021, + "step": 3512 + }, + { + "epoch": 4.562337662337662, + "grad_norm": 17.312041875553827, + "learning_rate": 1.1895954131156215e-05, + "loss": 0.0339, + "step": 3513 + }, + { + "epoch": 4.5636363636363635, + "grad_norm": 3.532127680027323, + "learning_rate": 1.1891824078178342e-05, + "loss": 0.008, + "step": 3514 + }, + { + "epoch": 4.564935064935065, + "grad_norm": 10.78379487789464, + "learning_rate": 1.1887693690500583e-05, + "loss": 0.0258, + "step": 3515 + }, + { + "epoch": 4.566233766233767, + "grad_norm": 3.1131574585018016, + "learning_rate": 1.1883562968853677e-05, + "loss": 0.014, + "step": 3516 + }, + { + "epoch": 4.567532467532468, + "grad_norm": 5.83864235546338, + "learning_rate": 1.1879431913968428e-05, + "loss": 0.0112, + "step": 3517 + }, + { + "epoch": 4.568831168831169, + "grad_norm": 10.580147612302815, + "learning_rate": 1.1875300526575704e-05, + "loss": 0.0226, + "step": 3518 + }, + { + "epoch": 4.57012987012987, + "grad_norm": 3.910934403724584, + "learning_rate": 1.1871168807406422e-05, + "loss": 0.0171, + "step": 3519 + }, + { + "epoch": 4.571428571428571, + "grad_norm": 6.466973027696468, + "learning_rate": 1.1867036757191563e-05, + "loss": 0.0349, + "step": 3520 + }, + { + "epoch": 4.572727272727272, + "grad_norm": 11.63618231077521, + "learning_rate": 1.1862904376662167e-05, + "loss": 0.0434, + "step": 3521 + }, + { + "epoch": 4.574025974025974, + "grad_norm": 7.259133437558549, + "learning_rate": 1.185877166654933e-05, + "loss": 0.0267, + "step": 3522 + }, + { + "epoch": 4.5753246753246755, + "grad_norm": 14.718125322575377, + "learning_rate": 1.1854638627584208e-05, + "loss": 0.0334, + "step": 3523 + }, + { + "epoch": 4.576623376623377, + "grad_norm": 9.817599108876337, + "learning_rate": 1.1850505260498013e-05, + "loss": 0.038, + "step": 3524 + }, + { + "epoch": 4.577922077922078, + "grad_norm": 13.471111171757757, + "learning_rate": 1.1846371566022021e-05, + "loss": 0.0155, + "step": 3525 + }, + { + "epoch": 4.579220779220779, + "grad_norm": 24.292506971008724, + "learning_rate": 1.1842237544887558e-05, + "loss": 0.0205, + "step": 3526 + }, + { + "epoch": 4.580519480519481, + "grad_norm": 13.60200062452595, + "learning_rate": 1.1838103197826011e-05, + "loss": 0.044, + "step": 3527 + }, + { + "epoch": 4.581818181818182, + "grad_norm": 12.592157548459578, + "learning_rate": 1.183396852556883e-05, + "loss": 0.0163, + "step": 3528 + }, + { + "epoch": 4.583116883116883, + "grad_norm": 8.760390696631, + "learning_rate": 1.1829833528847513e-05, + "loss": 0.0242, + "step": 3529 + }, + { + "epoch": 4.584415584415584, + "grad_norm": 10.261994758990799, + "learning_rate": 1.182569820839362e-05, + "loss": 0.0394, + "step": 3530 + }, + { + "epoch": 4.585714285714285, + "grad_norm": 14.178254353047315, + "learning_rate": 1.1821562564938768e-05, + "loss": 0.023, + "step": 3531 + }, + { + "epoch": 4.5870129870129865, + "grad_norm": 8.628808892116439, + "learning_rate": 1.1817426599214637e-05, + "loss": 0.0294, + "step": 3532 + }, + { + "epoch": 4.588311688311689, + "grad_norm": 12.15753519320131, + "learning_rate": 1.1813290311952951e-05, + "loss": 0.0236, + "step": 3533 + }, + { + "epoch": 4.58961038961039, + "grad_norm": 8.146420406577715, + "learning_rate": 1.1809153703885506e-05, + "loss": 0.0481, + "step": 3534 + }, + { + "epoch": 4.590909090909091, + "grad_norm": 6.949582360798295, + "learning_rate": 1.180501677574414e-05, + "loss": 0.0311, + "step": 3535 + }, + { + "epoch": 4.592207792207792, + "grad_norm": 10.235644983258032, + "learning_rate": 1.1800879528260761e-05, + "loss": 0.0399, + "step": 3536 + }, + { + "epoch": 4.593506493506494, + "grad_norm": 6.223068751487788, + "learning_rate": 1.1796741962167322e-05, + "loss": 0.0207, + "step": 3537 + }, + { + "epoch": 4.594805194805195, + "grad_norm": 8.140961690873056, + "learning_rate": 1.179260407819584e-05, + "loss": 0.0137, + "step": 3538 + }, + { + "epoch": 4.596103896103896, + "grad_norm": 4.431048082347942, + "learning_rate": 1.1788465877078385e-05, + "loss": 0.0077, + "step": 3539 + }, + { + "epoch": 4.597402597402597, + "grad_norm": 8.70746671742676, + "learning_rate": 1.1784327359547087e-05, + "loss": 0.0087, + "step": 3540 + }, + { + "epoch": 4.5987012987012985, + "grad_norm": 4.079649902681223, + "learning_rate": 1.1780188526334127e-05, + "loss": 0.0298, + "step": 3541 + }, + { + "epoch": 4.6, + "grad_norm": 3.1862481738265784, + "learning_rate": 1.1776049378171742e-05, + "loss": 0.0323, + "step": 3542 + }, + { + "epoch": 4.601298701298702, + "grad_norm": 5.7384362958096435, + "learning_rate": 1.177190991579223e-05, + "loss": 0.0475, + "step": 3543 + }, + { + "epoch": 4.602597402597403, + "grad_norm": 6.218431313011847, + "learning_rate": 1.1767770139927936e-05, + "loss": 0.0237, + "step": 3544 + }, + { + "epoch": 4.603896103896104, + "grad_norm": 13.843998429715551, + "learning_rate": 1.176363005131127e-05, + "loss": 0.0612, + "step": 3545 + }, + { + "epoch": 4.605194805194805, + "grad_norm": 6.957443144733935, + "learning_rate": 1.175948965067469e-05, + "loss": 0.0346, + "step": 3546 + }, + { + "epoch": 4.606493506493506, + "grad_norm": 4.619992142355319, + "learning_rate": 1.1755348938750719e-05, + "loss": 0.0102, + "step": 3547 + }, + { + "epoch": 4.607792207792208, + "grad_norm": 4.794771219276056, + "learning_rate": 1.1751207916271918e-05, + "loss": 0.0132, + "step": 3548 + }, + { + "epoch": 4.609090909090909, + "grad_norm": 3.617203285563049, + "learning_rate": 1.1747066583970921e-05, + "loss": 0.0117, + "step": 3549 + }, + { + "epoch": 4.6103896103896105, + "grad_norm": 6.461101176985594, + "learning_rate": 1.1742924942580405e-05, + "loss": 0.046, + "step": 3550 + }, + { + "epoch": 4.611688311688312, + "grad_norm": 7.639140843344469, + "learning_rate": 1.1738782992833109e-05, + "loss": 0.0139, + "step": 3551 + }, + { + "epoch": 4.612987012987013, + "grad_norm": 8.113528842784298, + "learning_rate": 1.1734640735461815e-05, + "loss": 0.0141, + "step": 3552 + }, + { + "epoch": 4.614285714285714, + "grad_norm": 4.068535767725405, + "learning_rate": 1.1730498171199379e-05, + "loss": 0.0335, + "step": 3553 + }, + { + "epoch": 4.615584415584416, + "grad_norm": 12.067607816142162, + "learning_rate": 1.1726355300778693e-05, + "loss": 0.0236, + "step": 3554 + }, + { + "epoch": 4.616883116883117, + "grad_norm": 7.621032060632052, + "learning_rate": 1.172221212493271e-05, + "loss": 0.0212, + "step": 3555 + }, + { + "epoch": 4.618181818181818, + "grad_norm": 8.722949038378607, + "learning_rate": 1.1718068644394443e-05, + "loss": 0.0501, + "step": 3556 + }, + { + "epoch": 4.619480519480519, + "grad_norm": 6.967648988757764, + "learning_rate": 1.1713924859896944e-05, + "loss": 0.0221, + "step": 3557 + }, + { + "epoch": 4.62077922077922, + "grad_norm": 10.096535433454862, + "learning_rate": 1.1709780772173334e-05, + "loss": 0.0364, + "step": 3558 + }, + { + "epoch": 4.6220779220779225, + "grad_norm": 6.518856660832095, + "learning_rate": 1.1705636381956777e-05, + "loss": 0.0213, + "step": 3559 + }, + { + "epoch": 4.623376623376624, + "grad_norm": 3.387309244548851, + "learning_rate": 1.1701491689980503e-05, + "loss": 0.0124, + "step": 3560 + }, + { + "epoch": 4.624675324675325, + "grad_norm": 7.409719204859372, + "learning_rate": 1.1697346696977779e-05, + "loss": 0.0187, + "step": 3561 + }, + { + "epoch": 4.625974025974026, + "grad_norm": 4.210081898378394, + "learning_rate": 1.1693201403681938e-05, + "loss": 0.02, + "step": 3562 + }, + { + "epoch": 4.627272727272727, + "grad_norm": 4.083527857283041, + "learning_rate": 1.1689055810826358e-05, + "loss": 0.0102, + "step": 3563 + }, + { + "epoch": 4.628571428571428, + "grad_norm": 6.624011758783772, + "learning_rate": 1.1684909919144477e-05, + "loss": 0.0246, + "step": 3564 + }, + { + "epoch": 4.62987012987013, + "grad_norm": 4.881355544574863, + "learning_rate": 1.1680763729369783e-05, + "loss": 0.013, + "step": 3565 + }, + { + "epoch": 4.631168831168831, + "grad_norm": 6.104527932863266, + "learning_rate": 1.1676617242235815e-05, + "loss": 0.0397, + "step": 3566 + }, + { + "epoch": 4.632467532467532, + "grad_norm": 4.153967781191553, + "learning_rate": 1.1672470458476166e-05, + "loss": 0.0415, + "step": 3567 + }, + { + "epoch": 4.6337662337662335, + "grad_norm": 5.8696325491040175, + "learning_rate": 1.1668323378824482e-05, + "loss": 0.0146, + "step": 3568 + }, + { + "epoch": 4.6350649350649356, + "grad_norm": 3.8523852414769255, + "learning_rate": 1.166417600401446e-05, + "loss": 0.0182, + "step": 3569 + }, + { + "epoch": 4.636363636363637, + "grad_norm": 5.364508302596048, + "learning_rate": 1.1660028334779847e-05, + "loss": 0.0126, + "step": 3570 + }, + { + "epoch": 4.637662337662338, + "grad_norm": 6.473938044650396, + "learning_rate": 1.1655880371854454e-05, + "loss": 0.0363, + "step": 3571 + }, + { + "epoch": 4.638961038961039, + "grad_norm": 7.037028528153, + "learning_rate": 1.1651732115972126e-05, + "loss": 0.0317, + "step": 3572 + }, + { + "epoch": 4.64025974025974, + "grad_norm": 7.885947340788396, + "learning_rate": 1.1647583567866772e-05, + "loss": 0.0338, + "step": 3573 + }, + { + "epoch": 4.641558441558441, + "grad_norm": 7.739617843796201, + "learning_rate": 1.1643434728272349e-05, + "loss": 0.0272, + "step": 3574 + }, + { + "epoch": 4.642857142857143, + "grad_norm": 5.945226375616342, + "learning_rate": 1.1639285597922871e-05, + "loss": 0.0337, + "step": 3575 + }, + { + "epoch": 4.644155844155844, + "grad_norm": 11.35002608727304, + "learning_rate": 1.163513617755239e-05, + "loss": 0.0325, + "step": 3576 + }, + { + "epoch": 4.6454545454545455, + "grad_norm": 7.946144039446131, + "learning_rate": 1.1630986467895024e-05, + "loss": 0.0511, + "step": 3577 + }, + { + "epoch": 4.646753246753247, + "grad_norm": 13.906120130776864, + "learning_rate": 1.1626836469684935e-05, + "loss": 0.0461, + "step": 3578 + }, + { + "epoch": 4.648051948051948, + "grad_norm": 3.1810598471983083, + "learning_rate": 1.1622686183656338e-05, + "loss": 0.0188, + "step": 3579 + }, + { + "epoch": 4.64935064935065, + "grad_norm": 5.379808450132123, + "learning_rate": 1.1618535610543491e-05, + "loss": 0.0197, + "step": 3580 + }, + { + "epoch": 4.650649350649351, + "grad_norm": 17.523571763026577, + "learning_rate": 1.1614384751080721e-05, + "loss": 0.0438, + "step": 3581 + }, + { + "epoch": 4.651948051948052, + "grad_norm": 6.20608691074639, + "learning_rate": 1.1610233606002388e-05, + "loss": 0.0564, + "step": 3582 + }, + { + "epoch": 4.653246753246753, + "grad_norm": 17.670294889366815, + "learning_rate": 1.1606082176042908e-05, + "loss": 0.0329, + "step": 3583 + }, + { + "epoch": 4.654545454545454, + "grad_norm": 10.675766250146937, + "learning_rate": 1.1601930461936752e-05, + "loss": 0.016, + "step": 3584 + }, + { + "epoch": 4.6558441558441555, + "grad_norm": 13.425853850127993, + "learning_rate": 1.1597778464418435e-05, + "loss": 0.0321, + "step": 3585 + }, + { + "epoch": 4.6571428571428575, + "grad_norm": 15.051003451794253, + "learning_rate": 1.1593626184222527e-05, + "loss": 0.0341, + "step": 3586 + }, + { + "epoch": 4.658441558441559, + "grad_norm": 7.429756743529231, + "learning_rate": 1.1589473622083642e-05, + "loss": 0.0388, + "step": 3587 + }, + { + "epoch": 4.65974025974026, + "grad_norm": 10.03701095488468, + "learning_rate": 1.1585320778736452e-05, + "loss": 0.0226, + "step": 3588 + }, + { + "epoch": 4.661038961038961, + "grad_norm": 24.766052943511614, + "learning_rate": 1.1581167654915673e-05, + "loss": 0.0311, + "step": 3589 + }, + { + "epoch": 4.662337662337662, + "grad_norm": 5.750103651827445, + "learning_rate": 1.1577014251356073e-05, + "loss": 0.0152, + "step": 3590 + }, + { + "epoch": 4.663636363636364, + "grad_norm": 26.407084253086584, + "learning_rate": 1.1572860568792466e-05, + "loss": 0.0325, + "step": 3591 + }, + { + "epoch": 4.664935064935065, + "grad_norm": 11.37411785649728, + "learning_rate": 1.156870660795972e-05, + "loss": 0.0406, + "step": 3592 + }, + { + "epoch": 4.666233766233766, + "grad_norm": 19.99000863612259, + "learning_rate": 1.1564552369592748e-05, + "loss": 0.0218, + "step": 3593 + }, + { + "epoch": 4.667532467532467, + "grad_norm": 22.6596629882395, + "learning_rate": 1.1560397854426519e-05, + "loss": 0.0146, + "step": 3594 + }, + { + "epoch": 4.6688311688311686, + "grad_norm": 5.374875156263335, + "learning_rate": 1.155624306319604e-05, + "loss": 0.0219, + "step": 3595 + }, + { + "epoch": 4.67012987012987, + "grad_norm": 26.95150337712216, + "learning_rate": 1.1552087996636381e-05, + "loss": 0.0432, + "step": 3596 + }, + { + "epoch": 4.671428571428572, + "grad_norm": 8.932862143821204, + "learning_rate": 1.1547932655482644e-05, + "loss": 0.0224, + "step": 3597 + }, + { + "epoch": 4.672727272727273, + "grad_norm": 15.99948512950156, + "learning_rate": 1.1543777040469994e-05, + "loss": 0.0288, + "step": 3598 + }, + { + "epoch": 4.674025974025974, + "grad_norm": 26.560626927156168, + "learning_rate": 1.1539621152333635e-05, + "loss": 0.0191, + "step": 3599 + }, + { + "epoch": 4.675324675324675, + "grad_norm": 7.1069592563008674, + "learning_rate": 1.1535464991808826e-05, + "loss": 0.0234, + "step": 3600 + }, + { + "epoch": 4.676623376623376, + "grad_norm": 23.767017506060874, + "learning_rate": 1.1531308559630876e-05, + "loss": 0.0307, + "step": 3601 + }, + { + "epoch": 4.677922077922078, + "grad_norm": 22.732312908691608, + "learning_rate": 1.1527151856535125e-05, + "loss": 0.0311, + "step": 3602 + }, + { + "epoch": 4.679220779220779, + "grad_norm": 14.787367837842753, + "learning_rate": 1.1522994883256989e-05, + "loss": 0.0431, + "step": 3603 + }, + { + "epoch": 4.6805194805194805, + "grad_norm": 24.75812216432321, + "learning_rate": 1.1518837640531898e-05, + "loss": 0.0289, + "step": 3604 + }, + { + "epoch": 4.681818181818182, + "grad_norm": 9.537819099223505, + "learning_rate": 1.1514680129095365e-05, + "loss": 0.0354, + "step": 3605 + }, + { + "epoch": 4.683116883116883, + "grad_norm": 16.72476763223255, + "learning_rate": 1.1510522349682922e-05, + "loss": 0.0208, + "step": 3606 + }, + { + "epoch": 4.684415584415584, + "grad_norm": 29.99368184647098, + "learning_rate": 1.1506364303030166e-05, + "loss": 0.0317, + "step": 3607 + }, + { + "epoch": 4.685714285714286, + "grad_norm": 6.927175252369358, + "learning_rate": 1.1502205989872735e-05, + "loss": 0.0173, + "step": 3608 + }, + { + "epoch": 4.687012987012987, + "grad_norm": 25.343413196073037, + "learning_rate": 1.1498047410946307e-05, + "loss": 0.0145, + "step": 3609 + }, + { + "epoch": 4.688311688311688, + "grad_norm": 8.692286074237199, + "learning_rate": 1.149388856698662e-05, + "loss": 0.0142, + "step": 3610 + }, + { + "epoch": 4.689610389610389, + "grad_norm": 13.7659173044065, + "learning_rate": 1.1489729458729453e-05, + "loss": 0.015, + "step": 3611 + }, + { + "epoch": 4.690909090909091, + "grad_norm": 16.345463545519674, + "learning_rate": 1.148557008691063e-05, + "loss": 0.0176, + "step": 3612 + }, + { + "epoch": 4.6922077922077925, + "grad_norm": 8.373015882120583, + "learning_rate": 1.1481410452266026e-05, + "loss": 0.0537, + "step": 3613 + }, + { + "epoch": 4.693506493506494, + "grad_norm": 12.18239693075628, + "learning_rate": 1.1477250555531559e-05, + "loss": 0.0197, + "step": 3614 + }, + { + "epoch": 4.694805194805195, + "grad_norm": 11.828418271286758, + "learning_rate": 1.1473090397443191e-05, + "loss": 0.0208, + "step": 3615 + }, + { + "epoch": 4.696103896103896, + "grad_norm": 3.025234871772514, + "learning_rate": 1.1468929978736939e-05, + "loss": 0.0092, + "step": 3616 + }, + { + "epoch": 4.697402597402597, + "grad_norm": 13.357629867944166, + "learning_rate": 1.1464769300148857e-05, + "loss": 0.0077, + "step": 3617 + }, + { + "epoch": 4.698701298701299, + "grad_norm": 8.729710678805564, + "learning_rate": 1.1460608362415051e-05, + "loss": 0.0276, + "step": 3618 + }, + { + "epoch": 4.7, + "grad_norm": 6.27328579646645, + "learning_rate": 1.1456447166271668e-05, + "loss": 0.0153, + "step": 3619 + }, + { + "epoch": 4.701298701298701, + "grad_norm": 10.903954912907045, + "learning_rate": 1.1452285712454905e-05, + "loss": 0.0519, + "step": 3620 + }, + { + "epoch": 4.702597402597402, + "grad_norm": 8.4283734856693, + "learning_rate": 1.1448124001701002e-05, + "loss": 0.0097, + "step": 3621 + }, + { + "epoch": 4.703896103896104, + "grad_norm": 11.255216612430699, + "learning_rate": 1.1443962034746246e-05, + "loss": 0.0323, + "step": 3622 + }, + { + "epoch": 4.705194805194806, + "grad_norm": 7.214266686105094, + "learning_rate": 1.143979981232697e-05, + "loss": 0.0352, + "step": 3623 + }, + { + "epoch": 4.706493506493507, + "grad_norm": 4.930338248515531, + "learning_rate": 1.1435637335179545e-05, + "loss": 0.0149, + "step": 3624 + }, + { + "epoch": 4.707792207792208, + "grad_norm": 10.614818363954056, + "learning_rate": 1.1431474604040402e-05, + "loss": 0.0311, + "step": 3625 + }, + { + "epoch": 4.709090909090909, + "grad_norm": 6.770443106519271, + "learning_rate": 1.1427311619646e-05, + "loss": 0.0056, + "step": 3626 + }, + { + "epoch": 4.71038961038961, + "grad_norm": 6.615545484024201, + "learning_rate": 1.1423148382732854e-05, + "loss": 0.0113, + "step": 3627 + }, + { + "epoch": 4.711688311688311, + "grad_norm": 5.327288037106743, + "learning_rate": 1.141898489403752e-05, + "loss": 0.0221, + "step": 3628 + }, + { + "epoch": 4.712987012987013, + "grad_norm": 3.2044960165505945, + "learning_rate": 1.1414821154296598e-05, + "loss": 0.0149, + "step": 3629 + }, + { + "epoch": 4.714285714285714, + "grad_norm": 7.62950312651285, + "learning_rate": 1.1410657164246732e-05, + "loss": 0.0272, + "step": 3630 + }, + { + "epoch": 4.7155844155844155, + "grad_norm": 2.9826404598970218, + "learning_rate": 1.1406492924624614e-05, + "loss": 0.0037, + "step": 3631 + }, + { + "epoch": 4.716883116883117, + "grad_norm": 6.028847742795414, + "learning_rate": 1.1402328436166981e-05, + "loss": 0.0204, + "step": 3632 + }, + { + "epoch": 4.718181818181818, + "grad_norm": 7.21850534781178, + "learning_rate": 1.1398163699610602e-05, + "loss": 0.0124, + "step": 3633 + }, + { + "epoch": 4.71948051948052, + "grad_norm": 4.194605747473362, + "learning_rate": 1.1393998715692307e-05, + "loss": 0.0313, + "step": 3634 + }, + { + "epoch": 4.720779220779221, + "grad_norm": 6.464104933290254, + "learning_rate": 1.1389833485148955e-05, + "loss": 0.0176, + "step": 3635 + }, + { + "epoch": 4.722077922077922, + "grad_norm": 9.62704284254681, + "learning_rate": 1.138566800871746e-05, + "loss": 0.0519, + "step": 3636 + }, + { + "epoch": 4.723376623376623, + "grad_norm": 4.587896249478887, + "learning_rate": 1.1381502287134772e-05, + "loss": 0.0047, + "step": 3637 + }, + { + "epoch": 4.724675324675324, + "grad_norm": 5.3645289386445505, + "learning_rate": 1.1377336321137889e-05, + "loss": 0.0198, + "step": 3638 + }, + { + "epoch": 4.7259740259740255, + "grad_norm": 9.93891468485946, + "learning_rate": 1.1373170111463843e-05, + "loss": 0.0075, + "step": 3639 + }, + { + "epoch": 4.7272727272727275, + "grad_norm": 4.507363058650456, + "learning_rate": 1.1369003658849727e-05, + "loss": 0.0291, + "step": 3640 + }, + { + "epoch": 4.728571428571429, + "grad_norm": 2.4276443822703366, + "learning_rate": 1.1364836964032658e-05, + "loss": 0.0096, + "step": 3641 + }, + { + "epoch": 4.72987012987013, + "grad_norm": 4.519556547788064, + "learning_rate": 1.1360670027749806e-05, + "loss": 0.0099, + "step": 3642 + }, + { + "epoch": 4.731168831168831, + "grad_norm": 4.975389979730784, + "learning_rate": 1.1356502850738382e-05, + "loss": 0.0129, + "step": 3643 + }, + { + "epoch": 4.732467532467533, + "grad_norm": 5.831213002806466, + "learning_rate": 1.1352335433735644e-05, + "loss": 0.0166, + "step": 3644 + }, + { + "epoch": 4.733766233766234, + "grad_norm": 7.453698330015692, + "learning_rate": 1.134816777747888e-05, + "loss": 0.0122, + "step": 3645 + }, + { + "epoch": 4.735064935064935, + "grad_norm": 6.191250335517364, + "learning_rate": 1.1343999882705434e-05, + "loss": 0.0251, + "step": 3646 + }, + { + "epoch": 4.736363636363636, + "grad_norm": 2.785944785063652, + "learning_rate": 1.1339831750152684e-05, + "loss": 0.0178, + "step": 3647 + }, + { + "epoch": 4.7376623376623375, + "grad_norm": 9.142680082356197, + "learning_rate": 1.1335663380558054e-05, + "loss": 0.0223, + "step": 3648 + }, + { + "epoch": 4.738961038961039, + "grad_norm": 5.328082360428813, + "learning_rate": 1.1331494774659007e-05, + "loss": 0.0428, + "step": 3649 + }, + { + "epoch": 4.740259740259741, + "grad_norm": 9.2385940503366, + "learning_rate": 1.1327325933193045e-05, + "loss": 0.0107, + "step": 3650 + }, + { + "epoch": 4.741558441558442, + "grad_norm": 5.08007693532044, + "learning_rate": 1.1323156856897723e-05, + "loss": 0.0168, + "step": 3651 + }, + { + "epoch": 4.742857142857143, + "grad_norm": 5.372572098730338, + "learning_rate": 1.1318987546510627e-05, + "loss": 0.0152, + "step": 3652 + }, + { + "epoch": 4.744155844155844, + "grad_norm": 10.857638302338781, + "learning_rate": 1.131481800276939e-05, + "loss": 0.0161, + "step": 3653 + }, + { + "epoch": 4.745454545454545, + "grad_norm": 3.6742359719006905, + "learning_rate": 1.1310648226411681e-05, + "loss": 0.0195, + "step": 3654 + }, + { + "epoch": 4.746753246753247, + "grad_norm": 9.87657702850541, + "learning_rate": 1.1306478218175217e-05, + "loss": 0.0316, + "step": 3655 + }, + { + "epoch": 4.748051948051948, + "grad_norm": 8.265122413897325, + "learning_rate": 1.1302307978797744e-05, + "loss": 0.0064, + "step": 3656 + }, + { + "epoch": 4.749350649350649, + "grad_norm": 5.391408079939828, + "learning_rate": 1.129813750901707e-05, + "loss": 0.0227, + "step": 3657 + }, + { + "epoch": 4.750649350649351, + "grad_norm": 8.882372953700727, + "learning_rate": 1.129396680957102e-05, + "loss": 0.0256, + "step": 3658 + }, + { + "epoch": 4.751948051948052, + "grad_norm": 10.663513443510674, + "learning_rate": 1.1289795881197477e-05, + "loss": 0.0325, + "step": 3659 + }, + { + "epoch": 4.753246753246753, + "grad_norm": 6.964049647103429, + "learning_rate": 1.1285624724634352e-05, + "loss": 0.0067, + "step": 3660 + }, + { + "epoch": 4.754545454545455, + "grad_norm": 8.006907214159886, + "learning_rate": 1.128145334061961e-05, + "loss": 0.0245, + "step": 3661 + }, + { + "epoch": 4.755844155844156, + "grad_norm": 5.4384822369043535, + "learning_rate": 1.1277281729891241e-05, + "loss": 0.0526, + "step": 3662 + }, + { + "epoch": 4.757142857142857, + "grad_norm": 15.630945928384019, + "learning_rate": 1.127310989318729e-05, + "loss": 0.0198, + "step": 3663 + }, + { + "epoch": 4.758441558441558, + "grad_norm": 9.352869510655669, + "learning_rate": 1.126893783124583e-05, + "loss": 0.0132, + "step": 3664 + }, + { + "epoch": 4.759740259740259, + "grad_norm": 14.802483654219484, + "learning_rate": 1.1264765544804982e-05, + "loss": 0.0263, + "step": 3665 + }, + { + "epoch": 4.761038961038961, + "grad_norm": 16.537461754955714, + "learning_rate": 1.1260593034602896e-05, + "loss": 0.0307, + "step": 3666 + }, + { + "epoch": 4.7623376623376625, + "grad_norm": 8.74666588947174, + "learning_rate": 1.1256420301377778e-05, + "loss": 0.009, + "step": 3667 + }, + { + "epoch": 4.763636363636364, + "grad_norm": 17.445030319309936, + "learning_rate": 1.125224734586786e-05, + "loss": 0.0093, + "step": 3668 + }, + { + "epoch": 4.764935064935065, + "grad_norm": 2.220817456587295, + "learning_rate": 1.1248074168811415e-05, + "loss": 0.006, + "step": 3669 + }, + { + "epoch": 4.766233766233766, + "grad_norm": 13.255209106712533, + "learning_rate": 1.1243900770946767e-05, + "loss": 0.0403, + "step": 3670 + }, + { + "epoch": 4.767532467532467, + "grad_norm": 16.605026592174227, + "learning_rate": 1.123972715301226e-05, + "loss": 0.0383, + "step": 3671 + }, + { + "epoch": 4.768831168831169, + "grad_norm": 11.580244238842953, + "learning_rate": 1.1235553315746293e-05, + "loss": 0.066, + "step": 3672 + }, + { + "epoch": 4.77012987012987, + "grad_norm": 24.328816714158375, + "learning_rate": 1.1231379259887294e-05, + "loss": 0.0474, + "step": 3673 + }, + { + "epoch": 4.771428571428571, + "grad_norm": 9.585829312765, + "learning_rate": 1.1227204986173739e-05, + "loss": 0.017, + "step": 3674 + }, + { + "epoch": 4.7727272727272725, + "grad_norm": 22.06097928906854, + "learning_rate": 1.1223030495344127e-05, + "loss": 0.0473, + "step": 3675 + }, + { + "epoch": 4.7740259740259745, + "grad_norm": 13.472705814543339, + "learning_rate": 1.1218855788137016e-05, + "loss": 0.0128, + "step": 3676 + }, + { + "epoch": 4.775324675324676, + "grad_norm": 12.121065793944172, + "learning_rate": 1.1214680865290988e-05, + "loss": 0.0127, + "step": 3677 + }, + { + "epoch": 4.776623376623377, + "grad_norm": 21.648674035849766, + "learning_rate": 1.121050572754466e-05, + "loss": 0.0146, + "step": 3678 + }, + { + "epoch": 4.777922077922078, + "grad_norm": 3.96234965344649, + "learning_rate": 1.1206330375636705e-05, + "loss": 0.0392, + "step": 3679 + }, + { + "epoch": 4.779220779220779, + "grad_norm": 16.27064309994212, + "learning_rate": 1.1202154810305812e-05, + "loss": 0.0206, + "step": 3680 + }, + { + "epoch": 4.78051948051948, + "grad_norm": 16.932188135664262, + "learning_rate": 1.1197979032290728e-05, + "loss": 0.0514, + "step": 3681 + }, + { + "epoch": 4.781818181818182, + "grad_norm": 13.03615252298017, + "learning_rate": 1.119380304233022e-05, + "loss": 0.0383, + "step": 3682 + }, + { + "epoch": 4.783116883116883, + "grad_norm": 21.738019919528792, + "learning_rate": 1.1189626841163109e-05, + "loss": 0.0563, + "step": 3683 + }, + { + "epoch": 4.7844155844155845, + "grad_norm": 14.946473044720692, + "learning_rate": 1.1185450429528236e-05, + "loss": 0.03, + "step": 3684 + }, + { + "epoch": 4.785714285714286, + "grad_norm": 9.056524916391169, + "learning_rate": 1.1181273808164494e-05, + "loss": 0.0348, + "step": 3685 + }, + { + "epoch": 4.787012987012987, + "grad_norm": 16.68879887699053, + "learning_rate": 1.1177096977810803e-05, + "loss": 0.0192, + "step": 3686 + }, + { + "epoch": 4.788311688311689, + "grad_norm": 6.896750166300825, + "learning_rate": 1.117291993920613e-05, + "loss": 0.0137, + "step": 3687 + }, + { + "epoch": 4.78961038961039, + "grad_norm": 9.166496766509141, + "learning_rate": 1.1168742693089468e-05, + "loss": 0.042, + "step": 3688 + }, + { + "epoch": 4.790909090909091, + "grad_norm": 12.22076152143448, + "learning_rate": 1.1164565240199854e-05, + "loss": 0.0092, + "step": 3689 + }, + { + "epoch": 4.792207792207792, + "grad_norm": 4.487467912065037, + "learning_rate": 1.1160387581276362e-05, + "loss": 0.0198, + "step": 3690 + }, + { + "epoch": 4.793506493506493, + "grad_norm": 15.056275388986785, + "learning_rate": 1.115620971705809e-05, + "loss": 0.0163, + "step": 3691 + }, + { + "epoch": 4.794805194805194, + "grad_norm": 7.198295409646036, + "learning_rate": 1.1152031648284195e-05, + "loss": 0.0235, + "step": 3692 + }, + { + "epoch": 4.796103896103896, + "grad_norm": 18.56821566492657, + "learning_rate": 1.1147853375693849e-05, + "loss": 0.0433, + "step": 3693 + }, + { + "epoch": 4.797402597402598, + "grad_norm": 5.0157677942329215, + "learning_rate": 1.1143674900026271e-05, + "loss": 0.0191, + "step": 3694 + }, + { + "epoch": 4.798701298701299, + "grad_norm": 18.071634772344048, + "learning_rate": 1.1139496222020712e-05, + "loss": 0.0463, + "step": 3695 + }, + { + "epoch": 4.8, + "grad_norm": 8.393541383446482, + "learning_rate": 1.1135317342416464e-05, + "loss": 0.0291, + "step": 3696 + }, + { + "epoch": 4.801298701298701, + "grad_norm": 8.506426435552823, + "learning_rate": 1.1131138261952845e-05, + "loss": 0.0363, + "step": 3697 + }, + { + "epoch": 4.802597402597403, + "grad_norm": 15.698355902128462, + "learning_rate": 1.1126958981369215e-05, + "loss": 0.0334, + "step": 3698 + }, + { + "epoch": 4.803896103896104, + "grad_norm": 3.9106621686828427, + "learning_rate": 1.1122779501404977e-05, + "loss": 0.0063, + "step": 3699 + }, + { + "epoch": 4.805194805194805, + "grad_norm": 13.058395695106551, + "learning_rate": 1.1118599822799551e-05, + "loss": 0.038, + "step": 3700 + }, + { + "epoch": 4.806493506493506, + "grad_norm": 9.600308597431136, + "learning_rate": 1.1114419946292407e-05, + "loss": 0.0406, + "step": 3701 + }, + { + "epoch": 4.8077922077922075, + "grad_norm": 2.652520882083966, + "learning_rate": 1.1110239872623045e-05, + "loss": 0.0118, + "step": 3702 + }, + { + "epoch": 4.809090909090909, + "grad_norm": 14.262762454119441, + "learning_rate": 1.1106059602530998e-05, + "loss": 0.0167, + "step": 3703 + }, + { + "epoch": 4.810389610389611, + "grad_norm": 7.105846754487564, + "learning_rate": 1.1101879136755838e-05, + "loss": 0.0165, + "step": 3704 + }, + { + "epoch": 4.811688311688312, + "grad_norm": 4.114455264359345, + "learning_rate": 1.1097698476037172e-05, + "loss": 0.012, + "step": 3705 + }, + { + "epoch": 4.812987012987013, + "grad_norm": 10.190434717028152, + "learning_rate": 1.1093517621114628e-05, + "loss": 0.0137, + "step": 3706 + }, + { + "epoch": 4.814285714285714, + "grad_norm": 6.600753195722137, + "learning_rate": 1.1089336572727893e-05, + "loss": 0.0404, + "step": 3707 + }, + { + "epoch": 4.815584415584416, + "grad_norm": 11.617118988453223, + "learning_rate": 1.1085155331616664e-05, + "loss": 0.0087, + "step": 3708 + }, + { + "epoch": 4.816883116883117, + "grad_norm": 9.521386050391758, + "learning_rate": 1.1080973898520691e-05, + "loss": 0.03, + "step": 3709 + }, + { + "epoch": 4.818181818181818, + "grad_norm": 7.003201803120101, + "learning_rate": 1.1076792274179739e-05, + "loss": 0.0379, + "step": 3710 + }, + { + "epoch": 4.8194805194805195, + "grad_norm": 5.581038296434202, + "learning_rate": 1.107261045933363e-05, + "loss": 0.0064, + "step": 3711 + }, + { + "epoch": 4.820779220779221, + "grad_norm": 4.363276437547504, + "learning_rate": 1.1068428454722198e-05, + "loss": 0.0101, + "step": 3712 + }, + { + "epoch": 4.822077922077922, + "grad_norm": 2.7249849289410566, + "learning_rate": 1.1064246261085323e-05, + "loss": 0.015, + "step": 3713 + }, + { + "epoch": 4.823376623376624, + "grad_norm": 4.835504487244092, + "learning_rate": 1.1060063879162913e-05, + "loss": 0.0291, + "step": 3714 + }, + { + "epoch": 4.824675324675325, + "grad_norm": 7.197958762494662, + "learning_rate": 1.1055881309694914e-05, + "loss": 0.0074, + "step": 3715 + }, + { + "epoch": 4.825974025974026, + "grad_norm": 6.744799610620505, + "learning_rate": 1.1051698553421305e-05, + "loss": 0.0223, + "step": 3716 + }, + { + "epoch": 4.827272727272727, + "grad_norm": 4.9621983043270435, + "learning_rate": 1.1047515611082087e-05, + "loss": 0.017, + "step": 3717 + }, + { + "epoch": 4.828571428571428, + "grad_norm": 1.9979197143489706, + "learning_rate": 1.1043332483417311e-05, + "loss": 0.012, + "step": 3718 + }, + { + "epoch": 4.82987012987013, + "grad_norm": 5.47448240303869, + "learning_rate": 1.1039149171167046e-05, + "loss": 0.0261, + "step": 3719 + }, + { + "epoch": 4.8311688311688314, + "grad_norm": 6.959592064619279, + "learning_rate": 1.1034965675071404e-05, + "loss": 0.0216, + "step": 3720 + }, + { + "epoch": 4.832467532467533, + "grad_norm": 7.271500489971827, + "learning_rate": 1.1030781995870525e-05, + "loss": 0.0297, + "step": 3721 + }, + { + "epoch": 4.833766233766234, + "grad_norm": 3.6387739374221373, + "learning_rate": 1.1026598134304582e-05, + "loss": 0.0399, + "step": 3722 + }, + { + "epoch": 4.835064935064935, + "grad_norm": 9.772285069036972, + "learning_rate": 1.102241409111378e-05, + "loss": 0.0148, + "step": 3723 + }, + { + "epoch": 4.836363636363636, + "grad_norm": 5.654707765054354, + "learning_rate": 1.1018229867038358e-05, + "loss": 0.0203, + "step": 3724 + }, + { + "epoch": 4.837662337662338, + "grad_norm": 10.279880263900925, + "learning_rate": 1.1014045462818578e-05, + "loss": 0.0626, + "step": 3725 + }, + { + "epoch": 4.838961038961039, + "grad_norm": 9.683546058713913, + "learning_rate": 1.1009860879194751e-05, + "loss": 0.0229, + "step": 3726 + }, + { + "epoch": 4.84025974025974, + "grad_norm": 4.383362838452509, + "learning_rate": 1.1005676116907202e-05, + "loss": 0.0252, + "step": 3727 + }, + { + "epoch": 4.841558441558441, + "grad_norm": 4.574100120416554, + "learning_rate": 1.1001491176696304e-05, + "loss": 0.0132, + "step": 3728 + }, + { + "epoch": 4.8428571428571425, + "grad_norm": 3.7761485349675183, + "learning_rate": 1.0997306059302444e-05, + "loss": 0.0109, + "step": 3729 + }, + { + "epoch": 4.8441558441558445, + "grad_norm": 3.8823122609072342, + "learning_rate": 1.0993120765466056e-05, + "loss": 0.0269, + "step": 3730 + }, + { + "epoch": 4.845454545454546, + "grad_norm": 4.260270050587224, + "learning_rate": 1.0988935295927594e-05, + "loss": 0.0171, + "step": 3731 + }, + { + "epoch": 4.846753246753247, + "grad_norm": 6.8538869549806645, + "learning_rate": 1.0984749651427552e-05, + "loss": 0.0121, + "step": 3732 + }, + { + "epoch": 4.848051948051948, + "grad_norm": 5.8813284300379385, + "learning_rate": 1.0980563832706445e-05, + "loss": 0.0184, + "step": 3733 + }, + { + "epoch": 4.849350649350649, + "grad_norm": 2.219304390867442, + "learning_rate": 1.0976377840504828e-05, + "loss": 0.0227, + "step": 3734 + }, + { + "epoch": 4.85064935064935, + "grad_norm": 7.45230321251896, + "learning_rate": 1.0972191675563286e-05, + "loss": 0.0217, + "step": 3735 + }, + { + "epoch": 4.851948051948052, + "grad_norm": 15.144553972705108, + "learning_rate": 1.0968005338622426e-05, + "loss": 0.0421, + "step": 3736 + }, + { + "epoch": 4.853246753246753, + "grad_norm": 4.484930586452027, + "learning_rate": 1.0963818830422896e-05, + "loss": 0.0201, + "step": 3737 + }, + { + "epoch": 4.8545454545454545, + "grad_norm": 9.41092192545252, + "learning_rate": 1.0959632151705366e-05, + "loss": 0.0096, + "step": 3738 + }, + { + "epoch": 4.855844155844156, + "grad_norm": 10.590861095688505, + "learning_rate": 1.0955445303210541e-05, + "loss": 0.0237, + "step": 3739 + }, + { + "epoch": 4.857142857142857, + "grad_norm": 4.987114101548821, + "learning_rate": 1.0951258285679154e-05, + "loss": 0.0178, + "step": 3740 + }, + { + "epoch": 4.858441558441559, + "grad_norm": 10.731252689799026, + "learning_rate": 1.0947071099851972e-05, + "loss": 0.0238, + "step": 3741 + }, + { + "epoch": 4.85974025974026, + "grad_norm": 13.802786930948976, + "learning_rate": 1.094288374646978e-05, + "loss": 0.0086, + "step": 3742 + }, + { + "epoch": 4.861038961038961, + "grad_norm": 5.894739838129115, + "learning_rate": 1.0938696226273412e-05, + "loss": 0.0136, + "step": 3743 + }, + { + "epoch": 4.862337662337662, + "grad_norm": 16.979417644935335, + "learning_rate": 1.0934508540003713e-05, + "loss": 0.0362, + "step": 3744 + }, + { + "epoch": 4.863636363636363, + "grad_norm": 8.706141293282462, + "learning_rate": 1.0930320688401563e-05, + "loss": 0.0061, + "step": 3745 + }, + { + "epoch": 4.8649350649350644, + "grad_norm": 11.519587548312842, + "learning_rate": 1.092613267220788e-05, + "loss": 0.0197, + "step": 3746 + }, + { + "epoch": 4.8662337662337665, + "grad_norm": 16.2647681041912, + "learning_rate": 1.0921944492163601e-05, + "loss": 0.0277, + "step": 3747 + }, + { + "epoch": 4.867532467532468, + "grad_norm": 9.058696378615073, + "learning_rate": 1.0917756149009694e-05, + "loss": 0.0212, + "step": 3748 + }, + { + "epoch": 4.868831168831169, + "grad_norm": 22.657604393845904, + "learning_rate": 1.0913567643487157e-05, + "loss": 0.0162, + "step": 3749 + }, + { + "epoch": 4.87012987012987, + "grad_norm": 7.520467332027682, + "learning_rate": 1.0909378976337022e-05, + "loss": 0.0372, + "step": 3750 + }, + { + "epoch": 4.871428571428572, + "grad_norm": 3.257760983679164, + "learning_rate": 1.0905190148300335e-05, + "loss": 0.0063, + "step": 3751 + }, + { + "epoch": 4.872727272727273, + "grad_norm": 14.941438483843882, + "learning_rate": 1.090100116011819e-05, + "loss": 0.0373, + "step": 3752 + }, + { + "epoch": 4.874025974025974, + "grad_norm": 7.2302584040423925, + "learning_rate": 1.0896812012531693e-05, + "loss": 0.0174, + "step": 3753 + }, + { + "epoch": 4.875324675324675, + "grad_norm": 10.541323807554187, + "learning_rate": 1.0892622706281985e-05, + "loss": 0.0106, + "step": 3754 + }, + { + "epoch": 4.876623376623376, + "grad_norm": 13.847137675941182, + "learning_rate": 1.0888433242110235e-05, + "loss": 0.0227, + "step": 3755 + }, + { + "epoch": 4.8779220779220775, + "grad_norm": 3.3568528456103315, + "learning_rate": 1.0884243620757639e-05, + "loss": 0.0157, + "step": 3756 + }, + { + "epoch": 4.87922077922078, + "grad_norm": 10.805485825511088, + "learning_rate": 1.0880053842965426e-05, + "loss": 0.0086, + "step": 3757 + }, + { + "epoch": 4.880519480519481, + "grad_norm": 14.164742163821883, + "learning_rate": 1.087586390947484e-05, + "loss": 0.0267, + "step": 3758 + }, + { + "epoch": 4.881818181818182, + "grad_norm": 5.1615121655021765, + "learning_rate": 1.0871673821027166e-05, + "loss": 0.0321, + "step": 3759 + }, + { + "epoch": 4.883116883116883, + "grad_norm": 7.851104675185822, + "learning_rate": 1.0867483578363708e-05, + "loss": 0.0304, + "step": 3760 + }, + { + "epoch": 4.884415584415584, + "grad_norm": 9.424137635781488, + "learning_rate": 1.08632931822258e-05, + "loss": 0.0317, + "step": 3761 + }, + { + "epoch": 4.885714285714286, + "grad_norm": 3.5194852431311694, + "learning_rate": 1.0859102633354805e-05, + "loss": 0.0044, + "step": 3762 + }, + { + "epoch": 4.887012987012987, + "grad_norm": 12.396143627571194, + "learning_rate": 1.0854911932492114e-05, + "loss": 0.0477, + "step": 3763 + }, + { + "epoch": 4.888311688311688, + "grad_norm": 3.015375670291402, + "learning_rate": 1.0850721080379134e-05, + "loss": 0.0105, + "step": 3764 + }, + { + "epoch": 4.8896103896103895, + "grad_norm": 15.241322670088458, + "learning_rate": 1.0846530077757314e-05, + "loss": 0.0101, + "step": 3765 + }, + { + "epoch": 4.890909090909091, + "grad_norm": 11.145791544547587, + "learning_rate": 1.0842338925368121e-05, + "loss": 0.0424, + "step": 3766 + }, + { + "epoch": 4.892207792207792, + "grad_norm": 9.45632378698483, + "learning_rate": 1.083814762395305e-05, + "loss": 0.025, + "step": 3767 + }, + { + "epoch": 4.893506493506494, + "grad_norm": 18.12388452369029, + "learning_rate": 1.0833956174253622e-05, + "loss": 0.0371, + "step": 3768 + }, + { + "epoch": 4.894805194805195, + "grad_norm": 5.539708949333756, + "learning_rate": 1.0829764577011384e-05, + "loss": 0.0131, + "step": 3769 + }, + { + "epoch": 4.896103896103896, + "grad_norm": 20.377683934525717, + "learning_rate": 1.082557283296791e-05, + "loss": 0.0356, + "step": 3770 + }, + { + "epoch": 4.897402597402597, + "grad_norm": 12.9934141207493, + "learning_rate": 1.0821380942864801e-05, + "loss": 0.0158, + "step": 3771 + }, + { + "epoch": 4.898701298701298, + "grad_norm": 13.12989993797903, + "learning_rate": 1.081718890744368e-05, + "loss": 0.0257, + "step": 3772 + }, + { + "epoch": 4.9, + "grad_norm": 17.125662187807347, + "learning_rate": 1.0812996727446204e-05, + "loss": 0.02, + "step": 3773 + }, + { + "epoch": 4.9012987012987015, + "grad_norm": 2.69061231758399, + "learning_rate": 1.0808804403614044e-05, + "loss": 0.0215, + "step": 3774 + }, + { + "epoch": 4.902597402597403, + "grad_norm": 12.198845237722145, + "learning_rate": 1.0804611936688903e-05, + "loss": 0.0296, + "step": 3775 + }, + { + "epoch": 4.903896103896104, + "grad_norm": 16.4921269957881, + "learning_rate": 1.080041932741251e-05, + "loss": 0.0204, + "step": 3776 + }, + { + "epoch": 4.905194805194805, + "grad_norm": 7.055652912515113, + "learning_rate": 1.0796226576526619e-05, + "loss": 0.0162, + "step": 3777 + }, + { + "epoch": 4.906493506493506, + "grad_norm": 14.683260129418041, + "learning_rate": 1.0792033684773006e-05, + "loss": 0.0119, + "step": 3778 + }, + { + "epoch": 4.907792207792208, + "grad_norm": 10.059865684961586, + "learning_rate": 1.0787840652893475e-05, + "loss": 0.0186, + "step": 3779 + }, + { + "epoch": 4.909090909090909, + "grad_norm": 9.417389301293843, + "learning_rate": 1.0783647481629856e-05, + "loss": 0.0112, + "step": 3780 + }, + { + "epoch": 4.91038961038961, + "grad_norm": 13.325847042884334, + "learning_rate": 1.0779454171723994e-05, + "loss": 0.0392, + "step": 3781 + }, + { + "epoch": 4.911688311688311, + "grad_norm": 6.481148567476602, + "learning_rate": 1.0775260723917771e-05, + "loss": 0.0148, + "step": 3782 + }, + { + "epoch": 4.9129870129870135, + "grad_norm": 6.902532652877539, + "learning_rate": 1.0771067138953086e-05, + "loss": 0.0055, + "step": 3783 + }, + { + "epoch": 4.914285714285715, + "grad_norm": 12.075816202616267, + "learning_rate": 1.076687341757187e-05, + "loss": 0.0162, + "step": 3784 + }, + { + "epoch": 4.915584415584416, + "grad_norm": 7.823947278056269, + "learning_rate": 1.0762679560516067e-05, + "loss": 0.0238, + "step": 3785 + }, + { + "epoch": 4.916883116883117, + "grad_norm": 7.272625642019768, + "learning_rate": 1.0758485568527648e-05, + "loss": 0.0177, + "step": 3786 + }, + { + "epoch": 4.918181818181818, + "grad_norm": 15.29574500900296, + "learning_rate": 1.0754291442348618e-05, + "loss": 0.0332, + "step": 3787 + }, + { + "epoch": 4.919480519480519, + "grad_norm": 4.759383299109013, + "learning_rate": 1.0750097182720993e-05, + "loss": 0.0279, + "step": 3788 + }, + { + "epoch": 4.920779220779221, + "grad_norm": 10.770196489415728, + "learning_rate": 1.0745902790386821e-05, + "loss": 0.0431, + "step": 3789 + }, + { + "epoch": 4.922077922077922, + "grad_norm": 14.630054442380464, + "learning_rate": 1.0741708266088166e-05, + "loss": 0.0462, + "step": 3790 + }, + { + "epoch": 4.923376623376623, + "grad_norm": 4.241079262652927, + "learning_rate": 1.0737513610567126e-05, + "loss": 0.0643, + "step": 3791 + }, + { + "epoch": 4.9246753246753245, + "grad_norm": 12.73160191937376, + "learning_rate": 1.0733318824565809e-05, + "loss": 0.0298, + "step": 3792 + }, + { + "epoch": 4.925974025974026, + "grad_norm": 12.642053305364826, + "learning_rate": 1.0729123908826357e-05, + "loss": 0.0074, + "step": 3793 + }, + { + "epoch": 4.927272727272728, + "grad_norm": 6.083122300716859, + "learning_rate": 1.0724928864090932e-05, + "loss": 0.0285, + "step": 3794 + }, + { + "epoch": 4.928571428571429, + "grad_norm": 13.732641333169724, + "learning_rate": 1.0720733691101713e-05, + "loss": 0.0272, + "step": 3795 + }, + { + "epoch": 4.92987012987013, + "grad_norm": 8.716707056333151, + "learning_rate": 1.0716538390600908e-05, + "loss": 0.0202, + "step": 3796 + }, + { + "epoch": 4.931168831168831, + "grad_norm": 13.276879782785787, + "learning_rate": 1.0712342963330751e-05, + "loss": 0.0244, + "step": 3797 + }, + { + "epoch": 4.932467532467532, + "grad_norm": 17.44124265795434, + "learning_rate": 1.0708147410033485e-05, + "loss": 0.0324, + "step": 3798 + }, + { + "epoch": 4.933766233766233, + "grad_norm": 6.051923136727763, + "learning_rate": 1.0703951731451392e-05, + "loss": 0.0338, + "step": 3799 + }, + { + "epoch": 4.935064935064935, + "grad_norm": 12.192156186655195, + "learning_rate": 1.069975592832676e-05, + "loss": 0.0259, + "step": 3800 + }, + { + "epoch": 4.9363636363636365, + "grad_norm": 10.379114060859635, + "learning_rate": 1.0695560001401913e-05, + "loss": 0.0173, + "step": 3801 + }, + { + "epoch": 4.937662337662338, + "grad_norm": 6.6800368281057505, + "learning_rate": 1.069136395141919e-05, + "loss": 0.0223, + "step": 3802 + }, + { + "epoch": 4.938961038961039, + "grad_norm": 7.2607882874679905, + "learning_rate": 1.0687167779120949e-05, + "loss": 0.0291, + "step": 3803 + }, + { + "epoch": 4.94025974025974, + "grad_norm": 5.983036160291205, + "learning_rate": 1.0682971485249576e-05, + "loss": 0.0095, + "step": 3804 + }, + { + "epoch": 4.941558441558442, + "grad_norm": 4.067818444364586, + "learning_rate": 1.0678775070547474e-05, + "loss": 0.0095, + "step": 3805 + }, + { + "epoch": 4.942857142857143, + "grad_norm": 9.89238188907776, + "learning_rate": 1.0674578535757071e-05, + "loss": 0.017, + "step": 3806 + }, + { + "epoch": 4.944155844155844, + "grad_norm": 10.952244127361935, + "learning_rate": 1.0670381881620814e-05, + "loss": 0.0254, + "step": 3807 + }, + { + "epoch": 4.945454545454545, + "grad_norm": 8.697822570128414, + "learning_rate": 1.0666185108881171e-05, + "loss": 0.0246, + "step": 3808 + }, + { + "epoch": 4.9467532467532465, + "grad_norm": 18.810790242521545, + "learning_rate": 1.066198821828063e-05, + "loss": 0.0321, + "step": 3809 + }, + { + "epoch": 4.948051948051948, + "grad_norm": 3.626188182340169, + "learning_rate": 1.0657791210561705e-05, + "loss": 0.0268, + "step": 3810 + }, + { + "epoch": 4.94935064935065, + "grad_norm": 13.889767959637544, + "learning_rate": 1.0653594086466925e-05, + "loss": 0.0294, + "step": 3811 + }, + { + "epoch": 4.950649350649351, + "grad_norm": 10.764500572130672, + "learning_rate": 1.0649396846738838e-05, + "loss": 0.0182, + "step": 3812 + }, + { + "epoch": 4.951948051948052, + "grad_norm": 6.04993803395585, + "learning_rate": 1.0645199492120022e-05, + "loss": 0.0106, + "step": 3813 + }, + { + "epoch": 4.953246753246753, + "grad_norm": 16.503700786115257, + "learning_rate": 1.0641002023353062e-05, + "loss": 0.0161, + "step": 3814 + }, + { + "epoch": 4.954545454545455, + "grad_norm": 6.372954697200006, + "learning_rate": 1.0636804441180579e-05, + "loss": 0.0335, + "step": 3815 + }, + { + "epoch": 4.955844155844156, + "grad_norm": 11.330447602973837, + "learning_rate": 1.0632606746345203e-05, + "loss": 0.0238, + "step": 3816 + }, + { + "epoch": 4.957142857142857, + "grad_norm": 12.044949729803148, + "learning_rate": 1.0628408939589582e-05, + "loss": 0.0343, + "step": 3817 + }, + { + "epoch": 4.958441558441558, + "grad_norm": 6.775594702750413, + "learning_rate": 1.0624211021656392e-05, + "loss": 0.0477, + "step": 3818 + }, + { + "epoch": 4.95974025974026, + "grad_norm": 14.71670914979003, + "learning_rate": 1.0620012993288327e-05, + "loss": 0.022, + "step": 3819 + }, + { + "epoch": 4.961038961038961, + "grad_norm": 11.507453674978477, + "learning_rate": 1.0615814855228095e-05, + "loss": 0.0345, + "step": 3820 + }, + { + "epoch": 4.962337662337663, + "grad_norm": 7.43039626701883, + "learning_rate": 1.0611616608218429e-05, + "loss": 0.02, + "step": 3821 + }, + { + "epoch": 4.963636363636364, + "grad_norm": 12.910324616326918, + "learning_rate": 1.0607418253002077e-05, + "loss": 0.0697, + "step": 3822 + }, + { + "epoch": 4.964935064935065, + "grad_norm": 13.424215264289105, + "learning_rate": 1.0603219790321812e-05, + "loss": 0.0241, + "step": 3823 + }, + { + "epoch": 4.966233766233766, + "grad_norm": 11.062505750106286, + "learning_rate": 1.0599021220920418e-05, + "loss": 0.0379, + "step": 3824 + }, + { + "epoch": 4.967532467532467, + "grad_norm": 12.884528913003267, + "learning_rate": 1.0594822545540707e-05, + "loss": 0.0267, + "step": 3825 + }, + { + "epoch": 4.968831168831169, + "grad_norm": 7.213551195141757, + "learning_rate": 1.0590623764925503e-05, + "loss": 0.0201, + "step": 3826 + }, + { + "epoch": 4.97012987012987, + "grad_norm": 6.457332716880885, + "learning_rate": 1.0586424879817647e-05, + "loss": 0.0085, + "step": 3827 + }, + { + "epoch": 4.9714285714285715, + "grad_norm": 11.85849924296318, + "learning_rate": 1.058222589096001e-05, + "loss": 0.0309, + "step": 3828 + }, + { + "epoch": 4.972727272727273, + "grad_norm": 3.96070576547216, + "learning_rate": 1.0578026799095464e-05, + "loss": 0.0137, + "step": 3829 + }, + { + "epoch": 4.974025974025974, + "grad_norm": 7.613745630634611, + "learning_rate": 1.0573827604966917e-05, + "loss": 0.0281, + "step": 3830 + }, + { + "epoch": 4.975324675324675, + "grad_norm": 10.069949964634457, + "learning_rate": 1.0569628309317282e-05, + "loss": 0.0304, + "step": 3831 + }, + { + "epoch": 4.976623376623377, + "grad_norm": 4.63437525566351, + "learning_rate": 1.0565428912889495e-05, + "loss": 0.0059, + "step": 3832 + }, + { + "epoch": 4.977922077922078, + "grad_norm": 8.3135914371947, + "learning_rate": 1.0561229416426516e-05, + "loss": 0.0091, + "step": 3833 + }, + { + "epoch": 4.979220779220779, + "grad_norm": 9.567299853120486, + "learning_rate": 1.0557029820671305e-05, + "loss": 0.0315, + "step": 3834 + }, + { + "epoch": 4.98051948051948, + "grad_norm": 2.8148977430122866, + "learning_rate": 1.055283012636686e-05, + "loss": 0.0051, + "step": 3835 + }, + { + "epoch": 4.9818181818181815, + "grad_norm": 8.659649344984974, + "learning_rate": 1.0548630334256185e-05, + "loss": 0.0148, + "step": 3836 + }, + { + "epoch": 4.9831168831168835, + "grad_norm": 6.349283617193724, + "learning_rate": 1.05444304450823e-05, + "loss": 0.0247, + "step": 3837 + }, + { + "epoch": 4.984415584415585, + "grad_norm": 4.61484542682628, + "learning_rate": 1.0540230459588252e-05, + "loss": 0.0106, + "step": 3838 + }, + { + "epoch": 4.985714285714286, + "grad_norm": 7.349155920270798, + "learning_rate": 1.0536030378517097e-05, + "loss": 0.0474, + "step": 3839 + }, + { + "epoch": 4.987012987012987, + "grad_norm": 8.24384462332522, + "learning_rate": 1.0531830202611904e-05, + "loss": 0.0186, + "step": 3840 + }, + { + "epoch": 4.988311688311688, + "grad_norm": 6.037262408328486, + "learning_rate": 1.052762993261577e-05, + "loss": 0.0119, + "step": 3841 + }, + { + "epoch": 4.989610389610389, + "grad_norm": 7.024766634112399, + "learning_rate": 1.0523429569271803e-05, + "loss": 0.0152, + "step": 3842 + }, + { + "epoch": 4.990909090909091, + "grad_norm": 5.512070055958907, + "learning_rate": 1.0519229113323129e-05, + "loss": 0.0189, + "step": 3843 + }, + { + "epoch": 4.992207792207792, + "grad_norm": 11.076661864260913, + "learning_rate": 1.0515028565512882e-05, + "loss": 0.0253, + "step": 3844 + }, + { + "epoch": 4.9935064935064934, + "grad_norm": 12.36621014948698, + "learning_rate": 1.0510827926584227e-05, + "loss": 0.0212, + "step": 3845 + }, + { + "epoch": 4.994805194805195, + "grad_norm": 8.005527237348028, + "learning_rate": 1.0506627197280332e-05, + "loss": 0.0197, + "step": 3846 + }, + { + "epoch": 4.996103896103897, + "grad_norm": 4.676933233207954, + "learning_rate": 1.050242637834439e-05, + "loss": 0.0272, + "step": 3847 + }, + { + "epoch": 4.997402597402598, + "grad_norm": 5.680543414291961, + "learning_rate": 1.0498225470519605e-05, + "loss": 0.0215, + "step": 3848 + }, + { + "epoch": 4.998701298701299, + "grad_norm": 10.402045264412903, + "learning_rate": 1.0494024474549201e-05, + "loss": 0.0187, + "step": 3849 + }, + { + "epoch": 5.0, + "grad_norm": 7.353957516479436, + "learning_rate": 1.0489823391176409e-05, + "loss": 0.0066, + "step": 3850 + }, + { + "epoch": 5.0, + "eval_accuracy": 0.9275687815833801, + "eval_f1": 0.9116762491628412, + "eval_loss": 0.15282918512821198, + "eval_precision": 0.9048585314341141, + "eval_recall": 0.9284650858141826, + "eval_runtime": 12.9588, + "eval_samples_per_second": 137.435, + "eval_steps_per_second": 1.08, + "step": 3850 + }, + { + "epoch": 5.001298701298701, + "grad_norm": 10.540392993790391, + "learning_rate": 1.0485622221144485e-05, + "loss": 0.0263, + "step": 3851 + }, + { + "epoch": 5.002597402597402, + "grad_norm": 4.923840131917241, + "learning_rate": 1.0481420965196694e-05, + "loss": 0.014, + "step": 3852 + }, + { + "epoch": 5.003896103896104, + "grad_norm": 11.094290100939375, + "learning_rate": 1.047721962407632e-05, + "loss": 0.0223, + "step": 3853 + }, + { + "epoch": 5.005194805194805, + "grad_norm": 5.354324773206372, + "learning_rate": 1.0473018198526661e-05, + "loss": 0.0038, + "step": 3854 + }, + { + "epoch": 5.0064935064935066, + "grad_norm": 3.3616718844388265, + "learning_rate": 1.0468816689291027e-05, + "loss": 0.0094, + "step": 3855 + }, + { + "epoch": 5.007792207792208, + "grad_norm": 7.578340785675294, + "learning_rate": 1.0464615097112748e-05, + "loss": 0.0116, + "step": 3856 + }, + { + "epoch": 5.009090909090909, + "grad_norm": 3.333448201852125, + "learning_rate": 1.0460413422735167e-05, + "loss": 0.0208, + "step": 3857 + }, + { + "epoch": 5.01038961038961, + "grad_norm": 5.405636828148465, + "learning_rate": 1.0456211666901638e-05, + "loss": 0.0392, + "step": 3858 + }, + { + "epoch": 5.011688311688312, + "grad_norm": 2.3140565414588696, + "learning_rate": 1.0452009830355535e-05, + "loss": 0.0133, + "step": 3859 + }, + { + "epoch": 5.012987012987013, + "grad_norm": 1.7168447780183793, + "learning_rate": 1.044780791384024e-05, + "loss": 0.0038, + "step": 3860 + }, + { + "epoch": 5.014285714285714, + "grad_norm": 4.680108989019957, + "learning_rate": 1.0443605918099154e-05, + "loss": 0.016, + "step": 3861 + }, + { + "epoch": 5.015584415584415, + "grad_norm": 4.545036532602537, + "learning_rate": 1.043940384387569e-05, + "loss": 0.0105, + "step": 3862 + }, + { + "epoch": 5.0168831168831165, + "grad_norm": 3.367309878777259, + "learning_rate": 1.0435201691913275e-05, + "loss": 0.0222, + "step": 3863 + }, + { + "epoch": 5.0181818181818185, + "grad_norm": 5.41123705702868, + "learning_rate": 1.0430999462955354e-05, + "loss": 0.0231, + "step": 3864 + }, + { + "epoch": 5.01948051948052, + "grad_norm": 5.193272941962208, + "learning_rate": 1.0426797157745377e-05, + "loss": 0.0115, + "step": 3865 + }, + { + "epoch": 5.020779220779221, + "grad_norm": 4.942710155879454, + "learning_rate": 1.0422594777026814e-05, + "loss": 0.0318, + "step": 3866 + }, + { + "epoch": 5.022077922077922, + "grad_norm": 4.25896701189043, + "learning_rate": 1.0418392321543143e-05, + "loss": 0.0189, + "step": 3867 + }, + { + "epoch": 5.023376623376623, + "grad_norm": 5.000311286629529, + "learning_rate": 1.0414189792037864e-05, + "loss": 0.008, + "step": 3868 + }, + { + "epoch": 5.024675324675325, + "grad_norm": 4.278853887004265, + "learning_rate": 1.0409987189254485e-05, + "loss": 0.012, + "step": 3869 + }, + { + "epoch": 5.025974025974026, + "grad_norm": 6.112693106237239, + "learning_rate": 1.0405784513936523e-05, + "loss": 0.0237, + "step": 3870 + }, + { + "epoch": 5.027272727272727, + "grad_norm": 5.843704868838991, + "learning_rate": 1.0401581766827516e-05, + "loss": 0.0138, + "step": 3871 + }, + { + "epoch": 5.0285714285714285, + "grad_norm": 7.431683742462649, + "learning_rate": 1.0397378948671007e-05, + "loss": 0.0205, + "step": 3872 + }, + { + "epoch": 5.02987012987013, + "grad_norm": 7.845855510761767, + "learning_rate": 1.0393176060210556e-05, + "loss": 0.0333, + "step": 3873 + }, + { + "epoch": 5.031168831168831, + "grad_norm": 2.8607802749655686, + "learning_rate": 1.0388973102189732e-05, + "loss": 0.0102, + "step": 3874 + }, + { + "epoch": 5.032467532467533, + "grad_norm": 6.833448488876962, + "learning_rate": 1.0384770075352126e-05, + "loss": 0.0071, + "step": 3875 + }, + { + "epoch": 5.033766233766234, + "grad_norm": 1.9505442076283215, + "learning_rate": 1.0380566980441325e-05, + "loss": 0.0089, + "step": 3876 + }, + { + "epoch": 5.035064935064935, + "grad_norm": 6.114583248945031, + "learning_rate": 1.0376363818200943e-05, + "loss": 0.008, + "step": 3877 + }, + { + "epoch": 5.036363636363636, + "grad_norm": 3.6801978748815274, + "learning_rate": 1.0372160589374597e-05, + "loss": 0.0236, + "step": 3878 + }, + { + "epoch": 5.037662337662337, + "grad_norm": 9.531160257784109, + "learning_rate": 1.0367957294705922e-05, + "loss": 0.0337, + "step": 3879 + }, + { + "epoch": 5.038961038961039, + "grad_norm": 9.749999510312533, + "learning_rate": 1.0363753934938559e-05, + "loss": 0.017, + "step": 3880 + }, + { + "epoch": 5.04025974025974, + "grad_norm": 8.12520492267456, + "learning_rate": 1.0359550510816158e-05, + "loss": 0.0345, + "step": 3881 + }, + { + "epoch": 5.041558441558442, + "grad_norm": 3.3231648328448062, + "learning_rate": 1.0355347023082394e-05, + "loss": 0.0262, + "step": 3882 + }, + { + "epoch": 5.042857142857143, + "grad_norm": 9.988446176375952, + "learning_rate": 1.0351143472480935e-05, + "loss": 0.0098, + "step": 3883 + }, + { + "epoch": 5.044155844155844, + "grad_norm": 5.733266843086631, + "learning_rate": 1.0346939859755481e-05, + "loss": 0.0173, + "step": 3884 + }, + { + "epoch": 5.045454545454546, + "grad_norm": 5.121506532683435, + "learning_rate": 1.0342736185649722e-05, + "loss": 0.0106, + "step": 3885 + }, + { + "epoch": 5.046753246753247, + "grad_norm": 11.556803445686805, + "learning_rate": 1.0338532450907373e-05, + "loss": 0.0235, + "step": 3886 + }, + { + "epoch": 5.048051948051948, + "grad_norm": 8.092182341260488, + "learning_rate": 1.0334328656272153e-05, + "loss": 0.0174, + "step": 3887 + }, + { + "epoch": 5.049350649350649, + "grad_norm": 2.430780509861474, + "learning_rate": 1.0330124802487795e-05, + "loss": 0.0039, + "step": 3888 + }, + { + "epoch": 5.05064935064935, + "grad_norm": 5.229135563374575, + "learning_rate": 1.0325920890298039e-05, + "loss": 0.0118, + "step": 3889 + }, + { + "epoch": 5.0519480519480515, + "grad_norm": 3.5223433745051316, + "learning_rate": 1.032171692044664e-05, + "loss": 0.0087, + "step": 3890 + }, + { + "epoch": 5.0532467532467535, + "grad_norm": 5.122180401177097, + "learning_rate": 1.0317512893677361e-05, + "loss": 0.0246, + "step": 3891 + }, + { + "epoch": 5.054545454545455, + "grad_norm": 6.435471420750089, + "learning_rate": 1.0313308810733976e-05, + "loss": 0.015, + "step": 3892 + }, + { + "epoch": 5.055844155844156, + "grad_norm": 5.846415517916124, + "learning_rate": 1.0309104672360264e-05, + "loss": 0.0147, + "step": 3893 + }, + { + "epoch": 5.057142857142857, + "grad_norm": 4.191357816639436, + "learning_rate": 1.0304900479300017e-05, + "loss": 0.0169, + "step": 3894 + }, + { + "epoch": 5.058441558441558, + "grad_norm": 6.855995035746627, + "learning_rate": 1.0300696232297041e-05, + "loss": 0.0135, + "step": 3895 + }, + { + "epoch": 5.05974025974026, + "grad_norm": 6.06386265740449, + "learning_rate": 1.0296491932095146e-05, + "loss": 0.0123, + "step": 3896 + }, + { + "epoch": 5.061038961038961, + "grad_norm": 6.984629607663004, + "learning_rate": 1.0292287579438154e-05, + "loss": 0.0101, + "step": 3897 + }, + { + "epoch": 5.062337662337662, + "grad_norm": 3.2249683515773397, + "learning_rate": 1.0288083175069892e-05, + "loss": 0.0241, + "step": 3898 + }, + { + "epoch": 5.0636363636363635, + "grad_norm": 9.002533491031981, + "learning_rate": 1.0283878719734203e-05, + "loss": 0.0244, + "step": 3899 + }, + { + "epoch": 5.064935064935065, + "grad_norm": 1.5554945955202877, + "learning_rate": 1.0279674214174937e-05, + "loss": 0.0125, + "step": 3900 + }, + { + "epoch": 5.066233766233767, + "grad_norm": 8.054048504460527, + "learning_rate": 1.0275469659135947e-05, + "loss": 0.0113, + "step": 3901 + }, + { + "epoch": 5.067532467532468, + "grad_norm": 6.610636066965114, + "learning_rate": 1.0271265055361106e-05, + "loss": 0.0225, + "step": 3902 + }, + { + "epoch": 5.068831168831169, + "grad_norm": 5.453380767847269, + "learning_rate": 1.026706040359428e-05, + "loss": 0.0105, + "step": 3903 + }, + { + "epoch": 5.07012987012987, + "grad_norm": 9.500411087051834, + "learning_rate": 1.0262855704579358e-05, + "loss": 0.0274, + "step": 3904 + }, + { + "epoch": 5.071428571428571, + "grad_norm": 8.17883355247147, + "learning_rate": 1.0258650959060234e-05, + "loss": 0.0262, + "step": 3905 + }, + { + "epoch": 5.072727272727272, + "grad_norm": 9.518695035170353, + "learning_rate": 1.0254446167780803e-05, + "loss": 0.0283, + "step": 3906 + }, + { + "epoch": 5.074025974025974, + "grad_norm": 10.816510968988121, + "learning_rate": 1.0250241331484973e-05, + "loss": 0.004, + "step": 3907 + }, + { + "epoch": 5.0753246753246755, + "grad_norm": 2.079397090905998, + "learning_rate": 1.0246036450916665e-05, + "loss": 0.0115, + "step": 3908 + }, + { + "epoch": 5.076623376623377, + "grad_norm": 8.958991335526047, + "learning_rate": 1.0241831526819796e-05, + "loss": 0.0128, + "step": 3909 + }, + { + "epoch": 5.077922077922078, + "grad_norm": 9.729803042625708, + "learning_rate": 1.0237626559938305e-05, + "loss": 0.0135, + "step": 3910 + }, + { + "epoch": 5.079220779220779, + "grad_norm": 2.7399100248320853, + "learning_rate": 1.0233421551016123e-05, + "loss": 0.0081, + "step": 3911 + }, + { + "epoch": 5.080519480519481, + "grad_norm": 7.772719630375479, + "learning_rate": 1.0229216500797206e-05, + "loss": 0.005, + "step": 3912 + }, + { + "epoch": 5.081818181818182, + "grad_norm": 7.357806273573311, + "learning_rate": 1.0225011410025498e-05, + "loss": 0.0177, + "step": 3913 + }, + { + "epoch": 5.083116883116883, + "grad_norm": 1.9195776904341404, + "learning_rate": 1.0220806279444968e-05, + "loss": 0.0057, + "step": 3914 + }, + { + "epoch": 5.084415584415584, + "grad_norm": 5.303710846536632, + "learning_rate": 1.0216601109799577e-05, + "loss": 0.0065, + "step": 3915 + }, + { + "epoch": 5.085714285714285, + "grad_norm": 6.453329992889627, + "learning_rate": 1.021239590183331e-05, + "loss": 0.01, + "step": 3916 + }, + { + "epoch": 5.087012987012987, + "grad_norm": 2.8392658045501142, + "learning_rate": 1.0208190656290138e-05, + "loss": 0.0126, + "step": 3917 + }, + { + "epoch": 5.088311688311689, + "grad_norm": 8.631587198916847, + "learning_rate": 1.0203985373914056e-05, + "loss": 0.025, + "step": 3918 + }, + { + "epoch": 5.08961038961039, + "grad_norm": 7.003783045730876, + "learning_rate": 1.0199780055449054e-05, + "loss": 0.015, + "step": 3919 + }, + { + "epoch": 5.090909090909091, + "grad_norm": 2.7118337429010384, + "learning_rate": 1.0195574701639139e-05, + "loss": 0.0094, + "step": 3920 + }, + { + "epoch": 5.092207792207792, + "grad_norm": 12.796438013331702, + "learning_rate": 1.0191369313228319e-05, + "loss": 0.0168, + "step": 3921 + }, + { + "epoch": 5.093506493506493, + "grad_norm": 10.223816152854857, + "learning_rate": 1.0187163890960598e-05, + "loss": 0.0363, + "step": 3922 + }, + { + "epoch": 5.094805194805195, + "grad_norm": 9.133457983537733, + "learning_rate": 1.0182958435580007e-05, + "loss": 0.0139, + "step": 3923 + }, + { + "epoch": 5.096103896103896, + "grad_norm": 13.779852735260956, + "learning_rate": 1.0178752947830566e-05, + "loss": 0.0051, + "step": 3924 + }, + { + "epoch": 5.097402597402597, + "grad_norm": 3.494446237590064, + "learning_rate": 1.0174547428456308e-05, + "loss": 0.01, + "step": 3925 + }, + { + "epoch": 5.0987012987012985, + "grad_norm": 15.1651879060539, + "learning_rate": 1.0170341878201266e-05, + "loss": 0.0217, + "step": 3926 + }, + { + "epoch": 5.1, + "grad_norm": 15.588950392967549, + "learning_rate": 1.0166136297809487e-05, + "loss": 0.016, + "step": 3927 + }, + { + "epoch": 5.101298701298702, + "grad_norm": 10.069358176682938, + "learning_rate": 1.0161930688025018e-05, + "loss": 0.0158, + "step": 3928 + }, + { + "epoch": 5.102597402597403, + "grad_norm": 15.118572832516536, + "learning_rate": 1.0157725049591908e-05, + "loss": 0.0271, + "step": 3929 + }, + { + "epoch": 5.103896103896104, + "grad_norm": 8.040475165350657, + "learning_rate": 1.015351938325422e-05, + "loss": 0.026, + "step": 3930 + }, + { + "epoch": 5.105194805194805, + "grad_norm": 14.761629145668623, + "learning_rate": 1.0149313689756015e-05, + "loss": 0.0235, + "step": 3931 + }, + { + "epoch": 5.106493506493506, + "grad_norm": 4.682930435644529, + "learning_rate": 1.0145107969841359e-05, + "loss": 0.0076, + "step": 3932 + }, + { + "epoch": 5.107792207792208, + "grad_norm": 9.606564741621552, + "learning_rate": 1.0140902224254328e-05, + "loss": 0.0223, + "step": 3933 + }, + { + "epoch": 5.109090909090909, + "grad_norm": 11.42367522878412, + "learning_rate": 1.0136696453738995e-05, + "loss": 0.0124, + "step": 3934 + }, + { + "epoch": 5.1103896103896105, + "grad_norm": 11.075251990926574, + "learning_rate": 1.0132490659039442e-05, + "loss": 0.0479, + "step": 3935 + }, + { + "epoch": 5.111688311688312, + "grad_norm": 8.394954418984964, + "learning_rate": 1.0128284840899758e-05, + "loss": 0.023, + "step": 3936 + }, + { + "epoch": 5.112987012987013, + "grad_norm": 4.473935802297636, + "learning_rate": 1.0124079000064027e-05, + "loss": 0.0122, + "step": 3937 + }, + { + "epoch": 5.114285714285714, + "grad_norm": 2.9455491909487352, + "learning_rate": 1.011987313727635e-05, + "loss": 0.0054, + "step": 3938 + }, + { + "epoch": 5.115584415584416, + "grad_norm": 9.032306015623194, + "learning_rate": 1.0115667253280817e-05, + "loss": 0.019, + "step": 3939 + }, + { + "epoch": 5.116883116883117, + "grad_norm": 7.342906446210257, + "learning_rate": 1.0111461348821537e-05, + "loss": 0.0202, + "step": 3940 + }, + { + "epoch": 5.118181818181818, + "grad_norm": 4.460152376835974, + "learning_rate": 1.0107255424642606e-05, + "loss": 0.0061, + "step": 3941 + }, + { + "epoch": 5.119480519480519, + "grad_norm": 7.034651315920055, + "learning_rate": 1.010304948148814e-05, + "loss": 0.0073, + "step": 3942 + }, + { + "epoch": 5.12077922077922, + "grad_norm": 4.214523610957707, + "learning_rate": 1.0098843520102248e-05, + "loss": 0.0164, + "step": 3943 + }, + { + "epoch": 5.1220779220779225, + "grad_norm": 7.24781356482792, + "learning_rate": 1.0094637541229047e-05, + "loss": 0.0149, + "step": 3944 + }, + { + "epoch": 5.123376623376624, + "grad_norm": 11.301041547624761, + "learning_rate": 1.0090431545612652e-05, + "loss": 0.0124, + "step": 3945 + }, + { + "epoch": 5.124675324675325, + "grad_norm": 3.3774429884231427, + "learning_rate": 1.0086225533997187e-05, + "loss": 0.008, + "step": 3946 + }, + { + "epoch": 5.125974025974026, + "grad_norm": 12.184259380160427, + "learning_rate": 1.0082019507126775e-05, + "loss": 0.0106, + "step": 3947 + }, + { + "epoch": 5.127272727272727, + "grad_norm": 5.397884389618201, + "learning_rate": 1.0077813465745541e-05, + "loss": 0.003, + "step": 3948 + }, + { + "epoch": 5.128571428571428, + "grad_norm": 7.430071604858085, + "learning_rate": 1.0073607410597617e-05, + "loss": 0.0031, + "step": 3949 + }, + { + "epoch": 5.12987012987013, + "grad_norm": 6.984766218007683, + "learning_rate": 1.006940134242713e-05, + "loss": 0.0126, + "step": 3950 + }, + { + "epoch": 5.131168831168831, + "grad_norm": 2.8202817286256914, + "learning_rate": 1.006519526197822e-05, + "loss": 0.0175, + "step": 3951 + }, + { + "epoch": 5.132467532467532, + "grad_norm": 10.810955934768126, + "learning_rate": 1.0060989169995017e-05, + "loss": 0.0261, + "step": 3952 + }, + { + "epoch": 5.1337662337662335, + "grad_norm": 4.516167683149986, + "learning_rate": 1.0056783067221666e-05, + "loss": 0.012, + "step": 3953 + }, + { + "epoch": 5.135064935064935, + "grad_norm": 3.678524627936639, + "learning_rate": 1.0052576954402302e-05, + "loss": 0.0184, + "step": 3954 + }, + { + "epoch": 5.136363636363637, + "grad_norm": 5.915895363338697, + "learning_rate": 1.0048370832281069e-05, + "loss": 0.0237, + "step": 3955 + }, + { + "epoch": 5.137662337662338, + "grad_norm": 10.779687001387607, + "learning_rate": 1.0044164701602111e-05, + "loss": 0.044, + "step": 3956 + }, + { + "epoch": 5.138961038961039, + "grad_norm": 11.660005820543498, + "learning_rate": 1.003995856310957e-05, + "loss": 0.0237, + "step": 3957 + }, + { + "epoch": 5.14025974025974, + "grad_norm": 4.5445367300655155, + "learning_rate": 1.0035752417547595e-05, + "loss": 0.0114, + "step": 3958 + }, + { + "epoch": 5.141558441558441, + "grad_norm": 4.111098387047959, + "learning_rate": 1.0031546265660332e-05, + "loss": 0.0217, + "step": 3959 + }, + { + "epoch": 5.142857142857143, + "grad_norm": 4.16162336631834, + "learning_rate": 1.0027340108191934e-05, + "loss": 0.0384, + "step": 3960 + }, + { + "epoch": 5.144155844155844, + "grad_norm": 4.607111543352715, + "learning_rate": 1.0023133945886545e-05, + "loss": 0.0169, + "step": 3961 + }, + { + "epoch": 5.1454545454545455, + "grad_norm": 2.9177202618191616, + "learning_rate": 1.001892777948832e-05, + "loss": 0.0043, + "step": 3962 + }, + { + "epoch": 5.146753246753247, + "grad_norm": 7.038602620491121, + "learning_rate": 1.0014721609741405e-05, + "loss": 0.0267, + "step": 3963 + }, + { + "epoch": 5.148051948051948, + "grad_norm": 7.600948412301627, + "learning_rate": 1.0010515437389958e-05, + "loss": 0.0235, + "step": 3964 + }, + { + "epoch": 5.14935064935065, + "grad_norm": 8.295528738575268, + "learning_rate": 1.0006309263178128e-05, + "loss": 0.019, + "step": 3965 + }, + { + "epoch": 5.150649350649351, + "grad_norm": 14.003024831659188, + "learning_rate": 1.0002103087850068e-05, + "loss": 0.0485, + "step": 3966 + }, + { + "epoch": 5.151948051948052, + "grad_norm": 13.741887929842449, + "learning_rate": 9.997896912149935e-06, + "loss": 0.0243, + "step": 3967 + }, + { + "epoch": 5.153246753246753, + "grad_norm": 8.262067433508966, + "learning_rate": 9.993690736821874e-06, + "loss": 0.0061, + "step": 3968 + }, + { + "epoch": 5.154545454545454, + "grad_norm": 5.664342970659906, + "learning_rate": 9.989484562610044e-06, + "loss": 0.0099, + "step": 3969 + }, + { + "epoch": 5.1558441558441555, + "grad_norm": 14.142673734181347, + "learning_rate": 9.985278390258596e-06, + "loss": 0.0219, + "step": 3970 + }, + { + "epoch": 5.1571428571428575, + "grad_norm": 4.281388593559851, + "learning_rate": 9.981072220511685e-06, + "loss": 0.0114, + "step": 3971 + }, + { + "epoch": 5.158441558441559, + "grad_norm": 9.322462764222145, + "learning_rate": 9.976866054113458e-06, + "loss": 0.0099, + "step": 3972 + }, + { + "epoch": 5.15974025974026, + "grad_norm": 10.793891699874521, + "learning_rate": 9.972659891808069e-06, + "loss": 0.014, + "step": 3973 + }, + { + "epoch": 5.161038961038961, + "grad_norm": 5.380307731560935, + "learning_rate": 9.968453734339672e-06, + "loss": 0.0119, + "step": 3974 + }, + { + "epoch": 5.162337662337662, + "grad_norm": 9.423212665026325, + "learning_rate": 9.964247582452408e-06, + "loss": 0.0048, + "step": 3975 + }, + { + "epoch": 5.163636363636364, + "grad_norm": 8.110775292542126, + "learning_rate": 9.960041436890433e-06, + "loss": 0.0222, + "step": 3976 + }, + { + "epoch": 5.164935064935065, + "grad_norm": 3.2621291065183793, + "learning_rate": 9.95583529839789e-06, + "loss": 0.0032, + "step": 3977 + }, + { + "epoch": 5.166233766233766, + "grad_norm": 9.077763940605596, + "learning_rate": 9.951629167718936e-06, + "loss": 0.0174, + "step": 3978 + }, + { + "epoch": 5.167532467532467, + "grad_norm": 4.689086331433104, + "learning_rate": 9.9474230455977e-06, + "loss": 0.0029, + "step": 3979 + }, + { + "epoch": 5.1688311688311686, + "grad_norm": 2.20790618050226, + "learning_rate": 9.943216932778337e-06, + "loss": 0.003, + "step": 3980 + }, + { + "epoch": 5.17012987012987, + "grad_norm": 6.8640867965852985, + "learning_rate": 9.939010830004983e-06, + "loss": 0.0038, + "step": 3981 + }, + { + "epoch": 5.171428571428572, + "grad_norm": 5.070712366634945, + "learning_rate": 9.934804738021787e-06, + "loss": 0.0105, + "step": 3982 + }, + { + "epoch": 5.172727272727273, + "grad_norm": 8.26288924821529, + "learning_rate": 9.930598657572874e-06, + "loss": 0.0204, + "step": 3983 + }, + { + "epoch": 5.174025974025974, + "grad_norm": 6.056478651156009, + "learning_rate": 9.926392589402388e-06, + "loss": 0.0084, + "step": 3984 + }, + { + "epoch": 5.175324675324675, + "grad_norm": 7.144228875029622, + "learning_rate": 9.92218653425446e-06, + "loss": 0.0115, + "step": 3985 + }, + { + "epoch": 5.176623376623376, + "grad_norm": 3.590691564942955, + "learning_rate": 9.91798049287323e-06, + "loss": 0.0084, + "step": 3986 + }, + { + "epoch": 5.177922077922078, + "grad_norm": 2.958685185246526, + "learning_rate": 9.913774466002815e-06, + "loss": 0.0034, + "step": 3987 + }, + { + "epoch": 5.179220779220779, + "grad_norm": 6.96419272305559, + "learning_rate": 9.909568454387348e-06, + "loss": 0.0118, + "step": 3988 + }, + { + "epoch": 5.1805194805194805, + "grad_norm": 3.515743140692576, + "learning_rate": 9.905362458770957e-06, + "loss": 0.0163, + "step": 3989 + }, + { + "epoch": 5.181818181818182, + "grad_norm": 3.417044137589949, + "learning_rate": 9.901156479897754e-06, + "loss": 0.0131, + "step": 3990 + }, + { + "epoch": 5.183116883116883, + "grad_norm": 8.31682530201377, + "learning_rate": 9.896950518511863e-06, + "loss": 0.0126, + "step": 3991 + }, + { + "epoch": 5.184415584415585, + "grad_norm": 3.891578750671561, + "learning_rate": 9.892744575357395e-06, + "loss": 0.0189, + "step": 3992 + }, + { + "epoch": 5.185714285714286, + "grad_norm": 5.14883392407837, + "learning_rate": 9.88853865117847e-06, + "loss": 0.0029, + "step": 3993 + }, + { + "epoch": 5.187012987012987, + "grad_norm": 8.95654122734859, + "learning_rate": 9.884332746719186e-06, + "loss": 0.0142, + "step": 3994 + }, + { + "epoch": 5.188311688311688, + "grad_norm": 4.34328095042232, + "learning_rate": 9.880126862723653e-06, + "loss": 0.0307, + "step": 3995 + }, + { + "epoch": 5.189610389610389, + "grad_norm": 7.710710497843328, + "learning_rate": 9.875920999935973e-06, + "loss": 0.0231, + "step": 3996 + }, + { + "epoch": 5.190909090909091, + "grad_norm": 7.971884440923867, + "learning_rate": 9.871715159100245e-06, + "loss": 0.0298, + "step": 3997 + }, + { + "epoch": 5.1922077922077925, + "grad_norm": 6.639025575301852, + "learning_rate": 9.867509340960561e-06, + "loss": 0.0105, + "step": 3998 + }, + { + "epoch": 5.193506493506494, + "grad_norm": 3.865400096048831, + "learning_rate": 9.863303546261008e-06, + "loss": 0.0121, + "step": 3999 + }, + { + "epoch": 5.194805194805195, + "grad_norm": 8.544062361711324, + "learning_rate": 9.859097775745674e-06, + "loss": 0.0306, + "step": 4000 + }, + { + "epoch": 5.196103896103896, + "grad_norm": 7.835900255330742, + "learning_rate": 9.854892030158643e-06, + "loss": 0.0185, + "step": 4001 + }, + { + "epoch": 5.197402597402597, + "grad_norm": 8.699987002084802, + "learning_rate": 9.850686310243986e-06, + "loss": 0.0069, + "step": 4002 + }, + { + "epoch": 5.198701298701299, + "grad_norm": 3.301871542356353, + "learning_rate": 9.846480616745783e-06, + "loss": 0.0248, + "step": 4003 + }, + { + "epoch": 5.2, + "grad_norm": 5.822445751310222, + "learning_rate": 9.842274950408095e-06, + "loss": 0.0167, + "step": 4004 + }, + { + "epoch": 5.201298701298701, + "grad_norm": 4.276869000566195, + "learning_rate": 9.838069311974986e-06, + "loss": 0.0184, + "step": 4005 + }, + { + "epoch": 5.202597402597402, + "grad_norm": 3.7531256492940352, + "learning_rate": 9.833863702190516e-06, + "loss": 0.0189, + "step": 4006 + }, + { + "epoch": 5.203896103896104, + "grad_norm": 4.539595851744104, + "learning_rate": 9.829658121798735e-06, + "loss": 0.0098, + "step": 4007 + }, + { + "epoch": 5.205194805194806, + "grad_norm": 2.1329762414415114, + "learning_rate": 9.825452571543699e-06, + "loss": 0.0336, + "step": 4008 + }, + { + "epoch": 5.206493506493507, + "grad_norm": 3.093991454560163, + "learning_rate": 9.821247052169439e-06, + "loss": 0.0135, + "step": 4009 + }, + { + "epoch": 5.207792207792208, + "grad_norm": 2.9106054726253663, + "learning_rate": 9.817041564419996e-06, + "loss": 0.013, + "step": 4010 + }, + { + "epoch": 5.209090909090909, + "grad_norm": 2.0700202181156766, + "learning_rate": 9.812836109039402e-06, + "loss": 0.0042, + "step": 4011 + }, + { + "epoch": 5.21038961038961, + "grad_norm": 1.700252122879527, + "learning_rate": 9.808630686771688e-06, + "loss": 0.0158, + "step": 4012 + }, + { + "epoch": 5.211688311688311, + "grad_norm": 3.5871458906527254, + "learning_rate": 9.804425298360863e-06, + "loss": 0.0241, + "step": 4013 + }, + { + "epoch": 5.212987012987013, + "grad_norm": 7.8935451402395636, + "learning_rate": 9.800219944550946e-06, + "loss": 0.0059, + "step": 4014 + }, + { + "epoch": 5.214285714285714, + "grad_norm": 7.198943664386734, + "learning_rate": 9.79601462608595e-06, + "loss": 0.0574, + "step": 4015 + }, + { + "epoch": 5.2155844155844155, + "grad_norm": 2.4348214125804812, + "learning_rate": 9.791809343709866e-06, + "loss": 0.0081, + "step": 4016 + }, + { + "epoch": 5.216883116883117, + "grad_norm": 1.8718907046962927, + "learning_rate": 9.787604098166694e-06, + "loss": 0.0109, + "step": 4017 + }, + { + "epoch": 5.218181818181818, + "grad_norm": 3.722306258204774, + "learning_rate": 9.783398890200423e-06, + "loss": 0.022, + "step": 4018 + }, + { + "epoch": 5.21948051948052, + "grad_norm": 4.603547150111322, + "learning_rate": 9.779193720555038e-06, + "loss": 0.0143, + "step": 4019 + }, + { + "epoch": 5.220779220779221, + "grad_norm": 3.602109186048625, + "learning_rate": 9.774988589974505e-06, + "loss": 0.0166, + "step": 4020 + }, + { + "epoch": 5.222077922077922, + "grad_norm": 1.2441684002802682, + "learning_rate": 9.770783499202798e-06, + "loss": 0.0042, + "step": 4021 + }, + { + "epoch": 5.223376623376623, + "grad_norm": 4.001458653807839, + "learning_rate": 9.766578448983877e-06, + "loss": 0.0086, + "step": 4022 + }, + { + "epoch": 5.224675324675324, + "grad_norm": 3.6566647354388127, + "learning_rate": 9.762373440061701e-06, + "loss": 0.0185, + "step": 4023 + }, + { + "epoch": 5.225974025974026, + "grad_norm": 2.902758224986935, + "learning_rate": 9.758168473180206e-06, + "loss": 0.004, + "step": 4024 + }, + { + "epoch": 5.2272727272727275, + "grad_norm": 3.2156399611920543, + "learning_rate": 9.753963549083339e-06, + "loss": 0.0081, + "step": 4025 + }, + { + "epoch": 5.228571428571429, + "grad_norm": 4.560464907437346, + "learning_rate": 9.749758668515027e-06, + "loss": 0.0154, + "step": 4026 + }, + { + "epoch": 5.22987012987013, + "grad_norm": 4.255603957067852, + "learning_rate": 9.745553832219202e-06, + "loss": 0.0166, + "step": 4027 + }, + { + "epoch": 5.231168831168831, + "grad_norm": 3.9009970268841316, + "learning_rate": 9.741349040939769e-06, + "loss": 0.005, + "step": 4028 + }, + { + "epoch": 5.232467532467532, + "grad_norm": 2.417726992478451, + "learning_rate": 9.737144295420642e-06, + "loss": 0.0129, + "step": 4029 + }, + { + "epoch": 5.233766233766234, + "grad_norm": 4.099401823976346, + "learning_rate": 9.732939596405723e-06, + "loss": 0.0072, + "step": 4030 + }, + { + "epoch": 5.235064935064935, + "grad_norm": 7.062278470160992, + "learning_rate": 9.728734944638897e-06, + "loss": 0.0246, + "step": 4031 + }, + { + "epoch": 5.236363636363636, + "grad_norm": 4.5061393553044775, + "learning_rate": 9.724530340864054e-06, + "loss": 0.0066, + "step": 4032 + }, + { + "epoch": 5.2376623376623375, + "grad_norm": 6.4481357716170775, + "learning_rate": 9.720325785825065e-06, + "loss": 0.0358, + "step": 4033 + }, + { + "epoch": 5.238961038961039, + "grad_norm": 8.655419214609381, + "learning_rate": 9.7161212802658e-06, + "loss": 0.024, + "step": 4034 + }, + { + "epoch": 5.240259740259741, + "grad_norm": 2.6852449440581387, + "learning_rate": 9.711916824930112e-06, + "loss": 0.003, + "step": 4035 + }, + { + "epoch": 5.241558441558442, + "grad_norm": 1.253214654976695, + "learning_rate": 9.70771242056185e-06, + "loss": 0.01, + "step": 4036 + }, + { + "epoch": 5.242857142857143, + "grad_norm": 3.243542814327436, + "learning_rate": 9.703508067904857e-06, + "loss": 0.0054, + "step": 4037 + }, + { + "epoch": 5.244155844155844, + "grad_norm": 4.68721628526464, + "learning_rate": 9.699303767702964e-06, + "loss": 0.006, + "step": 4038 + }, + { + "epoch": 5.245454545454545, + "grad_norm": 2.9526228575514115, + "learning_rate": 9.695099520699986e-06, + "loss": 0.0058, + "step": 4039 + }, + { + "epoch": 5.246753246753247, + "grad_norm": 1.392056754553189, + "learning_rate": 9.69089532763974e-06, + "loss": 0.0029, + "step": 4040 + }, + { + "epoch": 5.248051948051948, + "grad_norm": 5.795486566614474, + "learning_rate": 9.686691189266026e-06, + "loss": 0.0192, + "step": 4041 + }, + { + "epoch": 5.249350649350649, + "grad_norm": 3.7731663064661576, + "learning_rate": 9.682487106322642e-06, + "loss": 0.0071, + "step": 4042 + }, + { + "epoch": 5.250649350649351, + "grad_norm": 3.2965440444756715, + "learning_rate": 9.678283079553362e-06, + "loss": 0.0096, + "step": 4043 + }, + { + "epoch": 5.251948051948052, + "grad_norm": 3.4883971469504433, + "learning_rate": 9.674079109701961e-06, + "loss": 0.0059, + "step": 4044 + }, + { + "epoch": 5.253246753246753, + "grad_norm": 2.783100189906548, + "learning_rate": 9.66987519751221e-06, + "loss": 0.0103, + "step": 4045 + }, + { + "epoch": 5.254545454545455, + "grad_norm": 6.3389070331414805, + "learning_rate": 9.66567134372785e-06, + "loss": 0.0213, + "step": 4046 + }, + { + "epoch": 5.255844155844156, + "grad_norm": 3.787523417325556, + "learning_rate": 9.66146754909263e-06, + "loss": 0.0171, + "step": 4047 + }, + { + "epoch": 5.257142857142857, + "grad_norm": 4.984856182976176, + "learning_rate": 9.65726381435028e-06, + "loss": 0.0087, + "step": 4048 + }, + { + "epoch": 5.258441558441558, + "grad_norm": 6.21577078872473, + "learning_rate": 9.653060140244524e-06, + "loss": 0.0307, + "step": 4049 + }, + { + "epoch": 5.259740259740259, + "grad_norm": 4.936141463238712, + "learning_rate": 9.648856527519066e-06, + "loss": 0.015, + "step": 4050 + }, + { + "epoch": 5.261038961038961, + "grad_norm": 4.2842546814143745, + "learning_rate": 9.64465297691761e-06, + "loss": 0.0235, + "step": 4051 + }, + { + "epoch": 5.2623376623376625, + "grad_norm": 2.0365704799817674, + "learning_rate": 9.640449489183844e-06, + "loss": 0.0221, + "step": 4052 + }, + { + "epoch": 5.263636363636364, + "grad_norm": 3.69550710651763, + "learning_rate": 9.636246065061448e-06, + "loss": 0.0082, + "step": 4053 + }, + { + "epoch": 5.264935064935065, + "grad_norm": 2.3790003078941515, + "learning_rate": 9.632042705294081e-06, + "loss": 0.0104, + "step": 4054 + }, + { + "epoch": 5.266233766233766, + "grad_norm": 2.127368661423431, + "learning_rate": 9.627839410625404e-06, + "loss": 0.0097, + "step": 4055 + }, + { + "epoch": 5.267532467532467, + "grad_norm": 4.030165510278963, + "learning_rate": 9.623636181799062e-06, + "loss": 0.0214, + "step": 4056 + }, + { + "epoch": 5.268831168831169, + "grad_norm": 3.8747057403291323, + "learning_rate": 9.619433019558677e-06, + "loss": 0.0102, + "step": 4057 + }, + { + "epoch": 5.27012987012987, + "grad_norm": 3.846947105204918, + "learning_rate": 9.615229924647877e-06, + "loss": 0.0093, + "step": 4058 + }, + { + "epoch": 5.271428571428571, + "grad_norm": 5.760767355785745, + "learning_rate": 9.611026897810268e-06, + "loss": 0.0157, + "step": 4059 + }, + { + "epoch": 5.2727272727272725, + "grad_norm": 5.0920632936316075, + "learning_rate": 9.606823939789449e-06, + "loss": 0.0078, + "step": 4060 + }, + { + "epoch": 5.274025974025974, + "grad_norm": 3.5883508667724997, + "learning_rate": 9.602621051328998e-06, + "loss": 0.0063, + "step": 4061 + }, + { + "epoch": 5.275324675324676, + "grad_norm": 3.200266732018834, + "learning_rate": 9.598418233172485e-06, + "loss": 0.0168, + "step": 4062 + }, + { + "epoch": 5.276623376623377, + "grad_norm": 2.2253081969725845, + "learning_rate": 9.594215486063477e-06, + "loss": 0.0075, + "step": 4063 + }, + { + "epoch": 5.277922077922078, + "grad_norm": 6.915718255004593, + "learning_rate": 9.590012810745516e-06, + "loss": 0.0095, + "step": 4064 + }, + { + "epoch": 5.279220779220779, + "grad_norm": 17.86986199293691, + "learning_rate": 9.585810207962138e-06, + "loss": 0.0266, + "step": 4065 + }, + { + "epoch": 5.28051948051948, + "grad_norm": 8.479853951125724, + "learning_rate": 9.581607678456859e-06, + "loss": 0.0164, + "step": 4066 + }, + { + "epoch": 5.281818181818182, + "grad_norm": 8.713200922657677, + "learning_rate": 9.577405222973191e-06, + "loss": 0.0267, + "step": 4067 + }, + { + "epoch": 5.283116883116883, + "grad_norm": 6.194218672194852, + "learning_rate": 9.573202842254628e-06, + "loss": 0.0241, + "step": 4068 + }, + { + "epoch": 5.2844155844155845, + "grad_norm": 11.178598367315042, + "learning_rate": 9.56900053704465e-06, + "loss": 0.0074, + "step": 4069 + }, + { + "epoch": 5.285714285714286, + "grad_norm": 9.833546084882341, + "learning_rate": 9.564798308086727e-06, + "loss": 0.011, + "step": 4070 + }, + { + "epoch": 5.287012987012987, + "grad_norm": 14.799232958143145, + "learning_rate": 9.560596156124314e-06, + "loss": 0.0359, + "step": 4071 + }, + { + "epoch": 5.288311688311689, + "grad_norm": 3.409070067016064, + "learning_rate": 9.55639408190085e-06, + "loss": 0.019, + "step": 4072 + }, + { + "epoch": 5.28961038961039, + "grad_norm": 7.955851232377883, + "learning_rate": 9.552192086159762e-06, + "loss": 0.024, + "step": 4073 + }, + { + "epoch": 5.290909090909091, + "grad_norm": 19.68272139545327, + "learning_rate": 9.547990169644467e-06, + "loss": 0.0395, + "step": 4074 + }, + { + "epoch": 5.292207792207792, + "grad_norm": 8.900780153132914, + "learning_rate": 9.543788333098365e-06, + "loss": 0.0121, + "step": 4075 + }, + { + "epoch": 5.293506493506493, + "grad_norm": 4.528255300001054, + "learning_rate": 9.539586577264835e-06, + "loss": 0.0215, + "step": 4076 + }, + { + "epoch": 5.294805194805194, + "grad_norm": 3.3700758110858993, + "learning_rate": 9.535384902887253e-06, + "loss": 0.0074, + "step": 4077 + }, + { + "epoch": 5.296103896103896, + "grad_norm": 10.79566929999895, + "learning_rate": 9.531183310708974e-06, + "loss": 0.0285, + "step": 4078 + }, + { + "epoch": 5.297402597402598, + "grad_norm": 14.911511454585384, + "learning_rate": 9.526981801473344e-06, + "loss": 0.0411, + "step": 4079 + }, + { + "epoch": 5.298701298701299, + "grad_norm": 1.235131655811358, + "learning_rate": 9.522780375923683e-06, + "loss": 0.0048, + "step": 4080 + }, + { + "epoch": 5.3, + "grad_norm": 15.397186929268166, + "learning_rate": 9.51857903480331e-06, + "loss": 0.0532, + "step": 4081 + }, + { + "epoch": 5.301298701298701, + "grad_norm": 6.541852147175073, + "learning_rate": 9.514377778855521e-06, + "loss": 0.0333, + "step": 4082 + }, + { + "epoch": 5.302597402597403, + "grad_norm": 8.978378666408066, + "learning_rate": 9.510176608823595e-06, + "loss": 0.0502, + "step": 4083 + }, + { + "epoch": 5.303896103896104, + "grad_norm": 6.29785088124662, + "learning_rate": 9.505975525450802e-06, + "loss": 0.0135, + "step": 4084 + }, + { + "epoch": 5.305194805194805, + "grad_norm": 7.830961715357552, + "learning_rate": 9.501774529480395e-06, + "loss": 0.0201, + "step": 4085 + }, + { + "epoch": 5.306493506493506, + "grad_norm": 9.813852092354368, + "learning_rate": 9.497573621655613e-06, + "loss": 0.0126, + "step": 4086 + }, + { + "epoch": 5.3077922077922075, + "grad_norm": 4.179547810044327, + "learning_rate": 9.493372802719671e-06, + "loss": 0.0052, + "step": 4087 + }, + { + "epoch": 5.309090909090909, + "grad_norm": 9.550762440921778, + "learning_rate": 9.489172073415777e-06, + "loss": 0.0293, + "step": 4088 + }, + { + "epoch": 5.310389610389611, + "grad_norm": 8.037049752814685, + "learning_rate": 9.48497143448712e-06, + "loss": 0.0093, + "step": 4089 + }, + { + "epoch": 5.311688311688312, + "grad_norm": 4.504627658187615, + "learning_rate": 9.480770886676878e-06, + "loss": 0.0256, + "step": 4090 + }, + { + "epoch": 5.312987012987013, + "grad_norm": 7.807218992981229, + "learning_rate": 9.4765704307282e-06, + "loss": 0.0337, + "step": 4091 + }, + { + "epoch": 5.314285714285714, + "grad_norm": 7.155689043559664, + "learning_rate": 9.472370067384231e-06, + "loss": 0.0399, + "step": 4092 + }, + { + "epoch": 5.315584415584415, + "grad_norm": 8.575781969338161, + "learning_rate": 9.468169797388098e-06, + "loss": 0.0359, + "step": 4093 + }, + { + "epoch": 5.316883116883117, + "grad_norm": 5.661139032759142, + "learning_rate": 9.46396962148291e-06, + "loss": 0.0284, + "step": 4094 + }, + { + "epoch": 5.318181818181818, + "grad_norm": 2.2880946936858173, + "learning_rate": 9.459769540411751e-06, + "loss": 0.0168, + "step": 4095 + }, + { + "epoch": 5.3194805194805195, + "grad_norm": 8.746877882900813, + "learning_rate": 9.455569554917701e-06, + "loss": 0.0425, + "step": 4096 + }, + { + "epoch": 5.320779220779221, + "grad_norm": 4.410552681796364, + "learning_rate": 9.45136966574382e-06, + "loss": 0.0187, + "step": 4097 + }, + { + "epoch": 5.322077922077922, + "grad_norm": 5.003309627780692, + "learning_rate": 9.447169873633141e-06, + "loss": 0.0194, + "step": 4098 + }, + { + "epoch": 5.323376623376624, + "grad_norm": 5.8936684608316, + "learning_rate": 9.442970179328698e-06, + "loss": 0.0092, + "step": 4099 + }, + { + "epoch": 5.324675324675325, + "grad_norm": 4.081233901709783, + "learning_rate": 9.43877058357349e-06, + "loss": 0.0258, + "step": 4100 + }, + { + "epoch": 5.325974025974026, + "grad_norm": 7.776666951747243, + "learning_rate": 9.434571087110508e-06, + "loss": 0.0316, + "step": 4101 + }, + { + "epoch": 5.327272727272727, + "grad_norm": 7.7848111854894135, + "learning_rate": 9.430371690682721e-06, + "loss": 0.0348, + "step": 4102 + }, + { + "epoch": 5.328571428571428, + "grad_norm": 13.786295707241223, + "learning_rate": 9.426172395033085e-06, + "loss": 0.0176, + "step": 4103 + }, + { + "epoch": 5.32987012987013, + "grad_norm": 5.84072005773077, + "learning_rate": 9.421973200904538e-06, + "loss": 0.0282, + "step": 4104 + }, + { + "epoch": 5.3311688311688314, + "grad_norm": 9.902143922825038, + "learning_rate": 9.417774109039997e-06, + "loss": 0.0261, + "step": 4105 + }, + { + "epoch": 5.332467532467533, + "grad_norm": 7.215243615152865, + "learning_rate": 9.413575120182356e-06, + "loss": 0.0095, + "step": 4106 + }, + { + "epoch": 5.333766233766234, + "grad_norm": 8.690357328827853, + "learning_rate": 9.4093762350745e-06, + "loss": 0.0125, + "step": 4107 + }, + { + "epoch": 5.335064935064935, + "grad_norm": 4.344741878733138, + "learning_rate": 9.405177454459293e-06, + "loss": 0.0049, + "step": 4108 + }, + { + "epoch": 5.336363636363636, + "grad_norm": 4.281540879572002, + "learning_rate": 9.400978779079585e-06, + "loss": 0.0244, + "step": 4109 + }, + { + "epoch": 5.337662337662338, + "grad_norm": 5.456483662100169, + "learning_rate": 9.396780209678191e-06, + "loss": 0.0079, + "step": 4110 + }, + { + "epoch": 5.338961038961039, + "grad_norm": 7.5937835353384875, + "learning_rate": 9.392581746997924e-06, + "loss": 0.0066, + "step": 4111 + }, + { + "epoch": 5.34025974025974, + "grad_norm": 5.532998104564765, + "learning_rate": 9.388383391781576e-06, + "loss": 0.0209, + "step": 4112 + }, + { + "epoch": 5.341558441558441, + "grad_norm": 4.144143269026497, + "learning_rate": 9.384185144771908e-06, + "loss": 0.0038, + "step": 4113 + }, + { + "epoch": 5.3428571428571425, + "grad_norm": 8.575570200891525, + "learning_rate": 9.379987006711675e-06, + "loss": 0.0385, + "step": 4114 + }, + { + "epoch": 5.3441558441558445, + "grad_norm": 7.930009388188024, + "learning_rate": 9.375788978343608e-06, + "loss": 0.0251, + "step": 4115 + }, + { + "epoch": 5.345454545454546, + "grad_norm": 4.544270746286005, + "learning_rate": 9.371591060410423e-06, + "loss": 0.0213, + "step": 4116 + }, + { + "epoch": 5.346753246753247, + "grad_norm": 5.392614920979447, + "learning_rate": 9.367393253654802e-06, + "loss": 0.0124, + "step": 4117 + }, + { + "epoch": 5.348051948051948, + "grad_norm": 6.92134693284094, + "learning_rate": 9.363195558819423e-06, + "loss": 0.046, + "step": 4118 + }, + { + "epoch": 5.349350649350649, + "grad_norm": 4.68778338205952, + "learning_rate": 9.358997976646937e-06, + "loss": 0.0052, + "step": 4119 + }, + { + "epoch": 5.35064935064935, + "grad_norm": 9.646803934717509, + "learning_rate": 9.354800507879983e-06, + "loss": 0.0307, + "step": 4120 + }, + { + "epoch": 5.351948051948052, + "grad_norm": 2.174700258355227, + "learning_rate": 9.350603153261165e-06, + "loss": 0.0096, + "step": 4121 + }, + { + "epoch": 5.353246753246753, + "grad_norm": 7.535747585916531, + "learning_rate": 9.346405913533078e-06, + "loss": 0.0301, + "step": 4122 + }, + { + "epoch": 5.3545454545454545, + "grad_norm": 11.09923105674405, + "learning_rate": 9.3422087894383e-06, + "loss": 0.0082, + "step": 4123 + }, + { + "epoch": 5.355844155844156, + "grad_norm": 6.863123396810975, + "learning_rate": 9.338011781719373e-06, + "loss": 0.0231, + "step": 4124 + }, + { + "epoch": 5.357142857142857, + "grad_norm": 9.401577100934274, + "learning_rate": 9.33381489111883e-06, + "loss": 0.0323, + "step": 4125 + }, + { + "epoch": 5.358441558441559, + "grad_norm": 8.983500972554836, + "learning_rate": 9.329618118379187e-06, + "loss": 0.0119, + "step": 4126 + }, + { + "epoch": 5.35974025974026, + "grad_norm": 3.3815944568586165, + "learning_rate": 9.325421464242934e-06, + "loss": 0.019, + "step": 4127 + }, + { + "epoch": 5.361038961038961, + "grad_norm": 13.9551480554424, + "learning_rate": 9.32122492945253e-06, + "loss": 0.0269, + "step": 4128 + }, + { + "epoch": 5.362337662337662, + "grad_norm": 4.114505300137799, + "learning_rate": 9.317028514750427e-06, + "loss": 0.012, + "step": 4129 + }, + { + "epoch": 5.363636363636363, + "grad_norm": 4.736671220847225, + "learning_rate": 9.312832220879053e-06, + "loss": 0.0128, + "step": 4130 + }, + { + "epoch": 5.364935064935065, + "grad_norm": 15.069369217708887, + "learning_rate": 9.308636048580813e-06, + "loss": 0.0128, + "step": 4131 + }, + { + "epoch": 5.3662337662337665, + "grad_norm": 4.165364652763495, + "learning_rate": 9.30443999859809e-06, + "loss": 0.0104, + "step": 4132 + }, + { + "epoch": 5.367532467532468, + "grad_norm": 9.797797936773717, + "learning_rate": 9.300244071673243e-06, + "loss": 0.007, + "step": 4133 + }, + { + "epoch": 5.368831168831169, + "grad_norm": 7.7970428923713895, + "learning_rate": 9.296048268548611e-06, + "loss": 0.0118, + "step": 4134 + }, + { + "epoch": 5.37012987012987, + "grad_norm": 6.659865550734716, + "learning_rate": 9.291852589966517e-06, + "loss": 0.0183, + "step": 4135 + }, + { + "epoch": 5.371428571428572, + "grad_norm": 11.097224527004359, + "learning_rate": 9.287657036669252e-06, + "loss": 0.0142, + "step": 4136 + }, + { + "epoch": 5.372727272727273, + "grad_norm": 7.132403196347906, + "learning_rate": 9.283461609399094e-06, + "loss": 0.0361, + "step": 4137 + }, + { + "epoch": 5.374025974025974, + "grad_norm": 6.7677815753621875, + "learning_rate": 9.279266308898292e-06, + "loss": 0.0336, + "step": 4138 + }, + { + "epoch": 5.375324675324675, + "grad_norm": 9.74099199482934, + "learning_rate": 9.275071135909073e-06, + "loss": 0.0157, + "step": 4139 + }, + { + "epoch": 5.376623376623376, + "grad_norm": 3.801054008438779, + "learning_rate": 9.270876091173645e-06, + "loss": 0.0164, + "step": 4140 + }, + { + "epoch": 5.3779220779220775, + "grad_norm": 4.615676024670213, + "learning_rate": 9.266681175434191e-06, + "loss": 0.0038, + "step": 4141 + }, + { + "epoch": 5.37922077922078, + "grad_norm": 9.048617303113163, + "learning_rate": 9.262486389432879e-06, + "loss": 0.0169, + "step": 4142 + }, + { + "epoch": 5.380519480519481, + "grad_norm": 6.655496584407359, + "learning_rate": 9.258291733911836e-06, + "loss": 0.0111, + "step": 4143 + }, + { + "epoch": 5.381818181818182, + "grad_norm": 2.436541587244108, + "learning_rate": 9.25409720961318e-06, + "loss": 0.0111, + "step": 4144 + }, + { + "epoch": 5.383116883116883, + "grad_norm": 5.961355362747317, + "learning_rate": 9.249902817279007e-06, + "loss": 0.0147, + "step": 4145 + }, + { + "epoch": 5.384415584415584, + "grad_norm": 8.95302677122483, + "learning_rate": 9.245708557651387e-06, + "loss": 0.0033, + "step": 4146 + }, + { + "epoch": 5.385714285714286, + "grad_norm": 3.3756245196031585, + "learning_rate": 9.241514431472354e-06, + "loss": 0.0117, + "step": 4147 + }, + { + "epoch": 5.387012987012987, + "grad_norm": 9.151674799118695, + "learning_rate": 9.237320439483936e-06, + "loss": 0.006, + "step": 4148 + }, + { + "epoch": 5.388311688311688, + "grad_norm": 5.974402755543333, + "learning_rate": 9.233126582428131e-06, + "loss": 0.0034, + "step": 4149 + }, + { + "epoch": 5.3896103896103895, + "grad_norm": 4.088428454324548, + "learning_rate": 9.228932861046916e-06, + "loss": 0.0102, + "step": 4150 + }, + { + "epoch": 5.390909090909091, + "grad_norm": 6.971423878357229, + "learning_rate": 9.224739276082232e-06, + "loss": 0.0155, + "step": 4151 + }, + { + "epoch": 5.392207792207792, + "grad_norm": 11.540091989974572, + "learning_rate": 9.220545828276008e-06, + "loss": 0.0091, + "step": 4152 + }, + { + "epoch": 5.393506493506494, + "grad_norm": 10.196373554026408, + "learning_rate": 9.216352518370151e-06, + "loss": 0.006, + "step": 4153 + }, + { + "epoch": 5.394805194805195, + "grad_norm": 9.76317343666602, + "learning_rate": 9.212159347106527e-06, + "loss": 0.0455, + "step": 4154 + }, + { + "epoch": 5.396103896103896, + "grad_norm": 9.144720024251232, + "learning_rate": 9.207966315226995e-06, + "loss": 0.0315, + "step": 4155 + }, + { + "epoch": 5.397402597402597, + "grad_norm": 5.156719591802072, + "learning_rate": 9.203773423473381e-06, + "loss": 0.0127, + "step": 4156 + }, + { + "epoch": 5.398701298701298, + "grad_norm": 6.386628958031842, + "learning_rate": 9.199580672587494e-06, + "loss": 0.0036, + "step": 4157 + }, + { + "epoch": 5.4, + "grad_norm": 3.3934566635873695, + "learning_rate": 9.1953880633111e-06, + "loss": 0.0094, + "step": 4158 + }, + { + "epoch": 5.4012987012987015, + "grad_norm": 4.689758250627169, + "learning_rate": 9.19119559638596e-06, + "loss": 0.006, + "step": 4159 + }, + { + "epoch": 5.402597402597403, + "grad_norm": 5.408581889880576, + "learning_rate": 9.187003272553798e-06, + "loss": 0.0036, + "step": 4160 + }, + { + "epoch": 5.403896103896104, + "grad_norm": 3.2170646570590518, + "learning_rate": 9.182811092556323e-06, + "loss": 0.0124, + "step": 4161 + }, + { + "epoch": 5.405194805194805, + "grad_norm": 4.982775555574111, + "learning_rate": 9.178619057135202e-06, + "loss": 0.0087, + "step": 4162 + }, + { + "epoch": 5.406493506493507, + "grad_norm": 3.6531119985169775, + "learning_rate": 9.174427167032092e-06, + "loss": 0.0029, + "step": 4163 + }, + { + "epoch": 5.407792207792208, + "grad_norm": 5.368414180578101, + "learning_rate": 9.170235422988621e-06, + "loss": 0.0102, + "step": 4164 + }, + { + "epoch": 5.409090909090909, + "grad_norm": 5.229133927720921, + "learning_rate": 9.166043825746382e-06, + "loss": 0.0097, + "step": 4165 + }, + { + "epoch": 5.41038961038961, + "grad_norm": 2.0661252458923767, + "learning_rate": 9.161852376046953e-06, + "loss": 0.0168, + "step": 4166 + }, + { + "epoch": 5.411688311688311, + "grad_norm": 7.024653836796597, + "learning_rate": 9.15766107463188e-06, + "loss": 0.0239, + "step": 4167 + }, + { + "epoch": 5.4129870129870135, + "grad_norm": 5.635036884601249, + "learning_rate": 9.15346992224269e-06, + "loss": 0.0062, + "step": 4168 + }, + { + "epoch": 5.414285714285715, + "grad_norm": 4.301912263144003, + "learning_rate": 9.149278919620868e-06, + "loss": 0.0087, + "step": 4169 + }, + { + "epoch": 5.415584415584416, + "grad_norm": 8.169326336196985, + "learning_rate": 9.14508806750789e-06, + "loss": 0.0182, + "step": 4170 + }, + { + "epoch": 5.416883116883117, + "grad_norm": 4.0072694203997195, + "learning_rate": 9.140897366645197e-06, + "loss": 0.0122, + "step": 4171 + }, + { + "epoch": 5.418181818181818, + "grad_norm": 6.677747817919119, + "learning_rate": 9.136706817774203e-06, + "loss": 0.0193, + "step": 4172 + }, + { + "epoch": 5.419480519480519, + "grad_norm": 9.282834741722112, + "learning_rate": 9.132516421636296e-06, + "loss": 0.0175, + "step": 4173 + }, + { + "epoch": 5.420779220779221, + "grad_norm": 5.202736595941343, + "learning_rate": 9.128326178972837e-06, + "loss": 0.0115, + "step": 4174 + }, + { + "epoch": 5.422077922077922, + "grad_norm": 7.448996425616351, + "learning_rate": 9.12413609052516e-06, + "loss": 0.0079, + "step": 4175 + }, + { + "epoch": 5.423376623376623, + "grad_norm": 5.939582866642941, + "learning_rate": 9.119946157034579e-06, + "loss": 0.0234, + "step": 4176 + }, + { + "epoch": 5.4246753246753245, + "grad_norm": 3.7721504423004566, + "learning_rate": 9.115756379242363e-06, + "loss": 0.0271, + "step": 4177 + }, + { + "epoch": 5.425974025974026, + "grad_norm": 10.35429573360774, + "learning_rate": 9.111566757889766e-06, + "loss": 0.0264, + "step": 4178 + }, + { + "epoch": 5.427272727272728, + "grad_norm": 4.389687987273185, + "learning_rate": 9.107377293718021e-06, + "loss": 0.0055, + "step": 4179 + }, + { + "epoch": 5.428571428571429, + "grad_norm": 9.191677697716472, + "learning_rate": 9.103187987468312e-06, + "loss": 0.0239, + "step": 4180 + }, + { + "epoch": 5.42987012987013, + "grad_norm": 6.504334691781371, + "learning_rate": 9.098998839881812e-06, + "loss": 0.0236, + "step": 4181 + }, + { + "epoch": 5.431168831168831, + "grad_norm": 4.7061524551204155, + "learning_rate": 9.094809851699664e-06, + "loss": 0.0221, + "step": 4182 + }, + { + "epoch": 5.432467532467532, + "grad_norm": 8.934116060910917, + "learning_rate": 9.090621023662985e-06, + "loss": 0.0071, + "step": 4183 + }, + { + "epoch": 5.433766233766233, + "grad_norm": 8.10178155254805, + "learning_rate": 9.086432356512845e-06, + "loss": 0.0186, + "step": 4184 + }, + { + "epoch": 5.435064935064935, + "grad_norm": 6.612564020684433, + "learning_rate": 9.082243850990309e-06, + "loss": 0.0049, + "step": 4185 + }, + { + "epoch": 5.4363636363636365, + "grad_norm": 9.31411661193262, + "learning_rate": 9.0780555078364e-06, + "loss": 0.0105, + "step": 4186 + }, + { + "epoch": 5.437662337662338, + "grad_norm": 7.340277863433134, + "learning_rate": 9.073867327792124e-06, + "loss": 0.0235, + "step": 4187 + }, + { + "epoch": 5.438961038961039, + "grad_norm": 3.170579675159996, + "learning_rate": 9.069679311598439e-06, + "loss": 0.0147, + "step": 4188 + }, + { + "epoch": 5.44025974025974, + "grad_norm": 10.970280483274264, + "learning_rate": 9.06549145999629e-06, + "loss": 0.0181, + "step": 4189 + }, + { + "epoch": 5.441558441558442, + "grad_norm": 4.85750246929088, + "learning_rate": 9.061303773726588e-06, + "loss": 0.0151, + "step": 4190 + }, + { + "epoch": 5.442857142857143, + "grad_norm": 3.204064391596986, + "learning_rate": 9.057116253530221e-06, + "loss": 0.0137, + "step": 4191 + }, + { + "epoch": 5.444155844155844, + "grad_norm": 1.6375288289619976, + "learning_rate": 9.052928900148031e-06, + "loss": 0.0094, + "step": 4192 + }, + { + "epoch": 5.445454545454545, + "grad_norm": 6.750138736163015, + "learning_rate": 9.048741714320846e-06, + "loss": 0.0322, + "step": 4193 + }, + { + "epoch": 5.4467532467532465, + "grad_norm": 7.602003348707022, + "learning_rate": 9.044554696789462e-06, + "loss": 0.023, + "step": 4194 + }, + { + "epoch": 5.448051948051948, + "grad_norm": 2.137415644152167, + "learning_rate": 9.040367848294637e-06, + "loss": 0.0037, + "step": 4195 + }, + { + "epoch": 5.44935064935065, + "grad_norm": 7.73761315801847, + "learning_rate": 9.036181169577107e-06, + "loss": 0.0122, + "step": 4196 + }, + { + "epoch": 5.450649350649351, + "grad_norm": 4.657500276301457, + "learning_rate": 9.031994661377575e-06, + "loss": 0.0046, + "step": 4197 + }, + { + "epoch": 5.451948051948052, + "grad_norm": 9.813554539853133, + "learning_rate": 9.027808324436716e-06, + "loss": 0.0125, + "step": 4198 + }, + { + "epoch": 5.453246753246753, + "grad_norm": 12.980758231880792, + "learning_rate": 9.023622159495174e-06, + "loss": 0.0263, + "step": 4199 + }, + { + "epoch": 5.454545454545454, + "grad_norm": 5.198912087256125, + "learning_rate": 9.019436167293558e-06, + "loss": 0.006, + "step": 4200 + }, + { + "epoch": 5.455844155844156, + "grad_norm": 5.193616325014676, + "learning_rate": 9.015250348572452e-06, + "loss": 0.0026, + "step": 4201 + }, + { + "epoch": 5.457142857142857, + "grad_norm": 6.739390802454743, + "learning_rate": 9.01106470407241e-06, + "loss": 0.0094, + "step": 4202 + }, + { + "epoch": 5.458441558441558, + "grad_norm": 7.666692854853935, + "learning_rate": 9.006879234533948e-06, + "loss": 0.0079, + "step": 4203 + }, + { + "epoch": 5.45974025974026, + "grad_norm": 7.9066871113842945, + "learning_rate": 9.00269394069756e-06, + "loss": 0.0152, + "step": 4204 + }, + { + "epoch": 5.461038961038961, + "grad_norm": 8.566859937966678, + "learning_rate": 8.998508823303703e-06, + "loss": 0.0085, + "step": 4205 + }, + { + "epoch": 5.462337662337663, + "grad_norm": 5.203493614173241, + "learning_rate": 8.9943238830928e-06, + "loss": 0.0091, + "step": 4206 + }, + { + "epoch": 5.463636363636364, + "grad_norm": 6.092753359340959, + "learning_rate": 8.990139120805252e-06, + "loss": 0.0158, + "step": 4207 + }, + { + "epoch": 5.464935064935065, + "grad_norm": 7.486703855841767, + "learning_rate": 8.985954537181423e-06, + "loss": 0.0131, + "step": 4208 + }, + { + "epoch": 5.466233766233766, + "grad_norm": 4.066120432160411, + "learning_rate": 8.981770132961649e-06, + "loss": 0.0082, + "step": 4209 + }, + { + "epoch": 5.467532467532467, + "grad_norm": 12.960490033064282, + "learning_rate": 8.977585908886223e-06, + "loss": 0.0276, + "step": 4210 + }, + { + "epoch": 5.468831168831169, + "grad_norm": 6.2975289461982005, + "learning_rate": 8.97340186569542e-06, + "loss": 0.0107, + "step": 4211 + }, + { + "epoch": 5.47012987012987, + "grad_norm": 6.2551284357729005, + "learning_rate": 8.969218004129475e-06, + "loss": 0.0174, + "step": 4212 + }, + { + "epoch": 5.4714285714285715, + "grad_norm": 3.2804401114263766, + "learning_rate": 8.9650343249286e-06, + "loss": 0.0116, + "step": 4213 + }, + { + "epoch": 5.472727272727273, + "grad_norm": 2.8362579006826762, + "learning_rate": 8.960850828832958e-06, + "loss": 0.0191, + "step": 4214 + }, + { + "epoch": 5.474025974025974, + "grad_norm": 3.7448046767439327, + "learning_rate": 8.956667516582694e-06, + "loss": 0.0188, + "step": 4215 + }, + { + "epoch": 5.475324675324675, + "grad_norm": 1.5435093109648697, + "learning_rate": 8.952484388917914e-06, + "loss": 0.0114, + "step": 4216 + }, + { + "epoch": 5.476623376623377, + "grad_norm": 8.844271175748974, + "learning_rate": 8.948301446578702e-06, + "loss": 0.0204, + "step": 4217 + }, + { + "epoch": 5.477922077922078, + "grad_norm": 5.8298924073375185, + "learning_rate": 8.944118690305088e-06, + "loss": 0.0197, + "step": 4218 + }, + { + "epoch": 5.479220779220779, + "grad_norm": 8.560255250490076, + "learning_rate": 8.939936120837089e-06, + "loss": 0.0172, + "step": 4219 + }, + { + "epoch": 5.48051948051948, + "grad_norm": 4.5198559657970145, + "learning_rate": 8.935753738914684e-06, + "loss": 0.0313, + "step": 4220 + }, + { + "epoch": 5.4818181818181815, + "grad_norm": 6.586285127251748, + "learning_rate": 8.931571545277805e-06, + "loss": 0.029, + "step": 4221 + }, + { + "epoch": 5.4831168831168835, + "grad_norm": 8.928596000558676, + "learning_rate": 8.927389540666374e-06, + "loss": 0.0036, + "step": 4222 + }, + { + "epoch": 5.484415584415585, + "grad_norm": 4.348380660563461, + "learning_rate": 8.92320772582026e-06, + "loss": 0.0068, + "step": 4223 + }, + { + "epoch": 5.485714285714286, + "grad_norm": 11.490641850791928, + "learning_rate": 8.919026101479316e-06, + "loss": 0.0099, + "step": 4224 + }, + { + "epoch": 5.487012987012987, + "grad_norm": 6.934739219285602, + "learning_rate": 8.91484466838334e-06, + "loss": 0.0197, + "step": 4225 + }, + { + "epoch": 5.488311688311688, + "grad_norm": 2.3647240794160407, + "learning_rate": 8.91066342727211e-06, + "loss": 0.016, + "step": 4226 + }, + { + "epoch": 5.489610389610389, + "grad_norm": 9.20247133563294, + "learning_rate": 8.906482378885372e-06, + "loss": 0.0193, + "step": 4227 + }, + { + "epoch": 5.490909090909091, + "grad_norm": 4.6259777787117855, + "learning_rate": 8.902301523962835e-06, + "loss": 0.0095, + "step": 4228 + }, + { + "epoch": 5.492207792207792, + "grad_norm": 3.583148888832798, + "learning_rate": 8.898120863244164e-06, + "loss": 0.0051, + "step": 4229 + }, + { + "epoch": 5.4935064935064934, + "grad_norm": 5.219901658426453, + "learning_rate": 8.893940397469003e-06, + "loss": 0.0089, + "step": 4230 + }, + { + "epoch": 5.494805194805195, + "grad_norm": 6.3817688917803075, + "learning_rate": 8.889760127376958e-06, + "loss": 0.0127, + "step": 4231 + }, + { + "epoch": 5.496103896103896, + "grad_norm": 7.611301938389563, + "learning_rate": 8.885580053707595e-06, + "loss": 0.0373, + "step": 4232 + }, + { + "epoch": 5.497402597402598, + "grad_norm": 7.805227924863915, + "learning_rate": 8.881400177200452e-06, + "loss": 0.0034, + "step": 4233 + }, + { + "epoch": 5.498701298701299, + "grad_norm": 4.631896876977988, + "learning_rate": 8.877220498595026e-06, + "loss": 0.0062, + "step": 4234 + }, + { + "epoch": 5.5, + "grad_norm": 6.801630945788617, + "learning_rate": 8.873041018630786e-06, + "loss": 0.0402, + "step": 4235 + }, + { + "epoch": 5.501298701298701, + "grad_norm": 12.799021737992128, + "learning_rate": 8.868861738047158e-06, + "loss": 0.0229, + "step": 4236 + }, + { + "epoch": 5.502597402597402, + "grad_norm": 3.803566655127701, + "learning_rate": 8.864682657583538e-06, + "loss": 0.0031, + "step": 4237 + }, + { + "epoch": 5.503896103896103, + "grad_norm": 3.465213884422271, + "learning_rate": 8.86050377797929e-06, + "loss": 0.0071, + "step": 4238 + }, + { + "epoch": 5.505194805194805, + "grad_norm": 12.217318311979767, + "learning_rate": 8.856325099973732e-06, + "loss": 0.0131, + "step": 4239 + }, + { + "epoch": 5.5064935064935066, + "grad_norm": 13.687296276282712, + "learning_rate": 8.852146624306154e-06, + "loss": 0.017, + "step": 4240 + }, + { + "epoch": 5.507792207792208, + "grad_norm": 4.177690294678279, + "learning_rate": 8.847968351715808e-06, + "loss": 0.0205, + "step": 4241 + }, + { + "epoch": 5.509090909090909, + "grad_norm": 5.380743607220673, + "learning_rate": 8.843790282941909e-06, + "loss": 0.0103, + "step": 4242 + }, + { + "epoch": 5.510389610389611, + "grad_norm": 6.103327948871165, + "learning_rate": 8.839612418723645e-06, + "loss": 0.0192, + "step": 4243 + }, + { + "epoch": 5.511688311688312, + "grad_norm": 4.349252955964311, + "learning_rate": 8.835434759800148e-06, + "loss": 0.0167, + "step": 4244 + }, + { + "epoch": 5.512987012987013, + "grad_norm": 8.498520159703562, + "learning_rate": 8.831257306910534e-06, + "loss": 0.0217, + "step": 4245 + }, + { + "epoch": 5.514285714285714, + "grad_norm": 2.1264758594231354, + "learning_rate": 8.827080060793875e-06, + "loss": 0.0124, + "step": 4246 + }, + { + "epoch": 5.515584415584415, + "grad_norm": 4.041052854014161, + "learning_rate": 8.8229030221892e-06, + "loss": 0.0045, + "step": 4247 + }, + { + "epoch": 5.5168831168831165, + "grad_norm": 10.41079963018184, + "learning_rate": 8.818726191835509e-06, + "loss": 0.0323, + "step": 4248 + }, + { + "epoch": 5.5181818181818185, + "grad_norm": 8.218254901435941, + "learning_rate": 8.814549570471766e-06, + "loss": 0.0037, + "step": 4249 + }, + { + "epoch": 5.51948051948052, + "grad_norm": 7.367447963461757, + "learning_rate": 8.810373158836896e-06, + "loss": 0.025, + "step": 4250 + }, + { + "epoch": 5.520779220779221, + "grad_norm": 8.9300221967897, + "learning_rate": 8.806196957669781e-06, + "loss": 0.0186, + "step": 4251 + }, + { + "epoch": 5.522077922077922, + "grad_norm": 5.101802937103561, + "learning_rate": 8.802020967709275e-06, + "loss": 0.0533, + "step": 4252 + }, + { + "epoch": 5.523376623376623, + "grad_norm": 14.843077854491558, + "learning_rate": 8.797845189694188e-06, + "loss": 0.0447, + "step": 4253 + }, + { + "epoch": 5.524675324675325, + "grad_norm": 3.6191231185028396, + "learning_rate": 8.793669624363302e-06, + "loss": 0.0103, + "step": 4254 + }, + { + "epoch": 5.525974025974026, + "grad_norm": 2.661064406331444, + "learning_rate": 8.789494272455342e-06, + "loss": 0.0113, + "step": 4255 + }, + { + "epoch": 5.527272727272727, + "grad_norm": 8.730585385689185, + "learning_rate": 8.785319134709015e-06, + "loss": 0.0253, + "step": 4256 + }, + { + "epoch": 5.5285714285714285, + "grad_norm": 2.4187041016021, + "learning_rate": 8.781144211862984e-06, + "loss": 0.0045, + "step": 4257 + }, + { + "epoch": 5.52987012987013, + "grad_norm": 5.947051687526051, + "learning_rate": 8.776969504655875e-06, + "loss": 0.02, + "step": 4258 + }, + { + "epoch": 5.531168831168831, + "grad_norm": 5.863365729843592, + "learning_rate": 8.772795013826266e-06, + "loss": 0.0107, + "step": 4259 + }, + { + "epoch": 5.532467532467533, + "grad_norm": 3.749311843697149, + "learning_rate": 8.768620740112706e-06, + "loss": 0.0163, + "step": 4260 + }, + { + "epoch": 5.533766233766234, + "grad_norm": 4.620284552330405, + "learning_rate": 8.76444668425371e-06, + "loss": 0.0147, + "step": 4261 + }, + { + "epoch": 5.535064935064935, + "grad_norm": 6.384151919611918, + "learning_rate": 8.760272846987743e-06, + "loss": 0.0172, + "step": 4262 + }, + { + "epoch": 5.536363636363636, + "grad_norm": 7.1609276848774295, + "learning_rate": 8.756099229053236e-06, + "loss": 0.05, + "step": 4263 + }, + { + "epoch": 5.537662337662337, + "grad_norm": 5.9492062571972175, + "learning_rate": 8.751925831188584e-06, + "loss": 0.0108, + "step": 4264 + }, + { + "epoch": 5.538961038961039, + "grad_norm": 3.284243153620394, + "learning_rate": 8.747752654132144e-06, + "loss": 0.0252, + "step": 4265 + }, + { + "epoch": 5.54025974025974, + "grad_norm": 3.4470366000324906, + "learning_rate": 8.743579698622227e-06, + "loss": 0.0051, + "step": 4266 + }, + { + "epoch": 5.541558441558442, + "grad_norm": 5.017504711966994, + "learning_rate": 8.739406965397107e-06, + "loss": 0.0217, + "step": 4267 + }, + { + "epoch": 5.542857142857143, + "grad_norm": 6.1940508619287415, + "learning_rate": 8.735234455195023e-06, + "loss": 0.0077, + "step": 4268 + }, + { + "epoch": 5.544155844155844, + "grad_norm": 2.83024535029762, + "learning_rate": 8.731062168754174e-06, + "loss": 0.0181, + "step": 4269 + }, + { + "epoch": 5.545454545454545, + "grad_norm": 3.986677164037336, + "learning_rate": 8.726890106812713e-06, + "loss": 0.0113, + "step": 4270 + }, + { + "epoch": 5.546753246753247, + "grad_norm": 4.762116824859102, + "learning_rate": 8.72271827010876e-06, + "loss": 0.0126, + "step": 4271 + }, + { + "epoch": 5.548051948051948, + "grad_norm": 2.7508462480708684, + "learning_rate": 8.718546659380396e-06, + "loss": 0.0148, + "step": 4272 + }, + { + "epoch": 5.549350649350649, + "grad_norm": 4.291654677198973, + "learning_rate": 8.714375275365652e-06, + "loss": 0.02, + "step": 4273 + }, + { + "epoch": 5.55064935064935, + "grad_norm": 2.82038917476549, + "learning_rate": 8.710204118802526e-06, + "loss": 0.0167, + "step": 4274 + }, + { + "epoch": 5.551948051948052, + "grad_norm": 3.901761862229547, + "learning_rate": 8.706033190428981e-06, + "loss": 0.0197, + "step": 4275 + }, + { + "epoch": 5.5532467532467535, + "grad_norm": 4.019750605472879, + "learning_rate": 8.701862490982936e-06, + "loss": 0.0157, + "step": 4276 + }, + { + "epoch": 5.554545454545455, + "grad_norm": 6.060312470504485, + "learning_rate": 8.697692021202257e-06, + "loss": 0.0141, + "step": 4277 + }, + { + "epoch": 5.555844155844156, + "grad_norm": 5.172159663617861, + "learning_rate": 8.693521781824788e-06, + "loss": 0.0089, + "step": 4278 + }, + { + "epoch": 5.557142857142857, + "grad_norm": 6.081037981074208, + "learning_rate": 8.68935177358832e-06, + "loss": 0.0057, + "step": 4279 + }, + { + "epoch": 5.558441558441558, + "grad_norm": 5.418311612208101, + "learning_rate": 8.685181997230616e-06, + "loss": 0.0375, + "step": 4280 + }, + { + "epoch": 5.55974025974026, + "grad_norm": 5.283806541534198, + "learning_rate": 8.681012453489375e-06, + "loss": 0.0335, + "step": 4281 + }, + { + "epoch": 5.561038961038961, + "grad_norm": 4.165222896711696, + "learning_rate": 8.676843143102278e-06, + "loss": 0.0079, + "step": 4282 + }, + { + "epoch": 5.562337662337662, + "grad_norm": 4.7195513671492115, + "learning_rate": 8.672674066806955e-06, + "loss": 0.0436, + "step": 4283 + }, + { + "epoch": 5.5636363636363635, + "grad_norm": 4.54678206126649, + "learning_rate": 8.668505225341e-06, + "loss": 0.0287, + "step": 4284 + }, + { + "epoch": 5.564935064935065, + "grad_norm": 4.8268813509524575, + "learning_rate": 8.66433661944195e-06, + "loss": 0.0128, + "step": 4285 + }, + { + "epoch": 5.566233766233767, + "grad_norm": 3.5837885314464573, + "learning_rate": 8.660168249847318e-06, + "loss": 0.0127, + "step": 4286 + }, + { + "epoch": 5.567532467532468, + "grad_norm": 2.2002705005991663, + "learning_rate": 8.656000117294571e-06, + "loss": 0.0145, + "step": 4287 + }, + { + "epoch": 5.568831168831169, + "grad_norm": 7.984165734340331, + "learning_rate": 8.651832222521122e-06, + "loss": 0.0137, + "step": 4288 + }, + { + "epoch": 5.57012987012987, + "grad_norm": 5.880599139713601, + "learning_rate": 8.64766456626436e-06, + "loss": 0.0212, + "step": 4289 + }, + { + "epoch": 5.571428571428571, + "grad_norm": 2.3833070971213424, + "learning_rate": 8.643497149261616e-06, + "loss": 0.0025, + "step": 4290 + }, + { + "epoch": 5.572727272727272, + "grad_norm": 5.085162067424273, + "learning_rate": 8.639329972250199e-06, + "loss": 0.0103, + "step": 4291 + }, + { + "epoch": 5.574025974025974, + "grad_norm": 3.7200383607679983, + "learning_rate": 8.635163035967347e-06, + "loss": 0.0187, + "step": 4292 + }, + { + "epoch": 5.5753246753246755, + "grad_norm": 4.158419913980844, + "learning_rate": 8.630996341150277e-06, + "loss": 0.0252, + "step": 4293 + }, + { + "epoch": 5.576623376623377, + "grad_norm": 4.367956928472087, + "learning_rate": 8.626829888536157e-06, + "loss": 0.0032, + "step": 4294 + }, + { + "epoch": 5.577922077922078, + "grad_norm": 9.220626138524905, + "learning_rate": 8.622663678862118e-06, + "loss": 0.0142, + "step": 4295 + }, + { + "epoch": 5.579220779220779, + "grad_norm": 4.693484549899473, + "learning_rate": 8.61849771286523e-06, + "loss": 0.0051, + "step": 4296 + }, + { + "epoch": 5.580519480519481, + "grad_norm": 8.545652464153372, + "learning_rate": 8.614331991282541e-06, + "loss": 0.0119, + "step": 4297 + }, + { + "epoch": 5.581818181818182, + "grad_norm": 3.7565501906769287, + "learning_rate": 8.610166514851043e-06, + "loss": 0.0058, + "step": 4298 + }, + { + "epoch": 5.583116883116883, + "grad_norm": 6.729172150278371, + "learning_rate": 8.606001284307694e-06, + "loss": 0.0062, + "step": 4299 + }, + { + "epoch": 5.584415584415584, + "grad_norm": 9.791957844833062, + "learning_rate": 8.601836300389401e-06, + "loss": 0.0144, + "step": 4300 + }, + { + "epoch": 5.585714285714285, + "grad_norm": 12.324447200485166, + "learning_rate": 8.597671563833022e-06, + "loss": 0.0244, + "step": 4301 + }, + { + "epoch": 5.5870129870129865, + "grad_norm": 5.94218033360246, + "learning_rate": 8.593507075375388e-06, + "loss": 0.0125, + "step": 4302 + }, + { + "epoch": 5.588311688311689, + "grad_norm": 6.01693157329548, + "learning_rate": 8.58934283575327e-06, + "loss": 0.0048, + "step": 4303 + }, + { + "epoch": 5.58961038961039, + "grad_norm": 7.287973821014067, + "learning_rate": 8.585178845703404e-06, + "loss": 0.0119, + "step": 4304 + }, + { + "epoch": 5.590909090909091, + "grad_norm": 10.489571188618473, + "learning_rate": 8.581015105962484e-06, + "loss": 0.0053, + "step": 4305 + }, + { + "epoch": 5.592207792207792, + "grad_norm": 7.307041616752041, + "learning_rate": 8.576851617267151e-06, + "loss": 0.007, + "step": 4306 + }, + { + "epoch": 5.593506493506494, + "grad_norm": 7.989407124575179, + "learning_rate": 8.572688380354005e-06, + "loss": 0.036, + "step": 4307 + }, + { + "epoch": 5.594805194805195, + "grad_norm": 14.054578287600357, + "learning_rate": 8.568525395959601e-06, + "loss": 0.0185, + "step": 4308 + }, + { + "epoch": 5.596103896103896, + "grad_norm": 9.264826274333217, + "learning_rate": 8.564362664820455e-06, + "loss": 0.0112, + "step": 4309 + }, + { + "epoch": 5.597402597402597, + "grad_norm": 13.595082381145883, + "learning_rate": 8.560200187673036e-06, + "loss": 0.0378, + "step": 4310 + }, + { + "epoch": 5.5987012987012985, + "grad_norm": 11.650346913553426, + "learning_rate": 8.556037965253756e-06, + "loss": 0.0092, + "step": 4311 + }, + { + "epoch": 5.6, + "grad_norm": 3.6993456191777496, + "learning_rate": 8.551875998299e-06, + "loss": 0.034, + "step": 4312 + }, + { + "epoch": 5.601298701298702, + "grad_norm": 14.011141031459559, + "learning_rate": 8.5477142875451e-06, + "loss": 0.0269, + "step": 4313 + }, + { + "epoch": 5.602597402597403, + "grad_norm": 6.470067065791192, + "learning_rate": 8.543552833728336e-06, + "loss": 0.0507, + "step": 4314 + }, + { + "epoch": 5.603896103896104, + "grad_norm": 8.556611202340308, + "learning_rate": 8.539391637584952e-06, + "loss": 0.0053, + "step": 4315 + }, + { + "epoch": 5.605194805194805, + "grad_norm": 4.928961125203284, + "learning_rate": 8.535230699851145e-06, + "loss": 0.0194, + "step": 4316 + }, + { + "epoch": 5.606493506493506, + "grad_norm": 14.322345029156983, + "learning_rate": 8.531070021263066e-06, + "loss": 0.005, + "step": 4317 + }, + { + "epoch": 5.607792207792208, + "grad_norm": 8.27511827984989, + "learning_rate": 8.526909602556812e-06, + "loss": 0.0076, + "step": 4318 + }, + { + "epoch": 5.609090909090909, + "grad_norm": 6.61938723350351, + "learning_rate": 8.522749444468445e-06, + "loss": 0.0101, + "step": 4319 + }, + { + "epoch": 5.6103896103896105, + "grad_norm": 17.107511704945484, + "learning_rate": 8.518589547733976e-06, + "loss": 0.0326, + "step": 4320 + }, + { + "epoch": 5.611688311688312, + "grad_norm": 8.921744688971742, + "learning_rate": 8.514429913089374e-06, + "loss": 0.0169, + "step": 4321 + }, + { + "epoch": 5.612987012987013, + "grad_norm": 11.639795591382692, + "learning_rate": 8.510270541270552e-06, + "loss": 0.006, + "step": 4322 + }, + { + "epoch": 5.614285714285714, + "grad_norm": 15.0857421837321, + "learning_rate": 8.506111433013381e-06, + "loss": 0.0173, + "step": 4323 + }, + { + "epoch": 5.615584415584416, + "grad_norm": 3.2669861686058868, + "learning_rate": 8.501952589053694e-06, + "loss": 0.0319, + "step": 4324 + }, + { + "epoch": 5.616883116883117, + "grad_norm": 13.588738237379939, + "learning_rate": 8.497794010127272e-06, + "loss": 0.0356, + "step": 4325 + }, + { + "epoch": 5.618181818181818, + "grad_norm": 10.257433927154358, + "learning_rate": 8.493635696969836e-06, + "loss": 0.0254, + "step": 4326 + }, + { + "epoch": 5.619480519480519, + "grad_norm": 7.284868229418099, + "learning_rate": 8.489477650317076e-06, + "loss": 0.0199, + "step": 4327 + }, + { + "epoch": 5.62077922077922, + "grad_norm": 13.210391584814646, + "learning_rate": 8.48531987090464e-06, + "loss": 0.02, + "step": 4328 + }, + { + "epoch": 5.6220779220779225, + "grad_norm": 3.3620559850941945, + "learning_rate": 8.481162359468103e-06, + "loss": 0.0034, + "step": 4329 + }, + { + "epoch": 5.623376623376624, + "grad_norm": 6.873863547273562, + "learning_rate": 8.477005116743016e-06, + "loss": 0.0093, + "step": 4330 + }, + { + "epoch": 5.624675324675325, + "grad_norm": 5.464254510177294, + "learning_rate": 8.472848143464873e-06, + "loss": 0.0168, + "step": 4331 + }, + { + "epoch": 5.625974025974026, + "grad_norm": 2.145759920174658, + "learning_rate": 8.468691440369127e-06, + "loss": 0.0123, + "step": 4332 + }, + { + "epoch": 5.627272727272727, + "grad_norm": 5.178833871602706, + "learning_rate": 8.464535008191175e-06, + "loss": 0.0228, + "step": 4333 + }, + { + "epoch": 5.628571428571428, + "grad_norm": 8.685041635663936, + "learning_rate": 8.460378847666366e-06, + "loss": 0.0164, + "step": 4334 + }, + { + "epoch": 5.62987012987013, + "grad_norm": 7.527499591485661, + "learning_rate": 8.456222959530008e-06, + "loss": 0.0191, + "step": 4335 + }, + { + "epoch": 5.631168831168831, + "grad_norm": 4.422116486340864, + "learning_rate": 8.45206734451736e-06, + "loss": 0.0154, + "step": 4336 + }, + { + "epoch": 5.632467532467532, + "grad_norm": 3.1319195346711632, + "learning_rate": 8.447912003363622e-06, + "loss": 0.0038, + "step": 4337 + }, + { + "epoch": 5.6337662337662335, + "grad_norm": 5.199020543848418, + "learning_rate": 8.443756936803962e-06, + "loss": 0.0077, + "step": 4338 + }, + { + "epoch": 5.6350649350649356, + "grad_norm": 8.031741116493729, + "learning_rate": 8.439602145573483e-06, + "loss": 0.0255, + "step": 4339 + }, + { + "epoch": 5.636363636363637, + "grad_norm": 3.822346127440298, + "learning_rate": 8.435447630407253e-06, + "loss": 0.0033, + "step": 4340 + }, + { + "epoch": 5.637662337662338, + "grad_norm": 7.873917440731759, + "learning_rate": 8.431293392040283e-06, + "loss": 0.0034, + "step": 4341 + }, + { + "epoch": 5.638961038961039, + "grad_norm": 5.423779445749138, + "learning_rate": 8.427139431207536e-06, + "loss": 0.0197, + "step": 4342 + }, + { + "epoch": 5.64025974025974, + "grad_norm": 6.428073843409268, + "learning_rate": 8.422985748643932e-06, + "loss": 0.0039, + "step": 4343 + }, + { + "epoch": 5.641558441558441, + "grad_norm": 6.668552881854818, + "learning_rate": 8.41883234508433e-06, + "loss": 0.0168, + "step": 4344 + }, + { + "epoch": 5.642857142857143, + "grad_norm": 7.151068122010664, + "learning_rate": 8.41467922126355e-06, + "loss": 0.0221, + "step": 4345 + }, + { + "epoch": 5.644155844155844, + "grad_norm": 2.5536514482521464, + "learning_rate": 8.41052637791636e-06, + "loss": 0.0039, + "step": 4346 + }, + { + "epoch": 5.6454545454545455, + "grad_norm": 6.192123732633804, + "learning_rate": 8.406373815777478e-06, + "loss": 0.0147, + "step": 4347 + }, + { + "epoch": 5.646753246753247, + "grad_norm": 4.529763363695295, + "learning_rate": 8.402221535581568e-06, + "loss": 0.0108, + "step": 4348 + }, + { + "epoch": 5.648051948051948, + "grad_norm": 1.9292485393170469, + "learning_rate": 8.39806953806325e-06, + "loss": 0.0091, + "step": 4349 + }, + { + "epoch": 5.64935064935065, + "grad_norm": 6.478733142222264, + "learning_rate": 8.393917823957092e-06, + "loss": 0.0196, + "step": 4350 + }, + { + "epoch": 5.650649350649351, + "grad_norm": 5.697463999182433, + "learning_rate": 8.389766393997617e-06, + "loss": 0.0126, + "step": 4351 + }, + { + "epoch": 5.651948051948052, + "grad_norm": 1.5480790603552856, + "learning_rate": 8.38561524891928e-06, + "loss": 0.0165, + "step": 4352 + }, + { + "epoch": 5.653246753246753, + "grad_norm": 7.65542823616315, + "learning_rate": 8.381464389456509e-06, + "loss": 0.0034, + "step": 4353 + }, + { + "epoch": 5.654545454545454, + "grad_norm": 7.64417086768213, + "learning_rate": 8.377313816343669e-06, + "loss": 0.0098, + "step": 4354 + }, + { + "epoch": 5.6558441558441555, + "grad_norm": 12.258609701118651, + "learning_rate": 8.373163530315069e-06, + "loss": 0.0286, + "step": 4355 + }, + { + "epoch": 5.6571428571428575, + "grad_norm": 5.4187228337095865, + "learning_rate": 8.369013532104979e-06, + "loss": 0.0032, + "step": 4356 + }, + { + "epoch": 5.658441558441559, + "grad_norm": 2.6167213730435472, + "learning_rate": 8.364863822447612e-06, + "loss": 0.0033, + "step": 4357 + }, + { + "epoch": 5.65974025974026, + "grad_norm": 7.446743127135002, + "learning_rate": 8.360714402077134e-06, + "loss": 0.0166, + "step": 4358 + }, + { + "epoch": 5.661038961038961, + "grad_norm": 8.550045557036587, + "learning_rate": 8.356565271727653e-06, + "loss": 0.0078, + "step": 4359 + }, + { + "epoch": 5.662337662337662, + "grad_norm": 5.973601106043394, + "learning_rate": 8.35241643213323e-06, + "loss": 0.0123, + "step": 4360 + }, + { + "epoch": 5.663636363636364, + "grad_norm": 11.954912059672465, + "learning_rate": 8.348267884027876e-06, + "loss": 0.0169, + "step": 4361 + }, + { + "epoch": 5.664935064935065, + "grad_norm": 5.551489431492206, + "learning_rate": 8.344119628145552e-06, + "loss": 0.008, + "step": 4362 + }, + { + "epoch": 5.666233766233766, + "grad_norm": 7.780003062512353, + "learning_rate": 8.339971665220155e-06, + "loss": 0.0078, + "step": 4363 + }, + { + "epoch": 5.667532467532467, + "grad_norm": 4.331298689047107, + "learning_rate": 8.335823995985543e-06, + "loss": 0.004, + "step": 4364 + }, + { + "epoch": 5.6688311688311686, + "grad_norm": 3.807640742556662, + "learning_rate": 8.33167662117552e-06, + "loss": 0.0122, + "step": 4365 + }, + { + "epoch": 5.67012987012987, + "grad_norm": 5.790961516306531, + "learning_rate": 8.327529541523835e-06, + "loss": 0.0153, + "step": 4366 + }, + { + "epoch": 5.671428571428572, + "grad_norm": 12.830644096234503, + "learning_rate": 8.323382757764189e-06, + "loss": 0.0311, + "step": 4367 + }, + { + "epoch": 5.672727272727273, + "grad_norm": 6.454067359658479, + "learning_rate": 8.319236270630218e-06, + "loss": 0.0226, + "step": 4368 + }, + { + "epoch": 5.674025974025974, + "grad_norm": 11.164212422589442, + "learning_rate": 8.315090080855526e-06, + "loss": 0.0094, + "step": 4369 + }, + { + "epoch": 5.675324675324675, + "grad_norm": 5.990549055976564, + "learning_rate": 8.310944189173644e-06, + "loss": 0.0129, + "step": 4370 + }, + { + "epoch": 5.676623376623376, + "grad_norm": 3.314282768853619, + "learning_rate": 8.306798596318065e-06, + "loss": 0.005, + "step": 4371 + }, + { + "epoch": 5.677922077922078, + "grad_norm": 5.423964645473516, + "learning_rate": 8.302653303022224e-06, + "loss": 0.0152, + "step": 4372 + }, + { + "epoch": 5.679220779220779, + "grad_norm": 5.297575627586775, + "learning_rate": 8.298508310019502e-06, + "loss": 0.0125, + "step": 4373 + }, + { + "epoch": 5.6805194805194805, + "grad_norm": 6.511097712071234, + "learning_rate": 8.294363618043225e-06, + "loss": 0.019, + "step": 4374 + }, + { + "epoch": 5.681818181818182, + "grad_norm": 3.255156868425817, + "learning_rate": 8.29021922782667e-06, + "loss": 0.026, + "step": 4375 + }, + { + "epoch": 5.683116883116883, + "grad_norm": 8.755453968499037, + "learning_rate": 8.286075140103058e-06, + "loss": 0.0104, + "step": 4376 + }, + { + "epoch": 5.684415584415584, + "grad_norm": 7.467865196320131, + "learning_rate": 8.281931355605564e-06, + "loss": 0.0035, + "step": 4377 + }, + { + "epoch": 5.685714285714286, + "grad_norm": 6.999340997818356, + "learning_rate": 8.277787875067292e-06, + "loss": 0.0105, + "step": 4378 + }, + { + "epoch": 5.687012987012987, + "grad_norm": 7.221177122745981, + "learning_rate": 8.273644699221309e-06, + "loss": 0.051, + "step": 4379 + }, + { + "epoch": 5.688311688311688, + "grad_norm": 6.293584311501157, + "learning_rate": 8.269501828800626e-06, + "loss": 0.017, + "step": 4380 + }, + { + "epoch": 5.689610389610389, + "grad_norm": 5.65310373958141, + "learning_rate": 8.265359264538186e-06, + "loss": 0.0125, + "step": 4381 + }, + { + "epoch": 5.690909090909091, + "grad_norm": 9.29021195933242, + "learning_rate": 8.261217007166894e-06, + "loss": 0.0094, + "step": 4382 + }, + { + "epoch": 5.6922077922077925, + "grad_norm": 5.467182920284222, + "learning_rate": 8.257075057419595e-06, + "loss": 0.0169, + "step": 4383 + }, + { + "epoch": 5.693506493506494, + "grad_norm": 7.124283125137251, + "learning_rate": 8.252933416029082e-06, + "loss": 0.021, + "step": 4384 + }, + { + "epoch": 5.694805194805195, + "grad_norm": 16.223306173177598, + "learning_rate": 8.248792083728085e-06, + "loss": 0.0092, + "step": 4385 + }, + { + "epoch": 5.696103896103896, + "grad_norm": 4.028079411997529, + "learning_rate": 8.244651061249283e-06, + "loss": 0.0027, + "step": 4386 + }, + { + "epoch": 5.697402597402597, + "grad_norm": 11.921688936663042, + "learning_rate": 8.240510349325308e-06, + "loss": 0.0076, + "step": 4387 + }, + { + "epoch": 5.698701298701299, + "grad_norm": 12.143251237985378, + "learning_rate": 8.236369948688734e-06, + "loss": 0.0282, + "step": 4388 + }, + { + "epoch": 5.7, + "grad_norm": 3.846407201423879, + "learning_rate": 8.232229860072067e-06, + "loss": 0.0129, + "step": 4389 + }, + { + "epoch": 5.701298701298701, + "grad_norm": 13.730698310425623, + "learning_rate": 8.228090084207773e-06, + "loss": 0.0184, + "step": 4390 + }, + { + "epoch": 5.702597402597402, + "grad_norm": 10.297483099414947, + "learning_rate": 8.223950621828258e-06, + "loss": 0.0091, + "step": 4391 + }, + { + "epoch": 5.703896103896104, + "grad_norm": 7.310533549526117, + "learning_rate": 8.219811473665878e-06, + "loss": 0.0077, + "step": 4392 + }, + { + "epoch": 5.705194805194806, + "grad_norm": 6.916432089698865, + "learning_rate": 8.215672640452916e-06, + "loss": 0.011, + "step": 4393 + }, + { + "epoch": 5.706493506493507, + "grad_norm": 5.397294986826608, + "learning_rate": 8.211534122921615e-06, + "loss": 0.0104, + "step": 4394 + }, + { + "epoch": 5.707792207792208, + "grad_norm": 0.6468986070917153, + "learning_rate": 8.207395921804164e-06, + "loss": 0.0022, + "step": 4395 + }, + { + "epoch": 5.709090909090909, + "grad_norm": 9.627921795084104, + "learning_rate": 8.203258037832683e-06, + "loss": 0.0126, + "step": 4396 + }, + { + "epoch": 5.71038961038961, + "grad_norm": 2.6232943261074597, + "learning_rate": 8.199120471739242e-06, + "loss": 0.0062, + "step": 4397 + }, + { + "epoch": 5.711688311688311, + "grad_norm": 4.175181628215411, + "learning_rate": 8.194983224255861e-06, + "loss": 0.0247, + "step": 4398 + }, + { + "epoch": 5.712987012987013, + "grad_norm": 10.59349767284843, + "learning_rate": 8.190846296114497e-06, + "loss": 0.0067, + "step": 4399 + }, + { + "epoch": 5.714285714285714, + "grad_norm": 2.95234138111609, + "learning_rate": 8.186709688047052e-06, + "loss": 0.0175, + "step": 4400 + }, + { + "epoch": 5.7155844155844155, + "grad_norm": 4.531055953565727, + "learning_rate": 8.182573400785367e-06, + "loss": 0.0029, + "step": 4401 + }, + { + "epoch": 5.716883116883117, + "grad_norm": 10.65269925732378, + "learning_rate": 8.178437435061234e-06, + "loss": 0.0144, + "step": 4402 + }, + { + "epoch": 5.718181818181818, + "grad_norm": 5.01420351378211, + "learning_rate": 8.174301791606384e-06, + "loss": 0.0093, + "step": 4403 + }, + { + "epoch": 5.71948051948052, + "grad_norm": 4.430982380046002, + "learning_rate": 8.170166471152492e-06, + "loss": 0.0063, + "step": 4404 + }, + { + "epoch": 5.720779220779221, + "grad_norm": 7.370436896833906, + "learning_rate": 8.166031474431173e-06, + "loss": 0.0059, + "step": 4405 + }, + { + "epoch": 5.722077922077922, + "grad_norm": 3.6010417065581857, + "learning_rate": 8.16189680217399e-06, + "loss": 0.0052, + "step": 4406 + }, + { + "epoch": 5.723376623376623, + "grad_norm": 10.20086344350711, + "learning_rate": 8.157762455112445e-06, + "loss": 0.0226, + "step": 4407 + }, + { + "epoch": 5.724675324675324, + "grad_norm": 3.6635432384438524, + "learning_rate": 8.153628433977982e-06, + "loss": 0.0307, + "step": 4408 + }, + { + "epoch": 5.7259740259740255, + "grad_norm": 7.004170225029107, + "learning_rate": 8.149494739501987e-06, + "loss": 0.0241, + "step": 4409 + }, + { + "epoch": 5.7272727272727275, + "grad_norm": 5.073021032516835, + "learning_rate": 8.145361372415797e-06, + "loss": 0.0091, + "step": 4410 + }, + { + "epoch": 5.728571428571429, + "grad_norm": 4.409193826329913, + "learning_rate": 8.141228333450673e-06, + "loss": 0.0141, + "step": 4411 + }, + { + "epoch": 5.72987012987013, + "grad_norm": 8.065057216462968, + "learning_rate": 8.137095623337836e-06, + "loss": 0.0098, + "step": 4412 + }, + { + "epoch": 5.731168831168831, + "grad_norm": 11.546633671968777, + "learning_rate": 8.132963242808439e-06, + "loss": 0.0342, + "step": 4413 + }, + { + "epoch": 5.732467532467533, + "grad_norm": 5.2092767018281085, + "learning_rate": 8.128831192593583e-06, + "loss": 0.0092, + "step": 4414 + }, + { + "epoch": 5.733766233766234, + "grad_norm": 2.176537399979302, + "learning_rate": 8.1246994734243e-06, + "loss": 0.0092, + "step": 4415 + }, + { + "epoch": 5.735064935064935, + "grad_norm": 3.20742376787346, + "learning_rate": 8.120568086031573e-06, + "loss": 0.0097, + "step": 4416 + }, + { + "epoch": 5.736363636363636, + "grad_norm": 10.40111789806704, + "learning_rate": 8.116437031146325e-06, + "loss": 0.0189, + "step": 4417 + }, + { + "epoch": 5.7376623376623375, + "grad_norm": 6.326558215159109, + "learning_rate": 8.112306309499422e-06, + "loss": 0.0185, + "step": 4418 + }, + { + "epoch": 5.738961038961039, + "grad_norm": 5.593255433395189, + "learning_rate": 8.10817592182166e-06, + "loss": 0.0154, + "step": 4419 + }, + { + "epoch": 5.740259740259741, + "grad_norm": 5.719344072294139, + "learning_rate": 8.104045868843787e-06, + "loss": 0.0103, + "step": 4420 + }, + { + "epoch": 5.741558441558442, + "grad_norm": 5.61230393938917, + "learning_rate": 8.099916151296492e-06, + "loss": 0.0029, + "step": 4421 + }, + { + "epoch": 5.742857142857143, + "grad_norm": 5.421801561145029, + "learning_rate": 8.095786769910395e-06, + "loss": 0.0134, + "step": 4422 + }, + { + "epoch": 5.744155844155844, + "grad_norm": 2.940442970547086, + "learning_rate": 8.091657725416065e-06, + "loss": 0.0068, + "step": 4423 + }, + { + "epoch": 5.745454545454545, + "grad_norm": 5.60174970545664, + "learning_rate": 8.087529018544009e-06, + "loss": 0.0115, + "step": 4424 + }, + { + "epoch": 5.746753246753247, + "grad_norm": 4.492194305186476, + "learning_rate": 8.08340065002468e-06, + "loss": 0.0036, + "step": 4425 + }, + { + "epoch": 5.748051948051948, + "grad_norm": 8.097484606716009, + "learning_rate": 8.079272620588454e-06, + "loss": 0.0408, + "step": 4426 + }, + { + "epoch": 5.749350649350649, + "grad_norm": 2.312844869586151, + "learning_rate": 8.075144930965666e-06, + "loss": 0.0049, + "step": 4427 + }, + { + "epoch": 5.750649350649351, + "grad_norm": 4.17676362914042, + "learning_rate": 8.071017581886582e-06, + "loss": 0.0167, + "step": 4428 + }, + { + "epoch": 5.751948051948052, + "grad_norm": 9.983817877110766, + "learning_rate": 8.066890574081411e-06, + "loss": 0.0289, + "step": 4429 + }, + { + "epoch": 5.753246753246753, + "grad_norm": 2.720786450131976, + "learning_rate": 8.062763908280296e-06, + "loss": 0.012, + "step": 4430 + }, + { + "epoch": 5.754545454545455, + "grad_norm": 9.21266622424927, + "learning_rate": 8.058637585213324e-06, + "loss": 0.0254, + "step": 4431 + }, + { + "epoch": 5.755844155844156, + "grad_norm": 5.40537652645545, + "learning_rate": 8.054511605610521e-06, + "loss": 0.0039, + "step": 4432 + }, + { + "epoch": 5.757142857142857, + "grad_norm": 2.285405781157633, + "learning_rate": 8.050385970201854e-06, + "loss": 0.0161, + "step": 4433 + }, + { + "epoch": 5.758441558441558, + "grad_norm": 5.434707371206491, + "learning_rate": 8.046260679717225e-06, + "loss": 0.0107, + "step": 4434 + }, + { + "epoch": 5.759740259740259, + "grad_norm": 2.5147357859326798, + "learning_rate": 8.042135734886474e-06, + "loss": 0.0119, + "step": 4435 + }, + { + "epoch": 5.761038961038961, + "grad_norm": 1.428598311443439, + "learning_rate": 8.038011136439388e-06, + "loss": 0.0041, + "step": 4436 + }, + { + "epoch": 5.7623376623376625, + "grad_norm": 6.620120203833536, + "learning_rate": 8.033886885105683e-06, + "loss": 0.0221, + "step": 4437 + }, + { + "epoch": 5.763636363636364, + "grad_norm": 3.931680809111733, + "learning_rate": 8.02976298161502e-06, + "loss": 0.0138, + "step": 4438 + }, + { + "epoch": 5.764935064935065, + "grad_norm": 3.9539967247882757, + "learning_rate": 8.025639426696998e-06, + "loss": 0.0362, + "step": 4439 + }, + { + "epoch": 5.766233766233766, + "grad_norm": 5.803793406916199, + "learning_rate": 8.021516221081152e-06, + "loss": 0.0071, + "step": 4440 + }, + { + "epoch": 5.767532467532467, + "grad_norm": 6.896329304608257, + "learning_rate": 8.017393365496951e-06, + "loss": 0.0162, + "step": 4441 + }, + { + "epoch": 5.768831168831169, + "grad_norm": 7.340196652969827, + "learning_rate": 8.013270860673811e-06, + "loss": 0.0147, + "step": 4442 + }, + { + "epoch": 5.77012987012987, + "grad_norm": 6.8431528281330545, + "learning_rate": 8.009148707341084e-06, + "loss": 0.0308, + "step": 4443 + }, + { + "epoch": 5.771428571428571, + "grad_norm": 4.46180523902015, + "learning_rate": 8.00502690622806e-06, + "loss": 0.0073, + "step": 4444 + }, + { + "epoch": 5.7727272727272725, + "grad_norm": 8.965269292078268, + "learning_rate": 8.000905458063956e-06, + "loss": 0.0311, + "step": 4445 + }, + { + "epoch": 5.7740259740259745, + "grad_norm": 11.700730764641671, + "learning_rate": 7.99678436357794e-06, + "loss": 0.0071, + "step": 4446 + }, + { + "epoch": 5.775324675324676, + "grad_norm": 7.345881782494968, + "learning_rate": 7.99266362349911e-06, + "loss": 0.0045, + "step": 4447 + }, + { + "epoch": 5.776623376623377, + "grad_norm": 8.20042284004848, + "learning_rate": 7.988543238556512e-06, + "loss": 0.038, + "step": 4448 + }, + { + "epoch": 5.777922077922078, + "grad_norm": 2.2650750788088874, + "learning_rate": 7.98442320947911e-06, + "loss": 0.0041, + "step": 4449 + }, + { + "epoch": 5.779220779220779, + "grad_norm": 6.631439690865797, + "learning_rate": 7.98030353699582e-06, + "loss": 0.0227, + "step": 4450 + }, + { + "epoch": 5.78051948051948, + "grad_norm": 4.870166663601856, + "learning_rate": 7.976184221835497e-06, + "loss": 0.0181, + "step": 4451 + }, + { + "epoch": 5.781818181818182, + "grad_norm": 4.903940410182295, + "learning_rate": 7.972065264726917e-06, + "loss": 0.0201, + "step": 4452 + }, + { + "epoch": 5.783116883116883, + "grad_norm": 7.620137911925587, + "learning_rate": 7.967946666398807e-06, + "loss": 0.0323, + "step": 4453 + }, + { + "epoch": 5.7844155844155845, + "grad_norm": 3.7293873096447454, + "learning_rate": 7.963828427579826e-06, + "loss": 0.0063, + "step": 4454 + }, + { + "epoch": 5.785714285714286, + "grad_norm": 8.9757074816776, + "learning_rate": 7.959710548998572e-06, + "loss": 0.0189, + "step": 4455 + }, + { + "epoch": 5.787012987012987, + "grad_norm": 2.141979258373779, + "learning_rate": 7.95559303138357e-06, + "loss": 0.0052, + "step": 4456 + }, + { + "epoch": 5.788311688311689, + "grad_norm": 1.9390433014167416, + "learning_rate": 7.95147587546329e-06, + "loss": 0.0033, + "step": 4457 + }, + { + "epoch": 5.78961038961039, + "grad_norm": 9.681354828013342, + "learning_rate": 7.947359081966139e-06, + "loss": 0.0472, + "step": 4458 + }, + { + "epoch": 5.790909090909091, + "grad_norm": 13.043237344377113, + "learning_rate": 7.943242651620456e-06, + "loss": 0.0121, + "step": 4459 + }, + { + "epoch": 5.792207792207792, + "grad_norm": 7.1959788622389, + "learning_rate": 7.939126585154512e-06, + "loss": 0.0326, + "step": 4460 + }, + { + "epoch": 5.793506493506493, + "grad_norm": 5.132717036341811, + "learning_rate": 7.93501088329652e-06, + "loss": 0.0251, + "step": 4461 + }, + { + "epoch": 5.794805194805194, + "grad_norm": 6.507249591349562, + "learning_rate": 7.93089554677463e-06, + "loss": 0.0056, + "step": 4462 + }, + { + "epoch": 5.796103896103896, + "grad_norm": 2.4305189474761146, + "learning_rate": 7.92678057631692e-06, + "loss": 0.0189, + "step": 4463 + }, + { + "epoch": 5.797402597402598, + "grad_norm": 8.810726129708078, + "learning_rate": 7.922665972651407e-06, + "loss": 0.0158, + "step": 4464 + }, + { + "epoch": 5.798701298701299, + "grad_norm": 6.202048573772077, + "learning_rate": 7.918551736506042e-06, + "loss": 0.0102, + "step": 4465 + }, + { + "epoch": 5.8, + "grad_norm": 3.6366967615976113, + "learning_rate": 7.914437868608717e-06, + "loss": 0.0042, + "step": 4466 + }, + { + "epoch": 5.801298701298701, + "grad_norm": 6.874173439908574, + "learning_rate": 7.91032436968725e-06, + "loss": 0.0076, + "step": 4467 + }, + { + "epoch": 5.802597402597403, + "grad_norm": 9.097847138498377, + "learning_rate": 7.9062112404694e-06, + "loss": 0.0141, + "step": 4468 + }, + { + "epoch": 5.803896103896104, + "grad_norm": 6.958913023807964, + "learning_rate": 7.902098481682856e-06, + "loss": 0.0046, + "step": 4469 + }, + { + "epoch": 5.805194805194805, + "grad_norm": 9.526092879980574, + "learning_rate": 7.897986094055246e-06, + "loss": 0.0184, + "step": 4470 + }, + { + "epoch": 5.806493506493506, + "grad_norm": 3.9269184974120623, + "learning_rate": 7.893874078314127e-06, + "loss": 0.0297, + "step": 4471 + }, + { + "epoch": 5.8077922077922075, + "grad_norm": 17.58141645950017, + "learning_rate": 7.889762435186995e-06, + "loss": 0.0169, + "step": 4472 + }, + { + "epoch": 5.809090909090909, + "grad_norm": 5.48624581066505, + "learning_rate": 7.885651165401281e-06, + "loss": 0.0163, + "step": 4473 + }, + { + "epoch": 5.810389610389611, + "grad_norm": 5.0035450881472965, + "learning_rate": 7.881540269684345e-06, + "loss": 0.0129, + "step": 4474 + }, + { + "epoch": 5.811688311688312, + "grad_norm": 5.281301194191435, + "learning_rate": 7.877429748763481e-06, + "loss": 0.0061, + "step": 4475 + }, + { + "epoch": 5.812987012987013, + "grad_norm": 4.135771343244054, + "learning_rate": 7.87331960336592e-06, + "loss": 0.0059, + "step": 4476 + }, + { + "epoch": 5.814285714285714, + "grad_norm": 4.50380776016052, + "learning_rate": 7.869209834218832e-06, + "loss": 0.0109, + "step": 4477 + }, + { + "epoch": 5.815584415584416, + "grad_norm": 1.501208353799874, + "learning_rate": 7.865100442049302e-06, + "loss": 0.0042, + "step": 4478 + }, + { + "epoch": 5.816883116883117, + "grad_norm": 5.556882788716785, + "learning_rate": 7.860991427584368e-06, + "loss": 0.0261, + "step": 4479 + }, + { + "epoch": 5.818181818181818, + "grad_norm": 3.1500576593722194, + "learning_rate": 7.85688279155099e-06, + "loss": 0.0048, + "step": 4480 + }, + { + "epoch": 5.8194805194805195, + "grad_norm": 5.129987693784527, + "learning_rate": 7.852774534676073e-06, + "loss": 0.0073, + "step": 4481 + }, + { + "epoch": 5.820779220779221, + "grad_norm": 3.819527066547053, + "learning_rate": 7.848666657686432e-06, + "loss": 0.0193, + "step": 4482 + }, + { + "epoch": 5.822077922077922, + "grad_norm": 10.030957676811783, + "learning_rate": 7.844559161308836e-06, + "loss": 0.0223, + "step": 4483 + }, + { + "epoch": 5.823376623376624, + "grad_norm": 13.21331650757265, + "learning_rate": 7.840452046269982e-06, + "loss": 0.0276, + "step": 4484 + }, + { + "epoch": 5.824675324675325, + "grad_norm": 6.339545151202306, + "learning_rate": 7.836345313296499e-06, + "loss": 0.0207, + "step": 4485 + }, + { + "epoch": 5.825974025974026, + "grad_norm": 7.3619047922925205, + "learning_rate": 7.832238963114938e-06, + "loss": 0.0297, + "step": 4486 + }, + { + "epoch": 5.827272727272727, + "grad_norm": 5.205589511518018, + "learning_rate": 7.828132996451795e-06, + "loss": 0.0112, + "step": 4487 + }, + { + "epoch": 5.828571428571428, + "grad_norm": 3.547793808430587, + "learning_rate": 7.824027414033496e-06, + "loss": 0.0066, + "step": 4488 + }, + { + "epoch": 5.82987012987013, + "grad_norm": 5.219016001264999, + "learning_rate": 7.8199222165864e-06, + "loss": 0.0094, + "step": 4489 + }, + { + "epoch": 5.8311688311688314, + "grad_norm": 4.663496385155467, + "learning_rate": 7.815817404836787e-06, + "loss": 0.0139, + "step": 4490 + }, + { + "epoch": 5.832467532467533, + "grad_norm": 6.408473384386471, + "learning_rate": 7.81171297951088e-06, + "loss": 0.0088, + "step": 4491 + }, + { + "epoch": 5.833766233766234, + "grad_norm": 8.43454028944918, + "learning_rate": 7.807608941334836e-06, + "loss": 0.0277, + "step": 4492 + }, + { + "epoch": 5.835064935064935, + "grad_norm": 6.889102450828028, + "learning_rate": 7.80350529103473e-06, + "loss": 0.0099, + "step": 4493 + }, + { + "epoch": 5.836363636363636, + "grad_norm": 4.606443101771159, + "learning_rate": 7.799402029336579e-06, + "loss": 0.0437, + "step": 4494 + }, + { + "epoch": 5.837662337662338, + "grad_norm": 3.734210136104079, + "learning_rate": 7.795299156966327e-06, + "loss": 0.0134, + "step": 4495 + }, + { + "epoch": 5.838961038961039, + "grad_norm": 11.43644907719936, + "learning_rate": 7.791196674649859e-06, + "loss": 0.0236, + "step": 4496 + }, + { + "epoch": 5.84025974025974, + "grad_norm": 5.182557384762238, + "learning_rate": 7.78709458311297e-06, + "loss": 0.0343, + "step": 4497 + }, + { + "epoch": 5.841558441558441, + "grad_norm": 2.1695159143170675, + "learning_rate": 7.782992883081405e-06, + "loss": 0.009, + "step": 4498 + }, + { + "epoch": 5.8428571428571425, + "grad_norm": 6.8969703883764, + "learning_rate": 7.778891575280834e-06, + "loss": 0.0409, + "step": 4499 + }, + { + "epoch": 5.8441558441558445, + "grad_norm": 5.732997004560354, + "learning_rate": 7.774790660436857e-06, + "loss": 0.0061, + "step": 4500 + }, + { + "epoch": 5.845454545454546, + "grad_norm": 1.9657407129402442, + "learning_rate": 7.770690139275003e-06, + "loss": 0.0094, + "step": 4501 + }, + { + "epoch": 5.846753246753247, + "grad_norm": 11.19356025987263, + "learning_rate": 7.766590012520731e-06, + "loss": 0.0404, + "step": 4502 + }, + { + "epoch": 5.848051948051948, + "grad_norm": 5.04427380655409, + "learning_rate": 7.762490280899437e-06, + "loss": 0.0284, + "step": 4503 + }, + { + "epoch": 5.849350649350649, + "grad_norm": 7.851188813527713, + "learning_rate": 7.758390945136436e-06, + "loss": 0.0177, + "step": 4504 + }, + { + "epoch": 5.85064935064935, + "grad_norm": 3.7730891761921983, + "learning_rate": 7.754292005956982e-06, + "loss": 0.0051, + "step": 4505 + }, + { + "epoch": 5.851948051948052, + "grad_norm": 5.487033274760324, + "learning_rate": 7.750193464086257e-06, + "loss": 0.007, + "step": 4506 + }, + { + "epoch": 5.853246753246753, + "grad_norm": 7.97011841633135, + "learning_rate": 7.74609532024937e-06, + "loss": 0.0118, + "step": 4507 + }, + { + "epoch": 5.8545454545454545, + "grad_norm": 7.166742361219302, + "learning_rate": 7.741997575171358e-06, + "loss": 0.0073, + "step": 4508 + }, + { + "epoch": 5.855844155844156, + "grad_norm": 7.11860765260176, + "learning_rate": 7.737900229577194e-06, + "loss": 0.0204, + "step": 4509 + }, + { + "epoch": 5.857142857142857, + "grad_norm": 3.2650669997264385, + "learning_rate": 7.733803284191778e-06, + "loss": 0.0138, + "step": 4510 + }, + { + "epoch": 5.858441558441559, + "grad_norm": 4.734220594947321, + "learning_rate": 7.729706739739939e-06, + "loss": 0.0104, + "step": 4511 + }, + { + "epoch": 5.85974025974026, + "grad_norm": 7.269408578503539, + "learning_rate": 7.725610596946427e-06, + "loss": 0.0154, + "step": 4512 + }, + { + "epoch": 5.861038961038961, + "grad_norm": 13.118179041758282, + "learning_rate": 7.721514856535934e-06, + "loss": 0.0257, + "step": 4513 + }, + { + "epoch": 5.862337662337662, + "grad_norm": 6.979608539276991, + "learning_rate": 7.717419519233074e-06, + "loss": 0.0283, + "step": 4514 + }, + { + "epoch": 5.863636363636363, + "grad_norm": 9.083237432545406, + "learning_rate": 7.713324585762394e-06, + "loss": 0.0132, + "step": 4515 + }, + { + "epoch": 5.8649350649350644, + "grad_norm": 11.635772710578548, + "learning_rate": 7.709230056848356e-06, + "loss": 0.0157, + "step": 4516 + }, + { + "epoch": 5.8662337662337665, + "grad_norm": 4.516776629369751, + "learning_rate": 7.705135933215369e-06, + "loss": 0.0078, + "step": 4517 + }, + { + "epoch": 5.867532467532468, + "grad_norm": 5.557462173967797, + "learning_rate": 7.701042215587762e-06, + "loss": 0.0258, + "step": 4518 + }, + { + "epoch": 5.868831168831169, + "grad_norm": 7.9366553157352, + "learning_rate": 7.696948904689788e-06, + "loss": 0.0195, + "step": 4519 + }, + { + "epoch": 5.87012987012987, + "grad_norm": 5.2767668771856515, + "learning_rate": 7.692856001245632e-06, + "loss": 0.0071, + "step": 4520 + }, + { + "epoch": 5.871428571428572, + "grad_norm": 8.14376491303672, + "learning_rate": 7.688763505979409e-06, + "loss": 0.0268, + "step": 4521 + }, + { + "epoch": 5.872727272727273, + "grad_norm": 13.08190023609905, + "learning_rate": 7.684671419615164e-06, + "loss": 0.0366, + "step": 4522 + }, + { + "epoch": 5.874025974025974, + "grad_norm": 6.425840944903476, + "learning_rate": 7.680579742876855e-06, + "loss": 0.0127, + "step": 4523 + }, + { + "epoch": 5.875324675324675, + "grad_norm": 11.593235660744265, + "learning_rate": 7.676488476488383e-06, + "loss": 0.0042, + "step": 4524 + }, + { + "epoch": 5.876623376623376, + "grad_norm": 5.917495527758007, + "learning_rate": 7.672397621173573e-06, + "loss": 0.0412, + "step": 4525 + }, + { + "epoch": 5.8779220779220775, + "grad_norm": 9.549310086439872, + "learning_rate": 7.66830717765618e-06, + "loss": 0.0121, + "step": 4526 + }, + { + "epoch": 5.87922077922078, + "grad_norm": 10.592403298132744, + "learning_rate": 7.664217146659872e-06, + "loss": 0.0144, + "step": 4527 + }, + { + "epoch": 5.880519480519481, + "grad_norm": 4.745904822937406, + "learning_rate": 7.660127528908257e-06, + "loss": 0.0059, + "step": 4528 + }, + { + "epoch": 5.881818181818182, + "grad_norm": 13.013936749622797, + "learning_rate": 7.656038325124869e-06, + "loss": 0.0082, + "step": 4529 + }, + { + "epoch": 5.883116883116883, + "grad_norm": 20.669706803981587, + "learning_rate": 7.651949536033167e-06, + "loss": 0.0433, + "step": 4530 + }, + { + "epoch": 5.884415584415584, + "grad_norm": 10.242516681962869, + "learning_rate": 7.647861162356533e-06, + "loss": 0.0395, + "step": 4531 + }, + { + "epoch": 5.885714285714286, + "grad_norm": 26.279649175537283, + "learning_rate": 7.643773204818278e-06, + "loss": 0.0235, + "step": 4532 + }, + { + "epoch": 5.887012987012987, + "grad_norm": 7.75311184232076, + "learning_rate": 7.639685664141644e-06, + "loss": 0.0157, + "step": 4533 + }, + { + "epoch": 5.888311688311688, + "grad_norm": 22.057115172601353, + "learning_rate": 7.635598541049791e-06, + "loss": 0.0169, + "step": 4534 + }, + { + "epoch": 5.8896103896103895, + "grad_norm": 16.81852103733963, + "learning_rate": 7.631511836265813e-06, + "loss": 0.0199, + "step": 4535 + }, + { + "epoch": 5.890909090909091, + "grad_norm": 5.398804101399987, + "learning_rate": 7.627425550512724e-06, + "loss": 0.0286, + "step": 4536 + }, + { + "epoch": 5.892207792207792, + "grad_norm": 15.505879158441436, + "learning_rate": 7.623339684513467e-06, + "loss": 0.028, + "step": 4537 + }, + { + "epoch": 5.893506493506494, + "grad_norm": 15.81914197975372, + "learning_rate": 7.619254238990908e-06, + "loss": 0.0078, + "step": 4538 + }, + { + "epoch": 5.894805194805195, + "grad_norm": 3.262879770359461, + "learning_rate": 7.615169214667842e-06, + "loss": 0.0132, + "step": 4539 + }, + { + "epoch": 5.896103896103896, + "grad_norm": 16.1674741483486, + "learning_rate": 7.61108461226699e-06, + "loss": 0.0132, + "step": 4540 + }, + { + "epoch": 5.897402597402597, + "grad_norm": 12.301938803790476, + "learning_rate": 7.607000432510995e-06, + "loss": 0.0179, + "step": 4541 + }, + { + "epoch": 5.898701298701298, + "grad_norm": 8.204763442233123, + "learning_rate": 7.602916676122424e-06, + "loss": 0.0184, + "step": 4542 + }, + { + "epoch": 5.9, + "grad_norm": 8.979868726342556, + "learning_rate": 7.598833343823773e-06, + "loss": 0.0322, + "step": 4543 + }, + { + "epoch": 5.9012987012987015, + "grad_norm": 9.942815382156757, + "learning_rate": 7.594750436337467e-06, + "loss": 0.0195, + "step": 4544 + }, + { + "epoch": 5.902597402597403, + "grad_norm": 2.3454970931655628, + "learning_rate": 7.590667954385841e-06, + "loss": 0.0079, + "step": 4545 + }, + { + "epoch": 5.903896103896104, + "grad_norm": 7.136127247219019, + "learning_rate": 7.58658589869117e-06, + "loss": 0.0151, + "step": 4546 + }, + { + "epoch": 5.905194805194805, + "grad_norm": 8.503367272460894, + "learning_rate": 7.5825042699756455e-06, + "loss": 0.0035, + "step": 4547 + }, + { + "epoch": 5.906493506493506, + "grad_norm": 1.1217458369541922, + "learning_rate": 7.578423068961391e-06, + "loss": 0.0024, + "step": 4548 + }, + { + "epoch": 5.907792207792208, + "grad_norm": 12.735885885633804, + "learning_rate": 7.574342296370443e-06, + "loss": 0.0266, + "step": 4549 + }, + { + "epoch": 5.909090909090909, + "grad_norm": 7.548198751441535, + "learning_rate": 7.570261952924769e-06, + "loss": 0.0158, + "step": 4550 + }, + { + "epoch": 5.91038961038961, + "grad_norm": 8.304354491584643, + "learning_rate": 7.5661820393462605e-06, + "loss": 0.0149, + "step": 4551 + }, + { + "epoch": 5.911688311688311, + "grad_norm": 10.622337527463532, + "learning_rate": 7.5621025563567385e-06, + "loss": 0.0132, + "step": 4552 + }, + { + "epoch": 5.9129870129870135, + "grad_norm": 10.9822301346494, + "learning_rate": 7.558023504677931e-06, + "loss": 0.0485, + "step": 4553 + }, + { + "epoch": 5.914285714285715, + "grad_norm": 12.26879315250954, + "learning_rate": 7.553944885031505e-06, + "loss": 0.0152, + "step": 4554 + }, + { + "epoch": 5.915584415584416, + "grad_norm": 16.463550881562593, + "learning_rate": 7.549866698139048e-06, + "loss": 0.0227, + "step": 4555 + }, + { + "epoch": 5.916883116883117, + "grad_norm": 11.235379911824063, + "learning_rate": 7.545788944722072e-06, + "loss": 0.0403, + "step": 4556 + }, + { + "epoch": 5.918181818181818, + "grad_norm": 8.740603274980712, + "learning_rate": 7.541711625502002e-06, + "loss": 0.0247, + "step": 4557 + }, + { + "epoch": 5.919480519480519, + "grad_norm": 6.402058869698402, + "learning_rate": 7.537634741200197e-06, + "loss": 0.0129, + "step": 4558 + }, + { + "epoch": 5.920779220779221, + "grad_norm": 4.818860145950452, + "learning_rate": 7.533558292537941e-06, + "loss": 0.0184, + "step": 4559 + }, + { + "epoch": 5.922077922077922, + "grad_norm": 3.812459544812327, + "learning_rate": 7.529482280236426e-06, + "loss": 0.0032, + "step": 4560 + }, + { + "epoch": 5.923376623376623, + "grad_norm": 8.420667160881402, + "learning_rate": 7.525406705016783e-06, + "loss": 0.0241, + "step": 4561 + }, + { + "epoch": 5.9246753246753245, + "grad_norm": 4.375715654136629, + "learning_rate": 7.521331567600059e-06, + "loss": 0.0051, + "step": 4562 + }, + { + "epoch": 5.925974025974026, + "grad_norm": 13.237175882617276, + "learning_rate": 7.517256868707226e-06, + "loss": 0.0453, + "step": 4563 + }, + { + "epoch": 5.927272727272728, + "grad_norm": 2.6708115347862575, + "learning_rate": 7.513182609059169e-06, + "loss": 0.0127, + "step": 4564 + }, + { + "epoch": 5.928571428571429, + "grad_norm": 4.228724701750841, + "learning_rate": 7.5091087893767054e-06, + "loss": 0.0416, + "step": 4565 + }, + { + "epoch": 5.92987012987013, + "grad_norm": 6.052755825627974, + "learning_rate": 7.505035410380574e-06, + "loss": 0.0088, + "step": 4566 + }, + { + "epoch": 5.931168831168831, + "grad_norm": 3.3139461342334546, + "learning_rate": 7.5009624727914355e-06, + "loss": 0.0169, + "step": 4567 + }, + { + "epoch": 5.932467532467532, + "grad_norm": 9.085947022479102, + "learning_rate": 7.4968899773298665e-06, + "loss": 0.018, + "step": 4568 + }, + { + "epoch": 5.933766233766233, + "grad_norm": 6.82000334889356, + "learning_rate": 7.492817924716369e-06, + "loss": 0.0253, + "step": 4569 + }, + { + "epoch": 5.935064935064935, + "grad_norm": 5.559925474257855, + "learning_rate": 7.488746315671371e-06, + "loss": 0.006, + "step": 4570 + }, + { + "epoch": 5.9363636363636365, + "grad_norm": 4.278024786150062, + "learning_rate": 7.4846751509152135e-06, + "loss": 0.0116, + "step": 4571 + }, + { + "epoch": 5.937662337662338, + "grad_norm": 7.155463164457208, + "learning_rate": 7.480604431168166e-06, + "loss": 0.0071, + "step": 4572 + }, + { + "epoch": 5.938961038961039, + "grad_norm": 6.709976578975189, + "learning_rate": 7.476534157150418e-06, + "loss": 0.0288, + "step": 4573 + }, + { + "epoch": 5.94025974025974, + "grad_norm": 10.137441515006975, + "learning_rate": 7.4724643295820765e-06, + "loss": 0.0239, + "step": 4574 + }, + { + "epoch": 5.941558441558442, + "grad_norm": 6.056171539045843, + "learning_rate": 7.468394949183172e-06, + "loss": 0.0239, + "step": 4575 + }, + { + "epoch": 5.942857142857143, + "grad_norm": 3.5024502919095575, + "learning_rate": 7.464326016673656e-06, + "loss": 0.0122, + "step": 4576 + }, + { + "epoch": 5.944155844155844, + "grad_norm": 7.973944508661797, + "learning_rate": 7.4602575327734004e-06, + "loss": 0.0177, + "step": 4577 + }, + { + "epoch": 5.945454545454545, + "grad_norm": 7.507169777047244, + "learning_rate": 7.456189498202202e-06, + "loss": 0.0102, + "step": 4578 + }, + { + "epoch": 5.9467532467532465, + "grad_norm": 3.342113624419092, + "learning_rate": 7.452121913679768e-06, + "loss": 0.0329, + "step": 4579 + }, + { + "epoch": 5.948051948051948, + "grad_norm": 8.574500559421711, + "learning_rate": 7.448054779925733e-06, + "loss": 0.0188, + "step": 4580 + }, + { + "epoch": 5.94935064935065, + "grad_norm": 6.125685913357178, + "learning_rate": 7.443988097659654e-06, + "loss": 0.0155, + "step": 4581 + }, + { + "epoch": 5.950649350649351, + "grad_norm": 6.4019413614436615, + "learning_rate": 7.439921867601006e-06, + "loss": 0.0122, + "step": 4582 + }, + { + "epoch": 5.951948051948052, + "grad_norm": 4.7355157119172375, + "learning_rate": 7.435856090469175e-06, + "loss": 0.0027, + "step": 4583 + }, + { + "epoch": 5.953246753246753, + "grad_norm": 5.919267723148423, + "learning_rate": 7.43179076698348e-06, + "loss": 0.008, + "step": 4584 + }, + { + "epoch": 5.954545454545455, + "grad_norm": 4.25546773105198, + "learning_rate": 7.427725897863159e-06, + "loss": 0.0067, + "step": 4585 + }, + { + "epoch": 5.955844155844156, + "grad_norm": 5.468223841361123, + "learning_rate": 7.423661483827357e-06, + "loss": 0.0139, + "step": 4586 + }, + { + "epoch": 5.957142857142857, + "grad_norm": 6.7669695304285264, + "learning_rate": 7.41959752559515e-06, + "loss": 0.024, + "step": 4587 + }, + { + "epoch": 5.958441558441558, + "grad_norm": 4.8186094302725175, + "learning_rate": 7.415534023885529e-06, + "loss": 0.0197, + "step": 4588 + }, + { + "epoch": 5.95974025974026, + "grad_norm": 4.922182661181094, + "learning_rate": 7.411470979417412e-06, + "loss": 0.0054, + "step": 4589 + }, + { + "epoch": 5.961038961038961, + "grad_norm": 8.530741222158468, + "learning_rate": 7.407408392909619e-06, + "loss": 0.0451, + "step": 4590 + }, + { + "epoch": 5.962337662337663, + "grad_norm": 13.444034650068861, + "learning_rate": 7.403346265080904e-06, + "loss": 0.0384, + "step": 4591 + }, + { + "epoch": 5.963636363636364, + "grad_norm": 9.079109931756324, + "learning_rate": 7.399284596649935e-06, + "loss": 0.0096, + "step": 4592 + }, + { + "epoch": 5.964935064935065, + "grad_norm": 3.13683352308677, + "learning_rate": 7.395223388335304e-06, + "loss": 0.0136, + "step": 4593 + }, + { + "epoch": 5.966233766233766, + "grad_norm": 18.739385931001234, + "learning_rate": 7.391162640855506e-06, + "loss": 0.0294, + "step": 4594 + }, + { + "epoch": 5.967532467532467, + "grad_norm": 3.5214624218132826, + "learning_rate": 7.387102354928972e-06, + "loss": 0.0094, + "step": 4595 + }, + { + "epoch": 5.968831168831169, + "grad_norm": 8.42480344648172, + "learning_rate": 7.383042531274042e-06, + "loss": 0.026, + "step": 4596 + }, + { + "epoch": 5.97012987012987, + "grad_norm": 3.5259798829797604, + "learning_rate": 7.378983170608982e-06, + "loss": 0.0126, + "step": 4597 + }, + { + "epoch": 5.9714285714285715, + "grad_norm": 3.500562323334162, + "learning_rate": 7.37492427365196e-06, + "loss": 0.0197, + "step": 4598 + }, + { + "epoch": 5.972727272727273, + "grad_norm": 5.070399606037244, + "learning_rate": 7.37086584112108e-06, + "loss": 0.0171, + "step": 4599 + }, + { + "epoch": 5.974025974025974, + "grad_norm": 6.2374995289326085, + "learning_rate": 7.366807873734358e-06, + "loss": 0.0095, + "step": 4600 + }, + { + "epoch": 5.975324675324675, + "grad_norm": 5.6774611930480035, + "learning_rate": 7.3627503722097194e-06, + "loss": 0.0171, + "step": 4601 + }, + { + "epoch": 5.976623376623377, + "grad_norm": 7.462722390334393, + "learning_rate": 7.358693337265019e-06, + "loss": 0.0092, + "step": 4602 + }, + { + "epoch": 5.977922077922078, + "grad_norm": 1.5979158233669584, + "learning_rate": 7.354636769618022e-06, + "loss": 0.0032, + "step": 4603 + }, + { + "epoch": 5.979220779220779, + "grad_norm": 9.12213766706853, + "learning_rate": 7.3505806699864135e-06, + "loss": 0.025, + "step": 4604 + }, + { + "epoch": 5.98051948051948, + "grad_norm": 8.069232011462564, + "learning_rate": 7.346525039087794e-06, + "loss": 0.0152, + "step": 4605 + }, + { + "epoch": 5.9818181818181815, + "grad_norm": 4.669082102322424, + "learning_rate": 7.342469877639683e-06, + "loss": 0.0042, + "step": 4606 + }, + { + "epoch": 5.9831168831168835, + "grad_norm": 9.484995678836613, + "learning_rate": 7.338415186359518e-06, + "loss": 0.0124, + "step": 4607 + }, + { + "epoch": 5.984415584415585, + "grad_norm": 2.41750352591285, + "learning_rate": 7.3343609659646505e-06, + "loss": 0.0045, + "step": 4608 + }, + { + "epoch": 5.985714285714286, + "grad_norm": 3.714477395550312, + "learning_rate": 7.330307217172345e-06, + "loss": 0.0123, + "step": 4609 + }, + { + "epoch": 5.987012987012987, + "grad_norm": 4.6927531698341, + "learning_rate": 7.326253940699795e-06, + "loss": 0.055, + "step": 4610 + }, + { + "epoch": 5.988311688311688, + "grad_norm": 7.0790262044493035, + "learning_rate": 7.3222011372641e-06, + "loss": 0.0033, + "step": 4611 + }, + { + "epoch": 5.989610389610389, + "grad_norm": 6.940757112276588, + "learning_rate": 7.318148807582275e-06, + "loss": 0.0181, + "step": 4612 + }, + { + "epoch": 5.990909090909091, + "grad_norm": 3.827857556234704, + "learning_rate": 7.3140969523712545e-06, + "loss": 0.0135, + "step": 4613 + }, + { + "epoch": 5.992207792207792, + "grad_norm": 8.84607936604427, + "learning_rate": 7.310045572347895e-06, + "loss": 0.0138, + "step": 4614 + }, + { + "epoch": 5.9935064935064934, + "grad_norm": 7.18358733349879, + "learning_rate": 7.305994668228962e-06, + "loss": 0.0136, + "step": 4615 + }, + { + "epoch": 5.994805194805195, + "grad_norm": 4.806541593918106, + "learning_rate": 7.301944240731133e-06, + "loss": 0.0232, + "step": 4616 + }, + { + "epoch": 5.996103896103897, + "grad_norm": 9.824449718323988, + "learning_rate": 7.297894290571008e-06, + "loss": 0.0034, + "step": 4617 + }, + { + "epoch": 5.997402597402598, + "grad_norm": 14.118932246588393, + "learning_rate": 7.293844818465103e-06, + "loss": 0.0473, + "step": 4618 + }, + { + "epoch": 5.998701298701299, + "grad_norm": 3.1832102801925495, + "learning_rate": 7.289795825129851e-06, + "loss": 0.0107, + "step": 4619 + }, + { + "epoch": 6.0, + "grad_norm": 3.4628241481283846, + "learning_rate": 7.285747311281586e-06, + "loss": 0.0157, + "step": 4620 + }, + { + "epoch": 6.0, + "eval_accuracy": 0.9455362156092083, + "eval_f1": 0.935354822330183, + "eval_loss": 0.13828270137310028, + "eval_precision": 0.9279288946680332, + "eval_recall": 0.9465722020784099, + "eval_runtime": 13.0396, + "eval_samples_per_second": 136.584, + "eval_steps_per_second": 1.074, + "step": 4620 + }, + { + "epoch": 6.001298701298701, + "grad_norm": 3.5666399075699826, + "learning_rate": 7.2816992776365714e-06, + "loss": 0.0043, + "step": 4621 + }, + { + "epoch": 6.002597402597402, + "grad_norm": 3.5394353288821963, + "learning_rate": 7.277651724910986e-06, + "loss": 0.0132, + "step": 4622 + }, + { + "epoch": 6.003896103896104, + "grad_norm": 3.601402462201491, + "learning_rate": 7.2736046538209206e-06, + "loss": 0.0375, + "step": 4623 + }, + { + "epoch": 6.005194805194805, + "grad_norm": 3.965828232450422, + "learning_rate": 7.269558065082369e-06, + "loss": 0.0109, + "step": 4624 + }, + { + "epoch": 6.0064935064935066, + "grad_norm": 1.4824558458869221, + "learning_rate": 7.26551195941126e-06, + "loss": 0.0021, + "step": 4625 + }, + { + "epoch": 6.007792207792208, + "grad_norm": 1.8964844473323574, + "learning_rate": 7.261466337523427e-06, + "loss": 0.0025, + "step": 4626 + }, + { + "epoch": 6.009090909090909, + "grad_norm": 1.4109885596504634, + "learning_rate": 7.257421200134611e-06, + "loss": 0.0077, + "step": 4627 + }, + { + "epoch": 6.01038961038961, + "grad_norm": 0.620613156320488, + "learning_rate": 7.253376547960478e-06, + "loss": 0.0034, + "step": 4628 + }, + { + "epoch": 6.011688311688312, + "grad_norm": 1.5849132399105152, + "learning_rate": 7.249332381716605e-06, + "loss": 0.0033, + "step": 4629 + }, + { + "epoch": 6.012987012987013, + "grad_norm": 3.059216506296831, + "learning_rate": 7.2452887021184844e-06, + "loss": 0.0177, + "step": 4630 + }, + { + "epoch": 6.014285714285714, + "grad_norm": 1.6106207803665362, + "learning_rate": 7.241245509881516e-06, + "loss": 0.0089, + "step": 4631 + }, + { + "epoch": 6.015584415584415, + "grad_norm": 4.27739509359837, + "learning_rate": 7.23720280572102e-06, + "loss": 0.0103, + "step": 4632 + }, + { + "epoch": 6.0168831168831165, + "grad_norm": 3.82614179683853, + "learning_rate": 7.233160590352226e-06, + "loss": 0.0084, + "step": 4633 + }, + { + "epoch": 6.0181818181818185, + "grad_norm": 2.550190036870743, + "learning_rate": 7.229118864490285e-06, + "loss": 0.0025, + "step": 4634 + }, + { + "epoch": 6.01948051948052, + "grad_norm": 5.9514166311738, + "learning_rate": 7.225077628850254e-06, + "loss": 0.0157, + "step": 4635 + }, + { + "epoch": 6.020779220779221, + "grad_norm": 3.618241871512138, + "learning_rate": 7.221036884147099e-06, + "loss": 0.0105, + "step": 4636 + }, + { + "epoch": 6.022077922077922, + "grad_norm": 4.957393661174446, + "learning_rate": 7.216996631095709e-06, + "loss": 0.0099, + "step": 4637 + }, + { + "epoch": 6.023376623376623, + "grad_norm": 4.584315214464652, + "learning_rate": 7.2129568704108845e-06, + "loss": 0.0032, + "step": 4638 + }, + { + "epoch": 6.024675324675325, + "grad_norm": 4.135157890917664, + "learning_rate": 7.208917602807332e-06, + "loss": 0.0066, + "step": 4639 + }, + { + "epoch": 6.025974025974026, + "grad_norm": 6.806393440205881, + "learning_rate": 7.204878828999679e-06, + "loss": 0.0269, + "step": 4640 + }, + { + "epoch": 6.027272727272727, + "grad_norm": 11.273745306453081, + "learning_rate": 7.2008405497024615e-06, + "loss": 0.0206, + "step": 4641 + }, + { + "epoch": 6.0285714285714285, + "grad_norm": 2.7794231448728133, + "learning_rate": 7.196802765630124e-06, + "loss": 0.0031, + "step": 4642 + }, + { + "epoch": 6.02987012987013, + "grad_norm": 6.396648853479158, + "learning_rate": 7.192765477497032e-06, + "loss": 0.0069, + "step": 4643 + }, + { + "epoch": 6.031168831168831, + "grad_norm": 1.2527504547210169, + "learning_rate": 7.188728686017459e-06, + "loss": 0.0095, + "step": 4644 + }, + { + "epoch": 6.032467532467533, + "grad_norm": 4.717947686385261, + "learning_rate": 7.184692391905593e-06, + "loss": 0.0093, + "step": 4645 + }, + { + "epoch": 6.033766233766234, + "grad_norm": 5.158847324116731, + "learning_rate": 7.180656595875524e-06, + "loss": 0.0117, + "step": 4646 + }, + { + "epoch": 6.035064935064935, + "grad_norm": 5.777284125703964, + "learning_rate": 7.176621298641265e-06, + "loss": 0.0271, + "step": 4647 + }, + { + "epoch": 6.036363636363636, + "grad_norm": 8.008975027824453, + "learning_rate": 7.17258650091674e-06, + "loss": 0.0087, + "step": 4648 + }, + { + "epoch": 6.037662337662337, + "grad_norm": 6.969158571036636, + "learning_rate": 7.168552203415784e-06, + "loss": 0.0166, + "step": 4649 + }, + { + "epoch": 6.038961038961039, + "grad_norm": 3.7116282311940085, + "learning_rate": 7.164518406852132e-06, + "loss": 0.0168, + "step": 4650 + }, + { + "epoch": 6.04025974025974, + "grad_norm": 4.305230729695064, + "learning_rate": 7.160485111939447e-06, + "loss": 0.0073, + "step": 4651 + }, + { + "epoch": 6.041558441558442, + "grad_norm": 2.730574348701709, + "learning_rate": 7.156452319391298e-06, + "loss": 0.0061, + "step": 4652 + }, + { + "epoch": 6.042857142857143, + "grad_norm": 6.326192191089981, + "learning_rate": 7.152420029921157e-06, + "loss": 0.0106, + "step": 4653 + }, + { + "epoch": 6.044155844155844, + "grad_norm": 2.382645794862506, + "learning_rate": 7.148388244242414e-06, + "loss": 0.0096, + "step": 4654 + }, + { + "epoch": 6.045454545454546, + "grad_norm": 8.443225935257876, + "learning_rate": 7.144356963068373e-06, + "loss": 0.0424, + "step": 4655 + }, + { + "epoch": 6.046753246753247, + "grad_norm": 2.1542250505382414, + "learning_rate": 7.1403261871122466e-06, + "loss": 0.0022, + "step": 4656 + }, + { + "epoch": 6.048051948051948, + "grad_norm": 3.8903015493117126, + "learning_rate": 7.136295917087148e-06, + "loss": 0.0053, + "step": 4657 + }, + { + "epoch": 6.049350649350649, + "grad_norm": 6.416527775287179, + "learning_rate": 7.132266153706114e-06, + "loss": 0.0052, + "step": 4658 + }, + { + "epoch": 6.05064935064935, + "grad_norm": 3.1929092483895425, + "learning_rate": 7.128236897682084e-06, + "loss": 0.0102, + "step": 4659 + }, + { + "epoch": 6.0519480519480515, + "grad_norm": 15.6961324574367, + "learning_rate": 7.12420814972792e-06, + "loss": 0.0194, + "step": 4660 + }, + { + "epoch": 6.0532467532467535, + "grad_norm": 10.724976218035772, + "learning_rate": 7.120179910556372e-06, + "loss": 0.0361, + "step": 4661 + }, + { + "epoch": 6.054545454545455, + "grad_norm": 5.8510878908016695, + "learning_rate": 7.116152180880118e-06, + "loss": 0.014, + "step": 4662 + }, + { + "epoch": 6.055844155844156, + "grad_norm": 7.107425449515113, + "learning_rate": 7.112124961411741e-06, + "loss": 0.0171, + "step": 4663 + }, + { + "epoch": 6.057142857142857, + "grad_norm": 4.865762057478613, + "learning_rate": 7.108098252863738e-06, + "loss": 0.002, + "step": 4664 + }, + { + "epoch": 6.058441558441558, + "grad_norm": 5.545988301290102, + "learning_rate": 7.1040720559485e-06, + "loss": 0.0093, + "step": 4665 + }, + { + "epoch": 6.05974025974026, + "grad_norm": 5.632126702512004, + "learning_rate": 7.100046371378344e-06, + "loss": 0.0228, + "step": 4666 + }, + { + "epoch": 6.061038961038961, + "grad_norm": 6.869230248220813, + "learning_rate": 7.096021199865494e-06, + "loss": 0.0229, + "step": 4667 + }, + { + "epoch": 6.062337662337662, + "grad_norm": 6.301700621638261, + "learning_rate": 7.091996542122072e-06, + "loss": 0.0387, + "step": 4668 + }, + { + "epoch": 6.0636363636363635, + "grad_norm": 3.7484176855898848, + "learning_rate": 7.087972398860127e-06, + "loss": 0.0235, + "step": 4669 + }, + { + "epoch": 6.064935064935065, + "grad_norm": 10.141317340720201, + "learning_rate": 7.083948770791598e-06, + "loss": 0.0513, + "step": 4670 + }, + { + "epoch": 6.066233766233767, + "grad_norm": 6.881035362768756, + "learning_rate": 7.079925658628346e-06, + "loss": 0.0318, + "step": 4671 + }, + { + "epoch": 6.067532467532468, + "grad_norm": 10.120976844667638, + "learning_rate": 7.075903063082134e-06, + "loss": 0.0035, + "step": 4672 + }, + { + "epoch": 6.068831168831169, + "grad_norm": 5.375233696248492, + "learning_rate": 7.071880984864639e-06, + "loss": 0.0184, + "step": 4673 + }, + { + "epoch": 6.07012987012987, + "grad_norm": 11.244938748351029, + "learning_rate": 7.067859424687444e-06, + "loss": 0.0343, + "step": 4674 + }, + { + "epoch": 6.071428571428571, + "grad_norm": 5.578326380096563, + "learning_rate": 7.063838383262039e-06, + "loss": 0.0123, + "step": 4675 + }, + { + "epoch": 6.072727272727272, + "grad_norm": 5.25738968141956, + "learning_rate": 7.05981786129982e-06, + "loss": 0.0031, + "step": 4676 + }, + { + "epoch": 6.074025974025974, + "grad_norm": 9.176662622632776, + "learning_rate": 7.055797859512098e-06, + "loss": 0.0312, + "step": 4677 + }, + { + "epoch": 6.0753246753246755, + "grad_norm": 2.9159853046413495, + "learning_rate": 7.051778378610087e-06, + "loss": 0.0156, + "step": 4678 + }, + { + "epoch": 6.076623376623377, + "grad_norm": 8.926867085407736, + "learning_rate": 7.047759419304914e-06, + "loss": 0.005, + "step": 4679 + }, + { + "epoch": 6.077922077922078, + "grad_norm": 5.650255876305654, + "learning_rate": 7.043740982307603e-06, + "loss": 0.0099, + "step": 4680 + }, + { + "epoch": 6.079220779220779, + "grad_norm": 2.5479657828729994, + "learning_rate": 7.039723068329094e-06, + "loss": 0.0045, + "step": 4681 + }, + { + "epoch": 6.080519480519481, + "grad_norm": 4.4184639275077755, + "learning_rate": 7.03570567808024e-06, + "loss": 0.0536, + "step": 4682 + }, + { + "epoch": 6.081818181818182, + "grad_norm": 7.28834000349209, + "learning_rate": 7.0316888122717845e-06, + "loss": 0.0123, + "step": 4683 + }, + { + "epoch": 6.083116883116883, + "grad_norm": 2.8391609674783886, + "learning_rate": 7.027672471614392e-06, + "loss": 0.0111, + "step": 4684 + }, + { + "epoch": 6.084415584415584, + "grad_norm": 6.7320034518805505, + "learning_rate": 7.023656656818629e-06, + "loss": 0.0183, + "step": 4685 + }, + { + "epoch": 6.085714285714285, + "grad_norm": 7.689308854713697, + "learning_rate": 7.019641368594979e-06, + "loss": 0.0059, + "step": 4686 + }, + { + "epoch": 6.087012987012987, + "grad_norm": 3.515071104805564, + "learning_rate": 7.015626607653807e-06, + "loss": 0.0038, + "step": 4687 + }, + { + "epoch": 6.088311688311689, + "grad_norm": 5.568381940105322, + "learning_rate": 7.011612374705412e-06, + "loss": 0.0028, + "step": 4688 + }, + { + "epoch": 6.08961038961039, + "grad_norm": 13.347460630409476, + "learning_rate": 7.0075986704599856e-06, + "loss": 0.0152, + "step": 4689 + }, + { + "epoch": 6.090909090909091, + "grad_norm": 3.363150222865331, + "learning_rate": 7.003585495627634e-06, + "loss": 0.0045, + "step": 4690 + }, + { + "epoch": 6.092207792207792, + "grad_norm": 10.551359914498269, + "learning_rate": 6.999572850918357e-06, + "loss": 0.0107, + "step": 4691 + }, + { + "epoch": 6.093506493506493, + "grad_norm": 5.50797939841283, + "learning_rate": 6.995560737042069e-06, + "loss": 0.0101, + "step": 4692 + }, + { + "epoch": 6.094805194805195, + "grad_norm": 5.765391162545159, + "learning_rate": 6.991549154708599e-06, + "loss": 0.0133, + "step": 4693 + }, + { + "epoch": 6.096103896103896, + "grad_norm": 9.3569119232171, + "learning_rate": 6.9875381046276605e-06, + "loss": 0.0108, + "step": 4694 + }, + { + "epoch": 6.097402597402597, + "grad_norm": 1.6661611358411401, + "learning_rate": 6.983527587508889e-06, + "loss": 0.0031, + "step": 4695 + }, + { + "epoch": 6.0987012987012985, + "grad_norm": 4.803819707547482, + "learning_rate": 6.979517604061825e-06, + "loss": 0.0046, + "step": 4696 + }, + { + "epoch": 6.1, + "grad_norm": 2.57362921184888, + "learning_rate": 6.9755081549959135e-06, + "loss": 0.0178, + "step": 4697 + }, + { + "epoch": 6.101298701298702, + "grad_norm": 6.296339289682829, + "learning_rate": 6.971499241020495e-06, + "loss": 0.0105, + "step": 4698 + }, + { + "epoch": 6.102597402597403, + "grad_norm": 8.268071036235854, + "learning_rate": 6.9674908628448255e-06, + "loss": 0.0122, + "step": 4699 + }, + { + "epoch": 6.103896103896104, + "grad_norm": 9.475525441394367, + "learning_rate": 6.963483021178067e-06, + "loss": 0.0241, + "step": 4700 + }, + { + "epoch": 6.105194805194805, + "grad_norm": 4.575767097217805, + "learning_rate": 6.959475716729282e-06, + "loss": 0.0245, + "step": 4701 + }, + { + "epoch": 6.106493506493506, + "grad_norm": 5.769149618887501, + "learning_rate": 6.95546895020744e-06, + "loss": 0.0181, + "step": 4702 + }, + { + "epoch": 6.107792207792208, + "grad_norm": 6.834147711430904, + "learning_rate": 6.951462722321411e-06, + "loss": 0.0503, + "step": 4703 + }, + { + "epoch": 6.109090909090909, + "grad_norm": 4.626233211292643, + "learning_rate": 6.947457033779976e-06, + "loss": 0.0105, + "step": 4704 + }, + { + "epoch": 6.1103896103896105, + "grad_norm": 2.900060388662793, + "learning_rate": 6.943451885291821e-06, + "loss": 0.0174, + "step": 4705 + }, + { + "epoch": 6.111688311688312, + "grad_norm": 3.86782153120207, + "learning_rate": 6.939447277565528e-06, + "loss": 0.0076, + "step": 4706 + }, + { + "epoch": 6.112987012987013, + "grad_norm": 3.4866420868914134, + "learning_rate": 6.935443211309593e-06, + "loss": 0.0097, + "step": 4707 + }, + { + "epoch": 6.114285714285714, + "grad_norm": 4.429428919910894, + "learning_rate": 6.9314396872324105e-06, + "loss": 0.0044, + "step": 4708 + }, + { + "epoch": 6.115584415584416, + "grad_norm": 5.643211169800295, + "learning_rate": 6.927436706042276e-06, + "loss": 0.0161, + "step": 4709 + }, + { + "epoch": 6.116883116883117, + "grad_norm": 1.1110264417636901, + "learning_rate": 6.923434268447401e-06, + "loss": 0.0021, + "step": 4710 + }, + { + "epoch": 6.118181818181818, + "grad_norm": 8.861525114209908, + "learning_rate": 6.9194323751558875e-06, + "loss": 0.0283, + "step": 4711 + }, + { + "epoch": 6.119480519480519, + "grad_norm": 2.193092160405382, + "learning_rate": 6.915431026875756e-06, + "loss": 0.0033, + "step": 4712 + }, + { + "epoch": 6.12077922077922, + "grad_norm": 4.9251750113286965, + "learning_rate": 6.911430224314909e-06, + "loss": 0.033, + "step": 4713 + }, + { + "epoch": 6.1220779220779225, + "grad_norm": 3.066360975845354, + "learning_rate": 6.9074299681811715e-06, + "loss": 0.0102, + "step": 4714 + }, + { + "epoch": 6.123376623376624, + "grad_norm": 5.020542852115517, + "learning_rate": 6.903430259182267e-06, + "loss": 0.0056, + "step": 4715 + }, + { + "epoch": 6.124675324675325, + "grad_norm": 1.0157164195024633, + "learning_rate": 6.89943109802582e-06, + "loss": 0.0099, + "step": 4716 + }, + { + "epoch": 6.125974025974026, + "grad_norm": 5.932766053006979, + "learning_rate": 6.895432485419354e-06, + "loss": 0.0106, + "step": 4717 + }, + { + "epoch": 6.127272727272727, + "grad_norm": 3.6293861255875726, + "learning_rate": 6.891434422070304e-06, + "loss": 0.0108, + "step": 4718 + }, + { + "epoch": 6.128571428571428, + "grad_norm": 3.2565357750245587, + "learning_rate": 6.887436908686007e-06, + "loss": 0.0044, + "step": 4719 + }, + { + "epoch": 6.12987012987013, + "grad_norm": 10.195682576461946, + "learning_rate": 6.883439945973692e-06, + "loss": 0.0344, + "step": 4720 + }, + { + "epoch": 6.131168831168831, + "grad_norm": 5.9373292463323235, + "learning_rate": 6.879443534640502e-06, + "loss": 0.0113, + "step": 4721 + }, + { + "epoch": 6.132467532467532, + "grad_norm": 6.108479739621439, + "learning_rate": 6.875447675393479e-06, + "loss": 0.0161, + "step": 4722 + }, + { + "epoch": 6.1337662337662335, + "grad_norm": 2.232740708056649, + "learning_rate": 6.871452368939571e-06, + "loss": 0.0089, + "step": 4723 + }, + { + "epoch": 6.135064935064935, + "grad_norm": 5.368893018530377, + "learning_rate": 6.867457615985615e-06, + "loss": 0.0171, + "step": 4724 + }, + { + "epoch": 6.136363636363637, + "grad_norm": 1.1829491413511173, + "learning_rate": 6.8634634172383654e-06, + "loss": 0.0223, + "step": 4725 + }, + { + "epoch": 6.137662337662338, + "grad_norm": 2.7135209501326267, + "learning_rate": 6.859469773404471e-06, + "loss": 0.0031, + "step": 4726 + }, + { + "epoch": 6.138961038961039, + "grad_norm": 4.737178210094523, + "learning_rate": 6.855476685190487e-06, + "loss": 0.006, + "step": 4727 + }, + { + "epoch": 6.14025974025974, + "grad_norm": 6.322726565553391, + "learning_rate": 6.851484153302862e-06, + "loss": 0.0176, + "step": 4728 + }, + { + "epoch": 6.141558441558441, + "grad_norm": 3.3075481973904854, + "learning_rate": 6.847492178447952e-06, + "loss": 0.0248, + "step": 4729 + }, + { + "epoch": 6.142857142857143, + "grad_norm": 2.2683333009565376, + "learning_rate": 6.843500761332016e-06, + "loss": 0.0035, + "step": 4730 + }, + { + "epoch": 6.144155844155844, + "grad_norm": 5.2408403813534, + "learning_rate": 6.839509902661215e-06, + "loss": 0.007, + "step": 4731 + }, + { + "epoch": 6.1454545454545455, + "grad_norm": 5.502483626302622, + "learning_rate": 6.835519603141602e-06, + "loss": 0.0088, + "step": 4732 + }, + { + "epoch": 6.146753246753247, + "grad_norm": 5.084837525551477, + "learning_rate": 6.831529863479138e-06, + "loss": 0.0121, + "step": 4733 + }, + { + "epoch": 6.148051948051948, + "grad_norm": 6.464655882121024, + "learning_rate": 6.82754068437969e-06, + "loss": 0.0094, + "step": 4734 + }, + { + "epoch": 6.14935064935065, + "grad_norm": 1.8273412297570253, + "learning_rate": 6.823552066549013e-06, + "loss": 0.0091, + "step": 4735 + }, + { + "epoch": 6.150649350649351, + "grad_norm": 6.501520519422286, + "learning_rate": 6.819564010692776e-06, + "loss": 0.0071, + "step": 4736 + }, + { + "epoch": 6.151948051948052, + "grad_norm": 5.238722800354543, + "learning_rate": 6.8155765175165375e-06, + "loss": 0.0062, + "step": 4737 + }, + { + "epoch": 6.153246753246753, + "grad_norm": 5.772880916869714, + "learning_rate": 6.811589587725766e-06, + "loss": 0.0049, + "step": 4738 + }, + { + "epoch": 6.154545454545454, + "grad_norm": 2.879438766165047, + "learning_rate": 6.80760322202582e-06, + "loss": 0.0085, + "step": 4739 + }, + { + "epoch": 6.1558441558441555, + "grad_norm": 2.135658672349722, + "learning_rate": 6.8036174211219685e-06, + "loss": 0.0097, + "step": 4740 + }, + { + "epoch": 6.1571428571428575, + "grad_norm": 1.740940952330114, + "learning_rate": 6.799632185719376e-06, + "loss": 0.0098, + "step": 4741 + }, + { + "epoch": 6.158441558441559, + "grad_norm": 4.2707248353862015, + "learning_rate": 6.795647516523107e-06, + "loss": 0.0032, + "step": 4742 + }, + { + "epoch": 6.15974025974026, + "grad_norm": 5.870635820840249, + "learning_rate": 6.791663414238121e-06, + "loss": 0.0036, + "step": 4743 + }, + { + "epoch": 6.161038961038961, + "grad_norm": 6.525201900222113, + "learning_rate": 6.787679879569285e-06, + "loss": 0.009, + "step": 4744 + }, + { + "epoch": 6.162337662337662, + "grad_norm": 6.630950913089363, + "learning_rate": 6.783696913221363e-06, + "loss": 0.0232, + "step": 4745 + }, + { + "epoch": 6.163636363636364, + "grad_norm": 3.3837026507407884, + "learning_rate": 6.779714515899022e-06, + "loss": 0.0319, + "step": 4746 + }, + { + "epoch": 6.164935064935065, + "grad_norm": 2.5364407222978262, + "learning_rate": 6.775732688306815e-06, + "loss": 0.0021, + "step": 4747 + }, + { + "epoch": 6.166233766233766, + "grad_norm": 5.113331979771996, + "learning_rate": 6.77175143114921e-06, + "loss": 0.0022, + "step": 4748 + }, + { + "epoch": 6.167532467532467, + "grad_norm": 3.300577905966045, + "learning_rate": 6.76777074513057e-06, + "loss": 0.0133, + "step": 4749 + }, + { + "epoch": 6.1688311688311686, + "grad_norm": 2.9223017816251606, + "learning_rate": 6.763790630955146e-06, + "loss": 0.005, + "step": 4750 + }, + { + "epoch": 6.17012987012987, + "grad_norm": 5.186902342810746, + "learning_rate": 6.759811089327102e-06, + "loss": 0.0031, + "step": 4751 + }, + { + "epoch": 6.171428571428572, + "grad_norm": 3.926093865116436, + "learning_rate": 6.755832120950493e-06, + "loss": 0.0063, + "step": 4752 + }, + { + "epoch": 6.172727272727273, + "grad_norm": 1.669315680328768, + "learning_rate": 6.75185372652928e-06, + "loss": 0.0019, + "step": 4753 + }, + { + "epoch": 6.174025974025974, + "grad_norm": 6.227879993092723, + "learning_rate": 6.747875906767309e-06, + "loss": 0.0253, + "step": 4754 + }, + { + "epoch": 6.175324675324675, + "grad_norm": 5.73684759560029, + "learning_rate": 6.743898662368338e-06, + "loss": 0.0113, + "step": 4755 + }, + { + "epoch": 6.176623376623376, + "grad_norm": 1.1116716932837003, + "learning_rate": 6.739921994036014e-06, + "loss": 0.0106, + "step": 4756 + }, + { + "epoch": 6.177922077922078, + "grad_norm": 2.0207093825934743, + "learning_rate": 6.735945902473893e-06, + "loss": 0.0077, + "step": 4757 + }, + { + "epoch": 6.179220779220779, + "grad_norm": 9.370273901661742, + "learning_rate": 6.731970388385412e-06, + "loss": 0.0093, + "step": 4758 + }, + { + "epoch": 6.1805194805194805, + "grad_norm": 0.9237651180125187, + "learning_rate": 6.727995452473921e-06, + "loss": 0.0084, + "step": 4759 + }, + { + "epoch": 6.181818181818182, + "grad_norm": 14.369506848981413, + "learning_rate": 6.724021095442664e-06, + "loss": 0.0104, + "step": 4760 + }, + { + "epoch": 6.183116883116883, + "grad_norm": 2.627615217905403, + "learning_rate": 6.720047317994775e-06, + "loss": 0.0021, + "step": 4761 + }, + { + "epoch": 6.184415584415585, + "grad_norm": 3.256096276744683, + "learning_rate": 6.716074120833293e-06, + "loss": 0.008, + "step": 4762 + }, + { + "epoch": 6.185714285714286, + "grad_norm": 6.082274100937972, + "learning_rate": 6.712101504661156e-06, + "loss": 0.0039, + "step": 4763 + }, + { + "epoch": 6.187012987012987, + "grad_norm": 5.792794675738881, + "learning_rate": 6.708129470181197e-06, + "loss": 0.0101, + "step": 4764 + }, + { + "epoch": 6.188311688311688, + "grad_norm": 3.477388130116648, + "learning_rate": 6.7041580180961375e-06, + "loss": 0.0168, + "step": 4765 + }, + { + "epoch": 6.189610389610389, + "grad_norm": 8.935197123366297, + "learning_rate": 6.700187149108608e-06, + "loss": 0.0035, + "step": 4766 + }, + { + "epoch": 6.190909090909091, + "grad_norm": 7.863527385098982, + "learning_rate": 6.696216863921131e-06, + "loss": 0.0074, + "step": 4767 + }, + { + "epoch": 6.1922077922077925, + "grad_norm": 8.749131682115996, + "learning_rate": 6.6922471632361275e-06, + "loss": 0.0244, + "step": 4768 + }, + { + "epoch": 6.193506493506494, + "grad_norm": 9.688729581698794, + "learning_rate": 6.688278047755911e-06, + "loss": 0.0066, + "step": 4769 + }, + { + "epoch": 6.194805194805195, + "grad_norm": 2.0261270647375844, + "learning_rate": 6.6843095181826945e-06, + "loss": 0.0086, + "step": 4770 + }, + { + "epoch": 6.196103896103896, + "grad_norm": 9.570054521983293, + "learning_rate": 6.680341575218588e-06, + "loss": 0.0109, + "step": 4771 + }, + { + "epoch": 6.197402597402597, + "grad_norm": 5.934709075700057, + "learning_rate": 6.676374219565596e-06, + "loss": 0.0097, + "step": 4772 + }, + { + "epoch": 6.198701298701299, + "grad_norm": 9.339939040936052, + "learning_rate": 6.672407451925618e-06, + "loss": 0.034, + "step": 4773 + }, + { + "epoch": 6.2, + "grad_norm": 5.183371639356686, + "learning_rate": 6.6684412730004545e-06, + "loss": 0.0028, + "step": 4774 + }, + { + "epoch": 6.201298701298701, + "grad_norm": 4.6682246345301275, + "learning_rate": 6.664475683491797e-06, + "loss": 0.0138, + "step": 4775 + }, + { + "epoch": 6.202597402597402, + "grad_norm": 1.1589437598381864, + "learning_rate": 6.660510684101231e-06, + "loss": 0.0021, + "step": 4776 + }, + { + "epoch": 6.203896103896104, + "grad_norm": 4.981012033163918, + "learning_rate": 6.6565462755302434e-06, + "loss": 0.0169, + "step": 4777 + }, + { + "epoch": 6.205194805194806, + "grad_norm": 3.8223367008167624, + "learning_rate": 6.6525824584802145e-06, + "loss": 0.0088, + "step": 4778 + }, + { + "epoch": 6.206493506493507, + "grad_norm": 5.8650909790327, + "learning_rate": 6.648619233652423e-06, + "loss": 0.0286, + "step": 4779 + }, + { + "epoch": 6.207792207792208, + "grad_norm": 4.41872426129996, + "learning_rate": 6.6446566017480306e-06, + "loss": 0.0066, + "step": 4780 + }, + { + "epoch": 6.209090909090909, + "grad_norm": 4.592578892429151, + "learning_rate": 6.640694563468107e-06, + "loss": 0.0051, + "step": 4781 + }, + { + "epoch": 6.21038961038961, + "grad_norm": 3.66019460363239, + "learning_rate": 6.636733119513614e-06, + "loss": 0.0046, + "step": 4782 + }, + { + "epoch": 6.211688311688311, + "grad_norm": 3.7888766291767872, + "learning_rate": 6.6327722705854106e-06, + "loss": 0.0097, + "step": 4783 + }, + { + "epoch": 6.212987012987013, + "grad_norm": 2.0420309205233242, + "learning_rate": 6.628812017384236e-06, + "loss": 0.0092, + "step": 4784 + }, + { + "epoch": 6.214285714285714, + "grad_norm": 2.556985554762168, + "learning_rate": 6.62485236061074e-06, + "loss": 0.0334, + "step": 4785 + }, + { + "epoch": 6.2155844155844155, + "grad_norm": 10.440484358919013, + "learning_rate": 6.620893300965465e-06, + "loss": 0.0263, + "step": 4786 + }, + { + "epoch": 6.216883116883117, + "grad_norm": 6.543671619592385, + "learning_rate": 6.616934839148846e-06, + "loss": 0.0347, + "step": 4787 + }, + { + "epoch": 6.218181818181818, + "grad_norm": 8.643501886910256, + "learning_rate": 6.6129769758612005e-06, + "loss": 0.0142, + "step": 4788 + }, + { + "epoch": 6.21948051948052, + "grad_norm": 8.58316232611195, + "learning_rate": 6.6090197118027575e-06, + "loss": 0.005, + "step": 4789 + }, + { + "epoch": 6.220779220779221, + "grad_norm": 1.077791479019803, + "learning_rate": 6.605063047673634e-06, + "loss": 0.0024, + "step": 4790 + }, + { + "epoch": 6.222077922077922, + "grad_norm": 4.637450013201792, + "learning_rate": 6.601106984173835e-06, + "loss": 0.0045, + "step": 4791 + }, + { + "epoch": 6.223376623376623, + "grad_norm": 10.023872414213042, + "learning_rate": 6.597151522003265e-06, + "loss": 0.0187, + "step": 4792 + }, + { + "epoch": 6.224675324675324, + "grad_norm": 3.369317390684483, + "learning_rate": 6.5931966618617225e-06, + "loss": 0.0204, + "step": 4793 + }, + { + "epoch": 6.225974025974026, + "grad_norm": 9.616707550050187, + "learning_rate": 6.589242404448901e-06, + "loss": 0.0102, + "step": 4794 + }, + { + "epoch": 6.2272727272727275, + "grad_norm": 5.052045978342619, + "learning_rate": 6.585288750464377e-06, + "loss": 0.021, + "step": 4795 + }, + { + "epoch": 6.228571428571429, + "grad_norm": 5.3021679216478645, + "learning_rate": 6.581335700607632e-06, + "loss": 0.0081, + "step": 4796 + }, + { + "epoch": 6.22987012987013, + "grad_norm": 4.495802077173832, + "learning_rate": 6.577383255578034e-06, + "loss": 0.0052, + "step": 4797 + }, + { + "epoch": 6.231168831168831, + "grad_norm": 4.950804780971259, + "learning_rate": 6.573431416074852e-06, + "loss": 0.0101, + "step": 4798 + }, + { + "epoch": 6.232467532467532, + "grad_norm": 15.65919153976218, + "learning_rate": 6.569480182797234e-06, + "loss": 0.0386, + "step": 4799 + }, + { + "epoch": 6.233766233766234, + "grad_norm": 8.058692968269636, + "learning_rate": 6.5655295564442325e-06, + "loss": 0.0025, + "step": 4800 + }, + { + "epoch": 6.235064935064935, + "grad_norm": 9.296377538684542, + "learning_rate": 6.561579537714791e-06, + "loss": 0.0055, + "step": 4801 + }, + { + "epoch": 6.236363636363636, + "grad_norm": 2.7468140562467656, + "learning_rate": 6.557630127307739e-06, + "loss": 0.0025, + "step": 4802 + }, + { + "epoch": 6.2376623376623375, + "grad_norm": 6.276696263403466, + "learning_rate": 6.553681325921808e-06, + "loss": 0.0097, + "step": 4803 + }, + { + "epoch": 6.238961038961039, + "grad_norm": 7.903013798607147, + "learning_rate": 6.549733134255611e-06, + "loss": 0.0138, + "step": 4804 + }, + { + "epoch": 6.240259740259741, + "grad_norm": 8.163366545931428, + "learning_rate": 6.5457855530076645e-06, + "loss": 0.0207, + "step": 4805 + }, + { + "epoch": 6.241558441558442, + "grad_norm": 9.561241036744415, + "learning_rate": 6.541838582876365e-06, + "loss": 0.0102, + "step": 4806 + }, + { + "epoch": 6.242857142857143, + "grad_norm": 11.09526885826851, + "learning_rate": 6.537892224560009e-06, + "loss": 0.0162, + "step": 4807 + }, + { + "epoch": 6.244155844155844, + "grad_norm": 7.122257065261245, + "learning_rate": 6.533946478756789e-06, + "loss": 0.0199, + "step": 4808 + }, + { + "epoch": 6.245454545454545, + "grad_norm": 16.073416986127008, + "learning_rate": 6.530001346164776e-06, + "loss": 0.021, + "step": 4809 + }, + { + "epoch": 6.246753246753247, + "grad_norm": 8.080183338999603, + "learning_rate": 6.526056827481939e-06, + "loss": 0.0123, + "step": 4810 + }, + { + "epoch": 6.248051948051948, + "grad_norm": 2.4380283246399537, + "learning_rate": 6.522112923406141e-06, + "loss": 0.0157, + "step": 4811 + }, + { + "epoch": 6.249350649350649, + "grad_norm": 11.692561228665774, + "learning_rate": 6.518169634635135e-06, + "loss": 0.0156, + "step": 4812 + }, + { + "epoch": 6.250649350649351, + "grad_norm": 6.686727923267012, + "learning_rate": 6.514226961866567e-06, + "loss": 0.0137, + "step": 4813 + }, + { + "epoch": 6.251948051948052, + "grad_norm": 5.144067942035481, + "learning_rate": 6.510284905797966e-06, + "loss": 0.015, + "step": 4814 + }, + { + "epoch": 6.253246753246753, + "grad_norm": 1.597967129024345, + "learning_rate": 6.506343467126757e-06, + "loss": 0.009, + "step": 4815 + }, + { + "epoch": 6.254545454545455, + "grad_norm": 2.884798903530732, + "learning_rate": 6.502402646550261e-06, + "loss": 0.0029, + "step": 4816 + }, + { + "epoch": 6.255844155844156, + "grad_norm": 13.354126999457637, + "learning_rate": 6.49846244476568e-06, + "loss": 0.0322, + "step": 4817 + }, + { + "epoch": 6.257142857142857, + "grad_norm": 7.523697719786588, + "learning_rate": 6.494522862470112e-06, + "loss": 0.0042, + "step": 4818 + }, + { + "epoch": 6.258441558441558, + "grad_norm": 7.9120013572543435, + "learning_rate": 6.490583900360543e-06, + "loss": 0.0129, + "step": 4819 + }, + { + "epoch": 6.259740259740259, + "grad_norm": 5.9710580294303375, + "learning_rate": 6.48664555913386e-06, + "loss": 0.012, + "step": 4820 + }, + { + "epoch": 6.261038961038961, + "grad_norm": 5.60857710654767, + "learning_rate": 6.482707839486818e-06, + "loss": 0.0029, + "step": 4821 + }, + { + "epoch": 6.2623376623376625, + "grad_norm": 10.165406843384181, + "learning_rate": 6.478770742116081e-06, + "loss": 0.0032, + "step": 4822 + }, + { + "epoch": 6.263636363636364, + "grad_norm": 5.09061710963251, + "learning_rate": 6.474834267718198e-06, + "loss": 0.0032, + "step": 4823 + }, + { + "epoch": 6.264935064935065, + "grad_norm": 5.429565540712432, + "learning_rate": 6.4708984169896084e-06, + "loss": 0.0036, + "step": 4824 + }, + { + "epoch": 6.266233766233766, + "grad_norm": 7.51449966665674, + "learning_rate": 6.466963190626633e-06, + "loss": 0.0055, + "step": 4825 + }, + { + "epoch": 6.267532467532467, + "grad_norm": 4.554157868166777, + "learning_rate": 6.463028589325492e-06, + "loss": 0.0033, + "step": 4826 + }, + { + "epoch": 6.268831168831169, + "grad_norm": 2.0808522488338244, + "learning_rate": 6.459094613782292e-06, + "loss": 0.006, + "step": 4827 + }, + { + "epoch": 6.27012987012987, + "grad_norm": 6.418810299840018, + "learning_rate": 6.455161264693033e-06, + "loss": 0.013, + "step": 4828 + }, + { + "epoch": 6.271428571428571, + "grad_norm": 7.264674505863836, + "learning_rate": 6.4512285427535905e-06, + "loss": 0.0115, + "step": 4829 + }, + { + "epoch": 6.2727272727272725, + "grad_norm": 4.34641660622478, + "learning_rate": 6.447296448659744e-06, + "loss": 0.0113, + "step": 4830 + }, + { + "epoch": 6.274025974025974, + "grad_norm": 8.865745357882734, + "learning_rate": 6.443364983107156e-06, + "loss": 0.013, + "step": 4831 + }, + { + "epoch": 6.275324675324676, + "grad_norm": 4.129415604394781, + "learning_rate": 6.439434146791377e-06, + "loss": 0.0086, + "step": 4832 + }, + { + "epoch": 6.276623376623377, + "grad_norm": 6.55199125639661, + "learning_rate": 6.4355039404078455e-06, + "loss": 0.0096, + "step": 4833 + }, + { + "epoch": 6.277922077922078, + "grad_norm": 3.387929277015626, + "learning_rate": 6.431574364651893e-06, + "loss": 0.0112, + "step": 4834 + }, + { + "epoch": 6.279220779220779, + "grad_norm": 9.07714209498877, + "learning_rate": 6.427645420218738e-06, + "loss": 0.0333, + "step": 4835 + }, + { + "epoch": 6.28051948051948, + "grad_norm": 3.796337171796981, + "learning_rate": 6.423717107803482e-06, + "loss": 0.0059, + "step": 4836 + }, + { + "epoch": 6.281818181818182, + "grad_norm": 7.70609678138085, + "learning_rate": 6.41978942810112e-06, + "loss": 0.0394, + "step": 4837 + }, + { + "epoch": 6.283116883116883, + "grad_norm": 4.486551831812383, + "learning_rate": 6.415862381806533e-06, + "loss": 0.0025, + "step": 4838 + }, + { + "epoch": 6.2844155844155845, + "grad_norm": 3.3532460849282555, + "learning_rate": 6.411935969614495e-06, + "loss": 0.0175, + "step": 4839 + }, + { + "epoch": 6.285714285714286, + "grad_norm": 6.696810184303006, + "learning_rate": 6.408010192219658e-06, + "loss": 0.0133, + "step": 4840 + }, + { + "epoch": 6.287012987012987, + "grad_norm": 7.213389303776153, + "learning_rate": 6.404085050316568e-06, + "loss": 0.0033, + "step": 4841 + }, + { + "epoch": 6.288311688311689, + "grad_norm": 7.25662922779811, + "learning_rate": 6.400160544599665e-06, + "loss": 0.0174, + "step": 4842 + }, + { + "epoch": 6.28961038961039, + "grad_norm": 7.545152560596896, + "learning_rate": 6.396236675763259e-06, + "loss": 0.0054, + "step": 4843 + }, + { + "epoch": 6.290909090909091, + "grad_norm": 11.987178253246027, + "learning_rate": 6.392313444501561e-06, + "loss": 0.0311, + "step": 4844 + }, + { + "epoch": 6.292207792207792, + "grad_norm": 4.1201277298766525, + "learning_rate": 6.388390851508667e-06, + "loss": 0.0054, + "step": 4845 + }, + { + "epoch": 6.293506493506493, + "grad_norm": 5.5978061161301484, + "learning_rate": 6.3844688974785615e-06, + "loss": 0.0032, + "step": 4846 + }, + { + "epoch": 6.294805194805194, + "grad_norm": 8.804629619651076, + "learning_rate": 6.380547583105108e-06, + "loss": 0.0169, + "step": 4847 + }, + { + "epoch": 6.296103896103896, + "grad_norm": 7.935887507111981, + "learning_rate": 6.376626909082062e-06, + "loss": 0.0062, + "step": 4848 + }, + { + "epoch": 6.297402597402598, + "grad_norm": 4.483819255887192, + "learning_rate": 6.372706876103069e-06, + "loss": 0.0098, + "step": 4849 + }, + { + "epoch": 6.298701298701299, + "grad_norm": 6.646677362877124, + "learning_rate": 6.36878748486166e-06, + "loss": 0.007, + "step": 4850 + }, + { + "epoch": 6.3, + "grad_norm": 3.490957497950394, + "learning_rate": 6.364868736051244e-06, + "loss": 0.0052, + "step": 4851 + }, + { + "epoch": 6.301298701298701, + "grad_norm": 5.623088116209584, + "learning_rate": 6.360950630365126e-06, + "loss": 0.0025, + "step": 4852 + }, + { + "epoch": 6.302597402597403, + "grad_norm": 9.254086446800049, + "learning_rate": 6.357033168496491e-06, + "loss": 0.0057, + "step": 4853 + }, + { + "epoch": 6.303896103896104, + "grad_norm": 5.60677091109191, + "learning_rate": 6.353116351138423e-06, + "loss": 0.0236, + "step": 4854 + }, + { + "epoch": 6.305194805194805, + "grad_norm": 6.121871603697836, + "learning_rate": 6.34920017898387e-06, + "loss": 0.0023, + "step": 4855 + }, + { + "epoch": 6.306493506493506, + "grad_norm": 7.107283823659935, + "learning_rate": 6.345284652725681e-06, + "loss": 0.0038, + "step": 4856 + }, + { + "epoch": 6.3077922077922075, + "grad_norm": 2.0121457360872235, + "learning_rate": 6.341369773056594e-06, + "loss": 0.0077, + "step": 4857 + }, + { + "epoch": 6.309090909090909, + "grad_norm": 8.362777082707316, + "learning_rate": 6.337455540669215e-06, + "loss": 0.0105, + "step": 4858 + }, + { + "epoch": 6.310389610389611, + "grad_norm": 9.578248905031996, + "learning_rate": 6.333541956256055e-06, + "loss": 0.0084, + "step": 4859 + }, + { + "epoch": 6.311688311688312, + "grad_norm": 2.3681799668819856, + "learning_rate": 6.329629020509499e-06, + "loss": 0.0038, + "step": 4860 + }, + { + "epoch": 6.312987012987013, + "grad_norm": 10.21022644787356, + "learning_rate": 6.325716734121825e-06, + "loss": 0.023, + "step": 4861 + }, + { + "epoch": 6.314285714285714, + "grad_norm": 13.930529594965948, + "learning_rate": 6.321805097785183e-06, + "loss": 0.02, + "step": 4862 + }, + { + "epoch": 6.315584415584415, + "grad_norm": 1.3849170412413967, + "learning_rate": 6.317894112191621e-06, + "loss": 0.0082, + "step": 4863 + }, + { + "epoch": 6.316883116883117, + "grad_norm": 8.901633209983993, + "learning_rate": 6.313983778033066e-06, + "loss": 0.0033, + "step": 4864 + }, + { + "epoch": 6.318181818181818, + "grad_norm": 7.328550095839743, + "learning_rate": 6.3100740960013375e-06, + "loss": 0.0065, + "step": 4865 + }, + { + "epoch": 6.3194805194805195, + "grad_norm": 2.613724426716131, + "learning_rate": 6.306165066788121e-06, + "loss": 0.0077, + "step": 4866 + }, + { + "epoch": 6.320779220779221, + "grad_norm": 4.73769295479546, + "learning_rate": 6.302256691085007e-06, + "loss": 0.0038, + "step": 4867 + }, + { + "epoch": 6.322077922077922, + "grad_norm": 14.169647763572968, + "learning_rate": 6.298348969583461e-06, + "loss": 0.0192, + "step": 4868 + }, + { + "epoch": 6.323376623376624, + "grad_norm": 7.6982226417662245, + "learning_rate": 6.294441902974831e-06, + "loss": 0.0098, + "step": 4869 + }, + { + "epoch": 6.324675324675325, + "grad_norm": 8.378911246073343, + "learning_rate": 6.2905354919503534e-06, + "loss": 0.0106, + "step": 4870 + }, + { + "epoch": 6.325974025974026, + "grad_norm": 8.504913796035561, + "learning_rate": 6.286629737201149e-06, + "loss": 0.0132, + "step": 4871 + }, + { + "epoch": 6.327272727272727, + "grad_norm": 5.5152101089715195, + "learning_rate": 6.282724639418218e-06, + "loss": 0.0154, + "step": 4872 + }, + { + "epoch": 6.328571428571428, + "grad_norm": 1.0676009260125423, + "learning_rate": 6.278820199292446e-06, + "loss": 0.0018, + "step": 4873 + }, + { + "epoch": 6.32987012987013, + "grad_norm": 9.480642934822294, + "learning_rate": 6.274916417514605e-06, + "loss": 0.0046, + "step": 4874 + }, + { + "epoch": 6.3311688311688314, + "grad_norm": 3.517205176509037, + "learning_rate": 6.271013294775351e-06, + "loss": 0.0013, + "step": 4875 + }, + { + "epoch": 6.332467532467533, + "grad_norm": 2.4058947952503016, + "learning_rate": 6.267110831765219e-06, + "loss": 0.0025, + "step": 4876 + }, + { + "epoch": 6.333766233766234, + "grad_norm": 4.593924215747457, + "learning_rate": 6.263209029174626e-06, + "loss": 0.0109, + "step": 4877 + }, + { + "epoch": 6.335064935064935, + "grad_norm": 13.852882392565064, + "learning_rate": 6.25930788769388e-06, + "loss": 0.0482, + "step": 4878 + }, + { + "epoch": 6.336363636363636, + "grad_norm": 9.820896757250141, + "learning_rate": 6.2554074080131655e-06, + "loss": 0.0119, + "step": 4879 + }, + { + "epoch": 6.337662337662338, + "grad_norm": 8.535332816909749, + "learning_rate": 6.251507590822556e-06, + "loss": 0.0383, + "step": 4880 + }, + { + "epoch": 6.338961038961039, + "grad_norm": 7.985302880383314, + "learning_rate": 6.247608436811999e-06, + "loss": 0.0185, + "step": 4881 + }, + { + "epoch": 6.34025974025974, + "grad_norm": 9.751582883700658, + "learning_rate": 6.243709946671329e-06, + "loss": 0.032, + "step": 4882 + }, + { + "epoch": 6.341558441558441, + "grad_norm": 10.68886296370772, + "learning_rate": 6.239812121090271e-06, + "loss": 0.0266, + "step": 4883 + }, + { + "epoch": 6.3428571428571425, + "grad_norm": 5.557008712733501, + "learning_rate": 6.235914960758415e-06, + "loss": 0.042, + "step": 4884 + }, + { + "epoch": 6.3441558441558445, + "grad_norm": 5.025430573303012, + "learning_rate": 6.232018466365247e-06, + "loss": 0.0109, + "step": 4885 + }, + { + "epoch": 6.345454545454546, + "grad_norm": 5.60400591476872, + "learning_rate": 6.228122638600133e-06, + "loss": 0.0137, + "step": 4886 + }, + { + "epoch": 6.346753246753247, + "grad_norm": 2.3003992487115337, + "learning_rate": 6.224227478152322e-06, + "loss": 0.0046, + "step": 4887 + }, + { + "epoch": 6.348051948051948, + "grad_norm": 1.4399518452686362, + "learning_rate": 6.220332985710936e-06, + "loss": 0.0038, + "step": 4888 + }, + { + "epoch": 6.349350649350649, + "grad_norm": 2.6273991135373715, + "learning_rate": 6.216439161964989e-06, + "loss": 0.0206, + "step": 4889 + }, + { + "epoch": 6.35064935064935, + "grad_norm": 4.476998822434275, + "learning_rate": 6.21254600760337e-06, + "loss": 0.0049, + "step": 4890 + }, + { + "epoch": 6.351948051948052, + "grad_norm": 3.1830615158241793, + "learning_rate": 6.208653523314861e-06, + "loss": 0.0184, + "step": 4891 + }, + { + "epoch": 6.353246753246753, + "grad_norm": 7.126919491690643, + "learning_rate": 6.204761709788106e-06, + "loss": 0.0151, + "step": 4892 + }, + { + "epoch": 6.3545454545454545, + "grad_norm": 4.211681099253766, + "learning_rate": 6.200870567711647e-06, + "loss": 0.0027, + "step": 4893 + }, + { + "epoch": 6.355844155844156, + "grad_norm": 2.206936885934521, + "learning_rate": 6.1969800977739e-06, + "loss": 0.0016, + "step": 4894 + }, + { + "epoch": 6.357142857142857, + "grad_norm": 2.7002535392350304, + "learning_rate": 6.193090300663169e-06, + "loss": 0.0054, + "step": 4895 + }, + { + "epoch": 6.358441558441559, + "grad_norm": 4.250183058787329, + "learning_rate": 6.189201177067623e-06, + "loss": 0.005, + "step": 4896 + }, + { + "epoch": 6.35974025974026, + "grad_norm": 8.65361739004982, + "learning_rate": 6.18531272767533e-06, + "loss": 0.0267, + "step": 4897 + }, + { + "epoch": 6.361038961038961, + "grad_norm": 6.278596829995806, + "learning_rate": 6.181424953174231e-06, + "loss": 0.0191, + "step": 4898 + }, + { + "epoch": 6.362337662337662, + "grad_norm": 3.3365219585837123, + "learning_rate": 6.1775378542521465e-06, + "loss": 0.0029, + "step": 4899 + }, + { + "epoch": 6.363636363636363, + "grad_norm": 3.403638164405322, + "learning_rate": 6.173651431596776e-06, + "loss": 0.0114, + "step": 4900 + }, + { + "epoch": 6.364935064935065, + "grad_norm": 2.3661391151285147, + "learning_rate": 6.169765685895703e-06, + "loss": 0.0091, + "step": 4901 + }, + { + "epoch": 6.3662337662337665, + "grad_norm": 3.08454127108071, + "learning_rate": 6.165880617836395e-06, + "loss": 0.0066, + "step": 4902 + }, + { + "epoch": 6.367532467532468, + "grad_norm": 5.936737022452585, + "learning_rate": 6.16199622810619e-06, + "loss": 0.0199, + "step": 4903 + }, + { + "epoch": 6.368831168831169, + "grad_norm": 5.187941530441695, + "learning_rate": 6.158112517392314e-06, + "loss": 0.0046, + "step": 4904 + }, + { + "epoch": 6.37012987012987, + "grad_norm": 12.564986018589433, + "learning_rate": 6.154229486381865e-06, + "loss": 0.0073, + "step": 4905 + }, + { + "epoch": 6.371428571428572, + "grad_norm": 3.0865227571008047, + "learning_rate": 6.150347135761831e-06, + "loss": 0.0093, + "step": 4906 + }, + { + "epoch": 6.372727272727273, + "grad_norm": 1.1377716611822901, + "learning_rate": 6.146465466219072e-06, + "loss": 0.002, + "step": 4907 + }, + { + "epoch": 6.374025974025974, + "grad_norm": 2.5247416112039134, + "learning_rate": 6.142584478440328e-06, + "loss": 0.0134, + "step": 4908 + }, + { + "epoch": 6.375324675324675, + "grad_norm": 4.2232666593588295, + "learning_rate": 6.138704173112224e-06, + "loss": 0.0026, + "step": 4909 + }, + { + "epoch": 6.376623376623376, + "grad_norm": 3.1219104076437, + "learning_rate": 6.1348245509212555e-06, + "loss": 0.0177, + "step": 4910 + }, + { + "epoch": 6.3779220779220775, + "grad_norm": 8.008221060527994, + "learning_rate": 6.130945612553804e-06, + "loss": 0.0127, + "step": 4911 + }, + { + "epoch": 6.37922077922078, + "grad_norm": 1.7769526646479046, + "learning_rate": 6.127067358696128e-06, + "loss": 0.0023, + "step": 4912 + }, + { + "epoch": 6.380519480519481, + "grad_norm": 1.593285211490419, + "learning_rate": 6.123189790034368e-06, + "loss": 0.0028, + "step": 4913 + }, + { + "epoch": 6.381818181818182, + "grad_norm": 3.1176732034897587, + "learning_rate": 6.119312907254535e-06, + "loss": 0.0106, + "step": 4914 + }, + { + "epoch": 6.383116883116883, + "grad_norm": 6.417352678036016, + "learning_rate": 6.115436711042524e-06, + "loss": 0.0097, + "step": 4915 + }, + { + "epoch": 6.384415584415584, + "grad_norm": 5.43217533870259, + "learning_rate": 6.111561202084111e-06, + "loss": 0.0251, + "step": 4916 + }, + { + "epoch": 6.385714285714286, + "grad_norm": 7.356705554774608, + "learning_rate": 6.107686381064951e-06, + "loss": 0.0032, + "step": 4917 + }, + { + "epoch": 6.387012987012987, + "grad_norm": 3.5392165563681126, + "learning_rate": 6.103812248670567e-06, + "loss": 0.0027, + "step": 4918 + }, + { + "epoch": 6.388311688311688, + "grad_norm": 4.238358086484861, + "learning_rate": 6.099938805586367e-06, + "loss": 0.0042, + "step": 4919 + }, + { + "epoch": 6.3896103896103895, + "grad_norm": 6.507998432597933, + "learning_rate": 6.096066052497642e-06, + "loss": 0.0092, + "step": 4920 + }, + { + "epoch": 6.390909090909091, + "grad_norm": 4.7367156987520165, + "learning_rate": 6.092193990089559e-06, + "loss": 0.0158, + "step": 4921 + }, + { + "epoch": 6.392207792207792, + "grad_norm": 7.607220201324081, + "learning_rate": 6.0883226190471516e-06, + "loss": 0.0037, + "step": 4922 + }, + { + "epoch": 6.393506493506494, + "grad_norm": 8.890265304960849, + "learning_rate": 6.084451940055341e-06, + "loss": 0.0146, + "step": 4923 + }, + { + "epoch": 6.394805194805195, + "grad_norm": 3.320528940759403, + "learning_rate": 6.0805819537989316e-06, + "loss": 0.0073, + "step": 4924 + }, + { + "epoch": 6.396103896103896, + "grad_norm": 8.477004177998044, + "learning_rate": 6.076712660962591e-06, + "loss": 0.0109, + "step": 4925 + }, + { + "epoch": 6.397402597402597, + "grad_norm": 2.771711278112249, + "learning_rate": 6.072844062230871e-06, + "loss": 0.0022, + "step": 4926 + }, + { + "epoch": 6.398701298701298, + "grad_norm": 5.822886567349937, + "learning_rate": 6.068976158288204e-06, + "loss": 0.0191, + "step": 4927 + }, + { + "epoch": 6.4, + "grad_norm": 5.736097626810347, + "learning_rate": 6.065108949818899e-06, + "loss": 0.0135, + "step": 4928 + }, + { + "epoch": 6.4012987012987015, + "grad_norm": 4.93245122077385, + "learning_rate": 6.061242437507131e-06, + "loss": 0.017, + "step": 4929 + }, + { + "epoch": 6.402597402597403, + "grad_norm": 1.3012850813592383, + "learning_rate": 6.057376622036964e-06, + "loss": 0.0024, + "step": 4930 + }, + { + "epoch": 6.403896103896104, + "grad_norm": 4.2188499768109615, + "learning_rate": 6.053511504092335e-06, + "loss": 0.0099, + "step": 4931 + }, + { + "epoch": 6.405194805194805, + "grad_norm": 6.703900348300436, + "learning_rate": 6.049647084357062e-06, + "loss": 0.0163, + "step": 4932 + }, + { + "epoch": 6.406493506493507, + "grad_norm": 6.75561539474607, + "learning_rate": 6.045783363514828e-06, + "loss": 0.0071, + "step": 4933 + }, + { + "epoch": 6.407792207792208, + "grad_norm": 5.458175258775898, + "learning_rate": 6.041920342249199e-06, + "loss": 0.0211, + "step": 4934 + }, + { + "epoch": 6.409090909090909, + "grad_norm": 8.1268673721522, + "learning_rate": 6.038058021243621e-06, + "loss": 0.0116, + "step": 4935 + }, + { + "epoch": 6.41038961038961, + "grad_norm": 1.153424879009012, + "learning_rate": 6.034196401181414e-06, + "loss": 0.0124, + "step": 4936 + }, + { + "epoch": 6.411688311688311, + "grad_norm": 1.4837606061357382, + "learning_rate": 6.03033548274577e-06, + "loss": 0.0031, + "step": 4937 + }, + { + "epoch": 6.4129870129870135, + "grad_norm": 2.0350733605565345, + "learning_rate": 6.026475266619756e-06, + "loss": 0.0118, + "step": 4938 + }, + { + "epoch": 6.414285714285715, + "grad_norm": 10.652848065782146, + "learning_rate": 6.0226157534863236e-06, + "loss": 0.0286, + "step": 4939 + }, + { + "epoch": 6.415584415584416, + "grad_norm": 4.108465254596745, + "learning_rate": 6.018756944028291e-06, + "loss": 0.0236, + "step": 4940 + }, + { + "epoch": 6.416883116883117, + "grad_norm": 7.124920063935965, + "learning_rate": 6.014898838928357e-06, + "loss": 0.0084, + "step": 4941 + }, + { + "epoch": 6.418181818181818, + "grad_norm": 1.0291394215635583, + "learning_rate": 6.011041438869096e-06, + "loss": 0.0016, + "step": 4942 + }, + { + "epoch": 6.419480519480519, + "grad_norm": 7.052772192526333, + "learning_rate": 6.007184744532953e-06, + "loss": 0.0119, + "step": 4943 + }, + { + "epoch": 6.420779220779221, + "grad_norm": 2.61908728192196, + "learning_rate": 6.003328756602252e-06, + "loss": 0.0154, + "step": 4944 + }, + { + "epoch": 6.422077922077922, + "grad_norm": 7.013634786315303, + "learning_rate": 5.9994734757591875e-06, + "loss": 0.016, + "step": 4945 + }, + { + "epoch": 6.423376623376623, + "grad_norm": 2.08101170064285, + "learning_rate": 5.995618902685839e-06, + "loss": 0.0044, + "step": 4946 + }, + { + "epoch": 6.4246753246753245, + "grad_norm": 6.274630920352025, + "learning_rate": 5.991765038064155e-06, + "loss": 0.0138, + "step": 4947 + }, + { + "epoch": 6.425974025974026, + "grad_norm": 1.7705121814094236, + "learning_rate": 5.98791188257595e-06, + "loss": 0.0118, + "step": 4948 + }, + { + "epoch": 6.427272727272728, + "grad_norm": 3.97607556559282, + "learning_rate": 5.984059436902925e-06, + "loss": 0.0427, + "step": 4949 + }, + { + "epoch": 6.428571428571429, + "grad_norm": 5.808847014437884, + "learning_rate": 5.980207701726655e-06, + "loss": 0.0109, + "step": 4950 + }, + { + "epoch": 6.42987012987013, + "grad_norm": 2.4402161238286593, + "learning_rate": 5.976356677728578e-06, + "loss": 0.0118, + "step": 4951 + }, + { + "epoch": 6.431168831168831, + "grad_norm": 4.552627287128298, + "learning_rate": 5.972506365590019e-06, + "loss": 0.0107, + "step": 4952 + }, + { + "epoch": 6.432467532467532, + "grad_norm": 4.607578589344118, + "learning_rate": 5.96865676599217e-06, + "loss": 0.0091, + "step": 4953 + }, + { + "epoch": 6.433766233766233, + "grad_norm": 3.8117738229327296, + "learning_rate": 5.964807879616105e-06, + "loss": 0.0105, + "step": 4954 + }, + { + "epoch": 6.435064935064935, + "grad_norm": 6.118998616300519, + "learning_rate": 5.960959707142755e-06, + "loss": 0.0118, + "step": 4955 + }, + { + "epoch": 6.4363636363636365, + "grad_norm": 7.102028030174515, + "learning_rate": 5.9571122492529435e-06, + "loss": 0.0122, + "step": 4956 + }, + { + "epoch": 6.437662337662338, + "grad_norm": 6.595681472824495, + "learning_rate": 5.953265506627355e-06, + "loss": 0.0049, + "step": 4957 + }, + { + "epoch": 6.438961038961039, + "grad_norm": 1.5649823730428825, + "learning_rate": 5.949419479946556e-06, + "loss": 0.0047, + "step": 4958 + }, + { + "epoch": 6.44025974025974, + "grad_norm": 5.641843703375221, + "learning_rate": 5.945574169890979e-06, + "loss": 0.0302, + "step": 4959 + }, + { + "epoch": 6.441558441558442, + "grad_norm": 5.6281557916643346, + "learning_rate": 5.941729577140932e-06, + "loss": 0.0025, + "step": 4960 + }, + { + "epoch": 6.442857142857143, + "grad_norm": 7.733026901753861, + "learning_rate": 5.9378857023766e-06, + "loss": 0.0129, + "step": 4961 + }, + { + "epoch": 6.444155844155844, + "grad_norm": 5.625559438494942, + "learning_rate": 5.9340425462780405e-06, + "loss": 0.0033, + "step": 4962 + }, + { + "epoch": 6.445454545454545, + "grad_norm": 4.833149340042153, + "learning_rate": 5.930200109525174e-06, + "loss": 0.0291, + "step": 4963 + }, + { + "epoch": 6.4467532467532465, + "grad_norm": 2.6747897628548163, + "learning_rate": 5.926358392797804e-06, + "loss": 0.0028, + "step": 4964 + }, + { + "epoch": 6.448051948051948, + "grad_norm": 1.9994665671618532, + "learning_rate": 5.922517396775607e-06, + "loss": 0.0022, + "step": 4965 + }, + { + "epoch": 6.44935064935065, + "grad_norm": 7.069378688738637, + "learning_rate": 5.918677122138128e-06, + "loss": 0.0102, + "step": 4966 + }, + { + "epoch": 6.450649350649351, + "grad_norm": 7.11900321617579, + "learning_rate": 5.9148375695647795e-06, + "loss": 0.0208, + "step": 4967 + }, + { + "epoch": 6.451948051948052, + "grad_norm": 4.276858814250507, + "learning_rate": 5.9109987397348566e-06, + "loss": 0.0108, + "step": 4968 + }, + { + "epoch": 6.453246753246753, + "grad_norm": 6.948383080697875, + "learning_rate": 5.907160633327522e-06, + "loss": 0.0089, + "step": 4969 + }, + { + "epoch": 6.454545454545454, + "grad_norm": 3.5216848807571752, + "learning_rate": 5.903323251021808e-06, + "loss": 0.0237, + "step": 4970 + }, + { + "epoch": 6.455844155844156, + "grad_norm": 4.696894343868069, + "learning_rate": 5.899486593496625e-06, + "loss": 0.0042, + "step": 4971 + }, + { + "epoch": 6.457142857142857, + "grad_norm": 3.6224693777401678, + "learning_rate": 5.895650661430745e-06, + "loss": 0.0263, + "step": 4972 + }, + { + "epoch": 6.458441558441558, + "grad_norm": 4.777067148530515, + "learning_rate": 5.891815455502824e-06, + "loss": 0.0101, + "step": 4973 + }, + { + "epoch": 6.45974025974026, + "grad_norm": 5.913241914554619, + "learning_rate": 5.887980976391379e-06, + "loss": 0.0145, + "step": 4974 + }, + { + "epoch": 6.461038961038961, + "grad_norm": 2.3121873211637314, + "learning_rate": 5.8841472247748025e-06, + "loss": 0.0171, + "step": 4975 + }, + { + "epoch": 6.462337662337663, + "grad_norm": 3.379366234556156, + "learning_rate": 5.880314201331362e-06, + "loss": 0.0101, + "step": 4976 + }, + { + "epoch": 6.463636363636364, + "grad_norm": 2.692692669191053, + "learning_rate": 5.876481906739192e-06, + "loss": 0.0071, + "step": 4977 + }, + { + "epoch": 6.464935064935065, + "grad_norm": 2.855208300201893, + "learning_rate": 5.8726503416762955e-06, + "loss": 0.0035, + "step": 4978 + }, + { + "epoch": 6.466233766233766, + "grad_norm": 1.7468534553302923, + "learning_rate": 5.868819506820551e-06, + "loss": 0.0019, + "step": 4979 + }, + { + "epoch": 6.467532467532467, + "grad_norm": 5.171164574403623, + "learning_rate": 5.864989402849712e-06, + "loss": 0.0123, + "step": 4980 + }, + { + "epoch": 6.468831168831169, + "grad_norm": 2.8680987894892884, + "learning_rate": 5.861160030441389e-06, + "loss": 0.0115, + "step": 4981 + }, + { + "epoch": 6.47012987012987, + "grad_norm": 5.672327597009564, + "learning_rate": 5.8573313902730745e-06, + "loss": 0.009, + "step": 4982 + }, + { + "epoch": 6.4714285714285715, + "grad_norm": 5.227568868455298, + "learning_rate": 5.8535034830221296e-06, + "loss": 0.0131, + "step": 4983 + }, + { + "epoch": 6.472727272727273, + "grad_norm": 7.714547717164378, + "learning_rate": 5.849676309365786e-06, + "loss": 0.0162, + "step": 4984 + }, + { + "epoch": 6.474025974025974, + "grad_norm": 2.9398621677309222, + "learning_rate": 5.845849869981137e-06, + "loss": 0.0105, + "step": 4985 + }, + { + "epoch": 6.475324675324675, + "grad_norm": 3.048668758038415, + "learning_rate": 5.8420241655451585e-06, + "loss": 0.0172, + "step": 4986 + }, + { + "epoch": 6.476623376623377, + "grad_norm": 2.7554371432174776, + "learning_rate": 5.8381991967346904e-06, + "loss": 0.0025, + "step": 4987 + }, + { + "epoch": 6.477922077922078, + "grad_norm": 0.8008258248081228, + "learning_rate": 5.834374964226444e-06, + "loss": 0.0096, + "step": 4988 + }, + { + "epoch": 6.479220779220779, + "grad_norm": 2.4276762179167233, + "learning_rate": 5.8305514686969915e-06, + "loss": 0.0041, + "step": 4989 + }, + { + "epoch": 6.48051948051948, + "grad_norm": 4.888877830830518, + "learning_rate": 5.826728710822793e-06, + "loss": 0.0067, + "step": 4990 + }, + { + "epoch": 6.4818181818181815, + "grad_norm": 3.0114482298016654, + "learning_rate": 5.8229066912801625e-06, + "loss": 0.0049, + "step": 4991 + }, + { + "epoch": 6.4831168831168835, + "grad_norm": 6.002839823798016, + "learning_rate": 5.819085410745289e-06, + "loss": 0.0062, + "step": 4992 + }, + { + "epoch": 6.484415584415585, + "grad_norm": 6.839408678008397, + "learning_rate": 5.815264869894226e-06, + "loss": 0.0089, + "step": 4993 + }, + { + "epoch": 6.485714285714286, + "grad_norm": 2.7210094920807273, + "learning_rate": 5.811445069402908e-06, + "loss": 0.0017, + "step": 4994 + }, + { + "epoch": 6.487012987012987, + "grad_norm": 3.3650853807299694, + "learning_rate": 5.807626009947127e-06, + "loss": 0.0024, + "step": 4995 + }, + { + "epoch": 6.488311688311688, + "grad_norm": 4.512301993071972, + "learning_rate": 5.803807692202547e-06, + "loss": 0.0191, + "step": 4996 + }, + { + "epoch": 6.489610389610389, + "grad_norm": 4.541495643302186, + "learning_rate": 5.799990116844699e-06, + "loss": 0.0167, + "step": 4997 + }, + { + "epoch": 6.490909090909091, + "grad_norm": 5.431022750554109, + "learning_rate": 5.796173284548994e-06, + "loss": 0.0136, + "step": 4998 + }, + { + "epoch": 6.492207792207792, + "grad_norm": 1.9167218309785878, + "learning_rate": 5.7923571959906945e-06, + "loss": 0.0018, + "step": 4999 + }, + { + "epoch": 6.4935064935064934, + "grad_norm": 3.778147269986554, + "learning_rate": 5.788541851844943e-06, + "loss": 0.0174, + "step": 5000 + }, + { + "epoch": 6.494805194805195, + "grad_norm": 6.115530893673077, + "learning_rate": 5.784727252786741e-06, + "loss": 0.0124, + "step": 5001 + }, + { + "epoch": 6.496103896103896, + "grad_norm": 10.053343954374116, + "learning_rate": 5.780913399490975e-06, + "loss": 0.0199, + "step": 5002 + }, + { + "epoch": 6.497402597402598, + "grad_norm": 5.92441479080623, + "learning_rate": 5.777100292632383e-06, + "loss": 0.0049, + "step": 5003 + }, + { + "epoch": 6.498701298701299, + "grad_norm": 6.890604723971934, + "learning_rate": 5.7732879328855765e-06, + "loss": 0.0226, + "step": 5004 + }, + { + "epoch": 6.5, + "grad_norm": 6.100615749981247, + "learning_rate": 5.7694763209250295e-06, + "loss": 0.0075, + "step": 5005 + }, + { + "epoch": 6.501298701298701, + "grad_norm": 4.725801088341112, + "learning_rate": 5.765665457425102e-06, + "loss": 0.0308, + "step": 5006 + }, + { + "epoch": 6.502597402597402, + "grad_norm": 8.678296483203543, + "learning_rate": 5.761855343059995e-06, + "loss": 0.0034, + "step": 5007 + }, + { + "epoch": 6.503896103896103, + "grad_norm": 4.941389157917713, + "learning_rate": 5.7580459785037986e-06, + "loss": 0.0219, + "step": 5008 + }, + { + "epoch": 6.505194805194805, + "grad_norm": 1.7358538525711615, + "learning_rate": 5.754237364430457e-06, + "loss": 0.0089, + "step": 5009 + }, + { + "epoch": 6.5064935064935066, + "grad_norm": 2.395825259970337, + "learning_rate": 5.750429501513799e-06, + "loss": 0.0019, + "step": 5010 + }, + { + "epoch": 6.507792207792208, + "grad_norm": 8.276434820127943, + "learning_rate": 5.746622390427492e-06, + "loss": 0.021, + "step": 5011 + }, + { + "epoch": 6.509090909090909, + "grad_norm": 4.931456237944652, + "learning_rate": 5.742816031845099e-06, + "loss": 0.015, + "step": 5012 + }, + { + "epoch": 6.510389610389611, + "grad_norm": 3.067441623661771, + "learning_rate": 5.7390104264400285e-06, + "loss": 0.0086, + "step": 5013 + }, + { + "epoch": 6.511688311688312, + "grad_norm": 7.020267177392165, + "learning_rate": 5.735205574885579e-06, + "loss": 0.0033, + "step": 5014 + }, + { + "epoch": 6.512987012987013, + "grad_norm": 1.5253095499196019, + "learning_rate": 5.731401477854883e-06, + "loss": 0.0147, + "step": 5015 + }, + { + "epoch": 6.514285714285714, + "grad_norm": 5.165940087355389, + "learning_rate": 5.727598136020971e-06, + "loss": 0.0095, + "step": 5016 + }, + { + "epoch": 6.515584415584415, + "grad_norm": 5.499178700734452, + "learning_rate": 5.723795550056721e-06, + "loss": 0.0131, + "step": 5017 + }, + { + "epoch": 6.5168831168831165, + "grad_norm": 1.0509119504823061, + "learning_rate": 5.719993720634892e-06, + "loss": 0.0095, + "step": 5018 + }, + { + "epoch": 6.5181818181818185, + "grad_norm": 7.788765763408397, + "learning_rate": 5.716192648428087e-06, + "loss": 0.0297, + "step": 5019 + }, + { + "epoch": 6.51948051948052, + "grad_norm": 5.695204647065127, + "learning_rate": 5.7123923341088e-06, + "loss": 0.022, + "step": 5020 + }, + { + "epoch": 6.520779220779221, + "grad_norm": 5.166921002029682, + "learning_rate": 5.708592778349374e-06, + "loss": 0.008, + "step": 5021 + }, + { + "epoch": 6.522077922077922, + "grad_norm": 6.491316527283493, + "learning_rate": 5.7047939818220185e-06, + "loss": 0.0192, + "step": 5022 + }, + { + "epoch": 6.523376623376623, + "grad_norm": 1.8020650297735268, + "learning_rate": 5.700995945198825e-06, + "loss": 0.008, + "step": 5023 + }, + { + "epoch": 6.524675324675325, + "grad_norm": 6.369969269107715, + "learning_rate": 5.697198669151731e-06, + "loss": 0.0079, + "step": 5024 + }, + { + "epoch": 6.525974025974026, + "grad_norm": 9.876987886327274, + "learning_rate": 5.69340215435255e-06, + "loss": 0.0064, + "step": 5025 + }, + { + "epoch": 6.527272727272727, + "grad_norm": 0.958141664812273, + "learning_rate": 5.689606401472954e-06, + "loss": 0.0039, + "step": 5026 + }, + { + "epoch": 6.5285714285714285, + "grad_norm": 5.8647817713243375, + "learning_rate": 5.6858114111844895e-06, + "loss": 0.0029, + "step": 5027 + }, + { + "epoch": 6.52987012987013, + "grad_norm": 3.7857492008745606, + "learning_rate": 5.682017184158562e-06, + "loss": 0.0043, + "step": 5028 + }, + { + "epoch": 6.531168831168831, + "grad_norm": 8.030178194786563, + "learning_rate": 5.6782237210664405e-06, + "loss": 0.0226, + "step": 5029 + }, + { + "epoch": 6.532467532467533, + "grad_norm": 5.193248932335492, + "learning_rate": 5.674431022579259e-06, + "loss": 0.0178, + "step": 5030 + }, + { + "epoch": 6.533766233766234, + "grad_norm": 6.162579617811557, + "learning_rate": 5.670639089368027e-06, + "loss": 0.0043, + "step": 5031 + }, + { + "epoch": 6.535064935064935, + "grad_norm": 1.4996083887859217, + "learning_rate": 5.666847922103604e-06, + "loss": 0.0048, + "step": 5032 + }, + { + "epoch": 6.536363636363636, + "grad_norm": 1.9361439303285684, + "learning_rate": 5.663057521456721e-06, + "loss": 0.0026, + "step": 5033 + }, + { + "epoch": 6.537662337662337, + "grad_norm": 3.3950833446006006, + "learning_rate": 5.6592678880979676e-06, + "loss": 0.0127, + "step": 5034 + }, + { + "epoch": 6.538961038961039, + "grad_norm": 4.366807779609675, + "learning_rate": 5.655479022697812e-06, + "loss": 0.0188, + "step": 5035 + }, + { + "epoch": 6.54025974025974, + "grad_norm": 3.1104140659791755, + "learning_rate": 5.6516909259265715e-06, + "loss": 0.0023, + "step": 5036 + }, + { + "epoch": 6.541558441558442, + "grad_norm": 5.028913846131432, + "learning_rate": 5.647903598454435e-06, + "loss": 0.0143, + "step": 5037 + }, + { + "epoch": 6.542857142857143, + "grad_norm": 6.822431943359536, + "learning_rate": 5.644117040951448e-06, + "loss": 0.0259, + "step": 5038 + }, + { + "epoch": 6.544155844155844, + "grad_norm": 6.523379656917949, + "learning_rate": 5.6403312540875325e-06, + "loss": 0.0095, + "step": 5039 + }, + { + "epoch": 6.545454545454545, + "grad_norm": 3.118467619173201, + "learning_rate": 5.636546238532463e-06, + "loss": 0.0027, + "step": 5040 + }, + { + "epoch": 6.546753246753247, + "grad_norm": 3.625030899547642, + "learning_rate": 5.6327619949558806e-06, + "loss": 0.0055, + "step": 5041 + }, + { + "epoch": 6.548051948051948, + "grad_norm": 1.8022157893026907, + "learning_rate": 5.628978524027289e-06, + "loss": 0.0094, + "step": 5042 + }, + { + "epoch": 6.549350649350649, + "grad_norm": 5.215209636496135, + "learning_rate": 5.625195826416061e-06, + "loss": 0.0065, + "step": 5043 + }, + { + "epoch": 6.55064935064935, + "grad_norm": 6.059321553636256, + "learning_rate": 5.621413902791426e-06, + "loss": 0.0138, + "step": 5044 + }, + { + "epoch": 6.551948051948052, + "grad_norm": 2.6173799390272143, + "learning_rate": 5.61763275382248e-06, + "loss": 0.0036, + "step": 5045 + }, + { + "epoch": 6.5532467532467535, + "grad_norm": 6.033509075015975, + "learning_rate": 5.6138523801781744e-06, + "loss": 0.004, + "step": 5046 + }, + { + "epoch": 6.554545454545455, + "grad_norm": 6.198978056385289, + "learning_rate": 5.610072782527343e-06, + "loss": 0.0103, + "step": 5047 + }, + { + "epoch": 6.555844155844156, + "grad_norm": 4.1772025170800555, + "learning_rate": 5.6062939615386516e-06, + "loss": 0.0142, + "step": 5048 + }, + { + "epoch": 6.557142857142857, + "grad_norm": 6.512805675405583, + "learning_rate": 5.602515917880659e-06, + "loss": 0.0057, + "step": 5049 + }, + { + "epoch": 6.558441558441558, + "grad_norm": 8.790584488795682, + "learning_rate": 5.5987386522217645e-06, + "loss": 0.0073, + "step": 5050 + }, + { + "epoch": 6.55974025974026, + "grad_norm": 3.697809204690563, + "learning_rate": 5.594962165230249e-06, + "loss": 0.0059, + "step": 5051 + }, + { + "epoch": 6.561038961038961, + "grad_norm": 8.607636181997957, + "learning_rate": 5.59118645757424e-06, + "loss": 0.0076, + "step": 5052 + }, + { + "epoch": 6.562337662337662, + "grad_norm": 8.847824621851672, + "learning_rate": 5.587411529921731e-06, + "loss": 0.0054, + "step": 5053 + }, + { + "epoch": 6.5636363636363635, + "grad_norm": 6.429395229704972, + "learning_rate": 5.583637382940576e-06, + "loss": 0.0088, + "step": 5054 + }, + { + "epoch": 6.564935064935065, + "grad_norm": 1.672086613059725, + "learning_rate": 5.5798640172985045e-06, + "loss": 0.0162, + "step": 5055 + }, + { + "epoch": 6.566233766233767, + "grad_norm": 5.550152045868406, + "learning_rate": 5.5760914336630875e-06, + "loss": 0.0159, + "step": 5056 + }, + { + "epoch": 6.567532467532468, + "grad_norm": 4.4178709294720395, + "learning_rate": 5.572319632701773e-06, + "loss": 0.011, + "step": 5057 + }, + { + "epoch": 6.568831168831169, + "grad_norm": 3.3947688254390074, + "learning_rate": 5.56854861508186e-06, + "loss": 0.0116, + "step": 5058 + }, + { + "epoch": 6.57012987012987, + "grad_norm": 1.722152783259435, + "learning_rate": 5.564778381470512e-06, + "loss": 0.014, + "step": 5059 + }, + { + "epoch": 6.571428571428571, + "grad_norm": 0.8849487162403634, + "learning_rate": 5.561008932534766e-06, + "loss": 0.0028, + "step": 5060 + }, + { + "epoch": 6.572727272727272, + "grad_norm": 3.714314262525366, + "learning_rate": 5.557240268941497e-06, + "loss": 0.0047, + "step": 5061 + }, + { + "epoch": 6.574025974025974, + "grad_norm": 0.8071462969378108, + "learning_rate": 5.553472391357468e-06, + "loss": 0.0091, + "step": 5062 + }, + { + "epoch": 6.5753246753246755, + "grad_norm": 3.320689322845907, + "learning_rate": 5.549705300449272e-06, + "loss": 0.0185, + "step": 5063 + }, + { + "epoch": 6.576623376623377, + "grad_norm": 5.816359162909769, + "learning_rate": 5.545938996883393e-06, + "loss": 0.0263, + "step": 5064 + }, + { + "epoch": 6.577922077922078, + "grad_norm": 2.3606950704849807, + "learning_rate": 5.542173481326152e-06, + "loss": 0.002, + "step": 5065 + }, + { + "epoch": 6.579220779220779, + "grad_norm": 7.96571628782228, + "learning_rate": 5.538408754443753e-06, + "loss": 0.0122, + "step": 5066 + }, + { + "epoch": 6.580519480519481, + "grad_norm": 3.0868669021020416, + "learning_rate": 5.534644816902235e-06, + "loss": 0.0157, + "step": 5067 + }, + { + "epoch": 6.581818181818182, + "grad_norm": 1.6968974255506413, + "learning_rate": 5.53088166936752e-06, + "loss": 0.0028, + "step": 5068 + }, + { + "epoch": 6.583116883116883, + "grad_norm": 1.0626826600461172, + "learning_rate": 5.5271193125053735e-06, + "loss": 0.0016, + "step": 5069 + }, + { + "epoch": 6.584415584415584, + "grad_norm": 4.213321961964961, + "learning_rate": 5.523357746981441e-06, + "loss": 0.0089, + "step": 5070 + }, + { + "epoch": 6.585714285714285, + "grad_norm": 5.4243259191923405, + "learning_rate": 5.519596973461199e-06, + "loss": 0.0213, + "step": 5071 + }, + { + "epoch": 6.5870129870129865, + "grad_norm": 7.322928665461058, + "learning_rate": 5.515836992610011e-06, + "loss": 0.01, + "step": 5072 + }, + { + "epoch": 6.588311688311689, + "grad_norm": 2.708868996640396, + "learning_rate": 5.512077805093088e-06, + "loss": 0.0024, + "step": 5073 + }, + { + "epoch": 6.58961038961039, + "grad_norm": 6.349493218083205, + "learning_rate": 5.5083194115755e-06, + "loss": 0.0081, + "step": 5074 + }, + { + "epoch": 6.590909090909091, + "grad_norm": 2.1933259352857792, + "learning_rate": 5.504561812722177e-06, + "loss": 0.0094, + "step": 5075 + }, + { + "epoch": 6.592207792207792, + "grad_norm": 5.538238643712317, + "learning_rate": 5.500805009197916e-06, + "loss": 0.0024, + "step": 5076 + }, + { + "epoch": 6.593506493506494, + "grad_norm": 4.7199454415624675, + "learning_rate": 5.497049001667365e-06, + "loss": 0.0173, + "step": 5077 + }, + { + "epoch": 6.594805194805195, + "grad_norm": 7.272551425726004, + "learning_rate": 5.493293790795032e-06, + "loss": 0.0033, + "step": 5078 + }, + { + "epoch": 6.596103896103896, + "grad_norm": 8.75006774782739, + "learning_rate": 5.4895393772452835e-06, + "loss": 0.0153, + "step": 5079 + }, + { + "epoch": 6.597402597402597, + "grad_norm": 1.2785191366621005, + "learning_rate": 5.485785761682354e-06, + "loss": 0.0093, + "step": 5080 + }, + { + "epoch": 6.5987012987012985, + "grad_norm": 9.195633057178565, + "learning_rate": 5.482032944770326e-06, + "loss": 0.0122, + "step": 5081 + }, + { + "epoch": 6.6, + "grad_norm": 5.572920750798249, + "learning_rate": 5.478280927173145e-06, + "loss": 0.0103, + "step": 5082 + }, + { + "epoch": 6.601298701298702, + "grad_norm": 3.892651001189466, + "learning_rate": 5.4745297095546125e-06, + "loss": 0.0107, + "step": 5083 + }, + { + "epoch": 6.602597402597403, + "grad_norm": 6.635473323061378, + "learning_rate": 5.470779292578394e-06, + "loss": 0.0088, + "step": 5084 + }, + { + "epoch": 6.603896103896104, + "grad_norm": 9.11339884643751, + "learning_rate": 5.46702967690801e-06, + "loss": 0.0153, + "step": 5085 + }, + { + "epoch": 6.605194805194805, + "grad_norm": 2.1734702825875845, + "learning_rate": 5.463280863206839e-06, + "loss": 0.0039, + "step": 5086 + }, + { + "epoch": 6.606493506493506, + "grad_norm": 7.413287830309794, + "learning_rate": 5.459532852138112e-06, + "loss": 0.0097, + "step": 5087 + }, + { + "epoch": 6.607792207792208, + "grad_norm": 3.033925692733609, + "learning_rate": 5.455785644364933e-06, + "loss": 0.0188, + "step": 5088 + }, + { + "epoch": 6.609090909090909, + "grad_norm": 5.202786690531281, + "learning_rate": 5.452039240550251e-06, + "loss": 0.0252, + "step": 5089 + }, + { + "epoch": 6.6103896103896105, + "grad_norm": 11.005922857512841, + "learning_rate": 5.448293641356872e-06, + "loss": 0.0161, + "step": 5090 + }, + { + "epoch": 6.611688311688312, + "grad_norm": 4.722289922484737, + "learning_rate": 5.444548847447471e-06, + "loss": 0.0104, + "step": 5091 + }, + { + "epoch": 6.612987012987013, + "grad_norm": 2.986294922587326, + "learning_rate": 5.4408048594845744e-06, + "loss": 0.0104, + "step": 5092 + }, + { + "epoch": 6.614285714285714, + "grad_norm": 7.888761901584748, + "learning_rate": 5.437061678130559e-06, + "loss": 0.0139, + "step": 5093 + }, + { + "epoch": 6.615584415584416, + "grad_norm": 4.7275339438345085, + "learning_rate": 5.433319304047666e-06, + "loss": 0.0109, + "step": 5094 + }, + { + "epoch": 6.616883116883117, + "grad_norm": 0.8849886527488284, + "learning_rate": 5.4295777378979994e-06, + "loss": 0.0023, + "step": 5095 + }, + { + "epoch": 6.618181818181818, + "grad_norm": 3.903577645838044, + "learning_rate": 5.425836980343509e-06, + "loss": 0.0046, + "step": 5096 + }, + { + "epoch": 6.619480519480519, + "grad_norm": 7.603845635588286, + "learning_rate": 5.422097032046008e-06, + "loss": 0.0041, + "step": 5097 + }, + { + "epoch": 6.62077922077922, + "grad_norm": 2.084292658789164, + "learning_rate": 5.418357893667159e-06, + "loss": 0.0095, + "step": 5098 + }, + { + "epoch": 6.6220779220779225, + "grad_norm": 6.282861420591763, + "learning_rate": 5.4146195658685005e-06, + "loss": 0.0094, + "step": 5099 + }, + { + "epoch": 6.623376623376624, + "grad_norm": 6.661465829017695, + "learning_rate": 5.410882049311399e-06, + "loss": 0.0077, + "step": 5100 + }, + { + "epoch": 6.624675324675325, + "grad_norm": 7.42796298476896, + "learning_rate": 5.407145344657102e-06, + "loss": 0.0131, + "step": 5101 + }, + { + "epoch": 6.625974025974026, + "grad_norm": 6.369925169454945, + "learning_rate": 5.4034094525666996e-06, + "loss": 0.0143, + "step": 5102 + }, + { + "epoch": 6.627272727272727, + "grad_norm": 6.297564679901956, + "learning_rate": 5.39967437370115e-06, + "loss": 0.02, + "step": 5103 + }, + { + "epoch": 6.628571428571428, + "grad_norm": 3.6604963663050056, + "learning_rate": 5.395940108721249e-06, + "loss": 0.023, + "step": 5104 + }, + { + "epoch": 6.62987012987013, + "grad_norm": 2.2901463365837613, + "learning_rate": 5.3922066582876685e-06, + "loss": 0.0097, + "step": 5105 + }, + { + "epoch": 6.631168831168831, + "grad_norm": 7.491861715047327, + "learning_rate": 5.388474023060921e-06, + "loss": 0.012, + "step": 5106 + }, + { + "epoch": 6.632467532467532, + "grad_norm": 7.615778862813955, + "learning_rate": 5.384742203701392e-06, + "loss": 0.0271, + "step": 5107 + }, + { + "epoch": 6.6337662337662335, + "grad_norm": 7.11662672589764, + "learning_rate": 5.381011200869296e-06, + "loss": 0.0024, + "step": 5108 + }, + { + "epoch": 6.6350649350649356, + "grad_norm": 5.365382477413365, + "learning_rate": 5.377281015224731e-06, + "loss": 0.0108, + "step": 5109 + }, + { + "epoch": 6.636363636363637, + "grad_norm": 9.022160966885545, + "learning_rate": 5.373551647427631e-06, + "loss": 0.0151, + "step": 5110 + }, + { + "epoch": 6.637662337662338, + "grad_norm": 6.055544695701037, + "learning_rate": 5.369823098137803e-06, + "loss": 0.0176, + "step": 5111 + }, + { + "epoch": 6.638961038961039, + "grad_norm": 3.770609408391407, + "learning_rate": 5.366095368014883e-06, + "loss": 0.0235, + "step": 5112 + }, + { + "epoch": 6.64025974025974, + "grad_norm": 7.336822677170732, + "learning_rate": 5.36236845771839e-06, + "loss": 0.0028, + "step": 5113 + }, + { + "epoch": 6.641558441558441, + "grad_norm": 5.535367705759921, + "learning_rate": 5.358642367907683e-06, + "loss": 0.0109, + "step": 5114 + }, + { + "epoch": 6.642857142857143, + "grad_norm": 2.616289207684074, + "learning_rate": 5.3549170992419775e-06, + "loss": 0.0044, + "step": 5115 + }, + { + "epoch": 6.644155844155844, + "grad_norm": 6.409244802392799, + "learning_rate": 5.351192652380341e-06, + "loss": 0.0047, + "step": 5116 + }, + { + "epoch": 6.6454545454545455, + "grad_norm": 8.786988979173652, + "learning_rate": 5.347469027981707e-06, + "loss": 0.0123, + "step": 5117 + }, + { + "epoch": 6.646753246753247, + "grad_norm": 1.0612855248099629, + "learning_rate": 5.343746226704853e-06, + "loss": 0.0098, + "step": 5118 + }, + { + "epoch": 6.648051948051948, + "grad_norm": 5.782409431187182, + "learning_rate": 5.340024249208413e-06, + "loss": 0.0042, + "step": 5119 + }, + { + "epoch": 6.64935064935065, + "grad_norm": 5.080862883183068, + "learning_rate": 5.336303096150872e-06, + "loss": 0.0087, + "step": 5120 + }, + { + "epoch": 6.650649350649351, + "grad_norm": 4.538310203032246, + "learning_rate": 5.332582768190582e-06, + "loss": 0.0099, + "step": 5121 + }, + { + "epoch": 6.651948051948052, + "grad_norm": 3.3054052857306098, + "learning_rate": 5.3288632659857346e-06, + "loss": 0.0026, + "step": 5122 + }, + { + "epoch": 6.653246753246753, + "grad_norm": 8.258996681576917, + "learning_rate": 5.325144590194377e-06, + "loss": 0.0091, + "step": 5123 + }, + { + "epoch": 6.654545454545454, + "grad_norm": 3.6851817349797193, + "learning_rate": 5.321426741474426e-06, + "loss": 0.0172, + "step": 5124 + }, + { + "epoch": 6.6558441558441555, + "grad_norm": 1.9858808013882276, + "learning_rate": 5.317709720483632e-06, + "loss": 0.0028, + "step": 5125 + }, + { + "epoch": 6.6571428571428575, + "grad_norm": 7.038350610516827, + "learning_rate": 5.313993527879608e-06, + "loss": 0.007, + "step": 5126 + }, + { + "epoch": 6.658441558441559, + "grad_norm": 2.7043481315048052, + "learning_rate": 5.310278164319816e-06, + "loss": 0.0178, + "step": 5127 + }, + { + "epoch": 6.65974025974026, + "grad_norm": 6.992256831947467, + "learning_rate": 5.306563630461583e-06, + "loss": 0.0122, + "step": 5128 + }, + { + "epoch": 6.661038961038961, + "grad_norm": 4.291889284602628, + "learning_rate": 5.302849926962077e-06, + "loss": 0.0188, + "step": 5129 + }, + { + "epoch": 6.662337662337662, + "grad_norm": 2.2223561104454452, + "learning_rate": 5.299137054478325e-06, + "loss": 0.007, + "step": 5130 + }, + { + "epoch": 6.663636363636364, + "grad_norm": 2.8756861264558773, + "learning_rate": 5.295425013667197e-06, + "loss": 0.0021, + "step": 5131 + }, + { + "epoch": 6.664935064935065, + "grad_norm": 7.728282795810656, + "learning_rate": 5.291713805185436e-06, + "loss": 0.0218, + "step": 5132 + }, + { + "epoch": 6.666233766233766, + "grad_norm": 4.423127598721648, + "learning_rate": 5.288003429689621e-06, + "loss": 0.0171, + "step": 5133 + }, + { + "epoch": 6.667532467532467, + "grad_norm": 6.625638819392787, + "learning_rate": 5.2842938878361874e-06, + "loss": 0.0074, + "step": 5134 + }, + { + "epoch": 6.6688311688311686, + "grad_norm": 2.8321140006663437, + "learning_rate": 5.280585180281419e-06, + "loss": 0.0033, + "step": 5135 + }, + { + "epoch": 6.67012987012987, + "grad_norm": 9.439692581313494, + "learning_rate": 5.276877307681469e-06, + "loss": 0.0084, + "step": 5136 + }, + { + "epoch": 6.671428571428572, + "grad_norm": 4.829266491676187, + "learning_rate": 5.273170270692323e-06, + "loss": 0.0153, + "step": 5137 + }, + { + "epoch": 6.672727272727273, + "grad_norm": 1.4139644274654668, + "learning_rate": 5.26946406996983e-06, + "loss": 0.0104, + "step": 5138 + }, + { + "epoch": 6.674025974025974, + "grad_norm": 5.2440701087563815, + "learning_rate": 5.265758706169681e-06, + "loss": 0.0106, + "step": 5139 + }, + { + "epoch": 6.675324675324675, + "grad_norm": 2.3331603128401626, + "learning_rate": 5.26205417994744e-06, + "loss": 0.0163, + "step": 5140 + }, + { + "epoch": 6.676623376623376, + "grad_norm": 3.684952096303915, + "learning_rate": 5.258350491958493e-06, + "loss": 0.0167, + "step": 5141 + }, + { + "epoch": 6.677922077922078, + "grad_norm": 1.0264429770334393, + "learning_rate": 5.254647642858103e-06, + "loss": 0.0094, + "step": 5142 + }, + { + "epoch": 6.679220779220779, + "grad_norm": 3.1263657205151762, + "learning_rate": 5.250945633301367e-06, + "loss": 0.0191, + "step": 5143 + }, + { + "epoch": 6.6805194805194805, + "grad_norm": 2.4659627992077913, + "learning_rate": 5.247244463943256e-06, + "loss": 0.002, + "step": 5144 + }, + { + "epoch": 6.681818181818182, + "grad_norm": 0.9116365009749015, + "learning_rate": 5.243544135438559e-06, + "loss": 0.0088, + "step": 5145 + }, + { + "epoch": 6.683116883116883, + "grad_norm": 3.3226935366661654, + "learning_rate": 5.23984464844195e-06, + "loss": 0.0065, + "step": 5146 + }, + { + "epoch": 6.684415584415584, + "grad_norm": 0.9509250543552694, + "learning_rate": 5.236146003607927e-06, + "loss": 0.002, + "step": 5147 + }, + { + "epoch": 6.685714285714286, + "grad_norm": 2.5020122743549953, + "learning_rate": 5.232448201590865e-06, + "loss": 0.008, + "step": 5148 + }, + { + "epoch": 6.687012987012987, + "grad_norm": 8.17645451904154, + "learning_rate": 5.228751243044961e-06, + "loss": 0.0232, + "step": 5149 + }, + { + "epoch": 6.688311688311688, + "grad_norm": 4.098113911113251, + "learning_rate": 5.225055128624288e-06, + "loss": 0.0041, + "step": 5150 + }, + { + "epoch": 6.689610389610389, + "grad_norm": 5.456911088744775, + "learning_rate": 5.221359858982753e-06, + "loss": 0.0044, + "step": 5151 + }, + { + "epoch": 6.690909090909091, + "grad_norm": 2.058401078256658, + "learning_rate": 5.217665434774126e-06, + "loss": 0.0061, + "step": 5152 + }, + { + "epoch": 6.6922077922077925, + "grad_norm": 1.518832194499104, + "learning_rate": 5.213971856652019e-06, + "loss": 0.0016, + "step": 5153 + }, + { + "epoch": 6.693506493506494, + "grad_norm": 2.7610393278049026, + "learning_rate": 5.210279125269897e-06, + "loss": 0.0066, + "step": 5154 + }, + { + "epoch": 6.694805194805195, + "grad_norm": 5.016199976447658, + "learning_rate": 5.206587241281073e-06, + "loss": 0.0104, + "step": 5155 + }, + { + "epoch": 6.696103896103896, + "grad_norm": 2.3047710501254173, + "learning_rate": 5.202896205338709e-06, + "loss": 0.003, + "step": 5156 + }, + { + "epoch": 6.697402597402597, + "grad_norm": 1.0722968425790216, + "learning_rate": 5.199206018095828e-06, + "loss": 0.002, + "step": 5157 + }, + { + "epoch": 6.698701298701299, + "grad_norm": 2.3781040625359005, + "learning_rate": 5.19551668020529e-06, + "loss": 0.0032, + "step": 5158 + }, + { + "epoch": 6.7, + "grad_norm": 1.253038624286543, + "learning_rate": 5.19182819231981e-06, + "loss": 0.0018, + "step": 5159 + }, + { + "epoch": 6.701298701298701, + "grad_norm": 6.183551101385431, + "learning_rate": 5.18814055509195e-06, + "loss": 0.0207, + "step": 5160 + }, + { + "epoch": 6.702597402597402, + "grad_norm": 8.110621939708087, + "learning_rate": 5.184453769174128e-06, + "loss": 0.0206, + "step": 5161 + }, + { + "epoch": 6.703896103896104, + "grad_norm": 2.569747077033477, + "learning_rate": 5.1807678352186005e-06, + "loss": 0.0082, + "step": 5162 + }, + { + "epoch": 6.705194805194806, + "grad_norm": 4.814861660547237, + "learning_rate": 5.1770827538774935e-06, + "loss": 0.0213, + "step": 5163 + }, + { + "epoch": 6.706493506493507, + "grad_norm": 5.636940127713876, + "learning_rate": 5.173398525802751e-06, + "loss": 0.0147, + "step": 5164 + }, + { + "epoch": 6.707792207792208, + "grad_norm": 6.875364927327006, + "learning_rate": 5.169715151646194e-06, + "loss": 0.0109, + "step": 5165 + }, + { + "epoch": 6.709090909090909, + "grad_norm": 5.329252819272799, + "learning_rate": 5.166032632059477e-06, + "loss": 0.0069, + "step": 5166 + }, + { + "epoch": 6.71038961038961, + "grad_norm": 5.836384826192981, + "learning_rate": 5.162350967694116e-06, + "loss": 0.0045, + "step": 5167 + }, + { + "epoch": 6.711688311688311, + "grad_norm": 5.299980494568935, + "learning_rate": 5.158670159201457e-06, + "loss": 0.0058, + "step": 5168 + }, + { + "epoch": 6.712987012987013, + "grad_norm": 2.6211446020046307, + "learning_rate": 5.154990207232713e-06, + "loss": 0.0064, + "step": 5169 + }, + { + "epoch": 6.714285714285714, + "grad_norm": 8.791107214324297, + "learning_rate": 5.151311112438937e-06, + "loss": 0.0082, + "step": 5170 + }, + { + "epoch": 6.7155844155844155, + "grad_norm": 6.70829592955983, + "learning_rate": 5.1476328754710295e-06, + "loss": 0.0077, + "step": 5171 + }, + { + "epoch": 6.716883116883117, + "grad_norm": 3.0921089732941174, + "learning_rate": 5.143955496979738e-06, + "loss": 0.0071, + "step": 5172 + }, + { + "epoch": 6.718181818181818, + "grad_norm": 11.33321119092505, + "learning_rate": 5.140278977615669e-06, + "loss": 0.0068, + "step": 5173 + }, + { + "epoch": 6.71948051948052, + "grad_norm": 7.131608668224781, + "learning_rate": 5.136603318029264e-06, + "loss": 0.0151, + "step": 5174 + }, + { + "epoch": 6.720779220779221, + "grad_norm": 6.8177202179875, + "learning_rate": 5.132928518870819e-06, + "loss": 0.0038, + "step": 5175 + }, + { + "epoch": 6.722077922077922, + "grad_norm": 9.33600866163115, + "learning_rate": 5.1292545807904725e-06, + "loss": 0.0047, + "step": 5176 + }, + { + "epoch": 6.723376623376623, + "grad_norm": 4.669800088059324, + "learning_rate": 5.12558150443822e-06, + "loss": 0.0023, + "step": 5177 + }, + { + "epoch": 6.724675324675324, + "grad_norm": 3.4781668056684567, + "learning_rate": 5.1219092904638976e-06, + "loss": 0.0028, + "step": 5178 + }, + { + "epoch": 6.7259740259740255, + "grad_norm": 7.272863079913027, + "learning_rate": 5.11823793951719e-06, + "loss": 0.0023, + "step": 5179 + }, + { + "epoch": 6.7272727272727275, + "grad_norm": 4.332458435846631, + "learning_rate": 5.114567452247624e-06, + "loss": 0.0086, + "step": 5180 + }, + { + "epoch": 6.728571428571429, + "grad_norm": 10.35744538307044, + "learning_rate": 5.1108978293045915e-06, + "loss": 0.0282, + "step": 5181 + }, + { + "epoch": 6.72987012987013, + "grad_norm": 9.431553637086362, + "learning_rate": 5.1072290713373056e-06, + "loss": 0.0099, + "step": 5182 + }, + { + "epoch": 6.731168831168831, + "grad_norm": 6.738908251348592, + "learning_rate": 5.103561178994847e-06, + "loss": 0.0081, + "step": 5183 + }, + { + "epoch": 6.732467532467533, + "grad_norm": 3.3725978405181865, + "learning_rate": 5.099894152926132e-06, + "loss": 0.0093, + "step": 5184 + }, + { + "epoch": 6.733766233766234, + "grad_norm": 5.820936854750343, + "learning_rate": 5.096227993779934e-06, + "loss": 0.0164, + "step": 5185 + }, + { + "epoch": 6.735064935064935, + "grad_norm": 9.437482410045806, + "learning_rate": 5.092562702204863e-06, + "loss": 0.0305, + "step": 5186 + }, + { + "epoch": 6.736363636363636, + "grad_norm": 8.418849631888792, + "learning_rate": 5.088898278849379e-06, + "loss": 0.0037, + "step": 5187 + }, + { + "epoch": 6.7376623376623375, + "grad_norm": 8.39773696462973, + "learning_rate": 5.085234724361786e-06, + "loss": 0.0202, + "step": 5188 + }, + { + "epoch": 6.738961038961039, + "grad_norm": 4.212371525033004, + "learning_rate": 5.081572039390244e-06, + "loss": 0.0024, + "step": 5189 + }, + { + "epoch": 6.740259740259741, + "grad_norm": 11.98500897070253, + "learning_rate": 5.077910224582748e-06, + "loss": 0.0147, + "step": 5190 + }, + { + "epoch": 6.741558441558442, + "grad_norm": 1.9298523991502663, + "learning_rate": 5.074249280587141e-06, + "loss": 0.002, + "step": 5191 + }, + { + "epoch": 6.742857142857143, + "grad_norm": 3.5981353349571776, + "learning_rate": 5.070589208051113e-06, + "loss": 0.0171, + "step": 5192 + }, + { + "epoch": 6.744155844155844, + "grad_norm": 5.840144784263722, + "learning_rate": 5.066930007622209e-06, + "loss": 0.0118, + "step": 5193 + }, + { + "epoch": 6.745454545454545, + "grad_norm": 2.4225673624689215, + "learning_rate": 5.063271679947807e-06, + "loss": 0.0111, + "step": 5194 + }, + { + "epoch": 6.746753246753247, + "grad_norm": 2.38175623891339, + "learning_rate": 5.059614225675129e-06, + "loss": 0.0105, + "step": 5195 + }, + { + "epoch": 6.748051948051948, + "grad_norm": 10.966264947111002, + "learning_rate": 5.0559576454512635e-06, + "loss": 0.0182, + "step": 5196 + }, + { + "epoch": 6.749350649350649, + "grad_norm": 3.980341894110683, + "learning_rate": 5.052301939923112e-06, + "loss": 0.0141, + "step": 5197 + }, + { + "epoch": 6.750649350649351, + "grad_norm": 1.9819350631449972, + "learning_rate": 5.048647109737452e-06, + "loss": 0.0024, + "step": 5198 + }, + { + "epoch": 6.751948051948052, + "grad_norm": 4.580976668559584, + "learning_rate": 5.0449931555408846e-06, + "loss": 0.0126, + "step": 5199 + }, + { + "epoch": 6.753246753246753, + "grad_norm": 2.2343208328753668, + "learning_rate": 5.041340077979875e-06, + "loss": 0.0047, + "step": 5200 + }, + { + "epoch": 6.754545454545455, + "grad_norm": 4.820938474137985, + "learning_rate": 5.037687877700708e-06, + "loss": 0.0121, + "step": 5201 + }, + { + "epoch": 6.755844155844156, + "grad_norm": 5.2044920986480525, + "learning_rate": 5.034036555349538e-06, + "loss": 0.0272, + "step": 5202 + }, + { + "epoch": 6.757142857142857, + "grad_norm": 4.909396629059066, + "learning_rate": 5.030386111572347e-06, + "loss": 0.02, + "step": 5203 + }, + { + "epoch": 6.758441558441558, + "grad_norm": 9.051293350630958, + "learning_rate": 5.026736547014981e-06, + "loss": 0.017, + "step": 5204 + }, + { + "epoch": 6.759740259740259, + "grad_norm": 4.75816919413958, + "learning_rate": 5.023087862323102e-06, + "loss": 0.0057, + "step": 5205 + }, + { + "epoch": 6.761038961038961, + "grad_norm": 4.47038572950464, + "learning_rate": 5.019440058142242e-06, + "loss": 0.0265, + "step": 5206 + }, + { + "epoch": 6.7623376623376625, + "grad_norm": 9.37012990841407, + "learning_rate": 5.0157931351177635e-06, + "loss": 0.0189, + "step": 5207 + }, + { + "epoch": 6.763636363636364, + "grad_norm": 2.7550321053082527, + "learning_rate": 5.0121470938948805e-06, + "loss": 0.0054, + "step": 5208 + }, + { + "epoch": 6.764935064935065, + "grad_norm": 5.288050852136056, + "learning_rate": 5.008501935118639e-06, + "loss": 0.0043, + "step": 5209 + }, + { + "epoch": 6.766233766233766, + "grad_norm": 4.564752465321014, + "learning_rate": 5.004857659433949e-06, + "loss": 0.0084, + "step": 5210 + }, + { + "epoch": 6.767532467532467, + "grad_norm": 5.331445577240815, + "learning_rate": 5.001214267485547e-06, + "loss": 0.0035, + "step": 5211 + }, + { + "epoch": 6.768831168831169, + "grad_norm": 3.557677428222944, + "learning_rate": 4.9975717599180185e-06, + "loss": 0.0052, + "step": 5212 + }, + { + "epoch": 6.77012987012987, + "grad_norm": 6.838329129404177, + "learning_rate": 4.99393013737579e-06, + "loss": 0.0112, + "step": 5213 + }, + { + "epoch": 6.771428571428571, + "grad_norm": 6.367767407185818, + "learning_rate": 4.99028940050314e-06, + "loss": 0.004, + "step": 5214 + }, + { + "epoch": 6.7727272727272725, + "grad_norm": 2.3395503366637205, + "learning_rate": 4.9866495499441824e-06, + "loss": 0.0253, + "step": 5215 + }, + { + "epoch": 6.7740259740259745, + "grad_norm": 8.475457654312669, + "learning_rate": 4.983010586342876e-06, + "loss": 0.0069, + "step": 5216 + }, + { + "epoch": 6.775324675324676, + "grad_norm": 3.0930255759741057, + "learning_rate": 4.979372510343019e-06, + "loss": 0.0032, + "step": 5217 + }, + { + "epoch": 6.776623376623377, + "grad_norm": 0.5619649309370416, + "learning_rate": 4.975735322588266e-06, + "loss": 0.0013, + "step": 5218 + }, + { + "epoch": 6.777922077922078, + "grad_norm": 2.9386314038026065, + "learning_rate": 4.972099023722098e-06, + "loss": 0.0203, + "step": 5219 + }, + { + "epoch": 6.779220779220779, + "grad_norm": 5.4493523722450075, + "learning_rate": 4.968463614387848e-06, + "loss": 0.0027, + "step": 5220 + }, + { + "epoch": 6.78051948051948, + "grad_norm": 5.58403663043162, + "learning_rate": 4.964829095228686e-06, + "loss": 0.0185, + "step": 5221 + }, + { + "epoch": 6.781818181818182, + "grad_norm": 8.049192084168343, + "learning_rate": 4.961195466887635e-06, + "loss": 0.0068, + "step": 5222 + }, + { + "epoch": 6.783116883116883, + "grad_norm": 4.548008380121048, + "learning_rate": 4.95756273000755e-06, + "loss": 0.0125, + "step": 5223 + }, + { + "epoch": 6.7844155844155845, + "grad_norm": 6.563781881144845, + "learning_rate": 4.953930885231128e-06, + "loss": 0.0044, + "step": 5224 + }, + { + "epoch": 6.785714285714286, + "grad_norm": 2.2710778328920123, + "learning_rate": 4.950299933200917e-06, + "loss": 0.0179, + "step": 5225 + }, + { + "epoch": 6.787012987012987, + "grad_norm": 1.9040640707837695, + "learning_rate": 4.946669874559301e-06, + "loss": 0.0018, + "step": 5226 + }, + { + "epoch": 6.788311688311689, + "grad_norm": 6.117301088465032, + "learning_rate": 4.943040709948507e-06, + "loss": 0.0052, + "step": 5227 + }, + { + "epoch": 6.78961038961039, + "grad_norm": 14.83057949591686, + "learning_rate": 4.939412440010598e-06, + "loss": 0.0047, + "step": 5228 + }, + { + "epoch": 6.790909090909091, + "grad_norm": 19.38002435858757, + "learning_rate": 4.935785065387493e-06, + "loss": 0.0223, + "step": 5229 + }, + { + "epoch": 6.792207792207792, + "grad_norm": 6.47027612285691, + "learning_rate": 4.93215858672094e-06, + "loss": 0.0039, + "step": 5230 + }, + { + "epoch": 6.793506493506493, + "grad_norm": 5.355756447243751, + "learning_rate": 4.928533004652532e-06, + "loss": 0.0146, + "step": 5231 + }, + { + "epoch": 6.794805194805194, + "grad_norm": 2.6298632935752293, + "learning_rate": 4.924908319823702e-06, + "loss": 0.007, + "step": 5232 + }, + { + "epoch": 6.796103896103896, + "grad_norm": 7.30621708155836, + "learning_rate": 4.9212845328757304e-06, + "loss": 0.0105, + "step": 5233 + }, + { + "epoch": 6.797402597402598, + "grad_norm": 4.810434322824059, + "learning_rate": 4.9176616444497336e-06, + "loss": 0.01, + "step": 5234 + }, + { + "epoch": 6.798701298701299, + "grad_norm": 4.669261294774748, + "learning_rate": 4.9140396551866685e-06, + "loss": 0.0296, + "step": 5235 + }, + { + "epoch": 6.8, + "grad_norm": 9.115187301658988, + "learning_rate": 4.910418565727332e-06, + "loss": 0.0056, + "step": 5236 + }, + { + "epoch": 6.801298701298701, + "grad_norm": 4.527344501706739, + "learning_rate": 4.9067983767123736e-06, + "loss": 0.0145, + "step": 5237 + }, + { + "epoch": 6.802597402597403, + "grad_norm": 2.4863423808573875, + "learning_rate": 4.90317908878226e-06, + "loss": 0.0043, + "step": 5238 + }, + { + "epoch": 6.803896103896104, + "grad_norm": 4.6877935312032974, + "learning_rate": 4.899560702577323e-06, + "loss": 0.0068, + "step": 5239 + }, + { + "epoch": 6.805194805194805, + "grad_norm": 1.8264014372535908, + "learning_rate": 4.895943218737717e-06, + "loss": 0.0015, + "step": 5240 + }, + { + "epoch": 6.806493506493506, + "grad_norm": 2.487803591771957, + "learning_rate": 4.892326637903457e-06, + "loss": 0.0035, + "step": 5241 + }, + { + "epoch": 6.8077922077922075, + "grad_norm": 10.570183538090335, + "learning_rate": 4.888710960714369e-06, + "loss": 0.0126, + "step": 5242 + }, + { + "epoch": 6.809090909090909, + "grad_norm": 2.3665455128023294, + "learning_rate": 4.885096187810148e-06, + "loss": 0.0021, + "step": 5243 + }, + { + "epoch": 6.810389610389611, + "grad_norm": 4.536969480006531, + "learning_rate": 4.881482319830306e-06, + "loss": 0.0121, + "step": 5244 + }, + { + "epoch": 6.811688311688312, + "grad_norm": 2.5372494664738445, + "learning_rate": 4.877869357414219e-06, + "loss": 0.004, + "step": 5245 + }, + { + "epoch": 6.812987012987013, + "grad_norm": 5.590248348717552, + "learning_rate": 4.874257301201074e-06, + "loss": 0.0122, + "step": 5246 + }, + { + "epoch": 6.814285714285714, + "grad_norm": 2.496760250606383, + "learning_rate": 4.870646151829924e-06, + "loss": 0.0147, + "step": 5247 + }, + { + "epoch": 6.815584415584416, + "grad_norm": 7.410246190890166, + "learning_rate": 4.8670359099396466e-06, + "loss": 0.0171, + "step": 5248 + }, + { + "epoch": 6.816883116883117, + "grad_norm": 4.936133745455025, + "learning_rate": 4.8634265761689625e-06, + "loss": 0.0161, + "step": 5249 + }, + { + "epoch": 6.818181818181818, + "grad_norm": 9.235054987920833, + "learning_rate": 4.859818151156429e-06, + "loss": 0.0135, + "step": 5250 + }, + { + "epoch": 6.8194805194805195, + "grad_norm": 6.770450313268609, + "learning_rate": 4.856210635540452e-06, + "loss": 0.0109, + "step": 5251 + }, + { + "epoch": 6.820779220779221, + "grad_norm": 6.198998788899638, + "learning_rate": 4.8526040299592665e-06, + "loss": 0.016, + "step": 5252 + }, + { + "epoch": 6.822077922077922, + "grad_norm": 3.708268830132102, + "learning_rate": 4.848998335050951e-06, + "loss": 0.0055, + "step": 5253 + }, + { + "epoch": 6.823376623376624, + "grad_norm": 4.472499766524509, + "learning_rate": 4.845393551453417e-06, + "loss": 0.0139, + "step": 5254 + }, + { + "epoch": 6.824675324675325, + "grad_norm": 5.8095184359825, + "learning_rate": 4.8417896798044275e-06, + "loss": 0.0073, + "step": 5255 + }, + { + "epoch": 6.825974025974026, + "grad_norm": 10.059789252872442, + "learning_rate": 4.838186720741574e-06, + "loss": 0.0275, + "step": 5256 + }, + { + "epoch": 6.827272727272727, + "grad_norm": 5.664123014184107, + "learning_rate": 4.834584674902282e-06, + "loss": 0.0137, + "step": 5257 + }, + { + "epoch": 6.828571428571428, + "grad_norm": 4.155658889869491, + "learning_rate": 4.830983542923832e-06, + "loss": 0.0143, + "step": 5258 + }, + { + "epoch": 6.82987012987013, + "grad_norm": 6.438698057819534, + "learning_rate": 4.827383325443331e-06, + "loss": 0.0034, + "step": 5259 + }, + { + "epoch": 6.8311688311688314, + "grad_norm": 3.950742555659145, + "learning_rate": 4.823784023097724e-06, + "loss": 0.0089, + "step": 5260 + }, + { + "epoch": 6.832467532467533, + "grad_norm": 2.4496217402550164, + "learning_rate": 4.820185636523794e-06, + "loss": 0.0029, + "step": 5261 + }, + { + "epoch": 6.833766233766234, + "grad_norm": 6.466838555929319, + "learning_rate": 4.816588166358171e-06, + "loss": 0.0298, + "step": 5262 + }, + { + "epoch": 6.835064935064935, + "grad_norm": 6.303738353144945, + "learning_rate": 4.812991613237312e-06, + "loss": 0.0161, + "step": 5263 + }, + { + "epoch": 6.836363636363636, + "grad_norm": 7.993836218370171, + "learning_rate": 4.809395977797517e-06, + "loss": 0.013, + "step": 5264 + }, + { + "epoch": 6.837662337662338, + "grad_norm": 4.9376535495204354, + "learning_rate": 4.805801260674921e-06, + "loss": 0.0021, + "step": 5265 + }, + { + "epoch": 6.838961038961039, + "grad_norm": 5.242427825303009, + "learning_rate": 4.802207462505501e-06, + "loss": 0.0086, + "step": 5266 + }, + { + "epoch": 6.84025974025974, + "grad_norm": 9.977918195662046, + "learning_rate": 4.798614583925069e-06, + "loss": 0.0151, + "step": 5267 + }, + { + "epoch": 6.841558441558441, + "grad_norm": 9.377232407408533, + "learning_rate": 4.795022625569273e-06, + "loss": 0.0266, + "step": 5268 + }, + { + "epoch": 6.8428571428571425, + "grad_norm": 4.818319750193602, + "learning_rate": 4.791431588073595e-06, + "loss": 0.0046, + "step": 5269 + }, + { + "epoch": 6.8441558441558445, + "grad_norm": 6.870200555326816, + "learning_rate": 4.787841472073366e-06, + "loss": 0.0211, + "step": 5270 + }, + { + "epoch": 6.845454545454546, + "grad_norm": 2.773285425176257, + "learning_rate": 4.784252278203743e-06, + "loss": 0.0281, + "step": 5271 + }, + { + "epoch": 6.846753246753247, + "grad_norm": 6.249046749789598, + "learning_rate": 4.780664007099721e-06, + "loss": 0.0178, + "step": 5272 + }, + { + "epoch": 6.848051948051948, + "grad_norm": 8.072103211590974, + "learning_rate": 4.7770766593961324e-06, + "loss": 0.0194, + "step": 5273 + }, + { + "epoch": 6.849350649350649, + "grad_norm": 4.712251202674872, + "learning_rate": 4.773490235727653e-06, + "loss": 0.018, + "step": 5274 + }, + { + "epoch": 6.85064935064935, + "grad_norm": 3.0833552834747664, + "learning_rate": 4.76990473672879e-06, + "loss": 0.0033, + "step": 5275 + }, + { + "epoch": 6.851948051948052, + "grad_norm": 10.567473624745531, + "learning_rate": 4.766320163033882e-06, + "loss": 0.0273, + "step": 5276 + }, + { + "epoch": 6.853246753246753, + "grad_norm": 1.068243210870819, + "learning_rate": 4.762736515277107e-06, + "loss": 0.0019, + "step": 5277 + }, + { + "epoch": 6.8545454545454545, + "grad_norm": 6.069284992133067, + "learning_rate": 4.7591537940924935e-06, + "loss": 0.0402, + "step": 5278 + }, + { + "epoch": 6.855844155844156, + "grad_norm": 4.9006348421963795, + "learning_rate": 4.7555720001138776e-06, + "loss": 0.0094, + "step": 5279 + }, + { + "epoch": 6.857142857142857, + "grad_norm": 2.927787315175227, + "learning_rate": 4.751991133974956e-06, + "loss": 0.0034, + "step": 5280 + }, + { + "epoch": 6.858441558441559, + "grad_norm": 7.446373398103612, + "learning_rate": 4.748411196309248e-06, + "loss": 0.0105, + "step": 5281 + }, + { + "epoch": 6.85974025974026, + "grad_norm": 4.157934712445688, + "learning_rate": 4.744832187750124e-06, + "loss": 0.0077, + "step": 5282 + }, + { + "epoch": 6.861038961038961, + "grad_norm": 3.2675766062654814, + "learning_rate": 4.741254108930763e-06, + "loss": 0.0247, + "step": 5283 + }, + { + "epoch": 6.862337662337662, + "grad_norm": 2.3008152099335217, + "learning_rate": 4.737676960484208e-06, + "loss": 0.0133, + "step": 5284 + }, + { + "epoch": 6.863636363636363, + "grad_norm": 0.6583453270055097, + "learning_rate": 4.734100743043315e-06, + "loss": 0.0081, + "step": 5285 + }, + { + "epoch": 6.8649350649350644, + "grad_norm": 2.628426134334478, + "learning_rate": 4.730525457240796e-06, + "loss": 0.0038, + "step": 5286 + }, + { + "epoch": 6.8662337662337665, + "grad_norm": 1.7785583039383108, + "learning_rate": 4.726951103709182e-06, + "loss": 0.0021, + "step": 5287 + }, + { + "epoch": 6.867532467532468, + "grad_norm": 3.529420031692527, + "learning_rate": 4.723377683080845e-06, + "loss": 0.0205, + "step": 5288 + }, + { + "epoch": 6.868831168831169, + "grad_norm": 1.3189226372585579, + "learning_rate": 4.719805195987992e-06, + "loss": 0.0026, + "step": 5289 + }, + { + "epoch": 6.87012987012987, + "grad_norm": 5.398151986386059, + "learning_rate": 4.71623364306266e-06, + "loss": 0.0262, + "step": 5290 + }, + { + "epoch": 6.871428571428572, + "grad_norm": 5.234827278950468, + "learning_rate": 4.712663024936733e-06, + "loss": 0.0144, + "step": 5291 + }, + { + "epoch": 6.872727272727273, + "grad_norm": 4.703276972946264, + "learning_rate": 4.709093342241917e-06, + "loss": 0.0034, + "step": 5292 + }, + { + "epoch": 6.874025974025974, + "grad_norm": 2.3180383020681847, + "learning_rate": 4.705524595609758e-06, + "loss": 0.0019, + "step": 5293 + }, + { + "epoch": 6.875324675324675, + "grad_norm": 0.6813984783632717, + "learning_rate": 4.7019567856716305e-06, + "loss": 0.0027, + "step": 5294 + }, + { + "epoch": 6.876623376623376, + "grad_norm": 9.607136528513347, + "learning_rate": 4.698389913058759e-06, + "loss": 0.0075, + "step": 5295 + }, + { + "epoch": 6.8779220779220775, + "grad_norm": 0.8980842051418538, + "learning_rate": 4.694823978402181e-06, + "loss": 0.0019, + "step": 5296 + }, + { + "epoch": 6.87922077922078, + "grad_norm": 4.864893655326643, + "learning_rate": 4.691258982332791e-06, + "loss": 0.0298, + "step": 5297 + }, + { + "epoch": 6.880519480519481, + "grad_norm": 5.701038333973104, + "learning_rate": 4.68769492548129e-06, + "loss": 0.0146, + "step": 5298 + }, + { + "epoch": 6.881818181818182, + "grad_norm": 8.128260667509199, + "learning_rate": 4.684131808478238e-06, + "loss": 0.009, + "step": 5299 + }, + { + "epoch": 6.883116883116883, + "grad_norm": 6.02807264210718, + "learning_rate": 4.680569631954014e-06, + "loss": 0.0111, + "step": 5300 + }, + { + "epoch": 6.884415584415584, + "grad_norm": 6.824931020863645, + "learning_rate": 4.677008396538843e-06, + "loss": 0.0296, + "step": 5301 + }, + { + "epoch": 6.885714285714286, + "grad_norm": 4.596205958816024, + "learning_rate": 4.673448102862763e-06, + "loss": 0.0026, + "step": 5302 + }, + { + "epoch": 6.887012987012987, + "grad_norm": 4.809125508657909, + "learning_rate": 4.669888751555669e-06, + "loss": 0.0124, + "step": 5303 + }, + { + "epoch": 6.888311688311688, + "grad_norm": 4.859081436284976, + "learning_rate": 4.666330343247273e-06, + "loss": 0.0127, + "step": 5304 + }, + { + "epoch": 6.8896103896103895, + "grad_norm": 9.820781890119303, + "learning_rate": 4.662772878567128e-06, + "loss": 0.0215, + "step": 5305 + }, + { + "epoch": 6.890909090909091, + "grad_norm": 12.437713333175044, + "learning_rate": 4.659216358144613e-06, + "loss": 0.0122, + "step": 5306 + }, + { + "epoch": 6.892207792207792, + "grad_norm": 4.024076173273686, + "learning_rate": 4.65566078260895e-06, + "loss": 0.008, + "step": 5307 + }, + { + "epoch": 6.893506493506494, + "grad_norm": 10.055937252454303, + "learning_rate": 4.652106152589188e-06, + "loss": 0.0162, + "step": 5308 + }, + { + "epoch": 6.894805194805195, + "grad_norm": 8.909914969327064, + "learning_rate": 4.648552468714206e-06, + "loss": 0.0059, + "step": 5309 + }, + { + "epoch": 6.896103896103896, + "grad_norm": 3.0392825751230603, + "learning_rate": 4.644999731612717e-06, + "loss": 0.004, + "step": 5310 + }, + { + "epoch": 6.897402597402597, + "grad_norm": 4.963776932957483, + "learning_rate": 4.6414479419132755e-06, + "loss": 0.0098, + "step": 5311 + }, + { + "epoch": 6.898701298701298, + "grad_norm": 5.73333393668147, + "learning_rate": 4.6378971002442565e-06, + "loss": 0.0024, + "step": 5312 + }, + { + "epoch": 6.9, + "grad_norm": 1.6833491438433021, + "learning_rate": 4.634347207233871e-06, + "loss": 0.0022, + "step": 5313 + }, + { + "epoch": 6.9012987012987015, + "grad_norm": 5.765280758135272, + "learning_rate": 4.630798263510162e-06, + "loss": 0.0035, + "step": 5314 + }, + { + "epoch": 6.902597402597403, + "grad_norm": 6.698008508441055, + "learning_rate": 4.627250269701012e-06, + "loss": 0.0049, + "step": 5315 + }, + { + "epoch": 6.903896103896104, + "grad_norm": 6.674547869450685, + "learning_rate": 4.623703226434123e-06, + "loss": 0.0139, + "step": 5316 + }, + { + "epoch": 6.905194805194805, + "grad_norm": 2.778049492526108, + "learning_rate": 4.620157134337038e-06, + "loss": 0.0026, + "step": 5317 + }, + { + "epoch": 6.906493506493506, + "grad_norm": 2.058867960813897, + "learning_rate": 4.616611994037122e-06, + "loss": 0.0176, + "step": 5318 + }, + { + "epoch": 6.907792207792208, + "grad_norm": 11.687197715271447, + "learning_rate": 4.613067806161589e-06, + "loss": 0.0349, + "step": 5319 + }, + { + "epoch": 6.909090909090909, + "grad_norm": 3.193133006895866, + "learning_rate": 4.609524571337467e-06, + "loss": 0.0168, + "step": 5320 + }, + { + "epoch": 6.91038961038961, + "grad_norm": 6.35246062092081, + "learning_rate": 4.605982290191623e-06, + "loss": 0.0152, + "step": 5321 + }, + { + "epoch": 6.911688311688311, + "grad_norm": 6.6002381986054806, + "learning_rate": 4.602440963350752e-06, + "loss": 0.0168, + "step": 5322 + }, + { + "epoch": 6.9129870129870135, + "grad_norm": 9.245017964271552, + "learning_rate": 4.5989005914413874e-06, + "loss": 0.0194, + "step": 5323 + }, + { + "epoch": 6.914285714285715, + "grad_norm": 7.423663647278888, + "learning_rate": 4.595361175089887e-06, + "loss": 0.0147, + "step": 5324 + }, + { + "epoch": 6.915584415584416, + "grad_norm": 1.4192716224026578, + "learning_rate": 4.591822714922441e-06, + "loss": 0.01, + "step": 5325 + }, + { + "epoch": 6.916883116883117, + "grad_norm": 1.329162594785774, + "learning_rate": 4.588285211565065e-06, + "loss": 0.0019, + "step": 5326 + }, + { + "epoch": 6.918181818181818, + "grad_norm": 3.070730612700662, + "learning_rate": 4.5847486656436215e-06, + "loss": 0.0049, + "step": 5327 + }, + { + "epoch": 6.919480519480519, + "grad_norm": 3.8586698689609347, + "learning_rate": 4.581213077783787e-06, + "loss": 0.0033, + "step": 5328 + }, + { + "epoch": 6.920779220779221, + "grad_norm": 1.5209818100118506, + "learning_rate": 4.577678448611072e-06, + "loss": 0.0101, + "step": 5329 + }, + { + "epoch": 6.922077922077922, + "grad_norm": 2.345775617821083, + "learning_rate": 4.574144778750833e-06, + "loss": 0.0169, + "step": 5330 + }, + { + "epoch": 6.923376623376623, + "grad_norm": 3.739768752094077, + "learning_rate": 4.5706120688282255e-06, + "loss": 0.0104, + "step": 5331 + }, + { + "epoch": 6.9246753246753245, + "grad_norm": 3.9629053003710943, + "learning_rate": 4.567080319468267e-06, + "loss": 0.0113, + "step": 5332 + }, + { + "epoch": 6.925974025974026, + "grad_norm": 6.08053769865782, + "learning_rate": 4.563549531295784e-06, + "loss": 0.0043, + "step": 5333 + }, + { + "epoch": 6.927272727272728, + "grad_norm": 2.304567420100074, + "learning_rate": 4.56001970493545e-06, + "loss": 0.0096, + "step": 5334 + }, + { + "epoch": 6.928571428571429, + "grad_norm": 4.777510975696846, + "learning_rate": 4.556490841011746e-06, + "loss": 0.0106, + "step": 5335 + }, + { + "epoch": 6.92987012987013, + "grad_norm": 5.281035431647329, + "learning_rate": 4.552962940149005e-06, + "loss": 0.0107, + "step": 5336 + }, + { + "epoch": 6.931168831168831, + "grad_norm": 2.709842322216882, + "learning_rate": 4.549436002971374e-06, + "loss": 0.004, + "step": 5337 + }, + { + "epoch": 6.932467532467532, + "grad_norm": 3.799270435323958, + "learning_rate": 4.545910030102846e-06, + "loss": 0.0173, + "step": 5338 + }, + { + "epoch": 6.933766233766233, + "grad_norm": 8.36104390820704, + "learning_rate": 4.5423850221672196e-06, + "loss": 0.011, + "step": 5339 + }, + { + "epoch": 6.935064935064935, + "grad_norm": 11.548393657920217, + "learning_rate": 4.538860979788145e-06, + "loss": 0.0347, + "step": 5340 + }, + { + "epoch": 6.9363636363636365, + "grad_norm": 11.207356382791556, + "learning_rate": 4.535337903589087e-06, + "loss": 0.0166, + "step": 5341 + }, + { + "epoch": 6.937662337662338, + "grad_norm": 14.136789822483848, + "learning_rate": 4.531815794193357e-06, + "loss": 0.0175, + "step": 5342 + }, + { + "epoch": 6.938961038961039, + "grad_norm": 4.1560505566899595, + "learning_rate": 4.528294652224068e-06, + "loss": 0.0199, + "step": 5343 + }, + { + "epoch": 6.94025974025974, + "grad_norm": 3.0718841560092707, + "learning_rate": 4.524774478304187e-06, + "loss": 0.0146, + "step": 5344 + }, + { + "epoch": 6.941558441558442, + "grad_norm": 4.54523485855393, + "learning_rate": 4.521255273056499e-06, + "loss": 0.0086, + "step": 5345 + }, + { + "epoch": 6.942857142857143, + "grad_norm": 1.014279210011106, + "learning_rate": 4.517737037103615e-06, + "loss": 0.0093, + "step": 5346 + }, + { + "epoch": 6.944155844155844, + "grad_norm": 3.8090867238250086, + "learning_rate": 4.51421977106798e-06, + "loss": 0.0122, + "step": 5347 + }, + { + "epoch": 6.945454545454545, + "grad_norm": 11.429239104648303, + "learning_rate": 4.510703475571868e-06, + "loss": 0.0387, + "step": 5348 + }, + { + "epoch": 6.9467532467532465, + "grad_norm": 4.179232922120617, + "learning_rate": 4.507188151237379e-06, + "loss": 0.0148, + "step": 5349 + }, + { + "epoch": 6.948051948051948, + "grad_norm": 5.72560165359425, + "learning_rate": 4.503673798686439e-06, + "loss": 0.0247, + "step": 5350 + }, + { + "epoch": 6.94935064935065, + "grad_norm": 3.189278573881644, + "learning_rate": 4.500160418540801e-06, + "loss": 0.0216, + "step": 5351 + }, + { + "epoch": 6.950649350649351, + "grad_norm": 6.3799320527983, + "learning_rate": 4.496648011422057e-06, + "loss": 0.0054, + "step": 5352 + }, + { + "epoch": 6.951948051948052, + "grad_norm": 5.851621996274449, + "learning_rate": 4.493136577951615e-06, + "loss": 0.015, + "step": 5353 + }, + { + "epoch": 6.953246753246753, + "grad_norm": 4.924678837200494, + "learning_rate": 4.489626118750714e-06, + "loss": 0.009, + "step": 5354 + }, + { + "epoch": 6.954545454545455, + "grad_norm": 7.644119385905598, + "learning_rate": 4.486116634440419e-06, + "loss": 0.0071, + "step": 5355 + }, + { + "epoch": 6.955844155844156, + "grad_norm": 8.290969384223548, + "learning_rate": 4.482608125641633e-06, + "loss": 0.0034, + "step": 5356 + }, + { + "epoch": 6.957142857142857, + "grad_norm": 7.2992962043215694, + "learning_rate": 4.4791005929750735e-06, + "loss": 0.0081, + "step": 5357 + }, + { + "epoch": 6.958441558441558, + "grad_norm": 10.037209052516019, + "learning_rate": 4.475594037061285e-06, + "loss": 0.0271, + "step": 5358 + }, + { + "epoch": 6.95974025974026, + "grad_norm": 5.538375632323246, + "learning_rate": 4.4720884585206536e-06, + "loss": 0.0022, + "step": 5359 + }, + { + "epoch": 6.961038961038961, + "grad_norm": 4.319831348866869, + "learning_rate": 4.46858385797338e-06, + "loss": 0.0102, + "step": 5360 + }, + { + "epoch": 6.962337662337663, + "grad_norm": 5.514051070865744, + "learning_rate": 4.465080236039494e-06, + "loss": 0.014, + "step": 5361 + }, + { + "epoch": 6.963636363636364, + "grad_norm": 6.608576268978067, + "learning_rate": 4.4615775933388506e-06, + "loss": 0.0181, + "step": 5362 + }, + { + "epoch": 6.964935064935065, + "grad_norm": 8.696497100345395, + "learning_rate": 4.4580759304911405e-06, + "loss": 0.0107, + "step": 5363 + }, + { + "epoch": 6.966233766233766, + "grad_norm": 3.001635740550289, + "learning_rate": 4.454575248115872e-06, + "loss": 0.009, + "step": 5364 + }, + { + "epoch": 6.967532467532467, + "grad_norm": 6.632649736617921, + "learning_rate": 4.451075546832383e-06, + "loss": 0.0029, + "step": 5365 + }, + { + "epoch": 6.968831168831169, + "grad_norm": 9.175864356229413, + "learning_rate": 4.447576827259833e-06, + "loss": 0.0072, + "step": 5366 + }, + { + "epoch": 6.97012987012987, + "grad_norm": 3.3255785120157184, + "learning_rate": 4.4440790900172216e-06, + "loss": 0.0074, + "step": 5367 + }, + { + "epoch": 6.9714285714285715, + "grad_norm": 3.0132950449805, + "learning_rate": 4.440582335723359e-06, + "loss": 0.0028, + "step": 5368 + }, + { + "epoch": 6.972727272727273, + "grad_norm": 7.415245847088746, + "learning_rate": 4.437086564996891e-06, + "loss": 0.0054, + "step": 5369 + }, + { + "epoch": 6.974025974025974, + "grad_norm": 5.575683511418774, + "learning_rate": 4.433591778456281e-06, + "loss": 0.0109, + "step": 5370 + }, + { + "epoch": 6.975324675324675, + "grad_norm": 3.013284368394481, + "learning_rate": 4.430097976719834e-06, + "loss": 0.0095, + "step": 5371 + }, + { + "epoch": 6.976623376623377, + "grad_norm": 7.8592197737922, + "learning_rate": 4.426605160405657e-06, + "loss": 0.0125, + "step": 5372 + }, + { + "epoch": 6.977922077922078, + "grad_norm": 4.12630203977792, + "learning_rate": 4.423113330131708e-06, + "loss": 0.0404, + "step": 5373 + }, + { + "epoch": 6.979220779220779, + "grad_norm": 4.800735640657764, + "learning_rate": 4.419622486515749e-06, + "loss": 0.0112, + "step": 5374 + }, + { + "epoch": 6.98051948051948, + "grad_norm": 13.4945893047663, + "learning_rate": 4.416132630175388e-06, + "loss": 0.014, + "step": 5375 + }, + { + "epoch": 6.9818181818181815, + "grad_norm": 3.424444122281062, + "learning_rate": 4.412643761728035e-06, + "loss": 0.0042, + "step": 5376 + }, + { + "epoch": 6.9831168831168835, + "grad_norm": 5.898626256902112, + "learning_rate": 4.409155881790947e-06, + "loss": 0.0049, + "step": 5377 + }, + { + "epoch": 6.984415584415585, + "grad_norm": 8.361721525842329, + "learning_rate": 4.40566899098119e-06, + "loss": 0.0148, + "step": 5378 + }, + { + "epoch": 6.985714285714286, + "grad_norm": 1.3829598707339754, + "learning_rate": 4.402183089915672e-06, + "loss": 0.0023, + "step": 5379 + }, + { + "epoch": 6.987012987012987, + "grad_norm": 1.1027589087300749, + "learning_rate": 4.398698179211103e-06, + "loss": 0.0016, + "step": 5380 + }, + { + "epoch": 6.988311688311688, + "grad_norm": 0.5729237210077697, + "learning_rate": 4.395214259484039e-06, + "loss": 0.0083, + "step": 5381 + }, + { + "epoch": 6.989610389610389, + "grad_norm": 2.452065361740632, + "learning_rate": 4.3917313313508445e-06, + "loss": 0.0138, + "step": 5382 + }, + { + "epoch": 6.990909090909091, + "grad_norm": 2.435188523943704, + "learning_rate": 4.388249395427728e-06, + "loss": 0.005, + "step": 5383 + }, + { + "epoch": 6.992207792207792, + "grad_norm": 7.423357824841378, + "learning_rate": 4.384768452330698e-06, + "loss": 0.0062, + "step": 5384 + }, + { + "epoch": 6.9935064935064934, + "grad_norm": 10.470253530672458, + "learning_rate": 4.381288502675608e-06, + "loss": 0.0149, + "step": 5385 + }, + { + "epoch": 6.994805194805195, + "grad_norm": 4.860392073348555, + "learning_rate": 4.377809547078126e-06, + "loss": 0.0112, + "step": 5386 + }, + { + "epoch": 6.996103896103897, + "grad_norm": 4.617467126859106, + "learning_rate": 4.374331586153744e-06, + "loss": 0.0168, + "step": 5387 + }, + { + "epoch": 6.997402597402598, + "grad_norm": 4.9497019566321905, + "learning_rate": 4.370854620517777e-06, + "loss": 0.0019, + "step": 5388 + }, + { + "epoch": 6.998701298701299, + "grad_norm": 4.069529020358419, + "learning_rate": 4.367378650785374e-06, + "loss": 0.0028, + "step": 5389 + }, + { + "epoch": 7.0, + "grad_norm": 3.492718413033117, + "learning_rate": 4.363903677571499e-06, + "loss": 0.009, + "step": 5390 + }, + { + "epoch": 7.0, + "eval_accuracy": 0.9517125210555868, + "eval_f1": 0.9433588108954174, + "eval_loss": 0.1364794373512268, + "eval_precision": 0.9372772896707586, + "eval_recall": 0.951553720002035, + "eval_runtime": 12.7289, + "eval_samples_per_second": 139.918, + "eval_steps_per_second": 1.1, + "step": 5390 + }, + { + "epoch": 7.001298701298701, + "grad_norm": 1.261334216096367, + "learning_rate": 4.360429701490935e-06, + "loss": 0.0019, + "step": 5391 + }, + { + "epoch": 7.002597402597402, + "grad_norm": 6.42409876835353, + "learning_rate": 4.356956723158302e-06, + "loss": 0.0293, + "step": 5392 + }, + { + "epoch": 7.003896103896104, + "grad_norm": 4.75243139696617, + "learning_rate": 4.353484743188035e-06, + "loss": 0.0076, + "step": 5393 + }, + { + "epoch": 7.005194805194805, + "grad_norm": 5.075854370315024, + "learning_rate": 4.350013762194392e-06, + "loss": 0.0052, + "step": 5394 + }, + { + "epoch": 7.0064935064935066, + "grad_norm": 6.063316233404848, + "learning_rate": 4.346543780791452e-06, + "loss": 0.0269, + "step": 5395 + }, + { + "epoch": 7.007792207792208, + "grad_norm": 10.44017111923626, + "learning_rate": 4.34307479959313e-06, + "loss": 0.0119, + "step": 5396 + }, + { + "epoch": 7.009090909090909, + "grad_norm": 2.5249312205515375, + "learning_rate": 4.33960681921315e-06, + "loss": 0.005, + "step": 5397 + }, + { + "epoch": 7.01038961038961, + "grad_norm": 5.522703989891028, + "learning_rate": 4.336139840265066e-06, + "loss": 0.0039, + "step": 5398 + }, + { + "epoch": 7.011688311688312, + "grad_norm": 5.731470468819291, + "learning_rate": 4.332673863362247e-06, + "loss": 0.0057, + "step": 5399 + }, + { + "epoch": 7.012987012987013, + "grad_norm": 5.362179428643194, + "learning_rate": 4.329208889117899e-06, + "loss": 0.0136, + "step": 5400 + }, + { + "epoch": 7.014285714285714, + "grad_norm": 3.358223389116216, + "learning_rate": 4.325744918145039e-06, + "loss": 0.0099, + "step": 5401 + }, + { + "epoch": 7.015584415584415, + "grad_norm": 6.439181862402725, + "learning_rate": 4.3222819510565085e-06, + "loss": 0.0122, + "step": 5402 + }, + { + "epoch": 7.0168831168831165, + "grad_norm": 3.8468475577763916, + "learning_rate": 4.31881998846497e-06, + "loss": 0.0031, + "step": 5403 + }, + { + "epoch": 7.0181818181818185, + "grad_norm": 5.911907476303494, + "learning_rate": 4.31535903098292e-06, + "loss": 0.0053, + "step": 5404 + }, + { + "epoch": 7.01948051948052, + "grad_norm": 3.914548568518747, + "learning_rate": 4.311899079222661e-06, + "loss": 0.0096, + "step": 5405 + }, + { + "epoch": 7.020779220779221, + "grad_norm": 4.465024745525089, + "learning_rate": 4.308440133796327e-06, + "loss": 0.009, + "step": 5406 + }, + { + "epoch": 7.022077922077922, + "grad_norm": 4.776819268395243, + "learning_rate": 4.304982195315866e-06, + "loss": 0.0027, + "step": 5407 + }, + { + "epoch": 7.023376623376623, + "grad_norm": 2.333536241322254, + "learning_rate": 4.301525264393063e-06, + "loss": 0.0064, + "step": 5408 + }, + { + "epoch": 7.024675324675325, + "grad_norm": 5.3046563691852135, + "learning_rate": 4.298069341639512e-06, + "loss": 0.002, + "step": 5409 + }, + { + "epoch": 7.025974025974026, + "grad_norm": 4.1664039319021935, + "learning_rate": 4.294614427666628e-06, + "loss": 0.009, + "step": 5410 + }, + { + "epoch": 7.027272727272727, + "grad_norm": 4.037268494348003, + "learning_rate": 4.2911605230856525e-06, + "loss": 0.0036, + "step": 5411 + }, + { + "epoch": 7.0285714285714285, + "grad_norm": 2.3281077412423925, + "learning_rate": 4.2877076285076546e-06, + "loss": 0.0097, + "step": 5412 + }, + { + "epoch": 7.02987012987013, + "grad_norm": 5.873640133056513, + "learning_rate": 4.284255744543505e-06, + "loss": 0.0183, + "step": 5413 + }, + { + "epoch": 7.031168831168831, + "grad_norm": 0.9738936290125758, + "learning_rate": 4.280804871803918e-06, + "loss": 0.0029, + "step": 5414 + }, + { + "epoch": 7.032467532467533, + "grad_norm": 1.7701107066783364, + "learning_rate": 4.277355010899413e-06, + "loss": 0.0028, + "step": 5415 + }, + { + "epoch": 7.033766233766234, + "grad_norm": 4.9295292334491405, + "learning_rate": 4.273906162440345e-06, + "loss": 0.0084, + "step": 5416 + }, + { + "epoch": 7.035064935064935, + "grad_norm": 3.3839990111391542, + "learning_rate": 4.270458327036869e-06, + "loss": 0.0033, + "step": 5417 + }, + { + "epoch": 7.036363636363636, + "grad_norm": 2.8117666801918713, + "learning_rate": 4.267011505298984e-06, + "loss": 0.0019, + "step": 5418 + }, + { + "epoch": 7.037662337662337, + "grad_norm": 1.6157792143607963, + "learning_rate": 4.2635656978364905e-06, + "loss": 0.002, + "step": 5419 + }, + { + "epoch": 7.038961038961039, + "grad_norm": 4.112876553884632, + "learning_rate": 4.260120905259024e-06, + "loss": 0.0031, + "step": 5420 + }, + { + "epoch": 7.04025974025974, + "grad_norm": 3.4554423649558457, + "learning_rate": 4.256677128176032e-06, + "loss": 0.0065, + "step": 5421 + }, + { + "epoch": 7.041558441558442, + "grad_norm": 4.935870574858402, + "learning_rate": 4.2532343671967844e-06, + "loss": 0.0116, + "step": 5422 + }, + { + "epoch": 7.042857142857143, + "grad_norm": 4.7977961305037775, + "learning_rate": 4.249792622930368e-06, + "loss": 0.0044, + "step": 5423 + }, + { + "epoch": 7.044155844155844, + "grad_norm": 0.7948062221928068, + "learning_rate": 4.246351895985702e-06, + "loss": 0.0088, + "step": 5424 + }, + { + "epoch": 7.045454545454546, + "grad_norm": 2.0750522778810048, + "learning_rate": 4.242912186971509e-06, + "loss": 0.0029, + "step": 5425 + }, + { + "epoch": 7.046753246753247, + "grad_norm": 1.6713828251707716, + "learning_rate": 4.239473496496345e-06, + "loss": 0.0016, + "step": 5426 + }, + { + "epoch": 7.048051948051948, + "grad_norm": 10.476556424633436, + "learning_rate": 4.236035825168575e-06, + "loss": 0.0195, + "step": 5427 + }, + { + "epoch": 7.049350649350649, + "grad_norm": 1.8174805441003525, + "learning_rate": 4.232599173596388e-06, + "loss": 0.008, + "step": 5428 + }, + { + "epoch": 7.05064935064935, + "grad_norm": 2.055134557607944, + "learning_rate": 4.229163542387799e-06, + "loss": 0.0033, + "step": 5429 + }, + { + "epoch": 7.0519480519480515, + "grad_norm": 2.454310782430854, + "learning_rate": 4.225728932150631e-06, + "loss": 0.0014, + "step": 5430 + }, + { + "epoch": 7.0532467532467535, + "grad_norm": 2.5241596905009964, + "learning_rate": 4.222295343492544e-06, + "loss": 0.0031, + "step": 5431 + }, + { + "epoch": 7.054545454545455, + "grad_norm": 5.88321841766342, + "learning_rate": 4.218862777020989e-06, + "loss": 0.0187, + "step": 5432 + }, + { + "epoch": 7.055844155844156, + "grad_norm": 7.092438131198643, + "learning_rate": 4.215431233343265e-06, + "loss": 0.0072, + "step": 5433 + }, + { + "epoch": 7.057142857142857, + "grad_norm": 4.253360168307547, + "learning_rate": 4.21200071306647e-06, + "loss": 0.0095, + "step": 5434 + }, + { + "epoch": 7.058441558441558, + "grad_norm": 0.7198875180152473, + "learning_rate": 4.208571216797539e-06, + "loss": 0.0015, + "step": 5435 + }, + { + "epoch": 7.05974025974026, + "grad_norm": 3.617087494260773, + "learning_rate": 4.205142745143203e-06, + "loss": 0.0021, + "step": 5436 + }, + { + "epoch": 7.061038961038961, + "grad_norm": 2.5089904115295547, + "learning_rate": 4.2017152987100315e-06, + "loss": 0.0026, + "step": 5437 + }, + { + "epoch": 7.062337662337662, + "grad_norm": 2.5003074936593084, + "learning_rate": 4.198288878104406e-06, + "loss": 0.0092, + "step": 5438 + }, + { + "epoch": 7.0636363636363635, + "grad_norm": 8.478460592149558, + "learning_rate": 4.194863483932523e-06, + "loss": 0.0129, + "step": 5439 + }, + { + "epoch": 7.064935064935065, + "grad_norm": 5.691869081970183, + "learning_rate": 4.191439116800397e-06, + "loss": 0.0034, + "step": 5440 + }, + { + "epoch": 7.066233766233767, + "grad_norm": 1.8251814712680487, + "learning_rate": 4.188015777313874e-06, + "loss": 0.0047, + "step": 5441 + }, + { + "epoch": 7.067532467532468, + "grad_norm": 1.9488031812566997, + "learning_rate": 4.184593466078602e-06, + "loss": 0.0027, + "step": 5442 + }, + { + "epoch": 7.068831168831169, + "grad_norm": 11.380471367273651, + "learning_rate": 4.181172183700052e-06, + "loss": 0.0112, + "step": 5443 + }, + { + "epoch": 7.07012987012987, + "grad_norm": 3.2481427347820824, + "learning_rate": 4.177751930783514e-06, + "loss": 0.0029, + "step": 5444 + }, + { + "epoch": 7.071428571428571, + "grad_norm": 6.513735643522785, + "learning_rate": 4.174332707934102e-06, + "loss": 0.0126, + "step": 5445 + }, + { + "epoch": 7.072727272727272, + "grad_norm": 1.9197827247757195, + "learning_rate": 4.170914515756738e-06, + "loss": 0.0036, + "step": 5446 + }, + { + "epoch": 7.074025974025974, + "grad_norm": 4.424504193701267, + "learning_rate": 4.1674973548561655e-06, + "loss": 0.0041, + "step": 5447 + }, + { + "epoch": 7.0753246753246755, + "grad_norm": 4.938888489092671, + "learning_rate": 4.164081225836944e-06, + "loss": 0.0105, + "step": 5448 + }, + { + "epoch": 7.076623376623377, + "grad_norm": 2.6858042307060987, + "learning_rate": 4.160666129303458e-06, + "loss": 0.0101, + "step": 5449 + }, + { + "epoch": 7.077922077922078, + "grad_norm": 4.084157714724248, + "learning_rate": 4.1572520658598994e-06, + "loss": 0.0019, + "step": 5450 + }, + { + "epoch": 7.079220779220779, + "grad_norm": 3.1562193398756797, + "learning_rate": 4.153839036110281e-06, + "loss": 0.0017, + "step": 5451 + }, + { + "epoch": 7.080519480519481, + "grad_norm": 6.672029218270617, + "learning_rate": 4.150427040658432e-06, + "loss": 0.0202, + "step": 5452 + }, + { + "epoch": 7.081818181818182, + "grad_norm": 3.2913323879707157, + "learning_rate": 4.147016080108003e-06, + "loss": 0.0095, + "step": 5453 + }, + { + "epoch": 7.083116883116883, + "grad_norm": 4.794242486405889, + "learning_rate": 4.143606155062458e-06, + "loss": 0.0399, + "step": 5454 + }, + { + "epoch": 7.084415584415584, + "grad_norm": 3.1076097438690313, + "learning_rate": 4.140197266125075e-06, + "loss": 0.0029, + "step": 5455 + }, + { + "epoch": 7.085714285714285, + "grad_norm": 3.2366223305528887, + "learning_rate": 4.136789413898951e-06, + "loss": 0.0021, + "step": 5456 + }, + { + "epoch": 7.087012987012987, + "grad_norm": 1.7913431523507475, + "learning_rate": 4.133382598987007e-06, + "loss": 0.0015, + "step": 5457 + }, + { + "epoch": 7.088311688311689, + "grad_norm": 0.9727169086699161, + "learning_rate": 4.129976821991969e-06, + "loss": 0.0093, + "step": 5458 + }, + { + "epoch": 7.08961038961039, + "grad_norm": 2.046713107082005, + "learning_rate": 4.126572083516384e-06, + "loss": 0.0026, + "step": 5459 + }, + { + "epoch": 7.090909090909091, + "grad_norm": 1.1346998457975264, + "learning_rate": 4.123168384162614e-06, + "loss": 0.0015, + "step": 5460 + }, + { + "epoch": 7.092207792207792, + "grad_norm": 0.8264745225596093, + "learning_rate": 4.119765724532843e-06, + "loss": 0.0042, + "step": 5461 + }, + { + "epoch": 7.093506493506493, + "grad_norm": 2.247994328668336, + "learning_rate": 4.116364105229065e-06, + "loss": 0.002, + "step": 5462 + }, + { + "epoch": 7.094805194805195, + "grad_norm": 2.738756486296691, + "learning_rate": 4.112963526853086e-06, + "loss": 0.0241, + "step": 5463 + }, + { + "epoch": 7.096103896103896, + "grad_norm": 4.857523800832278, + "learning_rate": 4.109563990006543e-06, + "loss": 0.0103, + "step": 5464 + }, + { + "epoch": 7.097402597402597, + "grad_norm": 4.978967942120852, + "learning_rate": 4.106165495290873e-06, + "loss": 0.006, + "step": 5465 + }, + { + "epoch": 7.0987012987012985, + "grad_norm": 1.1072163171869611, + "learning_rate": 4.102768043307337e-06, + "loss": 0.0084, + "step": 5466 + }, + { + "epoch": 7.1, + "grad_norm": 3.5719924472457305, + "learning_rate": 4.099371634657005e-06, + "loss": 0.0034, + "step": 5467 + }, + { + "epoch": 7.101298701298702, + "grad_norm": 1.4835937516795772, + "learning_rate": 4.095976269940777e-06, + "loss": 0.0028, + "step": 5468 + }, + { + "epoch": 7.102597402597403, + "grad_norm": 6.769399795269067, + "learning_rate": 4.092581949759343e-06, + "loss": 0.0173, + "step": 5469 + }, + { + "epoch": 7.103896103896104, + "grad_norm": 1.498128176664603, + "learning_rate": 4.0891886747132356e-06, + "loss": 0.002, + "step": 5470 + }, + { + "epoch": 7.105194805194805, + "grad_norm": 10.924761443193029, + "learning_rate": 4.085796445402782e-06, + "loss": 0.0125, + "step": 5471 + }, + { + "epoch": 7.106493506493506, + "grad_norm": 5.665893325565886, + "learning_rate": 4.082405262428144e-06, + "loss": 0.007, + "step": 5472 + }, + { + "epoch": 7.107792207792208, + "grad_norm": 2.972141238453353, + "learning_rate": 4.079015126389272e-06, + "loss": 0.0058, + "step": 5473 + }, + { + "epoch": 7.109090909090909, + "grad_norm": 1.5185392273789229, + "learning_rate": 4.075626037885956e-06, + "loss": 0.0087, + "step": 5474 + }, + { + "epoch": 7.1103896103896105, + "grad_norm": 0.8365853392957727, + "learning_rate": 4.072237997517784e-06, + "loss": 0.0015, + "step": 5475 + }, + { + "epoch": 7.111688311688312, + "grad_norm": 0.4689063429203158, + "learning_rate": 4.0688510058841755e-06, + "loss": 0.0015, + "step": 5476 + }, + { + "epoch": 7.112987012987013, + "grad_norm": 1.320252847267332, + "learning_rate": 4.065465063584342e-06, + "loss": 0.0036, + "step": 5477 + }, + { + "epoch": 7.114285714285714, + "grad_norm": 3.4060512167973833, + "learning_rate": 4.062080171217329e-06, + "loss": 0.012, + "step": 5478 + }, + { + "epoch": 7.115584415584416, + "grad_norm": 1.881638746561231, + "learning_rate": 4.058696329381987e-06, + "loss": 0.01, + "step": 5479 + }, + { + "epoch": 7.116883116883117, + "grad_norm": 1.744645128566324, + "learning_rate": 4.055313538676982e-06, + "loss": 0.0095, + "step": 5480 + }, + { + "epoch": 7.118181818181818, + "grad_norm": 3.9770242093664474, + "learning_rate": 4.051931799700791e-06, + "loss": 0.0055, + "step": 5481 + }, + { + "epoch": 7.119480519480519, + "grad_norm": 2.296782835822497, + "learning_rate": 4.048551113051716e-06, + "loss": 0.0012, + "step": 5482 + }, + { + "epoch": 7.12077922077922, + "grad_norm": 7.6435851700118205, + "learning_rate": 4.0451714793278604e-06, + "loss": 0.0147, + "step": 5483 + }, + { + "epoch": 7.1220779220779225, + "grad_norm": 6.71288195781477, + "learning_rate": 4.041792899127147e-06, + "loss": 0.0154, + "step": 5484 + }, + { + "epoch": 7.123376623376624, + "grad_norm": 5.348622676727847, + "learning_rate": 4.0384153730473076e-06, + "loss": 0.0176, + "step": 5485 + }, + { + "epoch": 7.124675324675325, + "grad_norm": 4.1294078916023365, + "learning_rate": 4.035038901685898e-06, + "loss": 0.0112, + "step": 5486 + }, + { + "epoch": 7.125974025974026, + "grad_norm": 5.325533919706218, + "learning_rate": 4.0316634856402784e-06, + "loss": 0.0094, + "step": 5487 + }, + { + "epoch": 7.127272727272727, + "grad_norm": 5.25891319887096, + "learning_rate": 4.028289125507622e-06, + "loss": 0.0038, + "step": 5488 + }, + { + "epoch": 7.128571428571428, + "grad_norm": 13.603810978675021, + "learning_rate": 4.024915821884916e-06, + "loss": 0.03, + "step": 5489 + }, + { + "epoch": 7.12987012987013, + "grad_norm": 5.126200465225334, + "learning_rate": 4.021543575368968e-06, + "loss": 0.0021, + "step": 5490 + }, + { + "epoch": 7.131168831168831, + "grad_norm": 10.929873129139848, + "learning_rate": 4.018172386556392e-06, + "loss": 0.0193, + "step": 5491 + }, + { + "epoch": 7.132467532467532, + "grad_norm": 2.3487666900531377, + "learning_rate": 4.014802256043609e-06, + "loss": 0.0128, + "step": 5492 + }, + { + "epoch": 7.1337662337662335, + "grad_norm": 4.612706752391176, + "learning_rate": 4.011433184426869e-06, + "loss": 0.0079, + "step": 5493 + }, + { + "epoch": 7.135064935064935, + "grad_norm": 5.6242269971925225, + "learning_rate": 4.008065172302222e-06, + "loss": 0.0033, + "step": 5494 + }, + { + "epoch": 7.136363636363637, + "grad_norm": 4.790329001897355, + "learning_rate": 4.004698220265533e-06, + "loss": 0.0346, + "step": 5495 + }, + { + "epoch": 7.137662337662338, + "grad_norm": 6.721984633943841, + "learning_rate": 4.001332328912475e-06, + "loss": 0.0178, + "step": 5496 + }, + { + "epoch": 7.138961038961039, + "grad_norm": 9.473612389962753, + "learning_rate": 3.997967498838548e-06, + "loss": 0.0039, + "step": 5497 + }, + { + "epoch": 7.14025974025974, + "grad_norm": 5.37035427022655, + "learning_rate": 3.99460373063905e-06, + "loss": 0.0082, + "step": 5498 + }, + { + "epoch": 7.141558441558441, + "grad_norm": 7.642891131608647, + "learning_rate": 3.991241024909098e-06, + "loss": 0.0215, + "step": 5499 + }, + { + "epoch": 7.142857142857143, + "grad_norm": 4.606246283004425, + "learning_rate": 3.987879382243614e-06, + "loss": 0.0095, + "step": 5500 + }, + { + "epoch": 7.144155844155844, + "grad_norm": 3.7548207505296536, + "learning_rate": 3.984518803237343e-06, + "loss": 0.0052, + "step": 5501 + }, + { + "epoch": 7.1454545454545455, + "grad_norm": 3.372583497015311, + "learning_rate": 3.981159288484834e-06, + "loss": 0.0116, + "step": 5502 + }, + { + "epoch": 7.146753246753247, + "grad_norm": 5.677371341799693, + "learning_rate": 3.977800838580448e-06, + "loss": 0.0045, + "step": 5503 + }, + { + "epoch": 7.148051948051948, + "grad_norm": 1.62358373941733, + "learning_rate": 3.974443454118357e-06, + "loss": 0.0096, + "step": 5504 + }, + { + "epoch": 7.14935064935065, + "grad_norm": 3.5956798699966126, + "learning_rate": 3.9710871356925515e-06, + "loss": 0.0041, + "step": 5505 + }, + { + "epoch": 7.150649350649351, + "grad_norm": 16.983851345072946, + "learning_rate": 3.967731883896827e-06, + "loss": 0.0384, + "step": 5506 + }, + { + "epoch": 7.151948051948052, + "grad_norm": 4.045223122667803, + "learning_rate": 3.964377699324792e-06, + "loss": 0.0137, + "step": 5507 + }, + { + "epoch": 7.153246753246753, + "grad_norm": 3.1767690318004593, + "learning_rate": 3.96102458256986e-06, + "loss": 0.0022, + "step": 5508 + }, + { + "epoch": 7.154545454545454, + "grad_norm": 3.3559010218342022, + "learning_rate": 3.957672534225274e-06, + "loss": 0.0031, + "step": 5509 + }, + { + "epoch": 7.1558441558441555, + "grad_norm": 2.8273189186574563, + "learning_rate": 3.954321554884062e-06, + "loss": 0.0273, + "step": 5510 + }, + { + "epoch": 7.1571428571428575, + "grad_norm": 5.761892003384957, + "learning_rate": 3.950971645139086e-06, + "loss": 0.0042, + "step": 5511 + }, + { + "epoch": 7.158441558441559, + "grad_norm": 4.830243065448483, + "learning_rate": 3.9476228055830015e-06, + "loss": 0.0225, + "step": 5512 + }, + { + "epoch": 7.15974025974026, + "grad_norm": 5.446185683669611, + "learning_rate": 3.944275036808292e-06, + "loss": 0.0063, + "step": 5513 + }, + { + "epoch": 7.161038961038961, + "grad_norm": 4.871269581840201, + "learning_rate": 3.940928339407232e-06, + "loss": 0.0062, + "step": 5514 + }, + { + "epoch": 7.162337662337662, + "grad_norm": 1.5153172767373215, + "learning_rate": 3.9375827139719225e-06, + "loss": 0.0077, + "step": 5515 + }, + { + "epoch": 7.163636363636364, + "grad_norm": 5.370417619423953, + "learning_rate": 3.934238161094264e-06, + "loss": 0.0105, + "step": 5516 + }, + { + "epoch": 7.164935064935065, + "grad_norm": 3.155122924910463, + "learning_rate": 3.930894681365981e-06, + "loss": 0.0054, + "step": 5517 + }, + { + "epoch": 7.166233766233766, + "grad_norm": 2.482731213823952, + "learning_rate": 3.9275522753785865e-06, + "loss": 0.002, + "step": 5518 + }, + { + "epoch": 7.167532467532467, + "grad_norm": 5.169457371999814, + "learning_rate": 3.9242109437234255e-06, + "loss": 0.0053, + "step": 5519 + }, + { + "epoch": 7.1688311688311686, + "grad_norm": 3.6512176874123194, + "learning_rate": 3.92087068699164e-06, + "loss": 0.0134, + "step": 5520 + }, + { + "epoch": 7.17012987012987, + "grad_norm": 3.1610037656821492, + "learning_rate": 3.917531505774183e-06, + "loss": 0.0029, + "step": 5521 + }, + { + "epoch": 7.171428571428572, + "grad_norm": 1.356463332432368, + "learning_rate": 3.914193400661825e-06, + "loss": 0.0021, + "step": 5522 + }, + { + "epoch": 7.172727272727273, + "grad_norm": 3.074777028731449, + "learning_rate": 3.910856372245139e-06, + "loss": 0.0097, + "step": 5523 + }, + { + "epoch": 7.174025974025974, + "grad_norm": 3.8175211036864614, + "learning_rate": 3.9075204211145076e-06, + "loss": 0.0118, + "step": 5524 + }, + { + "epoch": 7.175324675324675, + "grad_norm": 4.0976330765683935, + "learning_rate": 3.904185547860122e-06, + "loss": 0.0056, + "step": 5525 + }, + { + "epoch": 7.176623376623376, + "grad_norm": 3.0154892991060556, + "learning_rate": 3.900851753071991e-06, + "loss": 0.0098, + "step": 5526 + }, + { + "epoch": 7.177922077922078, + "grad_norm": 1.0874786908532907, + "learning_rate": 3.897519037339924e-06, + "loss": 0.0019, + "step": 5527 + }, + { + "epoch": 7.179220779220779, + "grad_norm": 4.730782729509525, + "learning_rate": 3.894187401253542e-06, + "loss": 0.0043, + "step": 5528 + }, + { + "epoch": 7.1805194805194805, + "grad_norm": 3.9834381536844394, + "learning_rate": 3.890856845402272e-06, + "loss": 0.0097, + "step": 5529 + }, + { + "epoch": 7.181818181818182, + "grad_norm": 3.7798264601318796, + "learning_rate": 3.887527370375359e-06, + "loss": 0.0098, + "step": 5530 + }, + { + "epoch": 7.183116883116883, + "grad_norm": 5.765647908323951, + "learning_rate": 3.884198976761846e-06, + "loss": 0.0046, + "step": 5531 + }, + { + "epoch": 7.184415584415585, + "grad_norm": 4.036556219721578, + "learning_rate": 3.880871665150597e-06, + "loss": 0.006, + "step": 5532 + }, + { + "epoch": 7.185714285714286, + "grad_norm": 1.7239454994215488, + "learning_rate": 3.877545436130267e-06, + "loss": 0.0111, + "step": 5533 + }, + { + "epoch": 7.187012987012987, + "grad_norm": 4.231529662745346, + "learning_rate": 3.874220290289337e-06, + "loss": 0.0145, + "step": 5534 + }, + { + "epoch": 7.188311688311688, + "grad_norm": 2.7374296261084297, + "learning_rate": 3.870896228216086e-06, + "loss": 0.0036, + "step": 5535 + }, + { + "epoch": 7.189610389610389, + "grad_norm": 1.0706996101001776, + "learning_rate": 3.8675732504986044e-06, + "loss": 0.0122, + "step": 5536 + }, + { + "epoch": 7.190909090909091, + "grad_norm": 1.7036640958703975, + "learning_rate": 3.864251357724787e-06, + "loss": 0.0014, + "step": 5537 + }, + { + "epoch": 7.1922077922077925, + "grad_norm": 6.79299974248829, + "learning_rate": 3.860930550482347e-06, + "loss": 0.0062, + "step": 5538 + }, + { + "epoch": 7.193506493506494, + "grad_norm": 6.04685464699474, + "learning_rate": 3.857610829358797e-06, + "loss": 0.012, + "step": 5539 + }, + { + "epoch": 7.194805194805195, + "grad_norm": 0.6786833211429415, + "learning_rate": 3.854292194941457e-06, + "loss": 0.0092, + "step": 5540 + }, + { + "epoch": 7.196103896103896, + "grad_norm": 4.7851432056524175, + "learning_rate": 3.850974647817454e-06, + "loss": 0.0041, + "step": 5541 + }, + { + "epoch": 7.197402597402597, + "grad_norm": 3.4361624128698556, + "learning_rate": 3.847658188573733e-06, + "loss": 0.0018, + "step": 5542 + }, + { + "epoch": 7.198701298701299, + "grad_norm": 2.151791499258602, + "learning_rate": 3.844342817797036e-06, + "loss": 0.009, + "step": 5543 + }, + { + "epoch": 7.2, + "grad_norm": 7.377893011624007, + "learning_rate": 3.841028536073915e-06, + "loss": 0.0078, + "step": 5544 + }, + { + "epoch": 7.201298701298701, + "grad_norm": 4.120048848861622, + "learning_rate": 3.837715343990727e-06, + "loss": 0.0093, + "step": 5545 + }, + { + "epoch": 7.202597402597402, + "grad_norm": 10.770707723803547, + "learning_rate": 3.834403242133649e-06, + "loss": 0.0296, + "step": 5546 + }, + { + "epoch": 7.203896103896104, + "grad_norm": 1.4664224304268116, + "learning_rate": 3.83109223108864e-06, + "loss": 0.0014, + "step": 5547 + }, + { + "epoch": 7.205194805194806, + "grad_norm": 9.112136831851588, + "learning_rate": 3.827782311441494e-06, + "loss": 0.0149, + "step": 5548 + }, + { + "epoch": 7.206493506493507, + "grad_norm": 3.356291311404889, + "learning_rate": 3.8244734837777906e-06, + "loss": 0.0105, + "step": 5549 + }, + { + "epoch": 7.207792207792208, + "grad_norm": 3.0118725629880374, + "learning_rate": 3.821165748682935e-06, + "loss": 0.0079, + "step": 5550 + }, + { + "epoch": 7.209090909090909, + "grad_norm": 2.888751057636795, + "learning_rate": 3.817859106742118e-06, + "loss": 0.0147, + "step": 5551 + }, + { + "epoch": 7.21038961038961, + "grad_norm": 4.441469432712057, + "learning_rate": 3.814553558540354e-06, + "loss": 0.0101, + "step": 5552 + }, + { + "epoch": 7.211688311688311, + "grad_norm": 3.93725614210113, + "learning_rate": 3.8112491046624533e-06, + "loss": 0.0062, + "step": 5553 + }, + { + "epoch": 7.212987012987013, + "grad_norm": 0.9591231483345327, + "learning_rate": 3.8079457456930434e-06, + "loss": 0.0018, + "step": 5554 + }, + { + "epoch": 7.214285714285714, + "grad_norm": 2.3988115152690663, + "learning_rate": 3.8046434822165477e-06, + "loss": 0.0044, + "step": 5555 + }, + { + "epoch": 7.2155844155844155, + "grad_norm": 3.796772679035855, + "learning_rate": 3.801342314817201e-06, + "loss": 0.0172, + "step": 5556 + }, + { + "epoch": 7.216883116883117, + "grad_norm": 1.6681024320117412, + "learning_rate": 3.7980422440790386e-06, + "loss": 0.0116, + "step": 5557 + }, + { + "epoch": 7.218181818181818, + "grad_norm": 3.7675747697879554, + "learning_rate": 3.7947432705859146e-06, + "loss": 0.0026, + "step": 5558 + }, + { + "epoch": 7.21948051948052, + "grad_norm": 1.1045042368989482, + "learning_rate": 3.791445394921475e-06, + "loss": 0.0162, + "step": 5559 + }, + { + "epoch": 7.220779220779221, + "grad_norm": 0.9085010519042913, + "learning_rate": 3.7881486176691775e-06, + "loss": 0.0086, + "step": 5560 + }, + { + "epoch": 7.222077922077922, + "grad_norm": 11.151568839373763, + "learning_rate": 3.7848529394122857e-06, + "loss": 0.0169, + "step": 5561 + }, + { + "epoch": 7.223376623376623, + "grad_norm": 2.4772143847287955, + "learning_rate": 3.7815583607338656e-06, + "loss": 0.0028, + "step": 5562 + }, + { + "epoch": 7.224675324675324, + "grad_norm": 2.6871567337572464, + "learning_rate": 3.7782648822167966e-06, + "loss": 0.0058, + "step": 5563 + }, + { + "epoch": 7.225974025974026, + "grad_norm": 3.494194592531304, + "learning_rate": 3.7749725044437513e-06, + "loss": 0.0014, + "step": 5564 + }, + { + "epoch": 7.2272727272727275, + "grad_norm": 4.077195885590876, + "learning_rate": 3.771681227997226e-06, + "loss": 0.0245, + "step": 5565 + }, + { + "epoch": 7.228571428571429, + "grad_norm": 2.9636449333546793, + "learning_rate": 3.7683910534594957e-06, + "loss": 0.0014, + "step": 5566 + }, + { + "epoch": 7.22987012987013, + "grad_norm": 9.16655213942422, + "learning_rate": 3.7651019814126656e-06, + "loss": 0.01, + "step": 5567 + }, + { + "epoch": 7.231168831168831, + "grad_norm": 2.814645459705159, + "learning_rate": 3.7618140124386294e-06, + "loss": 0.0041, + "step": 5568 + }, + { + "epoch": 7.232467532467532, + "grad_norm": 3.0222127919379207, + "learning_rate": 3.7585271471191e-06, + "loss": 0.0086, + "step": 5569 + }, + { + "epoch": 7.233766233766234, + "grad_norm": 1.7748403579937329, + "learning_rate": 3.755241386035575e-06, + "loss": 0.0095, + "step": 5570 + }, + { + "epoch": 7.235064935064935, + "grad_norm": 3.5409183985927974, + "learning_rate": 3.7519567297693794e-06, + "loss": 0.0024, + "step": 5571 + }, + { + "epoch": 7.236363636363636, + "grad_norm": 1.3170241040805606, + "learning_rate": 3.748673178901623e-06, + "loss": 0.0016, + "step": 5572 + }, + { + "epoch": 7.2376623376623375, + "grad_norm": 2.218067909949003, + "learning_rate": 3.745390734013239e-06, + "loss": 0.0025, + "step": 5573 + }, + { + "epoch": 7.238961038961039, + "grad_norm": 4.420150859863516, + "learning_rate": 3.7421093956849418e-06, + "loss": 0.0032, + "step": 5574 + }, + { + "epoch": 7.240259740259741, + "grad_norm": 3.4761070338585873, + "learning_rate": 3.738829164497274e-06, + "loss": 0.0081, + "step": 5575 + }, + { + "epoch": 7.241558441558442, + "grad_norm": 6.982206320574588, + "learning_rate": 3.7355500410305667e-06, + "loss": 0.0108, + "step": 5576 + }, + { + "epoch": 7.242857142857143, + "grad_norm": 1.779689853145024, + "learning_rate": 3.7322720258649603e-06, + "loss": 0.0024, + "step": 5577 + }, + { + "epoch": 7.244155844155844, + "grad_norm": 0.79310317854462, + "learning_rate": 3.7289951195803954e-06, + "loss": 0.0013, + "step": 5578 + }, + { + "epoch": 7.245454545454545, + "grad_norm": 1.2364819526552706, + "learning_rate": 3.7257193227566246e-06, + "loss": 0.0039, + "step": 5579 + }, + { + "epoch": 7.246753246753247, + "grad_norm": 6.020201585034088, + "learning_rate": 3.722444635973196e-06, + "loss": 0.0075, + "step": 5580 + }, + { + "epoch": 7.248051948051948, + "grad_norm": 2.027786424018582, + "learning_rate": 3.7191710598094667e-06, + "loss": 0.0027, + "step": 5581 + }, + { + "epoch": 7.249350649350649, + "grad_norm": 1.9275244042107547, + "learning_rate": 3.715898594844588e-06, + "loss": 0.0018, + "step": 5582 + }, + { + "epoch": 7.250649350649351, + "grad_norm": 3.7103837600699516, + "learning_rate": 3.7126272416575316e-06, + "loss": 0.0062, + "step": 5583 + }, + { + "epoch": 7.251948051948052, + "grad_norm": 1.9626840387267226, + "learning_rate": 3.7093570008270573e-06, + "loss": 0.0021, + "step": 5584 + }, + { + "epoch": 7.253246753246753, + "grad_norm": 0.8596110433444789, + "learning_rate": 3.706087872931734e-06, + "loss": 0.0012, + "step": 5585 + }, + { + "epoch": 7.254545454545455, + "grad_norm": 0.4617049932318796, + "learning_rate": 3.7028198585499287e-06, + "loss": 0.0014, + "step": 5586 + }, + { + "epoch": 7.255844155844156, + "grad_norm": 4.887006493423827, + "learning_rate": 3.6995529582598235e-06, + "loss": 0.0131, + "step": 5587 + }, + { + "epoch": 7.257142857142857, + "grad_norm": 3.564340808544459, + "learning_rate": 3.6962871726393923e-06, + "loss": 0.0049, + "step": 5588 + }, + { + "epoch": 7.258441558441558, + "grad_norm": 2.2734316752071417, + "learning_rate": 3.6930225022664136e-06, + "loss": 0.0024, + "step": 5589 + }, + { + "epoch": 7.259740259740259, + "grad_norm": 3.9408266255692053, + "learning_rate": 3.6897589477184682e-06, + "loss": 0.002, + "step": 5590 + }, + { + "epoch": 7.261038961038961, + "grad_norm": 1.0667553741488458, + "learning_rate": 3.6864965095729476e-06, + "loss": 0.0013, + "step": 5591 + }, + { + "epoch": 7.2623376623376625, + "grad_norm": 8.160104445135069, + "learning_rate": 3.683235188407036e-06, + "loss": 0.0157, + "step": 5592 + }, + { + "epoch": 7.263636363636364, + "grad_norm": 2.206715287647699, + "learning_rate": 3.6799749847977206e-06, + "loss": 0.0031, + "step": 5593 + }, + { + "epoch": 7.264935064935065, + "grad_norm": 0.6882106186011232, + "learning_rate": 3.676715899321801e-06, + "loss": 0.0014, + "step": 5594 + }, + { + "epoch": 7.266233766233766, + "grad_norm": 1.124384120052089, + "learning_rate": 3.6734579325558674e-06, + "loss": 0.0031, + "step": 5595 + }, + { + "epoch": 7.267532467532467, + "grad_norm": 1.5062885156858101, + "learning_rate": 3.670201085076316e-06, + "loss": 0.0018, + "step": 5596 + }, + { + "epoch": 7.268831168831169, + "grad_norm": 3.469113996832555, + "learning_rate": 3.666945357459344e-06, + "loss": 0.005, + "step": 5597 + }, + { + "epoch": 7.27012987012987, + "grad_norm": 2.4044971304193448, + "learning_rate": 3.6636907502809573e-06, + "loss": 0.0017, + "step": 5598 + }, + { + "epoch": 7.271428571428571, + "grad_norm": 2.762689124313535, + "learning_rate": 3.660437264116955e-06, + "loss": 0.0092, + "step": 5599 + }, + { + "epoch": 7.2727272727272725, + "grad_norm": 1.1654466643101038, + "learning_rate": 3.6571848995429404e-06, + "loss": 0.0092, + "step": 5600 + }, + { + "epoch": 7.274025974025974, + "grad_norm": 7.884535202132486, + "learning_rate": 3.6539336571343177e-06, + "loss": 0.0084, + "step": 5601 + }, + { + "epoch": 7.275324675324676, + "grad_norm": 1.3719450705233034, + "learning_rate": 3.6506835374663017e-06, + "loss": 0.0017, + "step": 5602 + }, + { + "epoch": 7.276623376623377, + "grad_norm": 2.323889385325196, + "learning_rate": 3.647434541113889e-06, + "loss": 0.0198, + "step": 5603 + }, + { + "epoch": 7.277922077922078, + "grad_norm": 2.0227359361858746, + "learning_rate": 3.644186668651899e-06, + "loss": 0.0094, + "step": 5604 + }, + { + "epoch": 7.279220779220779, + "grad_norm": 1.7904217678530157, + "learning_rate": 3.6409399206549357e-06, + "loss": 0.0019, + "step": 5605 + }, + { + "epoch": 7.28051948051948, + "grad_norm": 6.220669521182732, + "learning_rate": 3.6376942976974205e-06, + "loss": 0.0284, + "step": 5606 + }, + { + "epoch": 7.281818181818182, + "grad_norm": 5.149602216740382, + "learning_rate": 3.634449800353552e-06, + "loss": 0.0117, + "step": 5607 + }, + { + "epoch": 7.283116883116883, + "grad_norm": 1.8444404065623934, + "learning_rate": 3.631206429197357e-06, + "loss": 0.0035, + "step": 5608 + }, + { + "epoch": 7.2844155844155845, + "grad_norm": 1.4194355005269128, + "learning_rate": 3.6279641848026403e-06, + "loss": 0.0023, + "step": 5609 + }, + { + "epoch": 7.285714285714286, + "grad_norm": 2.8687198876402134, + "learning_rate": 3.624723067743029e-06, + "loss": 0.01, + "step": 5610 + }, + { + "epoch": 7.287012987012987, + "grad_norm": 5.286717399817067, + "learning_rate": 3.621483078591923e-06, + "loss": 0.0126, + "step": 5611 + }, + { + "epoch": 7.288311688311689, + "grad_norm": 1.5943096028075603, + "learning_rate": 3.6182442179225507e-06, + "loss": 0.0033, + "step": 5612 + }, + { + "epoch": 7.28961038961039, + "grad_norm": 2.0500166824289927, + "learning_rate": 3.6150064863079203e-06, + "loss": 0.0016, + "step": 5613 + }, + { + "epoch": 7.290909090909091, + "grad_norm": 3.312889136507532, + "learning_rate": 3.6117698843208602e-06, + "loss": 0.0095, + "step": 5614 + }, + { + "epoch": 7.292207792207792, + "grad_norm": 10.391103850328774, + "learning_rate": 3.6085344125339727e-06, + "loss": 0.0285, + "step": 5615 + }, + { + "epoch": 7.293506493506493, + "grad_norm": 4.232821567499917, + "learning_rate": 3.605300071519685e-06, + "loss": 0.0145, + "step": 5616 + }, + { + "epoch": 7.294805194805194, + "grad_norm": 4.246808791446434, + "learning_rate": 3.60206686185021e-06, + "loss": 0.0079, + "step": 5617 + }, + { + "epoch": 7.296103896103896, + "grad_norm": 1.8300363165770241, + "learning_rate": 3.5988347840975656e-06, + "loss": 0.0027, + "step": 5618 + }, + { + "epoch": 7.297402597402598, + "grad_norm": 3.453187087682131, + "learning_rate": 3.5956038388335635e-06, + "loss": 0.0037, + "step": 5619 + }, + { + "epoch": 7.298701298701299, + "grad_norm": 1.00962269975956, + "learning_rate": 3.5923740266298267e-06, + "loss": 0.0099, + "step": 5620 + }, + { + "epoch": 7.3, + "grad_norm": 5.731589861992576, + "learning_rate": 3.5891453480577687e-06, + "loss": 0.0147, + "step": 5621 + }, + { + "epoch": 7.301298701298701, + "grad_norm": 0.8980048719465971, + "learning_rate": 3.585917803688603e-06, + "loss": 0.0085, + "step": 5622 + }, + { + "epoch": 7.302597402597403, + "grad_norm": 4.610473769877987, + "learning_rate": 3.582691394093343e-06, + "loss": 0.0102, + "step": 5623 + }, + { + "epoch": 7.303896103896104, + "grad_norm": 2.2056001507334644, + "learning_rate": 3.579466119842806e-06, + "loss": 0.0023, + "step": 5624 + }, + { + "epoch": 7.305194805194805, + "grad_norm": 2.655205446680385, + "learning_rate": 3.576241981507602e-06, + "loss": 0.0077, + "step": 5625 + }, + { + "epoch": 7.306493506493506, + "grad_norm": 2.738884527578916, + "learning_rate": 3.5730189796581415e-06, + "loss": 0.0022, + "step": 5626 + }, + { + "epoch": 7.3077922077922075, + "grad_norm": 1.555186976680867, + "learning_rate": 3.5697971148646402e-06, + "loss": 0.0022, + "step": 5627 + }, + { + "epoch": 7.309090909090909, + "grad_norm": 1.6937906122985618, + "learning_rate": 3.5665763876971037e-06, + "loss": 0.0019, + "step": 5628 + }, + { + "epoch": 7.310389610389611, + "grad_norm": 1.6599463476760998, + "learning_rate": 3.5633567987253425e-06, + "loss": 0.0036, + "step": 5629 + }, + { + "epoch": 7.311688311688312, + "grad_norm": 3.78677915756161, + "learning_rate": 3.560138348518959e-06, + "loss": 0.0054, + "step": 5630 + }, + { + "epoch": 7.312987012987013, + "grad_norm": 5.58216326510532, + "learning_rate": 3.556921037647364e-06, + "loss": 0.0141, + "step": 5631 + }, + { + "epoch": 7.314285714285714, + "grad_norm": 3.039225896504836, + "learning_rate": 3.55370486667976e-06, + "loss": 0.0018, + "step": 5632 + }, + { + "epoch": 7.315584415584415, + "grad_norm": 3.94913344904662, + "learning_rate": 3.5504898361851492e-06, + "loss": 0.0213, + "step": 5633 + }, + { + "epoch": 7.316883116883117, + "grad_norm": 1.3400942329309142, + "learning_rate": 3.547275946732327e-06, + "loss": 0.0082, + "step": 5634 + }, + { + "epoch": 7.318181818181818, + "grad_norm": 3.8408154513151143, + "learning_rate": 3.5440631988899e-06, + "loss": 0.0127, + "step": 5635 + }, + { + "epoch": 7.3194805194805195, + "grad_norm": 1.0809644635926587, + "learning_rate": 3.540851593226261e-06, + "loss": 0.0021, + "step": 5636 + }, + { + "epoch": 7.320779220779221, + "grad_norm": 2.475284318878767, + "learning_rate": 3.5376411303096047e-06, + "loss": 0.0091, + "step": 5637 + }, + { + "epoch": 7.322077922077922, + "grad_norm": 0.7719754082213225, + "learning_rate": 3.53443181070792e-06, + "loss": 0.0013, + "step": 5638 + }, + { + "epoch": 7.323376623376624, + "grad_norm": 1.4093338945351745, + "learning_rate": 3.531223634989004e-06, + "loss": 0.0012, + "step": 5639 + }, + { + "epoch": 7.324675324675325, + "grad_norm": 2.6015388791664145, + "learning_rate": 3.5280166037204398e-06, + "loss": 0.0022, + "step": 5640 + }, + { + "epoch": 7.325974025974026, + "grad_norm": 1.296109600493248, + "learning_rate": 3.524810717469613e-06, + "loss": 0.0022, + "step": 5641 + }, + { + "epoch": 7.327272727272727, + "grad_norm": 1.3229949176984375, + "learning_rate": 3.5216059768037046e-06, + "loss": 0.0111, + "step": 5642 + }, + { + "epoch": 7.328571428571428, + "grad_norm": 2.986937967780711, + "learning_rate": 3.518402382289704e-06, + "loss": 0.0049, + "step": 5643 + }, + { + "epoch": 7.32987012987013, + "grad_norm": 3.8173349481706498, + "learning_rate": 3.515199934494373e-06, + "loss": 0.0308, + "step": 5644 + }, + { + "epoch": 7.3311688311688314, + "grad_norm": 8.639370475638193, + "learning_rate": 3.511998633984297e-06, + "loss": 0.0216, + "step": 5645 + }, + { + "epoch": 7.332467532467533, + "grad_norm": 4.104894953180211, + "learning_rate": 3.5087984813258426e-06, + "loss": 0.0024, + "step": 5646 + }, + { + "epoch": 7.333766233766234, + "grad_norm": 3.7261823840325574, + "learning_rate": 3.5055994770851853e-06, + "loss": 0.0022, + "step": 5647 + }, + { + "epoch": 7.335064935064935, + "grad_norm": 1.2462220582552823, + "learning_rate": 3.5024016218282788e-06, + "loss": 0.0101, + "step": 5648 + }, + { + "epoch": 7.336363636363636, + "grad_norm": 9.789736434468455, + "learning_rate": 3.499204916120893e-06, + "loss": 0.0206, + "step": 5649 + }, + { + "epoch": 7.337662337662338, + "grad_norm": 5.448380209834105, + "learning_rate": 3.4960093605285817e-06, + "loss": 0.0163, + "step": 5650 + }, + { + "epoch": 7.338961038961039, + "grad_norm": 4.972821509493743, + "learning_rate": 3.4928149556167077e-06, + "loss": 0.0083, + "step": 5651 + }, + { + "epoch": 7.34025974025974, + "grad_norm": 5.543591357026569, + "learning_rate": 3.4896217019504106e-06, + "loss": 0.0177, + "step": 5652 + }, + { + "epoch": 7.341558441558441, + "grad_norm": 3.058521024142026, + "learning_rate": 3.4864296000946483e-06, + "loss": 0.0022, + "step": 5653 + }, + { + "epoch": 7.3428571428571425, + "grad_norm": 7.236347138555347, + "learning_rate": 3.483238650614158e-06, + "loss": 0.0191, + "step": 5654 + }, + { + "epoch": 7.3441558441558445, + "grad_norm": 1.8545398994379143, + "learning_rate": 3.4800488540734857e-06, + "loss": 0.009, + "step": 5655 + }, + { + "epoch": 7.345454545454546, + "grad_norm": 2.92131445422538, + "learning_rate": 3.476860211036963e-06, + "loss": 0.003, + "step": 5656 + }, + { + "epoch": 7.346753246753247, + "grad_norm": 6.362747376359274, + "learning_rate": 3.473672722068724e-06, + "loss": 0.0202, + "step": 5657 + }, + { + "epoch": 7.348051948051948, + "grad_norm": 0.5238308361800672, + "learning_rate": 3.4704863877326957e-06, + "loss": 0.0015, + "step": 5658 + }, + { + "epoch": 7.349350649350649, + "grad_norm": 3.5888122324966365, + "learning_rate": 3.467301208592598e-06, + "loss": 0.0068, + "step": 5659 + }, + { + "epoch": 7.35064935064935, + "grad_norm": 1.963936767091282, + "learning_rate": 3.464117185211958e-06, + "loss": 0.0147, + "step": 5660 + }, + { + "epoch": 7.351948051948052, + "grad_norm": 5.187551910892981, + "learning_rate": 3.4609343181540854e-06, + "loss": 0.0088, + "step": 5661 + }, + { + "epoch": 7.353246753246753, + "grad_norm": 3.541379712726672, + "learning_rate": 3.457752607982091e-06, + "loss": 0.0014, + "step": 5662 + }, + { + "epoch": 7.3545454545454545, + "grad_norm": 2.996361479781739, + "learning_rate": 3.454572055258877e-06, + "loss": 0.0015, + "step": 5663 + }, + { + "epoch": 7.355844155844156, + "grad_norm": 2.0100150569291557, + "learning_rate": 3.4513926605471504e-06, + "loss": 0.0055, + "step": 5664 + }, + { + "epoch": 7.357142857142857, + "grad_norm": 12.566440238545065, + "learning_rate": 3.4482144244093995e-06, + "loss": 0.0257, + "step": 5665 + }, + { + "epoch": 7.358441558441559, + "grad_norm": 4.717610535860202, + "learning_rate": 3.4450373474079278e-06, + "loss": 0.0213, + "step": 5666 + }, + { + "epoch": 7.35974025974026, + "grad_norm": 5.342847127190773, + "learning_rate": 3.4418614301048047e-06, + "loss": 0.009, + "step": 5667 + }, + { + "epoch": 7.361038961038961, + "grad_norm": 2.438069737948192, + "learning_rate": 3.438686673061923e-06, + "loss": 0.0024, + "step": 5668 + }, + { + "epoch": 7.362337662337662, + "grad_norm": 15.259015167924137, + "learning_rate": 3.4355130768409538e-06, + "loss": 0.0094, + "step": 5669 + }, + { + "epoch": 7.363636363636363, + "grad_norm": 4.067479513847216, + "learning_rate": 3.4323406420033665e-06, + "loss": 0.0086, + "step": 5670 + }, + { + "epoch": 7.364935064935065, + "grad_norm": 2.8867607510304425, + "learning_rate": 3.429169369110422e-06, + "loss": 0.0141, + "step": 5671 + }, + { + "epoch": 7.3662337662337665, + "grad_norm": 3.4044442641686814, + "learning_rate": 3.4259992587231874e-06, + "loss": 0.0172, + "step": 5672 + }, + { + "epoch": 7.367532467532468, + "grad_norm": 2.42953852266335, + "learning_rate": 3.4228303114025118e-06, + "loss": 0.0014, + "step": 5673 + }, + { + "epoch": 7.368831168831169, + "grad_norm": 2.0682787103657394, + "learning_rate": 3.419662527709042e-06, + "loss": 0.0012, + "step": 5674 + }, + { + "epoch": 7.37012987012987, + "grad_norm": 9.626804058030594, + "learning_rate": 3.4164959082032166e-06, + "loss": 0.0071, + "step": 5675 + }, + { + "epoch": 7.371428571428572, + "grad_norm": 6.682728484588193, + "learning_rate": 3.4133304534452783e-06, + "loss": 0.0154, + "step": 5676 + }, + { + "epoch": 7.372727272727273, + "grad_norm": 2.572900800699742, + "learning_rate": 3.4101661639952533e-06, + "loss": 0.0013, + "step": 5677 + }, + { + "epoch": 7.374025974025974, + "grad_norm": 5.010114617030156, + "learning_rate": 3.407003040412964e-06, + "loss": 0.0204, + "step": 5678 + }, + { + "epoch": 7.375324675324675, + "grad_norm": 4.177207362286198, + "learning_rate": 3.403841083258025e-06, + "loss": 0.0098, + "step": 5679 + }, + { + "epoch": 7.376623376623376, + "grad_norm": 4.051268237127237, + "learning_rate": 3.400680293089853e-06, + "loss": 0.0166, + "step": 5680 + }, + { + "epoch": 7.3779220779220775, + "grad_norm": 2.5514748848363644, + "learning_rate": 3.39752067046765e-06, + "loss": 0.0023, + "step": 5681 + }, + { + "epoch": 7.37922077922078, + "grad_norm": 3.19355790981096, + "learning_rate": 3.3943622159504132e-06, + "loss": 0.0109, + "step": 5682 + }, + { + "epoch": 7.380519480519481, + "grad_norm": 1.4781525089769878, + "learning_rate": 3.3912049300969308e-06, + "loss": 0.0073, + "step": 5683 + }, + { + "epoch": 7.381818181818182, + "grad_norm": 1.1545178068697544, + "learning_rate": 3.3880488134657965e-06, + "loss": 0.0091, + "step": 5684 + }, + { + "epoch": 7.383116883116883, + "grad_norm": 4.481977799038841, + "learning_rate": 3.384893866615374e-06, + "loss": 0.0114, + "step": 5685 + }, + { + "epoch": 7.384415584415584, + "grad_norm": 1.3019314836866043, + "learning_rate": 3.3817400901038442e-06, + "loss": 0.0015, + "step": 5686 + }, + { + "epoch": 7.385714285714286, + "grad_norm": 8.71786325113378, + "learning_rate": 3.378587484489164e-06, + "loss": 0.0147, + "step": 5687 + }, + { + "epoch": 7.387012987012987, + "grad_norm": 4.160648058373386, + "learning_rate": 3.3754360503290973e-06, + "loss": 0.0251, + "step": 5688 + }, + { + "epoch": 7.388311688311688, + "grad_norm": 2.571905535197575, + "learning_rate": 3.3722857881811875e-06, + "loss": 0.016, + "step": 5689 + }, + { + "epoch": 7.3896103896103895, + "grad_norm": 4.587217070133968, + "learning_rate": 3.3691366986027783e-06, + "loss": 0.0038, + "step": 5690 + }, + { + "epoch": 7.390909090909091, + "grad_norm": 2.8473103461184954, + "learning_rate": 3.3659887821509995e-06, + "loss": 0.0098, + "step": 5691 + }, + { + "epoch": 7.392207792207792, + "grad_norm": 2.3056766933922748, + "learning_rate": 3.3628420393827856e-06, + "loss": 0.013, + "step": 5692 + }, + { + "epoch": 7.393506493506494, + "grad_norm": 2.1282284988862963, + "learning_rate": 3.3596964708548505e-06, + "loss": 0.002, + "step": 5693 + }, + { + "epoch": 7.394805194805195, + "grad_norm": 4.090468767220106, + "learning_rate": 3.356552077123707e-06, + "loss": 0.0032, + "step": 5694 + }, + { + "epoch": 7.396103896103896, + "grad_norm": 3.646583928619727, + "learning_rate": 3.353408858745658e-06, + "loss": 0.0219, + "step": 5695 + }, + { + "epoch": 7.397402597402597, + "grad_norm": 4.800583921701423, + "learning_rate": 3.350266816276795e-06, + "loss": 0.0137, + "step": 5696 + }, + { + "epoch": 7.398701298701298, + "grad_norm": 2.18751160729801, + "learning_rate": 3.3471259502730136e-06, + "loss": 0.0103, + "step": 5697 + }, + { + "epoch": 7.4, + "grad_norm": 2.1941884054302325, + "learning_rate": 3.3439862612899852e-06, + "loss": 0.0077, + "step": 5698 + }, + { + "epoch": 7.4012987012987015, + "grad_norm": 5.67675518384743, + "learning_rate": 3.3408477498831917e-06, + "loss": 0.0122, + "step": 5699 + }, + { + "epoch": 7.402597402597403, + "grad_norm": 4.176762906454542, + "learning_rate": 3.337710416607882e-06, + "loss": 0.0119, + "step": 5700 + }, + { + "epoch": 7.403896103896104, + "grad_norm": 1.4782060282392353, + "learning_rate": 3.3345742620191203e-06, + "loss": 0.0023, + "step": 5701 + }, + { + "epoch": 7.405194805194805, + "grad_norm": 1.2830991003635597, + "learning_rate": 3.331439286671746e-06, + "loss": 0.0087, + "step": 5702 + }, + { + "epoch": 7.406493506493507, + "grad_norm": 1.614425481413077, + "learning_rate": 3.328305491120406e-06, + "loss": 0.009, + "step": 5703 + }, + { + "epoch": 7.407792207792208, + "grad_norm": 5.754788362597244, + "learning_rate": 3.3251728759195156e-06, + "loss": 0.006, + "step": 5704 + }, + { + "epoch": 7.409090909090909, + "grad_norm": 2.153255506419632, + "learning_rate": 3.322041441623305e-06, + "loss": 0.0106, + "step": 5705 + }, + { + "epoch": 7.41038961038961, + "grad_norm": 3.4348239270885457, + "learning_rate": 3.3189111887857773e-06, + "loss": 0.0055, + "step": 5706 + }, + { + "epoch": 7.411688311688311, + "grad_norm": 2.7919460101533042, + "learning_rate": 3.315782117960744e-06, + "loss": 0.0097, + "step": 5707 + }, + { + "epoch": 7.4129870129870135, + "grad_norm": 3.82562183196432, + "learning_rate": 3.312654229701785e-06, + "loss": 0.0192, + "step": 5708 + }, + { + "epoch": 7.414285714285715, + "grad_norm": 1.4019633515986447, + "learning_rate": 3.309527524562293e-06, + "loss": 0.0032, + "step": 5709 + }, + { + "epoch": 7.415584415584416, + "grad_norm": 2.2141236765440815, + "learning_rate": 3.306402003095438e-06, + "loss": 0.0137, + "step": 5710 + }, + { + "epoch": 7.416883116883117, + "grad_norm": 1.1403376624046933, + "learning_rate": 3.3032776658541865e-06, + "loss": 0.0013, + "step": 5711 + }, + { + "epoch": 7.418181818181818, + "grad_norm": 2.6610629265291683, + "learning_rate": 3.3001545133912892e-06, + "loss": 0.0022, + "step": 5712 + }, + { + "epoch": 7.419480519480519, + "grad_norm": 3.694458629992874, + "learning_rate": 3.297032546259299e-06, + "loss": 0.0141, + "step": 5713 + }, + { + "epoch": 7.420779220779221, + "grad_norm": 1.7870524559647865, + "learning_rate": 3.2939117650105457e-06, + "loss": 0.0017, + "step": 5714 + }, + { + "epoch": 7.422077922077922, + "grad_norm": 6.433715185094377, + "learning_rate": 3.290792170197159e-06, + "loss": 0.0124, + "step": 5715 + }, + { + "epoch": 7.423376623376623, + "grad_norm": 3.686318590338983, + "learning_rate": 3.2876737623710485e-06, + "loss": 0.0032, + "step": 5716 + }, + { + "epoch": 7.4246753246753245, + "grad_norm": 1.495502511719636, + "learning_rate": 3.284556542083928e-06, + "loss": 0.0053, + "step": 5717 + }, + { + "epoch": 7.425974025974026, + "grad_norm": 6.405441224412718, + "learning_rate": 3.2814405098872902e-06, + "loss": 0.0151, + "step": 5718 + }, + { + "epoch": 7.427272727272728, + "grad_norm": 1.4442461654326693, + "learning_rate": 3.2783256663324216e-06, + "loss": 0.0168, + "step": 5719 + }, + { + "epoch": 7.428571428571429, + "grad_norm": 4.215802783354324, + "learning_rate": 3.2752120119703923e-06, + "loss": 0.0189, + "step": 5720 + }, + { + "epoch": 7.42987012987013, + "grad_norm": 4.205833611688429, + "learning_rate": 3.2720995473520766e-06, + "loss": 0.0023, + "step": 5721 + }, + { + "epoch": 7.431168831168831, + "grad_norm": 1.7455545239330004, + "learning_rate": 3.268988273028123e-06, + "loss": 0.0017, + "step": 5722 + }, + { + "epoch": 7.432467532467532, + "grad_norm": 3.4867407506049553, + "learning_rate": 3.2658781895489776e-06, + "loss": 0.009, + "step": 5723 + }, + { + "epoch": 7.433766233766233, + "grad_norm": 13.660747941354913, + "learning_rate": 3.2627692974648693e-06, + "loss": 0.0326, + "step": 5724 + }, + { + "epoch": 7.435064935064935, + "grad_norm": 0.9788203250570121, + "learning_rate": 3.2596615973258284e-06, + "loss": 0.0097, + "step": 5725 + }, + { + "epoch": 7.4363636363636365, + "grad_norm": 2.7111481117923675, + "learning_rate": 3.256555089681661e-06, + "loss": 0.0023, + "step": 5726 + }, + { + "epoch": 7.437662337662338, + "grad_norm": 5.462211652848938, + "learning_rate": 3.2534497750819673e-06, + "loss": 0.0083, + "step": 5727 + }, + { + "epoch": 7.438961038961039, + "grad_norm": 2.4915461230871854, + "learning_rate": 3.2503456540761413e-06, + "loss": 0.0148, + "step": 5728 + }, + { + "epoch": 7.44025974025974, + "grad_norm": 4.067637174862464, + "learning_rate": 3.247242727213359e-06, + "loss": 0.0087, + "step": 5729 + }, + { + "epoch": 7.441558441558442, + "grad_norm": 2.621041485948525, + "learning_rate": 3.2441409950425873e-06, + "loss": 0.0019, + "step": 5730 + }, + { + "epoch": 7.442857142857143, + "grad_norm": 2.4564028571654037, + "learning_rate": 3.2410404581125787e-06, + "loss": 0.0018, + "step": 5731 + }, + { + "epoch": 7.444155844155844, + "grad_norm": 5.460700500969276, + "learning_rate": 3.2379411169718844e-06, + "loss": 0.0058, + "step": 5732 + }, + { + "epoch": 7.445454545454545, + "grad_norm": 3.3460383297085348, + "learning_rate": 3.2348429721688335e-06, + "loss": 0.0025, + "step": 5733 + }, + { + "epoch": 7.4467532467532465, + "grad_norm": 2.1602778020899884, + "learning_rate": 3.2317460242515476e-06, + "loss": 0.0116, + "step": 5734 + }, + { + "epoch": 7.448051948051948, + "grad_norm": 0.8345987464508529, + "learning_rate": 3.2286502737679315e-06, + "loss": 0.0017, + "step": 5735 + }, + { + "epoch": 7.44935064935065, + "grad_norm": 2.8256159557711356, + "learning_rate": 3.225555721265695e-06, + "loss": 0.0026, + "step": 5736 + }, + { + "epoch": 7.450649350649351, + "grad_norm": 2.122249874299065, + "learning_rate": 3.222462367292308e-06, + "loss": 0.0016, + "step": 5737 + }, + { + "epoch": 7.451948051948052, + "grad_norm": 1.1087683703284985, + "learning_rate": 3.2193702123950553e-06, + "loss": 0.001, + "step": 5738 + }, + { + "epoch": 7.453246753246753, + "grad_norm": 1.4652941859230078, + "learning_rate": 3.2162792571209924e-06, + "loss": 0.0015, + "step": 5739 + }, + { + "epoch": 7.454545454545454, + "grad_norm": 3.0377726244749197, + "learning_rate": 3.2131895020169756e-06, + "loss": 0.0151, + "step": 5740 + }, + { + "epoch": 7.455844155844156, + "grad_norm": 4.823757877718149, + "learning_rate": 3.2101009476296306e-06, + "loss": 0.0044, + "step": 5741 + }, + { + "epoch": 7.457142857142857, + "grad_norm": 3.5663894032402434, + "learning_rate": 3.2070135945053927e-06, + "loss": 0.002, + "step": 5742 + }, + { + "epoch": 7.458441558441558, + "grad_norm": 4.551362542925582, + "learning_rate": 3.2039274431904643e-06, + "loss": 0.0034, + "step": 5743 + }, + { + "epoch": 7.45974025974026, + "grad_norm": 1.087962489340484, + "learning_rate": 3.200842494230857e-06, + "loss": 0.0046, + "step": 5744 + }, + { + "epoch": 7.461038961038961, + "grad_norm": 1.983329893713052, + "learning_rate": 3.197758748172343e-06, + "loss": 0.0081, + "step": 5745 + }, + { + "epoch": 7.462337662337663, + "grad_norm": 3.082728994240199, + "learning_rate": 3.194676205560505e-06, + "loss": 0.0044, + "step": 5746 + }, + { + "epoch": 7.463636363636364, + "grad_norm": 1.5798978858984865, + "learning_rate": 3.1915948669406994e-06, + "loss": 0.0015, + "step": 5747 + }, + { + "epoch": 7.464935064935065, + "grad_norm": 1.161430215632477, + "learning_rate": 3.188514732858082e-06, + "loss": 0.0011, + "step": 5748 + }, + { + "epoch": 7.466233766233766, + "grad_norm": 0.6973456663385506, + "learning_rate": 3.185435803857575e-06, + "loss": 0.0074, + "step": 5749 + }, + { + "epoch": 7.467532467532467, + "grad_norm": 8.526764174062327, + "learning_rate": 3.1823580804839106e-06, + "loss": 0.0144, + "step": 5750 + }, + { + "epoch": 7.468831168831169, + "grad_norm": 2.7776848256705318, + "learning_rate": 3.1792815632815922e-06, + "loss": 0.0042, + "step": 5751 + }, + { + "epoch": 7.47012987012987, + "grad_norm": 0.8671043894082607, + "learning_rate": 3.1762062527949146e-06, + "loss": 0.0046, + "step": 5752 + }, + { + "epoch": 7.4714285714285715, + "grad_norm": 2.1028522051156333, + "learning_rate": 3.173132149567957e-06, + "loss": 0.004, + "step": 5753 + }, + { + "epoch": 7.472727272727273, + "grad_norm": 1.5124383144191171, + "learning_rate": 3.170059254144593e-06, + "loss": 0.002, + "step": 5754 + }, + { + "epoch": 7.474025974025974, + "grad_norm": 1.9679148656605205, + "learning_rate": 3.1669875670684735e-06, + "loss": 0.0099, + "step": 5755 + }, + { + "epoch": 7.475324675324675, + "grad_norm": 6.9912937928562116, + "learning_rate": 3.1639170888830382e-06, + "loss": 0.0172, + "step": 5756 + }, + { + "epoch": 7.476623376623377, + "grad_norm": 1.08454135486303, + "learning_rate": 3.1608478201315117e-06, + "loss": 0.0082, + "step": 5757 + }, + { + "epoch": 7.477922077922078, + "grad_norm": 3.4500359351284144, + "learning_rate": 3.1577797613569107e-06, + "loss": 0.0011, + "step": 5758 + }, + { + "epoch": 7.479220779220779, + "grad_norm": 4.536125288150091, + "learning_rate": 3.1547129131020317e-06, + "loss": 0.0119, + "step": 5759 + }, + { + "epoch": 7.48051948051948, + "grad_norm": 1.2091731003718722, + "learning_rate": 3.151647275909455e-06, + "loss": 0.002, + "step": 5760 + }, + { + "epoch": 7.4818181818181815, + "grad_norm": 4.012265180873122, + "learning_rate": 3.1485828503215588e-06, + "loss": 0.0016, + "step": 5761 + }, + { + "epoch": 7.4831168831168835, + "grad_norm": 8.138649818972679, + "learning_rate": 3.145519636880493e-06, + "loss": 0.0128, + "step": 5762 + }, + { + "epoch": 7.484415584415585, + "grad_norm": 1.9538699427323027, + "learning_rate": 3.142457636128199e-06, + "loss": 0.0103, + "step": 5763 + }, + { + "epoch": 7.485714285714286, + "grad_norm": 1.4783680124750482, + "learning_rate": 3.139396848606401e-06, + "loss": 0.0016, + "step": 5764 + }, + { + "epoch": 7.487012987012987, + "grad_norm": 3.111790241901183, + "learning_rate": 3.136337274856618e-06, + "loss": 0.0024, + "step": 5765 + }, + { + "epoch": 7.488311688311688, + "grad_norm": 2.5687837715920523, + "learning_rate": 3.1332789154201415e-06, + "loss": 0.0082, + "step": 5766 + }, + { + "epoch": 7.489610389610389, + "grad_norm": 0.3682291032467596, + "learning_rate": 3.1302217708380566e-06, + "loss": 0.0011, + "step": 5767 + }, + { + "epoch": 7.490909090909091, + "grad_norm": 3.070123056663128, + "learning_rate": 3.127165841651225e-06, + "loss": 0.0012, + "step": 5768 + }, + { + "epoch": 7.492207792207792, + "grad_norm": 5.352245722677235, + "learning_rate": 3.1241111284003078e-06, + "loss": 0.0056, + "step": 5769 + }, + { + "epoch": 7.4935064935064934, + "grad_norm": 9.925068093663333, + "learning_rate": 3.121057631625737e-06, + "loss": 0.0096, + "step": 5770 + }, + { + "epoch": 7.494805194805195, + "grad_norm": 4.598454434604356, + "learning_rate": 3.1180053518677354e-06, + "loss": 0.01, + "step": 5771 + }, + { + "epoch": 7.496103896103896, + "grad_norm": 3.3147698364359774, + "learning_rate": 3.114954289666308e-06, + "loss": 0.0048, + "step": 5772 + }, + { + "epoch": 7.497402597402598, + "grad_norm": 3.369339932479001, + "learning_rate": 3.11190444556125e-06, + "loss": 0.0052, + "step": 5773 + }, + { + "epoch": 7.498701298701299, + "grad_norm": 1.3691860100160238, + "learning_rate": 3.108855820092135e-06, + "loss": 0.0017, + "step": 5774 + }, + { + "epoch": 7.5, + "grad_norm": 4.964789182503435, + "learning_rate": 3.1058084137983234e-06, + "loss": 0.003, + "step": 5775 + }, + { + "epoch": 7.501298701298701, + "grad_norm": 4.390744090177057, + "learning_rate": 3.1027622272189572e-06, + "loss": 0.009, + "step": 5776 + }, + { + "epoch": 7.502597402597402, + "grad_norm": 1.9801274896414947, + "learning_rate": 3.0997172608929736e-06, + "loss": 0.0013, + "step": 5777 + }, + { + "epoch": 7.503896103896103, + "grad_norm": 3.688481131404839, + "learning_rate": 3.096673515359073e-06, + "loss": 0.0188, + "step": 5778 + }, + { + "epoch": 7.505194805194805, + "grad_norm": 8.594706249453557, + "learning_rate": 3.0936309911557615e-06, + "loss": 0.0213, + "step": 5779 + }, + { + "epoch": 7.5064935064935066, + "grad_norm": 3.0527834448172975, + "learning_rate": 3.090589688821314e-06, + "loss": 0.0095, + "step": 5780 + }, + { + "epoch": 7.507792207792208, + "grad_norm": 1.1510000249094143, + "learning_rate": 3.087549608893805e-06, + "loss": 0.0017, + "step": 5781 + }, + { + "epoch": 7.509090909090909, + "grad_norm": 4.537486731966134, + "learning_rate": 3.0845107519110682e-06, + "loss": 0.0129, + "step": 5782 + }, + { + "epoch": 7.510389610389611, + "grad_norm": 8.539109045977314, + "learning_rate": 3.081473118410747e-06, + "loss": 0.004, + "step": 5783 + }, + { + "epoch": 7.511688311688312, + "grad_norm": 2.7495069452986414, + "learning_rate": 3.0784367089302503e-06, + "loss": 0.0021, + "step": 5784 + }, + { + "epoch": 7.512987012987013, + "grad_norm": 9.878886809555294, + "learning_rate": 3.075401524006787e-06, + "loss": 0.0175, + "step": 5785 + }, + { + "epoch": 7.514285714285714, + "grad_norm": 1.8119045404754281, + "learning_rate": 3.0723675641773243e-06, + "loss": 0.0117, + "step": 5786 + }, + { + "epoch": 7.515584415584415, + "grad_norm": 1.659233549828857, + "learning_rate": 3.0693348299786407e-06, + "loss": 0.0027, + "step": 5787 + }, + { + "epoch": 7.5168831168831165, + "grad_norm": 0.8201109511246288, + "learning_rate": 3.066303321947276e-06, + "loss": 0.0148, + "step": 5788 + }, + { + "epoch": 7.5181818181818185, + "grad_norm": 2.137922435490477, + "learning_rate": 3.0632730406195686e-06, + "loss": 0.0102, + "step": 5789 + }, + { + "epoch": 7.51948051948052, + "grad_norm": 4.403125078868967, + "learning_rate": 3.0602439865316314e-06, + "loss": 0.0036, + "step": 5790 + }, + { + "epoch": 7.520779220779221, + "grad_norm": 1.5063605432512053, + "learning_rate": 3.0572161602193607e-06, + "loss": 0.0023, + "step": 5791 + }, + { + "epoch": 7.522077922077922, + "grad_norm": 2.7123102626901137, + "learning_rate": 3.054189562218438e-06, + "loss": 0.0026, + "step": 5792 + }, + { + "epoch": 7.523376623376623, + "grad_norm": 2.3696057296621214, + "learning_rate": 3.0511641930643234e-06, + "loss": 0.0027, + "step": 5793 + }, + { + "epoch": 7.524675324675325, + "grad_norm": 5.906649415951111, + "learning_rate": 3.048140053292268e-06, + "loss": 0.0053, + "step": 5794 + }, + { + "epoch": 7.525974025974026, + "grad_norm": 2.9499143203687717, + "learning_rate": 3.0451171434372963e-06, + "loss": 0.0083, + "step": 5795 + }, + { + "epoch": 7.527272727272727, + "grad_norm": 1.6395867549083494, + "learning_rate": 3.0420954640342203e-06, + "loss": 0.0015, + "step": 5796 + }, + { + "epoch": 7.5285714285714285, + "grad_norm": 6.625820045230881, + "learning_rate": 3.0390750156176297e-06, + "loss": 0.0065, + "step": 5797 + }, + { + "epoch": 7.52987012987013, + "grad_norm": 3.439866884689871, + "learning_rate": 3.0360557987219054e-06, + "loss": 0.0141, + "step": 5798 + }, + { + "epoch": 7.531168831168831, + "grad_norm": 2.626517225070244, + "learning_rate": 3.033037813881198e-06, + "loss": 0.0018, + "step": 5799 + }, + { + "epoch": 7.532467532467533, + "grad_norm": 6.9400689467113965, + "learning_rate": 3.030021061629458e-06, + "loss": 0.0217, + "step": 5800 + }, + { + "epoch": 7.533766233766234, + "grad_norm": 6.955948872202001, + "learning_rate": 3.0270055425003906e-06, + "loss": 0.0138, + "step": 5801 + }, + { + "epoch": 7.535064935064935, + "grad_norm": 4.958936475683977, + "learning_rate": 3.023991257027512e-06, + "loss": 0.0055, + "step": 5802 + }, + { + "epoch": 7.536363636363636, + "grad_norm": 3.5392562059923356, + "learning_rate": 3.0209782057440985e-06, + "loss": 0.0122, + "step": 5803 + }, + { + "epoch": 7.537662337662337, + "grad_norm": 7.375019226992633, + "learning_rate": 3.017966389183227e-06, + "loss": 0.0149, + "step": 5804 + }, + { + "epoch": 7.538961038961039, + "grad_norm": 3.8844467949922104, + "learning_rate": 3.0149558078777318e-06, + "loss": 0.0016, + "step": 5805 + }, + { + "epoch": 7.54025974025974, + "grad_norm": 2.07401007600205, + "learning_rate": 3.0119464623602534e-06, + "loss": 0.0111, + "step": 5806 + }, + { + "epoch": 7.541558441558442, + "grad_norm": 3.854246613234224, + "learning_rate": 3.0089383531631976e-06, + "loss": 0.0017, + "step": 5807 + }, + { + "epoch": 7.542857142857143, + "grad_norm": 4.273652878495732, + "learning_rate": 3.005931480818759e-06, + "loss": 0.0024, + "step": 5808 + }, + { + "epoch": 7.544155844155844, + "grad_norm": 7.888333960166507, + "learning_rate": 3.002925845858905e-06, + "loss": 0.0151, + "step": 5809 + }, + { + "epoch": 7.545454545454545, + "grad_norm": 1.3653946039310914, + "learning_rate": 2.9999214488153973e-06, + "loss": 0.0013, + "step": 5810 + }, + { + "epoch": 7.546753246753247, + "grad_norm": 4.213905127872333, + "learning_rate": 2.996918290219769e-06, + "loss": 0.003, + "step": 5811 + }, + { + "epoch": 7.548051948051948, + "grad_norm": 6.408894259345969, + "learning_rate": 2.993916370603335e-06, + "loss": 0.013, + "step": 5812 + }, + { + "epoch": 7.549350649350649, + "grad_norm": 1.6701208458698946, + "learning_rate": 2.9909156904971905e-06, + "loss": 0.0014, + "step": 5813 + }, + { + "epoch": 7.55064935064935, + "grad_norm": 1.0504484888809262, + "learning_rate": 2.9879162504322177e-06, + "loss": 0.0016, + "step": 5814 + }, + { + "epoch": 7.551948051948052, + "grad_norm": 9.194125789285355, + "learning_rate": 2.984918050939074e-06, + "loss": 0.0243, + "step": 5815 + }, + { + "epoch": 7.5532467532467535, + "grad_norm": 0.8301978851412198, + "learning_rate": 2.981921092548197e-06, + "loss": 0.0015, + "step": 5816 + }, + { + "epoch": 7.554545454545455, + "grad_norm": 8.660150867833222, + "learning_rate": 2.9789253757898017e-06, + "loss": 0.0222, + "step": 5817 + }, + { + "epoch": 7.555844155844156, + "grad_norm": 5.039555091819179, + "learning_rate": 2.9759309011938997e-06, + "loss": 0.0117, + "step": 5818 + }, + { + "epoch": 7.557142857142857, + "grad_norm": 2.518941154230151, + "learning_rate": 2.9729376692902567e-06, + "loss": 0.0043, + "step": 5819 + }, + { + "epoch": 7.558441558441558, + "grad_norm": 0.5909083808107414, + "learning_rate": 2.9699456806084427e-06, + "loss": 0.0019, + "step": 5820 + }, + { + "epoch": 7.55974025974026, + "grad_norm": 3.977466368503682, + "learning_rate": 2.9669549356777918e-06, + "loss": 0.0263, + "step": 5821 + }, + { + "epoch": 7.561038961038961, + "grad_norm": 2.3459914085867344, + "learning_rate": 2.9639654350274295e-06, + "loss": 0.0046, + "step": 5822 + }, + { + "epoch": 7.562337662337662, + "grad_norm": 7.867798223772266, + "learning_rate": 2.960977179186253e-06, + "loss": 0.0114, + "step": 5823 + }, + { + "epoch": 7.5636363636363635, + "grad_norm": 2.9353597630306174, + "learning_rate": 2.9579901686829417e-06, + "loss": 0.0096, + "step": 5824 + }, + { + "epoch": 7.564935064935065, + "grad_norm": 3.431111809361088, + "learning_rate": 2.9550044040459515e-06, + "loss": 0.004, + "step": 5825 + }, + { + "epoch": 7.566233766233767, + "grad_norm": 1.0883858599522218, + "learning_rate": 2.952019885803529e-06, + "loss": 0.0074, + "step": 5826 + }, + { + "epoch": 7.567532467532468, + "grad_norm": 1.419444864233371, + "learning_rate": 2.949036614483687e-06, + "loss": 0.0019, + "step": 5827 + }, + { + "epoch": 7.568831168831169, + "grad_norm": 0.9041829328221102, + "learning_rate": 2.946054590614226e-06, + "loss": 0.0064, + "step": 5828 + }, + { + "epoch": 7.57012987012987, + "grad_norm": 2.713339356217076, + "learning_rate": 2.943073814722719e-06, + "loss": 0.006, + "step": 5829 + }, + { + "epoch": 7.571428571428571, + "grad_norm": 1.35790945315805, + "learning_rate": 2.9400942873365268e-06, + "loss": 0.0018, + "step": 5830 + }, + { + "epoch": 7.572727272727272, + "grad_norm": 1.5593384027152901, + "learning_rate": 2.9371160089827842e-06, + "loss": 0.0045, + "step": 5831 + }, + { + "epoch": 7.574025974025974, + "grad_norm": 2.4041248886527775, + "learning_rate": 2.9341389801884012e-06, + "loss": 0.0275, + "step": 5832 + }, + { + "epoch": 7.5753246753246755, + "grad_norm": 0.5190292953954683, + "learning_rate": 2.9311632014800816e-06, + "loss": 0.0015, + "step": 5833 + }, + { + "epoch": 7.576623376623377, + "grad_norm": 0.7453432598195405, + "learning_rate": 2.9281886733842837e-06, + "loss": 0.0014, + "step": 5834 + }, + { + "epoch": 7.577922077922078, + "grad_norm": 2.8102470260869126, + "learning_rate": 2.9252153964272687e-06, + "loss": 0.006, + "step": 5835 + }, + { + "epoch": 7.579220779220779, + "grad_norm": 3.406213196574536, + "learning_rate": 2.9222433711350594e-06, + "loss": 0.0044, + "step": 5836 + }, + { + "epoch": 7.580519480519481, + "grad_norm": 3.7852770413138486, + "learning_rate": 2.919272598033475e-06, + "loss": 0.0105, + "step": 5837 + }, + { + "epoch": 7.581818181818182, + "grad_norm": 0.725716190314621, + "learning_rate": 2.9163030776480873e-06, + "loss": 0.0014, + "step": 5838 + }, + { + "epoch": 7.583116883116883, + "grad_norm": 5.774724281791764, + "learning_rate": 2.9133348105042713e-06, + "loss": 0.0114, + "step": 5839 + }, + { + "epoch": 7.584415584415584, + "grad_norm": 6.621492509455366, + "learning_rate": 2.9103677971271647e-06, + "loss": 0.006, + "step": 5840 + }, + { + "epoch": 7.585714285714285, + "grad_norm": 1.6113156702744462, + "learning_rate": 2.907402038041699e-06, + "loss": 0.0068, + "step": 5841 + }, + { + "epoch": 7.5870129870129865, + "grad_norm": 3.4937868794930904, + "learning_rate": 2.9044375337725595e-06, + "loss": 0.0131, + "step": 5842 + }, + { + "epoch": 7.588311688311689, + "grad_norm": 1.0431231412211377, + "learning_rate": 2.9014742848442335e-06, + "loss": 0.0013, + "step": 5843 + }, + { + "epoch": 7.58961038961039, + "grad_norm": 1.4803796039745907, + "learning_rate": 2.898512291780975e-06, + "loss": 0.0082, + "step": 5844 + }, + { + "epoch": 7.590909090909091, + "grad_norm": 5.907419977656032, + "learning_rate": 2.8955515551068147e-06, + "loss": 0.017, + "step": 5845 + }, + { + "epoch": 7.592207792207792, + "grad_norm": 2.8473437560617922, + "learning_rate": 2.8925920753455625e-06, + "loss": 0.0045, + "step": 5846 + }, + { + "epoch": 7.593506493506494, + "grad_norm": 4.197599620048244, + "learning_rate": 2.889633853020812e-06, + "loss": 0.0104, + "step": 5847 + }, + { + "epoch": 7.594805194805195, + "grad_norm": 7.993206483507294, + "learning_rate": 2.886676888655927e-06, + "loss": 0.0079, + "step": 5848 + }, + { + "epoch": 7.596103896103896, + "grad_norm": 0.5659990997056694, + "learning_rate": 2.8837211827740517e-06, + "loss": 0.0016, + "step": 5849 + }, + { + "epoch": 7.597402597402597, + "grad_norm": 1.618607985454341, + "learning_rate": 2.8807667358981016e-06, + "loss": 0.0076, + "step": 5850 + }, + { + "epoch": 7.5987012987012985, + "grad_norm": 5.5828459656970235, + "learning_rate": 2.8778135485507843e-06, + "loss": 0.01, + "step": 5851 + }, + { + "epoch": 7.6, + "grad_norm": 1.181514177313202, + "learning_rate": 2.8748616212545688e-06, + "loss": 0.0011, + "step": 5852 + }, + { + "epoch": 7.601298701298702, + "grad_norm": 3.857295386089176, + "learning_rate": 2.8719109545317102e-06, + "loss": 0.0193, + "step": 5853 + }, + { + "epoch": 7.602597402597403, + "grad_norm": 7.090249757276686, + "learning_rate": 2.8689615489042333e-06, + "loss": 0.0229, + "step": 5854 + }, + { + "epoch": 7.603896103896104, + "grad_norm": 0.9138065514800681, + "learning_rate": 2.86601340489395e-06, + "loss": 0.0013, + "step": 5855 + }, + { + "epoch": 7.605194805194805, + "grad_norm": 4.564247727376822, + "learning_rate": 2.863066523022443e-06, + "loss": 0.0061, + "step": 5856 + }, + { + "epoch": 7.606493506493506, + "grad_norm": 5.621788471357411, + "learning_rate": 2.86012090381107e-06, + "loss": 0.0029, + "step": 5857 + }, + { + "epoch": 7.607792207792208, + "grad_norm": 3.019311625864664, + "learning_rate": 2.8571765477809645e-06, + "loss": 0.0146, + "step": 5858 + }, + { + "epoch": 7.609090909090909, + "grad_norm": 3.690554700658155, + "learning_rate": 2.854233455453045e-06, + "loss": 0.0019, + "step": 5859 + }, + { + "epoch": 7.6103896103896105, + "grad_norm": 4.02214020346945, + "learning_rate": 2.8512916273479996e-06, + "loss": 0.0015, + "step": 5860 + }, + { + "epoch": 7.611688311688312, + "grad_norm": 3.9796038485221987, + "learning_rate": 2.84835106398629e-06, + "loss": 0.0067, + "step": 5861 + }, + { + "epoch": 7.612987012987013, + "grad_norm": 4.628375189439447, + "learning_rate": 2.845411765888164e-06, + "loss": 0.0118, + "step": 5862 + }, + { + "epoch": 7.614285714285714, + "grad_norm": 8.914967203543878, + "learning_rate": 2.8424737335736376e-06, + "loss": 0.0129, + "step": 5863 + }, + { + "epoch": 7.615584415584416, + "grad_norm": 4.242005291018762, + "learning_rate": 2.839536967562504e-06, + "loss": 0.01, + "step": 5864 + }, + { + "epoch": 7.616883116883117, + "grad_norm": 2.809295621145228, + "learning_rate": 2.8366014683743304e-06, + "loss": 0.0084, + "step": 5865 + }, + { + "epoch": 7.618181818181818, + "grad_norm": 4.692817958027848, + "learning_rate": 2.833667236528468e-06, + "loss": 0.0369, + "step": 5866 + }, + { + "epoch": 7.619480519480519, + "grad_norm": 6.19012632724313, + "learning_rate": 2.830734272544038e-06, + "loss": 0.0038, + "step": 5867 + }, + { + "epoch": 7.62077922077922, + "grad_norm": 49.07686599421166, + "learning_rate": 2.8278025769399354e-06, + "loss": 0.0394, + "step": 5868 + }, + { + "epoch": 7.6220779220779225, + "grad_norm": 0.9862302355584146, + "learning_rate": 2.824872150234832e-06, + "loss": 0.0088, + "step": 5869 + }, + { + "epoch": 7.623376623376624, + "grad_norm": 1.5840953118455094, + "learning_rate": 2.8219429929471808e-06, + "loss": 0.0111, + "step": 5870 + }, + { + "epoch": 7.624675324675325, + "grad_norm": 5.710399886641803, + "learning_rate": 2.819015105595204e-06, + "loss": 0.009, + "step": 5871 + }, + { + "epoch": 7.625974025974026, + "grad_norm": 6.1730103392427225, + "learning_rate": 2.8160884886969008e-06, + "loss": 0.0273, + "step": 5872 + }, + { + "epoch": 7.627272727272727, + "grad_norm": 5.944055766260683, + "learning_rate": 2.8131631427700422e-06, + "loss": 0.0059, + "step": 5873 + }, + { + "epoch": 7.628571428571428, + "grad_norm": 3.9577714987349273, + "learning_rate": 2.8102390683321877e-06, + "loss": 0.0076, + "step": 5874 + }, + { + "epoch": 7.62987012987013, + "grad_norm": 1.6227082020381347, + "learning_rate": 2.807316265900649e-06, + "loss": 0.0027, + "step": 5875 + }, + { + "epoch": 7.631168831168831, + "grad_norm": 8.435587134973224, + "learning_rate": 2.804394735992535e-06, + "loss": 0.0224, + "step": 5876 + }, + { + "epoch": 7.632467532467532, + "grad_norm": 3.085092586476463, + "learning_rate": 2.801474479124715e-06, + "loss": 0.0015, + "step": 5877 + }, + { + "epoch": 7.6337662337662335, + "grad_norm": 1.7399461519855524, + "learning_rate": 2.7985554958138473e-06, + "loss": 0.0035, + "step": 5878 + }, + { + "epoch": 7.6350649350649356, + "grad_norm": 2.409256794347447, + "learning_rate": 2.795637786576343e-06, + "loss": 0.0096, + "step": 5879 + }, + { + "epoch": 7.636363636363637, + "grad_norm": 3.427697139252926, + "learning_rate": 2.7927213519284104e-06, + "loss": 0.0071, + "step": 5880 + }, + { + "epoch": 7.637662337662338, + "grad_norm": 10.075061145486783, + "learning_rate": 2.7898061923860153e-06, + "loss": 0.0403, + "step": 5881 + }, + { + "epoch": 7.638961038961039, + "grad_norm": 4.1843794511284464, + "learning_rate": 2.786892308464917e-06, + "loss": 0.0048, + "step": 5882 + }, + { + "epoch": 7.64025974025974, + "grad_norm": 4.931717820445677, + "learning_rate": 2.7839797006806223e-06, + "loss": 0.0035, + "step": 5883 + }, + { + "epoch": 7.641558441558441, + "grad_norm": 3.3400306346210624, + "learning_rate": 2.7810683695484384e-06, + "loss": 0.0048, + "step": 5884 + }, + { + "epoch": 7.642857142857143, + "grad_norm": 6.815877858533215, + "learning_rate": 2.7781583155834325e-06, + "loss": 0.0098, + "step": 5885 + }, + { + "epoch": 7.644155844155844, + "grad_norm": 3.745241482459682, + "learning_rate": 2.7752495393004476e-06, + "loss": 0.0014, + "step": 5886 + }, + { + "epoch": 7.6454545454545455, + "grad_norm": 4.5472631734933895, + "learning_rate": 2.772342041214099e-06, + "loss": 0.0033, + "step": 5887 + }, + { + "epoch": 7.646753246753247, + "grad_norm": 2.1324535266736677, + "learning_rate": 2.7694358218387853e-06, + "loss": 0.0015, + "step": 5888 + }, + { + "epoch": 7.648051948051948, + "grad_norm": 2.719168890039553, + "learning_rate": 2.766530881688668e-06, + "loss": 0.0094, + "step": 5889 + }, + { + "epoch": 7.64935064935065, + "grad_norm": 2.5617001473738017, + "learning_rate": 2.7636272212776892e-06, + "loss": 0.0022, + "step": 5890 + }, + { + "epoch": 7.650649350649351, + "grad_norm": 2.386834251011846, + "learning_rate": 2.7607248411195573e-06, + "loss": 0.0013, + "step": 5891 + }, + { + "epoch": 7.651948051948052, + "grad_norm": 2.9990925711181777, + "learning_rate": 2.7578237417277653e-06, + "loss": 0.011, + "step": 5892 + }, + { + "epoch": 7.653246753246753, + "grad_norm": 3.6367365011268586, + "learning_rate": 2.754923923615571e-06, + "loss": 0.0102, + "step": 5893 + }, + { + "epoch": 7.654545454545454, + "grad_norm": 3.0089582538210053, + "learning_rate": 2.7520253872960024e-06, + "loss": 0.0014, + "step": 5894 + }, + { + "epoch": 7.6558441558441555, + "grad_norm": 3.50967970680855, + "learning_rate": 2.7491281332818743e-06, + "loss": 0.0085, + "step": 5895 + }, + { + "epoch": 7.6571428571428575, + "grad_norm": 1.3082106011828964, + "learning_rate": 2.7462321620857626e-06, + "loss": 0.0233, + "step": 5896 + }, + { + "epoch": 7.658441558441559, + "grad_norm": 1.7359529850762974, + "learning_rate": 2.743337474220019e-06, + "loss": 0.0099, + "step": 5897 + }, + { + "epoch": 7.65974025974026, + "grad_norm": 5.544962276819016, + "learning_rate": 2.7404440701967683e-06, + "loss": 0.0241, + "step": 5898 + }, + { + "epoch": 7.661038961038961, + "grad_norm": 1.7182993029814473, + "learning_rate": 2.7375519505279135e-06, + "loss": 0.0083, + "step": 5899 + }, + { + "epoch": 7.662337662337662, + "grad_norm": 5.0982851550902275, + "learning_rate": 2.7346611157251236e-06, + "loss": 0.0022, + "step": 5900 + }, + { + "epoch": 7.663636363636364, + "grad_norm": 4.473985175508148, + "learning_rate": 2.7317715662998423e-06, + "loss": 0.0064, + "step": 5901 + }, + { + "epoch": 7.664935064935065, + "grad_norm": 1.5649239004726416, + "learning_rate": 2.7288833027632834e-06, + "loss": 0.0014, + "step": 5902 + }, + { + "epoch": 7.666233766233766, + "grad_norm": 4.112376029131538, + "learning_rate": 2.725996325626442e-06, + "loss": 0.007, + "step": 5903 + }, + { + "epoch": 7.667532467532467, + "grad_norm": 1.5115001112459443, + "learning_rate": 2.7231106354000756e-06, + "loss": 0.0091, + "step": 5904 + }, + { + "epoch": 7.6688311688311686, + "grad_norm": 7.455949773148772, + "learning_rate": 2.7202262325947206e-06, + "loss": 0.0185, + "step": 5905 + }, + { + "epoch": 7.67012987012987, + "grad_norm": 3.4598598065438257, + "learning_rate": 2.7173431177206766e-06, + "loss": 0.0047, + "step": 5906 + }, + { + "epoch": 7.671428571428572, + "grad_norm": 8.71196213985597, + "learning_rate": 2.714461291288031e-06, + "loss": 0.0227, + "step": 5907 + }, + { + "epoch": 7.672727272727273, + "grad_norm": 0.8507265938387066, + "learning_rate": 2.71158075380663e-06, + "loss": 0.0017, + "step": 5908 + }, + { + "epoch": 7.674025974025974, + "grad_norm": 1.0843960647939561, + "learning_rate": 2.7087015057860943e-06, + "loss": 0.0055, + "step": 5909 + }, + { + "epoch": 7.675324675324675, + "grad_norm": 3.6398286292698456, + "learning_rate": 2.7058235477358162e-06, + "loss": 0.0043, + "step": 5910 + }, + { + "epoch": 7.676623376623376, + "grad_norm": 2.668616589963154, + "learning_rate": 2.702946880164967e-06, + "loss": 0.0038, + "step": 5911 + }, + { + "epoch": 7.677922077922078, + "grad_norm": 4.2101602287429225, + "learning_rate": 2.7000715035824832e-06, + "loss": 0.0237, + "step": 5912 + }, + { + "epoch": 7.679220779220779, + "grad_norm": 3.4027512758201404, + "learning_rate": 2.697197418497072e-06, + "loss": 0.0031, + "step": 5913 + }, + { + "epoch": 7.6805194805194805, + "grad_norm": 0.7779567970557508, + "learning_rate": 2.6943246254172116e-06, + "loss": 0.0016, + "step": 5914 + }, + { + "epoch": 7.681818181818182, + "grad_norm": 1.4620747995619634, + "learning_rate": 2.6914531248511633e-06, + "loss": 0.0028, + "step": 5915 + }, + { + "epoch": 7.683116883116883, + "grad_norm": 1.2676124056219367, + "learning_rate": 2.688582917306938e-06, + "loss": 0.0014, + "step": 5916 + }, + { + "epoch": 7.684415584415584, + "grad_norm": 2.233484126794543, + "learning_rate": 2.68571400329234e-06, + "loss": 0.0159, + "step": 5917 + }, + { + "epoch": 7.685714285714286, + "grad_norm": 1.8476240917361813, + "learning_rate": 2.6828463833149275e-06, + "loss": 0.0022, + "step": 5918 + }, + { + "epoch": 7.687012987012987, + "grad_norm": 5.496595749870752, + "learning_rate": 2.679980057882049e-06, + "loss": 0.013, + "step": 5919 + }, + { + "epoch": 7.688311688311688, + "grad_norm": 1.4386764617293057, + "learning_rate": 2.6771150275007986e-06, + "loss": 0.0087, + "step": 5920 + }, + { + "epoch": 7.689610389610389, + "grad_norm": 2.6367340128265795, + "learning_rate": 2.6742512926780637e-06, + "loss": 0.0164, + "step": 5921 + }, + { + "epoch": 7.690909090909091, + "grad_norm": 1.931135706653283, + "learning_rate": 2.671388853920489e-06, + "loss": 0.0024, + "step": 5922 + }, + { + "epoch": 7.6922077922077925, + "grad_norm": 2.068607496098983, + "learning_rate": 2.668527711734501e-06, + "loss": 0.0112, + "step": 5923 + }, + { + "epoch": 7.693506493506494, + "grad_norm": 1.6859228572811027, + "learning_rate": 2.665667866626286e-06, + "loss": 0.0082, + "step": 5924 + }, + { + "epoch": 7.694805194805195, + "grad_norm": 0.520566110464785, + "learning_rate": 2.6628093191018055e-06, + "loss": 0.0012, + "step": 5925 + }, + { + "epoch": 7.696103896103896, + "grad_norm": 5.601057713122163, + "learning_rate": 2.659952069666791e-06, + "loss": 0.0061, + "step": 5926 + }, + { + "epoch": 7.697402597402597, + "grad_norm": 0.8420780666232555, + "learning_rate": 2.657096118826744e-06, + "loss": 0.0013, + "step": 5927 + }, + { + "epoch": 7.698701298701299, + "grad_norm": 1.2449808726077127, + "learning_rate": 2.654241467086941e-06, + "loss": 0.0151, + "step": 5928 + }, + { + "epoch": 7.7, + "grad_norm": 2.7017004013421726, + "learning_rate": 2.6513881149524212e-06, + "loss": 0.0089, + "step": 5929 + }, + { + "epoch": 7.701298701298701, + "grad_norm": 1.6418939474962204, + "learning_rate": 2.648536062927999e-06, + "loss": 0.0022, + "step": 5930 + }, + { + "epoch": 7.702597402597402, + "grad_norm": 0.8368825527056254, + "learning_rate": 2.6456853115182513e-06, + "loss": 0.0014, + "step": 5931 + }, + { + "epoch": 7.703896103896104, + "grad_norm": 2.8438874450335163, + "learning_rate": 2.6428358612275383e-06, + "loss": 0.0126, + "step": 5932 + }, + { + "epoch": 7.705194805194806, + "grad_norm": 1.6317817712028264, + "learning_rate": 2.639987712559977e-06, + "loss": 0.0026, + "step": 5933 + }, + { + "epoch": 7.706493506493507, + "grad_norm": 1.2454753509166874, + "learning_rate": 2.6371408660194664e-06, + "loss": 0.0012, + "step": 5934 + }, + { + "epoch": 7.707792207792208, + "grad_norm": 0.7097080481843113, + "learning_rate": 2.634295322109658e-06, + "loss": 0.0013, + "step": 5935 + }, + { + "epoch": 7.709090909090909, + "grad_norm": 0.3494804174345036, + "learning_rate": 2.631451081333991e-06, + "loss": 0.0013, + "step": 5936 + }, + { + "epoch": 7.71038961038961, + "grad_norm": 0.8062218368132095, + "learning_rate": 2.62860814419566e-06, + "loss": 0.0093, + "step": 5937 + }, + { + "epoch": 7.711688311688311, + "grad_norm": 2.3809375775832224, + "learning_rate": 2.625766511197645e-06, + "loss": 0.0066, + "step": 5938 + }, + { + "epoch": 7.712987012987013, + "grad_norm": 1.0727394157777561, + "learning_rate": 2.6229261828426722e-06, + "loss": 0.0088, + "step": 5939 + }, + { + "epoch": 7.714285714285714, + "grad_norm": 2.9582466943197407, + "learning_rate": 2.620087159633259e-06, + "loss": 0.0047, + "step": 5940 + }, + { + "epoch": 7.7155844155844155, + "grad_norm": 2.578151458278859, + "learning_rate": 2.61724944207168e-06, + "loss": 0.0105, + "step": 5941 + }, + { + "epoch": 7.716883116883117, + "grad_norm": 0.9061479234617347, + "learning_rate": 2.6144130306599826e-06, + "loss": 0.0014, + "step": 5942 + }, + { + "epoch": 7.718181818181818, + "grad_norm": 3.212634911580662, + "learning_rate": 2.6115779258999784e-06, + "loss": 0.0064, + "step": 5943 + }, + { + "epoch": 7.71948051948052, + "grad_norm": 6.192600766307782, + "learning_rate": 2.608744128293258e-06, + "loss": 0.0157, + "step": 5944 + }, + { + "epoch": 7.720779220779221, + "grad_norm": 3.4714341044893824, + "learning_rate": 2.605911638341172e-06, + "loss": 0.0035, + "step": 5945 + }, + { + "epoch": 7.722077922077922, + "grad_norm": 3.627841292691617, + "learning_rate": 2.6030804565448408e-06, + "loss": 0.0012, + "step": 5946 + }, + { + "epoch": 7.723376623376623, + "grad_norm": 5.264670213293454, + "learning_rate": 2.6002505834051527e-06, + "loss": 0.0116, + "step": 5947 + }, + { + "epoch": 7.724675324675324, + "grad_norm": 4.43165465174005, + "learning_rate": 2.597422019422773e-06, + "loss": 0.0192, + "step": 5948 + }, + { + "epoch": 7.7259740259740255, + "grad_norm": 6.1229716861992465, + "learning_rate": 2.594594765098124e-06, + "loss": 0.0156, + "step": 5949 + }, + { + "epoch": 7.7272727272727275, + "grad_norm": 4.797396075114644, + "learning_rate": 2.591768820931402e-06, + "loss": 0.0117, + "step": 5950 + }, + { + "epoch": 7.728571428571429, + "grad_norm": 5.346904532114007, + "learning_rate": 2.5889441874225676e-06, + "loss": 0.0018, + "step": 5951 + }, + { + "epoch": 7.72987012987013, + "grad_norm": 4.784170994531838, + "learning_rate": 2.5861208650713586e-06, + "loss": 0.0219, + "step": 5952 + }, + { + "epoch": 7.731168831168831, + "grad_norm": 4.343926865412247, + "learning_rate": 2.583298854377272e-06, + "loss": 0.0032, + "step": 5953 + }, + { + "epoch": 7.732467532467533, + "grad_norm": 6.192350344555353, + "learning_rate": 2.5804781558395754e-06, + "loss": 0.002, + "step": 5954 + }, + { + "epoch": 7.733766233766234, + "grad_norm": 5.351126031318584, + "learning_rate": 2.5776587699573007e-06, + "loss": 0.0098, + "step": 5955 + }, + { + "epoch": 7.735064935064935, + "grad_norm": 4.7211417268400915, + "learning_rate": 2.574840697229257e-06, + "loss": 0.0129, + "step": 5956 + }, + { + "epoch": 7.736363636363636, + "grad_norm": 3.5288503809938607, + "learning_rate": 2.5720239381540146e-06, + "loss": 0.0026, + "step": 5957 + }, + { + "epoch": 7.7376623376623375, + "grad_norm": 5.715043411126832, + "learning_rate": 2.56920849322991e-06, + "loss": 0.0028, + "step": 5958 + }, + { + "epoch": 7.738961038961039, + "grad_norm": 5.292825303329866, + "learning_rate": 2.566394362955047e-06, + "loss": 0.0031, + "step": 5959 + }, + { + "epoch": 7.740259740259741, + "grad_norm": 1.5745506281559802, + "learning_rate": 2.5635815478273043e-06, + "loss": 0.0028, + "step": 5960 + }, + { + "epoch": 7.741558441558442, + "grad_norm": 8.22067221883882, + "learning_rate": 2.560770048344321e-06, + "loss": 0.0055, + "step": 5961 + }, + { + "epoch": 7.742857142857143, + "grad_norm": 6.121458623991645, + "learning_rate": 2.557959865003501e-06, + "loss": 0.0081, + "step": 5962 + }, + { + "epoch": 7.744155844155844, + "grad_norm": 5.229591039856647, + "learning_rate": 2.555150998302026e-06, + "loss": 0.0092, + "step": 5963 + }, + { + "epoch": 7.745454545454545, + "grad_norm": 0.8663390125375069, + "learning_rate": 2.552343448736836e-06, + "loss": 0.0082, + "step": 5964 + }, + { + "epoch": 7.746753246753247, + "grad_norm": 8.442183322302776, + "learning_rate": 2.5495372168046383e-06, + "loss": 0.0155, + "step": 5965 + }, + { + "epoch": 7.748051948051948, + "grad_norm": 5.393699885489191, + "learning_rate": 2.5467323030019084e-06, + "loss": 0.0019, + "step": 5966 + }, + { + "epoch": 7.749350649350649, + "grad_norm": 4.672662013981407, + "learning_rate": 2.5439287078248965e-06, + "loss": 0.0015, + "step": 5967 + }, + { + "epoch": 7.750649350649351, + "grad_norm": 2.7733256548108636, + "learning_rate": 2.541126431769602e-06, + "loss": 0.0036, + "step": 5968 + }, + { + "epoch": 7.751948051948052, + "grad_norm": 4.605705334156269, + "learning_rate": 2.5383254753318075e-06, + "loss": 0.0066, + "step": 5969 + }, + { + "epoch": 7.753246753246753, + "grad_norm": 4.274369722539711, + "learning_rate": 2.5355258390070523e-06, + "loss": 0.0037, + "step": 5970 + }, + { + "epoch": 7.754545454545455, + "grad_norm": 3.0217230918492413, + "learning_rate": 2.5327275232906546e-06, + "loss": 0.0193, + "step": 5971 + }, + { + "epoch": 7.755844155844156, + "grad_norm": 10.496915101873315, + "learning_rate": 2.5299305286776767e-06, + "loss": 0.0149, + "step": 5972 + }, + { + "epoch": 7.757142857142857, + "grad_norm": 4.729916133415913, + "learning_rate": 2.5271348556629695e-06, + "loss": 0.015, + "step": 5973 + }, + { + "epoch": 7.758441558441558, + "grad_norm": 1.5867243801223754, + "learning_rate": 2.5243405047411353e-06, + "loss": 0.0015, + "step": 5974 + }, + { + "epoch": 7.759740259740259, + "grad_norm": 3.36454651249042, + "learning_rate": 2.5215474764065574e-06, + "loss": 0.0026, + "step": 5975 + }, + { + "epoch": 7.761038961038961, + "grad_norm": 2.9257602343628375, + "learning_rate": 2.518755771153364e-06, + "loss": 0.0024, + "step": 5976 + }, + { + "epoch": 7.7623376623376625, + "grad_norm": 4.738024675553123, + "learning_rate": 2.5159653894754708e-06, + "loss": 0.0156, + "step": 5977 + }, + { + "epoch": 7.763636363636364, + "grad_norm": 4.923537715658842, + "learning_rate": 2.5131763318665424e-06, + "loss": 0.0097, + "step": 5978 + }, + { + "epoch": 7.764935064935065, + "grad_norm": 1.1944051032487621, + "learning_rate": 2.510388598820027e-06, + "loss": 0.0012, + "step": 5979 + }, + { + "epoch": 7.766233766233766, + "grad_norm": 1.9008799028310486, + "learning_rate": 2.507602190829115e-06, + "loss": 0.0021, + "step": 5980 + }, + { + "epoch": 7.767532467532467, + "grad_norm": 6.9447505624225, + "learning_rate": 2.5048171083867844e-06, + "loss": 0.0054, + "step": 5981 + }, + { + "epoch": 7.768831168831169, + "grad_norm": 0.7536087182123893, + "learning_rate": 2.5020333519857667e-06, + "loss": 0.0014, + "step": 5982 + }, + { + "epoch": 7.77012987012987, + "grad_norm": 3.108136686389559, + "learning_rate": 2.4992509221185613e-06, + "loss": 0.0164, + "step": 5983 + }, + { + "epoch": 7.771428571428571, + "grad_norm": 4.785816513233684, + "learning_rate": 2.4964698192774304e-06, + "loss": 0.0224, + "step": 5984 + }, + { + "epoch": 7.7727272727272725, + "grad_norm": 7.44460224045512, + "learning_rate": 2.493690043954411e-06, + "loss": 0.0214, + "step": 5985 + }, + { + "epoch": 7.7740259740259745, + "grad_norm": 4.510961893575152, + "learning_rate": 2.4909115966412957e-06, + "loss": 0.0092, + "step": 5986 + }, + { + "epoch": 7.775324675324676, + "grad_norm": 4.357040689563775, + "learning_rate": 2.4881344778296433e-06, + "loss": 0.0109, + "step": 5987 + }, + { + "epoch": 7.776623376623377, + "grad_norm": 2.9545576679801395, + "learning_rate": 2.485358688010778e-06, + "loss": 0.0019, + "step": 5988 + }, + { + "epoch": 7.777922077922078, + "grad_norm": 3.494421161695982, + "learning_rate": 2.482584227675796e-06, + "loss": 0.0167, + "step": 5989 + }, + { + "epoch": 7.779220779220779, + "grad_norm": 3.7471102244668133, + "learning_rate": 2.479811097315549e-06, + "loss": 0.0104, + "step": 5990 + }, + { + "epoch": 7.78051948051948, + "grad_norm": 5.28629371301082, + "learning_rate": 2.477039297420656e-06, + "loss": 0.0245, + "step": 5991 + }, + { + "epoch": 7.781818181818182, + "grad_norm": 1.413573736286715, + "learning_rate": 2.4742688284815e-06, + "loss": 0.0191, + "step": 5992 + }, + { + "epoch": 7.783116883116883, + "grad_norm": 0.9087443730192274, + "learning_rate": 2.471499690988235e-06, + "loss": 0.0013, + "step": 5993 + }, + { + "epoch": 7.7844155844155845, + "grad_norm": 9.286283790942935, + "learning_rate": 2.468731885430772e-06, + "loss": 0.008, + "step": 5994 + }, + { + "epoch": 7.785714285714286, + "grad_norm": 5.840037809644084, + "learning_rate": 2.465965412298784e-06, + "loss": 0.0125, + "step": 5995 + }, + { + "epoch": 7.787012987012987, + "grad_norm": 4.055576206702553, + "learning_rate": 2.4632002720817227e-06, + "loss": 0.0074, + "step": 5996 + }, + { + "epoch": 7.788311688311689, + "grad_norm": 3.375412638939208, + "learning_rate": 2.4604364652687875e-06, + "loss": 0.0091, + "step": 5997 + }, + { + "epoch": 7.78961038961039, + "grad_norm": 2.945235121286538, + "learning_rate": 2.457673992348951e-06, + "loss": 0.0104, + "step": 5998 + }, + { + "epoch": 7.790909090909091, + "grad_norm": 3.489653260891861, + "learning_rate": 2.4549128538109446e-06, + "loss": 0.0087, + "step": 5999 + }, + { + "epoch": 7.792207792207792, + "grad_norm": 8.09679666344503, + "learning_rate": 2.452153050143271e-06, + "loss": 0.0071, + "step": 6000 + }, + { + "epoch": 7.793506493506493, + "grad_norm": 1.7560576229758824, + "learning_rate": 2.44939458183419e-06, + "loss": 0.0013, + "step": 6001 + }, + { + "epoch": 7.794805194805194, + "grad_norm": 0.7611433198979576, + "learning_rate": 2.4466374493717283e-06, + "loss": 0.0013, + "step": 6002 + }, + { + "epoch": 7.796103896103896, + "grad_norm": 5.988736734725621, + "learning_rate": 2.4438816532436717e-06, + "loss": 0.0071, + "step": 6003 + }, + { + "epoch": 7.797402597402598, + "grad_norm": 5.907846399702933, + "learning_rate": 2.441127193937578e-06, + "loss": 0.0169, + "step": 6004 + }, + { + "epoch": 7.798701298701299, + "grad_norm": 2.4104784465822586, + "learning_rate": 2.4383740719407634e-06, + "loss": 0.0105, + "step": 6005 + }, + { + "epoch": 7.8, + "grad_norm": 7.16436502837699, + "learning_rate": 2.4356222877403056e-06, + "loss": 0.0113, + "step": 6006 + }, + { + "epoch": 7.801298701298701, + "grad_norm": 3.268014323152024, + "learning_rate": 2.432871841823047e-06, + "loss": 0.0118, + "step": 6007 + }, + { + "epoch": 7.802597402597403, + "grad_norm": 5.06165627285094, + "learning_rate": 2.4301227346756006e-06, + "loss": 0.0211, + "step": 6008 + }, + { + "epoch": 7.803896103896104, + "grad_norm": 5.352156620796673, + "learning_rate": 2.427374966784327e-06, + "loss": 0.0112, + "step": 6009 + }, + { + "epoch": 7.805194805194805, + "grad_norm": 2.325003486701861, + "learning_rate": 2.4246285386353654e-06, + "loss": 0.0098, + "step": 6010 + }, + { + "epoch": 7.806493506493506, + "grad_norm": 4.309356802362503, + "learning_rate": 2.421883450714606e-06, + "loss": 0.0131, + "step": 6011 + }, + { + "epoch": 7.8077922077922075, + "grad_norm": 1.4644504476501037, + "learning_rate": 2.419139703507718e-06, + "loss": 0.0072, + "step": 6012 + }, + { + "epoch": 7.809090909090909, + "grad_norm": 1.1061751604926156, + "learning_rate": 2.416397297500109e-06, + "loss": 0.0017, + "step": 6013 + }, + { + "epoch": 7.810389610389611, + "grad_norm": 1.7870500135950833, + "learning_rate": 2.413656233176973e-06, + "loss": 0.0097, + "step": 6014 + }, + { + "epoch": 7.811688311688312, + "grad_norm": 0.9461095238913929, + "learning_rate": 2.41091651102325e-06, + "loss": 0.0075, + "step": 6015 + }, + { + "epoch": 7.812987012987013, + "grad_norm": 2.3131358219901648, + "learning_rate": 2.4081781315236595e-06, + "loss": 0.0084, + "step": 6016 + }, + { + "epoch": 7.814285714285714, + "grad_norm": 2.5208351869231365, + "learning_rate": 2.4054410951626604e-06, + "loss": 0.0149, + "step": 6017 + }, + { + "epoch": 7.815584415584416, + "grad_norm": 3.188157655599653, + "learning_rate": 2.4027054024244954e-06, + "loss": 0.0066, + "step": 6018 + }, + { + "epoch": 7.816883116883117, + "grad_norm": 7.657004538433645, + "learning_rate": 2.3999710537931564e-06, + "loss": 0.0143, + "step": 6019 + }, + { + "epoch": 7.818181818181818, + "grad_norm": 3.0970385686902775, + "learning_rate": 2.3972380497524096e-06, + "loss": 0.0017, + "step": 6020 + }, + { + "epoch": 7.8194805194805195, + "grad_norm": 4.418084327166766, + "learning_rate": 2.3945063907857647e-06, + "loss": 0.0088, + "step": 6021 + }, + { + "epoch": 7.820779220779221, + "grad_norm": 1.4895406745343736, + "learning_rate": 2.391776077376512e-06, + "loss": 0.0013, + "step": 6022 + }, + { + "epoch": 7.822077922077922, + "grad_norm": 9.001290428416652, + "learning_rate": 2.389047110007695e-06, + "loss": 0.0365, + "step": 6023 + }, + { + "epoch": 7.823376623376624, + "grad_norm": 2.0249586734776264, + "learning_rate": 2.3863194891621155e-06, + "loss": 0.0015, + "step": 6024 + }, + { + "epoch": 7.824675324675325, + "grad_norm": 7.631751939014535, + "learning_rate": 2.3835932153223494e-06, + "loss": 0.0211, + "step": 6025 + }, + { + "epoch": 7.825974025974026, + "grad_norm": 0.855265751199128, + "learning_rate": 2.3808682889707236e-06, + "loss": 0.0118, + "step": 6026 + }, + { + "epoch": 7.827272727272727, + "grad_norm": 1.9291924136795289, + "learning_rate": 2.3781447105893286e-06, + "loss": 0.002, + "step": 6027 + }, + { + "epoch": 7.828571428571428, + "grad_norm": 0.5538931451965184, + "learning_rate": 2.3754224806600157e-06, + "loss": 0.0014, + "step": 6028 + }, + { + "epoch": 7.82987012987013, + "grad_norm": 1.5245419271683271, + "learning_rate": 2.3727015996644043e-06, + "loss": 0.0018, + "step": 6029 + }, + { + "epoch": 7.8311688311688314, + "grad_norm": 3.4406707956577978, + "learning_rate": 2.369982068083868e-06, + "loss": 0.0106, + "step": 6030 + }, + { + "epoch": 7.832467532467533, + "grad_norm": 2.706241345175187, + "learning_rate": 2.3672638863995444e-06, + "loss": 0.002, + "step": 6031 + }, + { + "epoch": 7.833766233766234, + "grad_norm": 1.4472820676910871, + "learning_rate": 2.3645470550923275e-06, + "loss": 0.0018, + "step": 6032 + }, + { + "epoch": 7.835064935064935, + "grad_norm": 0.9919175353074183, + "learning_rate": 2.3618315746428844e-06, + "loss": 0.0023, + "step": 6033 + }, + { + "epoch": 7.836363636363636, + "grad_norm": 4.756199857292935, + "learning_rate": 2.3591174455316322e-06, + "loss": 0.0261, + "step": 6034 + }, + { + "epoch": 7.837662337662338, + "grad_norm": 4.136422445824406, + "learning_rate": 2.356404668238751e-06, + "loss": 0.004, + "step": 6035 + }, + { + "epoch": 7.838961038961039, + "grad_norm": 4.749675106013333, + "learning_rate": 2.3536932432441816e-06, + "loss": 0.0099, + "step": 6036 + }, + { + "epoch": 7.84025974025974, + "grad_norm": 2.741374766957237, + "learning_rate": 2.350983171027632e-06, + "loss": 0.0089, + "step": 6037 + }, + { + "epoch": 7.841558441558441, + "grad_norm": 6.563670907284776, + "learning_rate": 2.348274452068564e-06, + "loss": 0.008, + "step": 6038 + }, + { + "epoch": 7.8428571428571425, + "grad_norm": 2.4355552905543982, + "learning_rate": 2.3455670868461998e-06, + "loss": 0.009, + "step": 6039 + }, + { + "epoch": 7.8441558441558445, + "grad_norm": 6.104093450367728, + "learning_rate": 2.342861075839523e-06, + "loss": 0.0169, + "step": 6040 + }, + { + "epoch": 7.845454545454546, + "grad_norm": 2.4335693554066276, + "learning_rate": 2.340156419527284e-06, + "loss": 0.0179, + "step": 6041 + }, + { + "epoch": 7.846753246753247, + "grad_norm": 2.4467310179812096, + "learning_rate": 2.3374531183879856e-06, + "loss": 0.0019, + "step": 6042 + }, + { + "epoch": 7.848051948051948, + "grad_norm": 4.069758450204276, + "learning_rate": 2.334751172899893e-06, + "loss": 0.011, + "step": 6043 + }, + { + "epoch": 7.849350649350649, + "grad_norm": 2.161777789924935, + "learning_rate": 2.33205058354103e-06, + "loss": 0.0067, + "step": 6044 + }, + { + "epoch": 7.85064935064935, + "grad_norm": 1.7683208674444024, + "learning_rate": 2.3293513507891874e-06, + "loss": 0.0018, + "step": 6045 + }, + { + "epoch": 7.851948051948052, + "grad_norm": 1.2900336634536518, + "learning_rate": 2.3266534751219084e-06, + "loss": 0.0083, + "step": 6046 + }, + { + "epoch": 7.853246753246753, + "grad_norm": 1.7796754835184743, + "learning_rate": 2.3239569570164987e-06, + "loss": 0.0017, + "step": 6047 + }, + { + "epoch": 7.8545454545454545, + "grad_norm": 2.1012056209002403, + "learning_rate": 2.3212617969500216e-06, + "loss": 0.003, + "step": 6048 + }, + { + "epoch": 7.855844155844156, + "grad_norm": 6.132159942822693, + "learning_rate": 2.3185679953993114e-06, + "loss": 0.0423, + "step": 6049 + }, + { + "epoch": 7.857142857142857, + "grad_norm": 4.71848096923287, + "learning_rate": 2.3158755528409415e-06, + "loss": 0.0075, + "step": 6050 + }, + { + "epoch": 7.858441558441559, + "grad_norm": 5.329175378575104, + "learning_rate": 2.313184469751265e-06, + "loss": 0.0015, + "step": 6051 + }, + { + "epoch": 7.85974025974026, + "grad_norm": 6.071519533343373, + "learning_rate": 2.3104947466063785e-06, + "loss": 0.0188, + "step": 6052 + }, + { + "epoch": 7.861038961038961, + "grad_norm": 2.0038642054640508, + "learning_rate": 2.3078063838821574e-06, + "loss": 0.0026, + "step": 6053 + }, + { + "epoch": 7.862337662337662, + "grad_norm": 3.921650217846772, + "learning_rate": 2.3051193820542118e-06, + "loss": 0.0036, + "step": 6054 + }, + { + "epoch": 7.863636363636363, + "grad_norm": 5.27355744459837, + "learning_rate": 2.3024337415979304e-06, + "loss": 0.0026, + "step": 6055 + }, + { + "epoch": 7.8649350649350644, + "grad_norm": 5.548069086497052, + "learning_rate": 2.299749462988451e-06, + "loss": 0.0147, + "step": 6056 + }, + { + "epoch": 7.8662337662337665, + "grad_norm": 3.4991395336944975, + "learning_rate": 2.297066546700679e-06, + "loss": 0.0183, + "step": 6057 + }, + { + "epoch": 7.867532467532468, + "grad_norm": 2.2741347120552713, + "learning_rate": 2.2943849932092697e-06, + "loss": 0.0079, + "step": 6058 + }, + { + "epoch": 7.868831168831169, + "grad_norm": 2.7239411792015598, + "learning_rate": 2.2917048029886436e-06, + "loss": 0.0103, + "step": 6059 + }, + { + "epoch": 7.87012987012987, + "grad_norm": 1.2696935322912162, + "learning_rate": 2.289025976512973e-06, + "loss": 0.0011, + "step": 6060 + }, + { + "epoch": 7.871428571428572, + "grad_norm": 0.7639421407251876, + "learning_rate": 2.286348514256199e-06, + "loss": 0.0013, + "step": 6061 + }, + { + "epoch": 7.872727272727273, + "grad_norm": 2.6310727625922876, + "learning_rate": 2.2836724166920134e-06, + "loss": 0.0149, + "step": 6062 + }, + { + "epoch": 7.874025974025974, + "grad_norm": 1.1049818341427933, + "learning_rate": 2.2809976842938695e-06, + "loss": 0.0159, + "step": 6063 + }, + { + "epoch": 7.875324675324675, + "grad_norm": 3.716906981354547, + "learning_rate": 2.278324317534979e-06, + "loss": 0.009, + "step": 6064 + }, + { + "epoch": 7.876623376623376, + "grad_norm": 4.227185356395917, + "learning_rate": 2.2756523168883083e-06, + "loss": 0.0086, + "step": 6065 + }, + { + "epoch": 7.8779220779220775, + "grad_norm": 1.8280303677093277, + "learning_rate": 2.2729816828265914e-06, + "loss": 0.0015, + "step": 6066 + }, + { + "epoch": 7.87922077922078, + "grad_norm": 0.6925033761426429, + "learning_rate": 2.270312415822309e-06, + "loss": 0.0013, + "step": 6067 + }, + { + "epoch": 7.880519480519481, + "grad_norm": 1.8404999778199713, + "learning_rate": 2.2676445163477123e-06, + "loss": 0.0017, + "step": 6068 + }, + { + "epoch": 7.881818181818182, + "grad_norm": 9.278146344492603, + "learning_rate": 2.2649779848747954e-06, + "loss": 0.007, + "step": 6069 + }, + { + "epoch": 7.883116883116883, + "grad_norm": 2.721992834493842, + "learning_rate": 2.2623128218753244e-06, + "loss": 0.0016, + "step": 6070 + }, + { + "epoch": 7.884415584415584, + "grad_norm": 1.5475678268355262, + "learning_rate": 2.259649027820814e-06, + "loss": 0.0064, + "step": 6071 + }, + { + "epoch": 7.885714285714286, + "grad_norm": 1.0015983569410614, + "learning_rate": 2.2569866031825483e-06, + "loss": 0.0078, + "step": 6072 + }, + { + "epoch": 7.887012987012987, + "grad_norm": 5.365650162817737, + "learning_rate": 2.2543255484315486e-06, + "loss": 0.0125, + "step": 6073 + }, + { + "epoch": 7.888311688311688, + "grad_norm": 4.687643844826618, + "learning_rate": 2.251665864038617e-06, + "loss": 0.0021, + "step": 6074 + }, + { + "epoch": 7.8896103896103895, + "grad_norm": 3.6285388400429617, + "learning_rate": 2.249007550474299e-06, + "loss": 0.0031, + "step": 6075 + }, + { + "epoch": 7.890909090909091, + "grad_norm": 1.7542604071290169, + "learning_rate": 2.2463506082089005e-06, + "loss": 0.0021, + "step": 6076 + }, + { + "epoch": 7.892207792207792, + "grad_norm": 5.74857361356176, + "learning_rate": 2.243695037712482e-06, + "loss": 0.0119, + "step": 6077 + }, + { + "epoch": 7.893506493506494, + "grad_norm": 3.7046986081184277, + "learning_rate": 2.2410408394548735e-06, + "loss": 0.0152, + "step": 6078 + }, + { + "epoch": 7.894805194805195, + "grad_norm": 2.1727113578725943, + "learning_rate": 2.238388013905647e-06, + "loss": 0.0018, + "step": 6079 + }, + { + "epoch": 7.896103896103896, + "grad_norm": 2.73762295695192, + "learning_rate": 2.23573656153414e-06, + "loss": 0.002, + "step": 6080 + }, + { + "epoch": 7.897402597402597, + "grad_norm": 1.432984152982598, + "learning_rate": 2.233086482809442e-06, + "loss": 0.0093, + "step": 6081 + }, + { + "epoch": 7.898701298701298, + "grad_norm": 2.5603011410543917, + "learning_rate": 2.2304377782004083e-06, + "loss": 0.0182, + "step": 6082 + }, + { + "epoch": 7.9, + "grad_norm": 0.746365888871478, + "learning_rate": 2.227790448175643e-06, + "loss": 0.0012, + "step": 6083 + }, + { + "epoch": 7.9012987012987015, + "grad_norm": 0.9184232784178485, + "learning_rate": 2.2251444932035094e-06, + "loss": 0.0018, + "step": 6084 + }, + { + "epoch": 7.902597402597403, + "grad_norm": 5.006393172413457, + "learning_rate": 2.2224999137521232e-06, + "loss": 0.0112, + "step": 6085 + }, + { + "epoch": 7.903896103896104, + "grad_norm": 6.6469171477876055, + "learning_rate": 2.2198567102893685e-06, + "loss": 0.0306, + "step": 6086 + }, + { + "epoch": 7.905194805194805, + "grad_norm": 3.2889035329115757, + "learning_rate": 2.2172148832828756e-06, + "loss": 0.0157, + "step": 6087 + }, + { + "epoch": 7.906493506493506, + "grad_norm": 5.321321304363839, + "learning_rate": 2.214574433200034e-06, + "loss": 0.0089, + "step": 6088 + }, + { + "epoch": 7.907792207792208, + "grad_norm": 7.679329695921203, + "learning_rate": 2.2119353605079876e-06, + "loss": 0.0097, + "step": 6089 + }, + { + "epoch": 7.909090909090909, + "grad_norm": 5.330789413900633, + "learning_rate": 2.2092976656736433e-06, + "loss": 0.0143, + "step": 6090 + }, + { + "epoch": 7.91038961038961, + "grad_norm": 3.6641261473428943, + "learning_rate": 2.2066613491636568e-06, + "loss": 0.0025, + "step": 6091 + }, + { + "epoch": 7.911688311688311, + "grad_norm": 7.112754935835955, + "learning_rate": 2.2040264114444443e-06, + "loss": 0.0079, + "step": 6092 + }, + { + "epoch": 7.9129870129870135, + "grad_norm": 2.2257783845005648, + "learning_rate": 2.201392852982174e-06, + "loss": 0.008, + "step": 6093 + }, + { + "epoch": 7.914285714285715, + "grad_norm": 5.558376439233083, + "learning_rate": 2.198760674242777e-06, + "loss": 0.0071, + "step": 6094 + }, + { + "epoch": 7.915584415584416, + "grad_norm": 3.957581555720251, + "learning_rate": 2.1961298756919337e-06, + "loss": 0.0072, + "step": 6095 + }, + { + "epoch": 7.916883116883117, + "grad_norm": 5.456201484709295, + "learning_rate": 2.1935004577950803e-06, + "loss": 0.0116, + "step": 6096 + }, + { + "epoch": 7.918181818181818, + "grad_norm": 1.9540573246541908, + "learning_rate": 2.190872421017417e-06, + "loss": 0.0087, + "step": 6097 + }, + { + "epoch": 7.919480519480519, + "grad_norm": 5.821453865369008, + "learning_rate": 2.188245765823891e-06, + "loss": 0.0118, + "step": 6098 + }, + { + "epoch": 7.920779220779221, + "grad_norm": 3.641748362331131, + "learning_rate": 2.185620492679208e-06, + "loss": 0.0129, + "step": 6099 + }, + { + "epoch": 7.922077922077922, + "grad_norm": 6.403015234196248, + "learning_rate": 2.1829966020478254e-06, + "loss": 0.0052, + "step": 6100 + }, + { + "epoch": 7.923376623376623, + "grad_norm": 1.5922561643431237, + "learning_rate": 2.1803740943939654e-06, + "loss": 0.0021, + "step": 6101 + }, + { + "epoch": 7.9246753246753245, + "grad_norm": 0.7388494805058692, + "learning_rate": 2.177752970181598e-06, + "loss": 0.0013, + "step": 6102 + }, + { + "epoch": 7.925974025974026, + "grad_norm": 3.5419971976771887, + "learning_rate": 2.1751332298744487e-06, + "loss": 0.0093, + "step": 6103 + }, + { + "epoch": 7.927272727272728, + "grad_norm": 3.557361672013312, + "learning_rate": 2.172514873935998e-06, + "loss": 0.0018, + "step": 6104 + }, + { + "epoch": 7.928571428571429, + "grad_norm": 2.7743796319886123, + "learning_rate": 2.169897902829492e-06, + "loss": 0.0028, + "step": 6105 + }, + { + "epoch": 7.92987012987013, + "grad_norm": 0.3846658401046506, + "learning_rate": 2.16728231701791e-06, + "loss": 0.001, + "step": 6106 + }, + { + "epoch": 7.931168831168831, + "grad_norm": 6.5631504681665245, + "learning_rate": 2.1646681169640105e-06, + "loss": 0.0155, + "step": 6107 + }, + { + "epoch": 7.932467532467532, + "grad_norm": 5.423768799987004, + "learning_rate": 2.162055303130287e-06, + "loss": 0.0042, + "step": 6108 + }, + { + "epoch": 7.933766233766233, + "grad_norm": 5.7180082231847935, + "learning_rate": 2.159443875979006e-06, + "loss": 0.0095, + "step": 6109 + }, + { + "epoch": 7.935064935064935, + "grad_norm": 5.099191107265713, + "learning_rate": 2.1568338359721675e-06, + "loss": 0.0045, + "step": 6110 + }, + { + "epoch": 7.9363636363636365, + "grad_norm": 3.8085605283905295, + "learning_rate": 2.1542251835715457e-06, + "loss": 0.0154, + "step": 6111 + }, + { + "epoch": 7.937662337662338, + "grad_norm": 4.017072954958327, + "learning_rate": 2.1516179192386568e-06, + "loss": 0.0064, + "step": 6112 + }, + { + "epoch": 7.938961038961039, + "grad_norm": 2.1779936153490165, + "learning_rate": 2.1490120434347827e-06, + "loss": 0.0049, + "step": 6113 + }, + { + "epoch": 7.94025974025974, + "grad_norm": 0.7091055637514028, + "learning_rate": 2.146407556620942e-06, + "loss": 0.0017, + "step": 6114 + }, + { + "epoch": 7.941558441558442, + "grad_norm": 2.4149069884304697, + "learning_rate": 2.1438044592579265e-06, + "loss": 0.0156, + "step": 6115 + }, + { + "epoch": 7.942857142857143, + "grad_norm": 5.302381381246561, + "learning_rate": 2.141202751806272e-06, + "loss": 0.007, + "step": 6116 + }, + { + "epoch": 7.944155844155844, + "grad_norm": 2.4170946098949, + "learning_rate": 2.1386024347262692e-06, + "loss": 0.0133, + "step": 6117 + }, + { + "epoch": 7.945454545454545, + "grad_norm": 4.492089647304752, + "learning_rate": 2.1360035084779608e-06, + "loss": 0.0031, + "step": 6118 + }, + { + "epoch": 7.9467532467532465, + "grad_norm": 0.980002866078944, + "learning_rate": 2.133405973521154e-06, + "loss": 0.0133, + "step": 6119 + }, + { + "epoch": 7.948051948051948, + "grad_norm": 3.2292030604145148, + "learning_rate": 2.130809830315399e-06, + "loss": 0.0048, + "step": 6120 + }, + { + "epoch": 7.94935064935065, + "grad_norm": 3.7995763118507466, + "learning_rate": 2.128215079320002e-06, + "loss": 0.0146, + "step": 6121 + }, + { + "epoch": 7.950649350649351, + "grad_norm": 0.6100654790092426, + "learning_rate": 2.1256217209940233e-06, + "loss": 0.0011, + "step": 6122 + }, + { + "epoch": 7.951948051948052, + "grad_norm": 0.5887713369982736, + "learning_rate": 2.123029755796282e-06, + "loss": 0.0014, + "step": 6123 + }, + { + "epoch": 7.953246753246753, + "grad_norm": 2.930561966593572, + "learning_rate": 2.1204391841853434e-06, + "loss": 0.0147, + "step": 6124 + }, + { + "epoch": 7.954545454545455, + "grad_norm": 0.8087184083963925, + "learning_rate": 2.1178500066195286e-06, + "loss": 0.0086, + "step": 6125 + }, + { + "epoch": 7.955844155844156, + "grad_norm": 3.072372618063828, + "learning_rate": 2.115262223556912e-06, + "loss": 0.0092, + "step": 6126 + }, + { + "epoch": 7.957142857142857, + "grad_norm": 1.5096055882806978, + "learning_rate": 2.1126758354553258e-06, + "loss": 0.0016, + "step": 6127 + }, + { + "epoch": 7.958441558441558, + "grad_norm": 1.3271774483009195, + "learning_rate": 2.11009084277235e-06, + "loss": 0.0082, + "step": 6128 + }, + { + "epoch": 7.95974025974026, + "grad_norm": 0.8768648026149877, + "learning_rate": 2.1075072459653144e-06, + "loss": 0.0012, + "step": 6129 + }, + { + "epoch": 7.961038961038961, + "grad_norm": 2.4666781938213043, + "learning_rate": 2.104925045491315e-06, + "loss": 0.0028, + "step": 6130 + }, + { + "epoch": 7.962337662337663, + "grad_norm": 4.116991135127574, + "learning_rate": 2.102344241807188e-06, + "loss": 0.011, + "step": 6131 + }, + { + "epoch": 7.963636363636364, + "grad_norm": 6.4232436964618245, + "learning_rate": 2.0997648353695277e-06, + "loss": 0.0049, + "step": 6132 + }, + { + "epoch": 7.964935064935065, + "grad_norm": 2.935013088389416, + "learning_rate": 2.0971868266346774e-06, + "loss": 0.0015, + "step": 6133 + }, + { + "epoch": 7.966233766233766, + "grad_norm": 1.3251769898013446, + "learning_rate": 2.0946102160587413e-06, + "loss": 0.0012, + "step": 6134 + }, + { + "epoch": 7.967532467532467, + "grad_norm": 1.3935233081893057, + "learning_rate": 2.092035004097568e-06, + "loss": 0.0082, + "step": 6135 + }, + { + "epoch": 7.968831168831169, + "grad_norm": 0.5743035966491, + "learning_rate": 2.089461191206764e-06, + "loss": 0.0083, + "step": 6136 + }, + { + "epoch": 7.97012987012987, + "grad_norm": 3.1962578226573375, + "learning_rate": 2.086888777841681e-06, + "loss": 0.004, + "step": 6137 + }, + { + "epoch": 7.9714285714285715, + "grad_norm": 1.9310204681663627, + "learning_rate": 2.084317764457434e-06, + "loss": 0.0078, + "step": 6138 + }, + { + "epoch": 7.972727272727273, + "grad_norm": 1.6893968481236927, + "learning_rate": 2.081748151508883e-06, + "loss": 0.0091, + "step": 6139 + }, + { + "epoch": 7.974025974025974, + "grad_norm": 1.2960987533647401, + "learning_rate": 2.079179939450642e-06, + "loss": 0.002, + "step": 6140 + }, + { + "epoch": 7.975324675324675, + "grad_norm": 3.730036494499076, + "learning_rate": 2.0766131287370715e-06, + "loss": 0.0065, + "step": 6141 + }, + { + "epoch": 7.976623376623377, + "grad_norm": 2.956124462405178, + "learning_rate": 2.074047719822296e-06, + "loss": 0.0205, + "step": 6142 + }, + { + "epoch": 7.977922077922078, + "grad_norm": 4.991928365550799, + "learning_rate": 2.071483713160185e-06, + "loss": 0.0063, + "step": 6143 + }, + { + "epoch": 7.979220779220779, + "grad_norm": 6.701575611381533, + "learning_rate": 2.068921109204358e-06, + "loss": 0.0138, + "step": 6144 + }, + { + "epoch": 7.98051948051948, + "grad_norm": 2.054929158803742, + "learning_rate": 2.0663599084081865e-06, + "loss": 0.0018, + "step": 6145 + }, + { + "epoch": 7.9818181818181815, + "grad_norm": 3.1383265000416602, + "learning_rate": 2.063800111224803e-06, + "loss": 0.0056, + "step": 6146 + }, + { + "epoch": 7.9831168831168835, + "grad_norm": 1.493526190908136, + "learning_rate": 2.0612417181070766e-06, + "loss": 0.0016, + "step": 6147 + }, + { + "epoch": 7.984415584415585, + "grad_norm": 1.6703321482877636, + "learning_rate": 2.0586847295076407e-06, + "loss": 0.0086, + "step": 6148 + }, + { + "epoch": 7.985714285714286, + "grad_norm": 3.165711050991919, + "learning_rate": 2.0561291458788736e-06, + "loss": 0.0021, + "step": 6149 + }, + { + "epoch": 7.987012987012987, + "grad_norm": 1.5246789601829873, + "learning_rate": 2.0535749676729123e-06, + "loss": 0.0088, + "step": 6150 + }, + { + "epoch": 7.988311688311688, + "grad_norm": 2.139940573285958, + "learning_rate": 2.0510221953416296e-06, + "loss": 0.0061, + "step": 6151 + }, + { + "epoch": 7.989610389610389, + "grad_norm": 2.746652047604029, + "learning_rate": 2.0484708293366697e-06, + "loss": 0.003, + "step": 6152 + }, + { + "epoch": 7.990909090909091, + "grad_norm": 4.87605703510757, + "learning_rate": 2.0459208701094103e-06, + "loss": 0.0173, + "step": 6153 + }, + { + "epoch": 7.992207792207792, + "grad_norm": 1.4574958904922952, + "learning_rate": 2.0433723181109964e-06, + "loss": 0.0076, + "step": 6154 + }, + { + "epoch": 7.9935064935064934, + "grad_norm": 1.4177190431837097, + "learning_rate": 2.0408251737923067e-06, + "loss": 0.0021, + "step": 6155 + }, + { + "epoch": 7.994805194805195, + "grad_norm": 2.7180336872541164, + "learning_rate": 2.038279437603986e-06, + "loss": 0.0011, + "step": 6156 + }, + { + "epoch": 7.996103896103897, + "grad_norm": 3.107942877814491, + "learning_rate": 2.035735109996423e-06, + "loss": 0.0115, + "step": 6157 + }, + { + "epoch": 7.997402597402598, + "grad_norm": 1.869349161523487, + "learning_rate": 2.0331921914197527e-06, + "loss": 0.0071, + "step": 6158 + }, + { + "epoch": 7.998701298701299, + "grad_norm": 2.8710704252730643, + "learning_rate": 2.0306506823238727e-06, + "loss": 0.0047, + "step": 6159 + }, + { + "epoch": 8.0, + "grad_norm": 3.7997767760520706, + "learning_rate": 2.0281105831584223e-06, + "loss": 0.004, + "step": 6160 + }, + { + "epoch": 8.0, + "eval_accuracy": 0.9382369455362156, + "eval_f1": 0.9255424907145756, + "eval_loss": 0.14838746190071106, + "eval_precision": 0.9183090597157983, + "eval_recall": 0.936851277384922, + "eval_runtime": 12.8924, + "eval_samples_per_second": 138.143, + "eval_steps_per_second": 1.086, + "step": 6160 + }, + { + "epoch": 8.001298701298701, + "grad_norm": 2.5631870611635357, + "learning_rate": 2.025571894372794e-06, + "loss": 0.0083, + "step": 6161 + }, + { + "epoch": 8.002597402597402, + "grad_norm": 4.273470969654459, + "learning_rate": 2.0230346164161263e-06, + "loss": 0.0087, + "step": 6162 + }, + { + "epoch": 8.003896103896103, + "grad_norm": 1.0232083991870053, + "learning_rate": 2.0204987497373184e-06, + "loss": 0.0018, + "step": 6163 + }, + { + "epoch": 8.005194805194805, + "grad_norm": 11.404978869232227, + "learning_rate": 2.017964294785011e-06, + "loss": 0.0153, + "step": 6164 + }, + { + "epoch": 8.006493506493506, + "grad_norm": 1.3348246970943178, + "learning_rate": 2.015431252007598e-06, + "loss": 0.0013, + "step": 6165 + }, + { + "epoch": 8.007792207792209, + "grad_norm": 13.811032329289876, + "learning_rate": 2.0128996218532206e-06, + "loss": 0.0135, + "step": 6166 + }, + { + "epoch": 8.00909090909091, + "grad_norm": 1.332084832866817, + "learning_rate": 2.0103694047697766e-06, + "loss": 0.0013, + "step": 6167 + }, + { + "epoch": 8.01038961038961, + "grad_norm": 2.640411449755205, + "learning_rate": 2.0078406012049056e-06, + "loss": 0.0015, + "step": 6168 + }, + { + "epoch": 8.011688311688312, + "grad_norm": 6.0230971173495655, + "learning_rate": 2.005313211606009e-06, + "loss": 0.0068, + "step": 6169 + }, + { + "epoch": 8.012987012987013, + "grad_norm": 6.049447646131978, + "learning_rate": 2.0027872364202194e-06, + "loss": 0.0148, + "step": 6170 + }, + { + "epoch": 8.014285714285714, + "grad_norm": 0.6615123752235041, + "learning_rate": 2.00026267609444e-06, + "loss": 0.0012, + "step": 6171 + }, + { + "epoch": 8.015584415584415, + "grad_norm": 2.382293831732463, + "learning_rate": 1.9977395310753087e-06, + "loss": 0.002, + "step": 6172 + }, + { + "epoch": 8.016883116883117, + "grad_norm": 2.8390460623948894, + "learning_rate": 1.9952178018092193e-06, + "loss": 0.0093, + "step": 6173 + }, + { + "epoch": 8.018181818181818, + "grad_norm": 1.3550845725255842, + "learning_rate": 1.9926974887423123e-06, + "loss": 0.0078, + "step": 6174 + }, + { + "epoch": 8.019480519480519, + "grad_norm": 11.676443938317798, + "learning_rate": 1.990178592320482e-06, + "loss": 0.02, + "step": 6175 + }, + { + "epoch": 8.02077922077922, + "grad_norm": 3.0501531526030594, + "learning_rate": 1.9876611129893696e-06, + "loss": 0.0018, + "step": 6176 + }, + { + "epoch": 8.022077922077923, + "grad_norm": 1.7386177170175892, + "learning_rate": 1.985145051194364e-06, + "loss": 0.0026, + "step": 6177 + }, + { + "epoch": 8.023376623376624, + "grad_norm": 8.593258618890752, + "learning_rate": 1.9826304073806013e-06, + "loss": 0.0117, + "step": 6178 + }, + { + "epoch": 8.024675324675325, + "grad_norm": 1.494039389895626, + "learning_rate": 1.9801171819929766e-06, + "loss": 0.0021, + "step": 6179 + }, + { + "epoch": 8.025974025974026, + "grad_norm": 2.2312495936191694, + "learning_rate": 1.977605375476125e-06, + "loss": 0.0039, + "step": 6180 + }, + { + "epoch": 8.027272727272727, + "grad_norm": 5.346645409228642, + "learning_rate": 1.975094988274433e-06, + "loss": 0.0055, + "step": 6181 + }, + { + "epoch": 8.028571428571428, + "grad_norm": 2.1012729669670325, + "learning_rate": 1.9725860208320325e-06, + "loss": 0.0077, + "step": 6182 + }, + { + "epoch": 8.02987012987013, + "grad_norm": 6.396179368937277, + "learning_rate": 1.970078473592817e-06, + "loss": 0.0131, + "step": 6183 + }, + { + "epoch": 8.03116883116883, + "grad_norm": 1.6061901176410962, + "learning_rate": 1.967572347000408e-06, + "loss": 0.0017, + "step": 6184 + }, + { + "epoch": 8.032467532467532, + "grad_norm": 3.817509876073159, + "learning_rate": 1.965067641498197e-06, + "loss": 0.0196, + "step": 6185 + }, + { + "epoch": 8.033766233766233, + "grad_norm": 1.8487184615324646, + "learning_rate": 1.9625643575293065e-06, + "loss": 0.0123, + "step": 6186 + }, + { + "epoch": 8.035064935064936, + "grad_norm": 1.9038842650034808, + "learning_rate": 1.9600624955366255e-06, + "loss": 0.0247, + "step": 6187 + }, + { + "epoch": 8.036363636363637, + "grad_norm": 5.852486216876051, + "learning_rate": 1.9575620559627696e-06, + "loss": 0.0069, + "step": 6188 + }, + { + "epoch": 8.037662337662338, + "grad_norm": 3.0737949812632968, + "learning_rate": 1.9550630392501225e-06, + "loss": 0.0045, + "step": 6189 + }, + { + "epoch": 8.03896103896104, + "grad_norm": 1.335518193217451, + "learning_rate": 1.9525654458408027e-06, + "loss": 0.0018, + "step": 6190 + }, + { + "epoch": 8.04025974025974, + "grad_norm": 1.3676539726299306, + "learning_rate": 1.950069276176687e-06, + "loss": 0.0078, + "step": 6191 + }, + { + "epoch": 8.041558441558442, + "grad_norm": 4.406696871638486, + "learning_rate": 1.947574530699394e-06, + "loss": 0.0033, + "step": 6192 + }, + { + "epoch": 8.042857142857143, + "grad_norm": 2.3045745672736038, + "learning_rate": 1.9450812098502925e-06, + "loss": 0.002, + "step": 6193 + }, + { + "epoch": 8.044155844155844, + "grad_norm": 1.3902055314671553, + "learning_rate": 1.942589314070494e-06, + "loss": 0.0014, + "step": 6194 + }, + { + "epoch": 8.045454545454545, + "grad_norm": 1.2434366107105692, + "learning_rate": 1.9400988438008705e-06, + "loss": 0.0024, + "step": 6195 + }, + { + "epoch": 8.046753246753246, + "grad_norm": 4.956373761017426, + "learning_rate": 1.9376097994820286e-06, + "loss": 0.0099, + "step": 6196 + }, + { + "epoch": 8.048051948051947, + "grad_norm": 5.959862790403066, + "learning_rate": 1.93512218155433e-06, + "loss": 0.0097, + "step": 6197 + }, + { + "epoch": 8.04935064935065, + "grad_norm": 13.93754448938105, + "learning_rate": 1.9326359904578796e-06, + "loss": 0.0135, + "step": 6198 + }, + { + "epoch": 8.050649350649351, + "grad_norm": 1.1956865344180183, + "learning_rate": 1.930151226632531e-06, + "loss": 0.0079, + "step": 6199 + }, + { + "epoch": 8.051948051948052, + "grad_norm": 1.6811122769221352, + "learning_rate": 1.9276678905178916e-06, + "loss": 0.0026, + "step": 6200 + }, + { + "epoch": 8.053246753246754, + "grad_norm": 8.526980987307132, + "learning_rate": 1.925185982553306e-06, + "loss": 0.0131, + "step": 6201 + }, + { + "epoch": 8.054545454545455, + "grad_norm": 2.433204132492946, + "learning_rate": 1.9227055031778794e-06, + "loss": 0.0086, + "step": 6202 + }, + { + "epoch": 8.055844155844156, + "grad_norm": 8.597968962086945, + "learning_rate": 1.9202264528304447e-06, + "loss": 0.0297, + "step": 6203 + }, + { + "epoch": 8.057142857142857, + "grad_norm": 3.0666473055919776, + "learning_rate": 1.917748831949602e-06, + "loss": 0.0093, + "step": 6204 + }, + { + "epoch": 8.058441558441558, + "grad_norm": 3.7155265888444866, + "learning_rate": 1.9152726409736834e-06, + "loss": 0.0024, + "step": 6205 + }, + { + "epoch": 8.05974025974026, + "grad_norm": 3.9127051426248762, + "learning_rate": 1.912797880340783e-06, + "loss": 0.0174, + "step": 6206 + }, + { + "epoch": 8.06103896103896, + "grad_norm": 6.750534330457494, + "learning_rate": 1.9103245504887245e-06, + "loss": 0.0192, + "step": 6207 + }, + { + "epoch": 8.062337662337661, + "grad_norm": 1.859968153384958, + "learning_rate": 1.907852651855093e-06, + "loss": 0.0019, + "step": 6208 + }, + { + "epoch": 8.063636363636364, + "grad_norm": 3.553086589997573, + "learning_rate": 1.9053821848772104e-06, + "loss": 0.0103, + "step": 6209 + }, + { + "epoch": 8.064935064935066, + "grad_norm": 2.221279987103835, + "learning_rate": 1.9029131499921561e-06, + "loss": 0.0014, + "step": 6210 + }, + { + "epoch": 8.066233766233767, + "grad_norm": 2.629261161180196, + "learning_rate": 1.9004455476367411e-06, + "loss": 0.0081, + "step": 6211 + }, + { + "epoch": 8.067532467532468, + "grad_norm": 1.558854885519962, + "learning_rate": 1.8979793782475376e-06, + "loss": 0.0088, + "step": 6212 + }, + { + "epoch": 8.068831168831169, + "grad_norm": 1.2393912711042445, + "learning_rate": 1.8955146422608572e-06, + "loss": 0.0014, + "step": 6213 + }, + { + "epoch": 8.07012987012987, + "grad_norm": 2.296977328909177, + "learning_rate": 1.893051340112758e-06, + "loss": 0.0025, + "step": 6214 + }, + { + "epoch": 8.071428571428571, + "grad_norm": 1.2106512388548374, + "learning_rate": 1.8905894722390427e-06, + "loss": 0.0032, + "step": 6215 + }, + { + "epoch": 8.072727272727272, + "grad_norm": 2.1002593697470466, + "learning_rate": 1.8881290390752682e-06, + "loss": 0.0053, + "step": 6216 + }, + { + "epoch": 8.074025974025973, + "grad_norm": 2.5504782065079703, + "learning_rate": 1.8856700410567297e-06, + "loss": 0.01, + "step": 6217 + }, + { + "epoch": 8.075324675324675, + "grad_norm": 3.6146217312955953, + "learning_rate": 1.8832124786184712e-06, + "loss": 0.0041, + "step": 6218 + }, + { + "epoch": 8.076623376623377, + "grad_norm": 3.717636113459934, + "learning_rate": 1.8807563521952787e-06, + "loss": 0.0044, + "step": 6219 + }, + { + "epoch": 8.077922077922079, + "grad_norm": 7.161262540296423, + "learning_rate": 1.8783016622216943e-06, + "loss": 0.0069, + "step": 6220 + }, + { + "epoch": 8.07922077922078, + "grad_norm": 0.9772937981528247, + "learning_rate": 1.8758484091319972e-06, + "loss": 0.0017, + "step": 6221 + }, + { + "epoch": 8.08051948051948, + "grad_norm": 1.8338253312858874, + "learning_rate": 1.8733965933602138e-06, + "loss": 0.0022, + "step": 6222 + }, + { + "epoch": 8.081818181818182, + "grad_norm": 0.6784390061217497, + "learning_rate": 1.870946215340116e-06, + "loss": 0.0012, + "step": 6223 + }, + { + "epoch": 8.083116883116883, + "grad_norm": 0.464179999377292, + "learning_rate": 1.8684972755052267e-06, + "loss": 0.0015, + "step": 6224 + }, + { + "epoch": 8.084415584415584, + "grad_norm": 0.9923128467022609, + "learning_rate": 1.8660497742888083e-06, + "loss": 0.0014, + "step": 6225 + }, + { + "epoch": 8.085714285714285, + "grad_norm": 0.3463286446322456, + "learning_rate": 1.8636037121238705e-06, + "loss": 0.0011, + "step": 6226 + }, + { + "epoch": 8.087012987012987, + "grad_norm": 1.7770161100482025, + "learning_rate": 1.8611590894431653e-06, + "loss": 0.0102, + "step": 6227 + }, + { + "epoch": 8.088311688311688, + "grad_norm": 3.956908017957818, + "learning_rate": 1.8587159066791982e-06, + "loss": 0.0061, + "step": 6228 + }, + { + "epoch": 8.089610389610389, + "grad_norm": 3.3369209413906664, + "learning_rate": 1.8562741642642135e-06, + "loss": 0.008, + "step": 6229 + }, + { + "epoch": 8.090909090909092, + "grad_norm": 1.4836769501611007, + "learning_rate": 1.8538338626301999e-06, + "loss": 0.0027, + "step": 6230 + }, + { + "epoch": 8.092207792207793, + "grad_norm": 1.9002348881377848, + "learning_rate": 1.8513950022088966e-06, + "loss": 0.0081, + "step": 6231 + }, + { + "epoch": 8.093506493506494, + "grad_norm": 2.9704110334604974, + "learning_rate": 1.8489575834317841e-06, + "loss": 0.0087, + "step": 6232 + }, + { + "epoch": 8.094805194805195, + "grad_norm": 2.414887170229808, + "learning_rate": 1.8465216067300885e-06, + "loss": 0.0011, + "step": 6233 + }, + { + "epoch": 8.096103896103896, + "grad_norm": 9.12797204737431, + "learning_rate": 1.844087072534776e-06, + "loss": 0.0128, + "step": 6234 + }, + { + "epoch": 8.097402597402597, + "grad_norm": 1.9259344399106146, + "learning_rate": 1.8416539812765689e-06, + "loss": 0.0015, + "step": 6235 + }, + { + "epoch": 8.098701298701299, + "grad_norm": 0.7194775672201146, + "learning_rate": 1.8392223333859261e-06, + "loss": 0.0013, + "step": 6236 + }, + { + "epoch": 8.1, + "grad_norm": 4.85500891378436, + "learning_rate": 1.8367921292930502e-06, + "loss": 0.0062, + "step": 6237 + }, + { + "epoch": 8.1012987012987, + "grad_norm": 4.0939894814397215, + "learning_rate": 1.8343633694278895e-06, + "loss": 0.0031, + "step": 6238 + }, + { + "epoch": 8.102597402597402, + "grad_norm": 1.4536695212187847, + "learning_rate": 1.831936054220147e-06, + "loss": 0.0019, + "step": 6239 + }, + { + "epoch": 8.103896103896103, + "grad_norm": 4.951802097780959, + "learning_rate": 1.829510184099249e-06, + "loss": 0.0111, + "step": 6240 + }, + { + "epoch": 8.105194805194806, + "grad_norm": 3.3346913452925935, + "learning_rate": 1.827085759494387e-06, + "loss": 0.0021, + "step": 6241 + }, + { + "epoch": 8.106493506493507, + "grad_norm": 2.3872700318348277, + "learning_rate": 1.824662780834483e-06, + "loss": 0.0015, + "step": 6242 + }, + { + "epoch": 8.107792207792208, + "grad_norm": 2.0765387176239325, + "learning_rate": 1.8222412485482154e-06, + "loss": 0.0082, + "step": 6243 + }, + { + "epoch": 8.10909090909091, + "grad_norm": 1.348741839099563, + "learning_rate": 1.8198211630639895e-06, + "loss": 0.0095, + "step": 6244 + }, + { + "epoch": 8.11038961038961, + "grad_norm": 12.28199702019141, + "learning_rate": 1.8174025248099736e-06, + "loss": 0.0154, + "step": 6245 + }, + { + "epoch": 8.111688311688312, + "grad_norm": 3.2988438093012356, + "learning_rate": 1.8149853342140644e-06, + "loss": 0.0035, + "step": 6246 + }, + { + "epoch": 8.112987012987013, + "grad_norm": 2.0127920966134947, + "learning_rate": 1.8125695917039178e-06, + "loss": 0.0012, + "step": 6247 + }, + { + "epoch": 8.114285714285714, + "grad_norm": 3.504128404614124, + "learning_rate": 1.8101552977069148e-06, + "loss": 0.005, + "step": 6248 + }, + { + "epoch": 8.115584415584415, + "grad_norm": 4.833126842038276, + "learning_rate": 1.8077424526501964e-06, + "loss": 0.0033, + "step": 6249 + }, + { + "epoch": 8.116883116883116, + "grad_norm": 5.935621407218237, + "learning_rate": 1.8053310569606375e-06, + "loss": 0.006, + "step": 6250 + }, + { + "epoch": 8.118181818181819, + "grad_norm": 11.582592529300307, + "learning_rate": 1.802921111064868e-06, + "loss": 0.0132, + "step": 6251 + }, + { + "epoch": 8.11948051948052, + "grad_norm": 2.5191470594736725, + "learning_rate": 1.8005126153892426e-06, + "loss": 0.0019, + "step": 6252 + }, + { + "epoch": 8.120779220779221, + "grad_norm": 0.849328715512456, + "learning_rate": 1.798105570359877e-06, + "loss": 0.0075, + "step": 6253 + }, + { + "epoch": 8.122077922077922, + "grad_norm": 1.4151465210226108, + "learning_rate": 1.7956999764026229e-06, + "loss": 0.0015, + "step": 6254 + }, + { + "epoch": 8.123376623376624, + "grad_norm": 3.1387711499018836, + "learning_rate": 1.7932958339430751e-06, + "loss": 0.0089, + "step": 6255 + }, + { + "epoch": 8.124675324675325, + "grad_norm": 2.79539217630009, + "learning_rate": 1.7908931434065691e-06, + "loss": 0.0087, + "step": 6256 + }, + { + "epoch": 8.125974025974026, + "grad_norm": 0.6149462848526274, + "learning_rate": 1.7884919052181926e-06, + "loss": 0.0014, + "step": 6257 + }, + { + "epoch": 8.127272727272727, + "grad_norm": 1.292614872793944, + "learning_rate": 1.786092119802768e-06, + "loss": 0.0011, + "step": 6258 + }, + { + "epoch": 8.128571428571428, + "grad_norm": 0.6039140363910595, + "learning_rate": 1.7836937875848625e-06, + "loss": 0.0084, + "step": 6259 + }, + { + "epoch": 8.12987012987013, + "grad_norm": 1.5878573198160228, + "learning_rate": 1.7812969089887867e-06, + "loss": 0.0092, + "step": 6260 + }, + { + "epoch": 8.13116883116883, + "grad_norm": 2.6845182805212895, + "learning_rate": 1.7789014844385965e-06, + "loss": 0.0013, + "step": 6261 + }, + { + "epoch": 8.132467532467533, + "grad_norm": 4.465245815208509, + "learning_rate": 1.776507514358089e-06, + "loss": 0.0077, + "step": 6262 + }, + { + "epoch": 8.133766233766234, + "grad_norm": 5.1829110801708955, + "learning_rate": 1.7741149991707974e-06, + "loss": 0.0038, + "step": 6263 + }, + { + "epoch": 8.135064935064936, + "grad_norm": 8.529723185697673, + "learning_rate": 1.7717239393000108e-06, + "loss": 0.0085, + "step": 6264 + }, + { + "epoch": 8.136363636363637, + "grad_norm": 2.4665634030555257, + "learning_rate": 1.7693343351687498e-06, + "loss": 0.0042, + "step": 6265 + }, + { + "epoch": 8.137662337662338, + "grad_norm": 3.7491981787676596, + "learning_rate": 1.7669461871997817e-06, + "loss": 0.0082, + "step": 6266 + }, + { + "epoch": 8.138961038961039, + "grad_norm": 1.354073033441667, + "learning_rate": 1.764559495815613e-06, + "loss": 0.0012, + "step": 6267 + }, + { + "epoch": 8.14025974025974, + "grad_norm": 0.8292989998658638, + "learning_rate": 1.7621742614385007e-06, + "loss": 0.014, + "step": 6268 + }, + { + "epoch": 8.141558441558441, + "grad_norm": 3.5185550620523296, + "learning_rate": 1.7597904844904345e-06, + "loss": 0.0015, + "step": 6269 + }, + { + "epoch": 8.142857142857142, + "grad_norm": 3.5198330367557604, + "learning_rate": 1.7574081653931507e-06, + "loss": 0.0014, + "step": 6270 + }, + { + "epoch": 8.144155844155843, + "grad_norm": 2.9946226652024346, + "learning_rate": 1.7550273045681243e-06, + "loss": 0.0017, + "step": 6271 + }, + { + "epoch": 8.145454545454545, + "grad_norm": 4.654090421053931, + "learning_rate": 1.7526479024365816e-06, + "loss": 0.0056, + "step": 6272 + }, + { + "epoch": 8.146753246753248, + "grad_norm": 3.509139671086144, + "learning_rate": 1.7502699594194805e-06, + "loss": 0.0242, + "step": 6273 + }, + { + "epoch": 8.148051948051949, + "grad_norm": 3.680219766254652, + "learning_rate": 1.7478934759375255e-06, + "loss": 0.0011, + "step": 6274 + }, + { + "epoch": 8.14935064935065, + "grad_norm": 5.5676263043465, + "learning_rate": 1.745518452411159e-06, + "loss": 0.0092, + "step": 6275 + }, + { + "epoch": 8.150649350649351, + "grad_norm": 4.2883749971968435, + "learning_rate": 1.743144889260574e-06, + "loss": 0.0025, + "step": 6276 + }, + { + "epoch": 8.151948051948052, + "grad_norm": 1.4174218552397588, + "learning_rate": 1.7407727869056945e-06, + "loss": 0.0033, + "step": 6277 + }, + { + "epoch": 8.153246753246753, + "grad_norm": 4.678120979952961, + "learning_rate": 1.7384021457661936e-06, + "loss": 0.0022, + "step": 6278 + }, + { + "epoch": 8.154545454545454, + "grad_norm": 4.783709622467911, + "learning_rate": 1.7360329662614784e-06, + "loss": 0.0016, + "step": 6279 + }, + { + "epoch": 8.155844155844155, + "grad_norm": 4.16568621371377, + "learning_rate": 1.7336652488107109e-06, + "loss": 0.0034, + "step": 6280 + }, + { + "epoch": 8.157142857142857, + "grad_norm": 4.926661342260446, + "learning_rate": 1.7312989938327752e-06, + "loss": 0.0093, + "step": 6281 + }, + { + "epoch": 8.158441558441558, + "grad_norm": 4.263317017363129, + "learning_rate": 1.728934201746314e-06, + "loss": 0.0044, + "step": 6282 + }, + { + "epoch": 8.15974025974026, + "grad_norm": 6.2099006506963725, + "learning_rate": 1.7265708729697006e-06, + "loss": 0.0075, + "step": 6283 + }, + { + "epoch": 8.161038961038962, + "grad_norm": 1.5500216700451335, + "learning_rate": 1.7242090079210593e-06, + "loss": 0.0011, + "step": 6284 + }, + { + "epoch": 8.162337662337663, + "grad_norm": 1.9578121209070305, + "learning_rate": 1.7218486070182404e-06, + "loss": 0.0075, + "step": 6285 + }, + { + "epoch": 8.163636363636364, + "grad_norm": 3.2881757303994905, + "learning_rate": 1.7194896706788512e-06, + "loss": 0.0022, + "step": 6286 + }, + { + "epoch": 8.164935064935065, + "grad_norm": 6.123375601563677, + "learning_rate": 1.7171321993202272e-06, + "loss": 0.0051, + "step": 6287 + }, + { + "epoch": 8.166233766233766, + "grad_norm": 6.361103109417934, + "learning_rate": 1.7147761933594575e-06, + "loss": 0.0148, + "step": 6288 + }, + { + "epoch": 8.167532467532467, + "grad_norm": 0.6599720986222115, + "learning_rate": 1.7124216532133553e-06, + "loss": 0.0015, + "step": 6289 + }, + { + "epoch": 8.168831168831169, + "grad_norm": 6.855318961991979, + "learning_rate": 1.7100685792984906e-06, + "loss": 0.0059, + "step": 6290 + }, + { + "epoch": 8.17012987012987, + "grad_norm": 3.2011100566523503, + "learning_rate": 1.7077169720311637e-06, + "loss": 0.0028, + "step": 6291 + }, + { + "epoch": 8.17142857142857, + "grad_norm": 2.616393755050941, + "learning_rate": 1.7053668318274219e-06, + "loss": 0.0097, + "step": 6292 + }, + { + "epoch": 8.172727272727272, + "grad_norm": 2.12356780086445, + "learning_rate": 1.7030181591030493e-06, + "loss": 0.0207, + "step": 6293 + }, + { + "epoch": 8.174025974025975, + "grad_norm": 2.1741591819284865, + "learning_rate": 1.70067095427357e-06, + "loss": 0.0088, + "step": 6294 + }, + { + "epoch": 8.175324675324676, + "grad_norm": 5.606916619897311, + "learning_rate": 1.6983252177542498e-06, + "loss": 0.0094, + "step": 6295 + }, + { + "epoch": 8.176623376623377, + "grad_norm": 1.2826760945327003, + "learning_rate": 1.695980949960091e-06, + "loss": 0.0027, + "step": 6296 + }, + { + "epoch": 8.177922077922078, + "grad_norm": 4.305931577675351, + "learning_rate": 1.6936381513058464e-06, + "loss": 0.0042, + "step": 6297 + }, + { + "epoch": 8.17922077922078, + "grad_norm": 1.586255361005633, + "learning_rate": 1.6912968222059977e-06, + "loss": 0.0324, + "step": 6298 + }, + { + "epoch": 8.18051948051948, + "grad_norm": 1.458790166918778, + "learning_rate": 1.6889569630747715e-06, + "loss": 0.0013, + "step": 6299 + }, + { + "epoch": 8.181818181818182, + "grad_norm": 1.0046453333357028, + "learning_rate": 1.6866185743261309e-06, + "loss": 0.0012, + "step": 6300 + }, + { + "epoch": 8.183116883116883, + "grad_norm": 0.5545077770384786, + "learning_rate": 1.6842816563737875e-06, + "loss": 0.0012, + "step": 6301 + }, + { + "epoch": 8.184415584415584, + "grad_norm": 2.2691397690505455, + "learning_rate": 1.6819462096311812e-06, + "loss": 0.0211, + "step": 6302 + }, + { + "epoch": 8.185714285714285, + "grad_norm": 2.9125416749170725, + "learning_rate": 1.679612234511504e-06, + "loss": 0.0158, + "step": 6303 + }, + { + "epoch": 8.187012987012986, + "grad_norm": 2.111523305972463, + "learning_rate": 1.6772797314276712e-06, + "loss": 0.0121, + "step": 6304 + }, + { + "epoch": 8.188311688311689, + "grad_norm": 4.3666867028527445, + "learning_rate": 1.674948700792356e-06, + "loss": 0.006, + "step": 6305 + }, + { + "epoch": 8.18961038961039, + "grad_norm": 2.113227991994401, + "learning_rate": 1.6726191430179583e-06, + "loss": 0.0021, + "step": 6306 + }, + { + "epoch": 8.190909090909091, + "grad_norm": 4.784276106369777, + "learning_rate": 1.670291058516622e-06, + "loss": 0.0075, + "step": 6307 + }, + { + "epoch": 8.192207792207792, + "grad_norm": 0.901765226822907, + "learning_rate": 1.6679644477002265e-06, + "loss": 0.0019, + "step": 6308 + }, + { + "epoch": 8.193506493506494, + "grad_norm": 3.067937559952356, + "learning_rate": 1.6656393109804003e-06, + "loss": 0.0068, + "step": 6309 + }, + { + "epoch": 8.194805194805195, + "grad_norm": 1.1690670156246632, + "learning_rate": 1.663315648768502e-06, + "loss": 0.0018, + "step": 6310 + }, + { + "epoch": 8.196103896103896, + "grad_norm": 0.40991836530830034, + "learning_rate": 1.6609934614756308e-06, + "loss": 0.0011, + "step": 6311 + }, + { + "epoch": 8.197402597402597, + "grad_norm": 1.9112417751439743, + "learning_rate": 1.658672749512623e-06, + "loss": 0.0039, + "step": 6312 + }, + { + "epoch": 8.198701298701298, + "grad_norm": 1.904362100602193, + "learning_rate": 1.6563535132900643e-06, + "loss": 0.0083, + "step": 6313 + }, + { + "epoch": 8.2, + "grad_norm": 1.2302251014893313, + "learning_rate": 1.6540357532182683e-06, + "loss": 0.0017, + "step": 6314 + }, + { + "epoch": 8.2012987012987, + "grad_norm": 1.8852857717479095, + "learning_rate": 1.6517194697072903e-06, + "loss": 0.0012, + "step": 6315 + }, + { + "epoch": 8.202597402597403, + "grad_norm": 0.7885789493976344, + "learning_rate": 1.6494046631669236e-06, + "loss": 0.003, + "step": 6316 + }, + { + "epoch": 8.203896103896104, + "grad_norm": 1.1602521239503711, + "learning_rate": 1.6470913340067064e-06, + "loss": 0.0018, + "step": 6317 + }, + { + "epoch": 8.205194805194806, + "grad_norm": 3.619203074027743, + "learning_rate": 1.6447794826359087e-06, + "loss": 0.0045, + "step": 6318 + }, + { + "epoch": 8.206493506493507, + "grad_norm": 1.1599647485245979, + "learning_rate": 1.6424691094635413e-06, + "loss": 0.0016, + "step": 6319 + }, + { + "epoch": 8.207792207792208, + "grad_norm": 3.5720421521704906, + "learning_rate": 1.640160214898351e-06, + "loss": 0.0144, + "step": 6320 + }, + { + "epoch": 8.209090909090909, + "grad_norm": 3.473119829067112, + "learning_rate": 1.6378527993488313e-06, + "loss": 0.0147, + "step": 6321 + }, + { + "epoch": 8.21038961038961, + "grad_norm": 10.032764250348224, + "learning_rate": 1.6355468632232007e-06, + "loss": 0.0167, + "step": 6322 + }, + { + "epoch": 8.211688311688311, + "grad_norm": 4.581872682532007, + "learning_rate": 1.6332424069294306e-06, + "loss": 0.0127, + "step": 6323 + }, + { + "epoch": 8.212987012987012, + "grad_norm": 2.5789171865799814, + "learning_rate": 1.630939430875217e-06, + "loss": 0.0061, + "step": 6324 + }, + { + "epoch": 8.214285714285714, + "grad_norm": 1.5722400693131156, + "learning_rate": 1.628637935468006e-06, + "loss": 0.0083, + "step": 6325 + }, + { + "epoch": 8.215584415584416, + "grad_norm": 0.7413520463732368, + "learning_rate": 1.6263379211149733e-06, + "loss": 0.0068, + "step": 6326 + }, + { + "epoch": 8.216883116883118, + "grad_norm": 1.3226060496731267, + "learning_rate": 1.6240393882230365e-06, + "loss": 0.0022, + "step": 6327 + }, + { + "epoch": 8.218181818181819, + "grad_norm": 4.246542174557424, + "learning_rate": 1.6217423371988471e-06, + "loss": 0.0046, + "step": 6328 + }, + { + "epoch": 8.21948051948052, + "grad_norm": 2.629082313594407, + "learning_rate": 1.6194467684488025e-06, + "loss": 0.0088, + "step": 6329 + }, + { + "epoch": 8.220779220779221, + "grad_norm": 9.526873227516804, + "learning_rate": 1.6171526823790295e-06, + "loss": 0.0161, + "step": 6330 + }, + { + "epoch": 8.222077922077922, + "grad_norm": 6.758410948424776, + "learning_rate": 1.6148600793953962e-06, + "loss": 0.0112, + "step": 6331 + }, + { + "epoch": 8.223376623376623, + "grad_norm": 3.147893300886733, + "learning_rate": 1.6125689599035088e-06, + "loss": 0.0181, + "step": 6332 + }, + { + "epoch": 8.224675324675324, + "grad_norm": 2.277667235884005, + "learning_rate": 1.6102793243087067e-06, + "loss": 0.0097, + "step": 6333 + }, + { + "epoch": 8.225974025974025, + "grad_norm": 1.727617321500896, + "learning_rate": 1.6079911730160758e-06, + "loss": 0.0101, + "step": 6334 + }, + { + "epoch": 8.227272727272727, + "grad_norm": 3.2266259985752463, + "learning_rate": 1.6057045064304289e-06, + "loss": 0.0012, + "step": 6335 + }, + { + "epoch": 8.228571428571428, + "grad_norm": 5.9082893345995044, + "learning_rate": 1.603419324956328e-06, + "loss": 0.0099, + "step": 6336 + }, + { + "epoch": 8.22987012987013, + "grad_norm": 5.335343273099192, + "learning_rate": 1.6011356289980562e-06, + "loss": 0.0065, + "step": 6337 + }, + { + "epoch": 8.231168831168832, + "grad_norm": 3.8256700615466324, + "learning_rate": 1.5988534189596494e-06, + "loss": 0.0108, + "step": 6338 + }, + { + "epoch": 8.232467532467533, + "grad_norm": 3.638939500306878, + "learning_rate": 1.59657269524487e-06, + "loss": 0.0016, + "step": 6339 + }, + { + "epoch": 8.233766233766234, + "grad_norm": 3.5893151928028173, + "learning_rate": 1.594293458257229e-06, + "loss": 0.0014, + "step": 6340 + }, + { + "epoch": 8.235064935064935, + "grad_norm": 1.056564090176049, + "learning_rate": 1.5920157083999577e-06, + "loss": 0.0013, + "step": 6341 + }, + { + "epoch": 8.236363636363636, + "grad_norm": 1.4473746617672363, + "learning_rate": 1.5897394460760406e-06, + "loss": 0.0144, + "step": 6342 + }, + { + "epoch": 8.237662337662337, + "grad_norm": 2.6390838434204364, + "learning_rate": 1.587464671688187e-06, + "loss": 0.0024, + "step": 6343 + }, + { + "epoch": 8.238961038961039, + "grad_norm": 4.018193946192787, + "learning_rate": 1.5851913856388556e-06, + "loss": 0.0021, + "step": 6344 + }, + { + "epoch": 8.24025974025974, + "grad_norm": 3.3390708714355437, + "learning_rate": 1.5829195883302239e-06, + "loss": 0.0015, + "step": 6345 + }, + { + "epoch": 8.24155844155844, + "grad_norm": 1.0538270319639393, + "learning_rate": 1.5806492801642249e-06, + "loss": 0.002, + "step": 6346 + }, + { + "epoch": 8.242857142857142, + "grad_norm": 4.875394069518504, + "learning_rate": 1.5783804615425159e-06, + "loss": 0.0097, + "step": 6347 + }, + { + "epoch": 8.244155844155845, + "grad_norm": 5.056812356960358, + "learning_rate": 1.5761131328664946e-06, + "loss": 0.002, + "step": 6348 + }, + { + "epoch": 8.245454545454546, + "grad_norm": 5.925347695539774, + "learning_rate": 1.5738472945372928e-06, + "loss": 0.0144, + "step": 6349 + }, + { + "epoch": 8.246753246753247, + "grad_norm": 4.560361316336791, + "learning_rate": 1.5715829469557853e-06, + "loss": 0.0171, + "step": 6350 + }, + { + "epoch": 8.248051948051948, + "grad_norm": 7.184311143279375, + "learning_rate": 1.569320090522576e-06, + "loss": 0.0074, + "step": 6351 + }, + { + "epoch": 8.24935064935065, + "grad_norm": 6.223378926623537, + "learning_rate": 1.5670587256380078e-06, + "loss": 0.0165, + "step": 6352 + }, + { + "epoch": 8.25064935064935, + "grad_norm": 4.963553611685732, + "learning_rate": 1.5647988527021563e-06, + "loss": 0.0013, + "step": 6353 + }, + { + "epoch": 8.251948051948052, + "grad_norm": 2.4968001313828667, + "learning_rate": 1.562540472114843e-06, + "loss": 0.0026, + "step": 6354 + }, + { + "epoch": 8.253246753246753, + "grad_norm": 3.102535266421428, + "learning_rate": 1.5602835842756136e-06, + "loss": 0.0037, + "step": 6355 + }, + { + "epoch": 8.254545454545454, + "grad_norm": 3.449799092029054, + "learning_rate": 1.5580281895837557e-06, + "loss": 0.0015, + "step": 6356 + }, + { + "epoch": 8.255844155844155, + "grad_norm": 3.181907978663455, + "learning_rate": 1.555774288438291e-06, + "loss": 0.0014, + "step": 6357 + }, + { + "epoch": 8.257142857142856, + "grad_norm": 1.8757452181270626, + "learning_rate": 1.5535218812379804e-06, + "loss": 0.0011, + "step": 6358 + }, + { + "epoch": 8.25844155844156, + "grad_norm": 2.875020105969289, + "learning_rate": 1.5512709683813165e-06, + "loss": 0.0147, + "step": 6359 + }, + { + "epoch": 8.25974025974026, + "grad_norm": 2.4764145465416347, + "learning_rate": 1.5490215502665297e-06, + "loss": 0.0091, + "step": 6360 + }, + { + "epoch": 8.261038961038961, + "grad_norm": 3.7309125136642485, + "learning_rate": 1.5467736272915802e-06, + "loss": 0.0038, + "step": 6361 + }, + { + "epoch": 8.262337662337663, + "grad_norm": 4.541504783252121, + "learning_rate": 1.5445271998541756e-06, + "loss": 0.0085, + "step": 6362 + }, + { + "epoch": 8.263636363636364, + "grad_norm": 15.207824993711215, + "learning_rate": 1.5422822683517492e-06, + "loss": 0.0181, + "step": 6363 + }, + { + "epoch": 8.264935064935065, + "grad_norm": 2.5973329788605173, + "learning_rate": 1.5400388331814686e-06, + "loss": 0.0111, + "step": 6364 + }, + { + "epoch": 8.266233766233766, + "grad_norm": 3.467070502305466, + "learning_rate": 1.5377968947402465e-06, + "loss": 0.002, + "step": 6365 + }, + { + "epoch": 8.267532467532467, + "grad_norm": 6.705589743130502, + "learning_rate": 1.5355564534247225e-06, + "loss": 0.0092, + "step": 6366 + }, + { + "epoch": 8.268831168831168, + "grad_norm": 13.105672887321981, + "learning_rate": 1.5333175096312724e-06, + "loss": 0.0205, + "step": 6367 + }, + { + "epoch": 8.27012987012987, + "grad_norm": 5.2261880194987365, + "learning_rate": 1.5310800637560074e-06, + "loss": 0.0083, + "step": 6368 + }, + { + "epoch": 8.271428571428572, + "grad_norm": 3.4801651486998892, + "learning_rate": 1.5288441161947776e-06, + "loss": 0.0097, + "step": 6369 + }, + { + "epoch": 8.272727272727273, + "grad_norm": 7.876706332415049, + "learning_rate": 1.526609667343163e-06, + "loss": 0.004, + "step": 6370 + }, + { + "epoch": 8.274025974025975, + "grad_norm": 6.30375645649304, + "learning_rate": 1.5243767175964818e-06, + "loss": 0.0086, + "step": 6371 + }, + { + "epoch": 8.275324675324676, + "grad_norm": 1.8692694555633733, + "learning_rate": 1.5221452673497816e-06, + "loss": 0.015, + "step": 6372 + }, + { + "epoch": 8.276623376623377, + "grad_norm": 2.0463298710503306, + "learning_rate": 1.5199153169978553e-06, + "loss": 0.0063, + "step": 6373 + }, + { + "epoch": 8.277922077922078, + "grad_norm": 1.3075599983922281, + "learning_rate": 1.5176868669352164e-06, + "loss": 0.0016, + "step": 6374 + }, + { + "epoch": 8.279220779220779, + "grad_norm": 5.161034326112554, + "learning_rate": 1.5154599175561257e-06, + "loss": 0.0039, + "step": 6375 + }, + { + "epoch": 8.28051948051948, + "grad_norm": 5.762088086783746, + "learning_rate": 1.5132344692545698e-06, + "loss": 0.0121, + "step": 6376 + }, + { + "epoch": 8.281818181818181, + "grad_norm": 5.767325398241772, + "learning_rate": 1.5110105224242789e-06, + "loss": 0.0159, + "step": 6377 + }, + { + "epoch": 8.283116883116882, + "grad_norm": 2.722624683752853, + "learning_rate": 1.5087880774587028e-06, + "loss": 0.014, + "step": 6378 + }, + { + "epoch": 8.284415584415584, + "grad_norm": 3.9700921139378393, + "learning_rate": 1.5065671347510414e-06, + "loss": 0.0157, + "step": 6379 + }, + { + "epoch": 8.285714285714286, + "grad_norm": 9.037540947744343, + "learning_rate": 1.5043476946942182e-06, + "loss": 0.0075, + "step": 6380 + }, + { + "epoch": 8.287012987012988, + "grad_norm": 5.4105893470981945, + "learning_rate": 1.5021297576809002e-06, + "loss": 0.0018, + "step": 6381 + }, + { + "epoch": 8.288311688311689, + "grad_norm": 6.900439561729553, + "learning_rate": 1.4999133241034747e-06, + "loss": 0.0152, + "step": 6382 + }, + { + "epoch": 8.28961038961039, + "grad_norm": 4.085601928986028, + "learning_rate": 1.497698394354078e-06, + "loss": 0.0032, + "step": 6383 + }, + { + "epoch": 8.290909090909091, + "grad_norm": 5.287145864646405, + "learning_rate": 1.4954849688245699e-06, + "loss": 0.0038, + "step": 6384 + }, + { + "epoch": 8.292207792207792, + "grad_norm": 6.247893451947501, + "learning_rate": 1.4932730479065515e-06, + "loss": 0.0067, + "step": 6385 + }, + { + "epoch": 8.293506493506493, + "grad_norm": 7.255722264649447, + "learning_rate": 1.491062631991348e-06, + "loss": 0.0088, + "step": 6386 + }, + { + "epoch": 8.294805194805194, + "grad_norm": 2.7833836119572988, + "learning_rate": 1.4888537214700304e-06, + "loss": 0.0077, + "step": 6387 + }, + { + "epoch": 8.296103896103896, + "grad_norm": 9.492066612119075, + "learning_rate": 1.4866463167333933e-06, + "loss": 0.01, + "step": 6388 + }, + { + "epoch": 8.297402597402597, + "grad_norm": 4.354454213204406, + "learning_rate": 1.4844404181719708e-06, + "loss": 0.0123, + "step": 6389 + }, + { + "epoch": 8.2987012987013, + "grad_norm": 4.140241450852374, + "learning_rate": 1.482236026176026e-06, + "loss": 0.002, + "step": 6390 + }, + { + "epoch": 8.3, + "grad_norm": 3.5505600248266638, + "learning_rate": 1.4800331411355618e-06, + "loss": 0.0088, + "step": 6391 + }, + { + "epoch": 8.301298701298702, + "grad_norm": 2.1948854957623007, + "learning_rate": 1.4778317634403082e-06, + "loss": 0.0011, + "step": 6392 + }, + { + "epoch": 8.302597402597403, + "grad_norm": 1.852675818209928, + "learning_rate": 1.475631893479731e-06, + "loss": 0.0175, + "step": 6393 + }, + { + "epoch": 8.303896103896104, + "grad_norm": 6.379997393395281, + "learning_rate": 1.4734335316430303e-06, + "loss": 0.0038, + "step": 6394 + }, + { + "epoch": 8.305194805194805, + "grad_norm": 4.7778376235367634, + "learning_rate": 1.47123667831914e-06, + "loss": 0.0046, + "step": 6395 + }, + { + "epoch": 8.306493506493506, + "grad_norm": 4.427027417780311, + "learning_rate": 1.4690413338967224e-06, + "loss": 0.0078, + "step": 6396 + }, + { + "epoch": 8.307792207792208, + "grad_norm": 2.6575285271273454, + "learning_rate": 1.4668474987641745e-06, + "loss": 0.0017, + "step": 6397 + }, + { + "epoch": 8.309090909090909, + "grad_norm": 3.073720199602591, + "learning_rate": 1.4646551733096338e-06, + "loss": 0.0012, + "step": 6398 + }, + { + "epoch": 8.31038961038961, + "grad_norm": 3.178612823072671, + "learning_rate": 1.4624643579209607e-06, + "loss": 0.0038, + "step": 6399 + }, + { + "epoch": 8.311688311688311, + "grad_norm": 5.188306923366621, + "learning_rate": 1.4602750529857523e-06, + "loss": 0.0014, + "step": 6400 + }, + { + "epoch": 8.312987012987014, + "grad_norm": 14.318521020303322, + "learning_rate": 1.4580872588913374e-06, + "loss": 0.0202, + "step": 6401 + }, + { + "epoch": 8.314285714285715, + "grad_norm": 5.128824820402371, + "learning_rate": 1.4559009760247821e-06, + "loss": 0.0285, + "step": 6402 + }, + { + "epoch": 8.315584415584416, + "grad_norm": 1.3893668453851824, + "learning_rate": 1.4537162047728804e-06, + "loss": 0.0015, + "step": 6403 + }, + { + "epoch": 8.316883116883117, + "grad_norm": 3.875407058067913, + "learning_rate": 1.4515329455221604e-06, + "loss": 0.0094, + "step": 6404 + }, + { + "epoch": 8.318181818181818, + "grad_norm": 3.6404587263222354, + "learning_rate": 1.4493511986588782e-06, + "loss": 0.0088, + "step": 6405 + }, + { + "epoch": 8.31948051948052, + "grad_norm": 2.9428472301979234, + "learning_rate": 1.4471709645690336e-06, + "loss": 0.0085, + "step": 6406 + }, + { + "epoch": 8.32077922077922, + "grad_norm": 5.457649374246802, + "learning_rate": 1.4449922436383467e-06, + "loss": 0.0063, + "step": 6407 + }, + { + "epoch": 8.322077922077922, + "grad_norm": 0.9923001719160036, + "learning_rate": 1.4428150362522774e-06, + "loss": 0.009, + "step": 6408 + }, + { + "epoch": 8.323376623376623, + "grad_norm": 3.558344641025288, + "learning_rate": 1.440639342796013e-06, + "loss": 0.0085, + "step": 6409 + }, + { + "epoch": 8.324675324675324, + "grad_norm": 3.6427135559111363, + "learning_rate": 1.4384651636544778e-06, + "loss": 0.0022, + "step": 6410 + }, + { + "epoch": 8.325974025974025, + "grad_norm": 2.9960250198094838, + "learning_rate": 1.436292499212325e-06, + "loss": 0.002, + "step": 6411 + }, + { + "epoch": 8.327272727272728, + "grad_norm": 2.8943866089379453, + "learning_rate": 1.4341213498539397e-06, + "loss": 0.0056, + "step": 6412 + }, + { + "epoch": 8.32857142857143, + "grad_norm": 4.330923812997877, + "learning_rate": 1.4319517159634389e-06, + "loss": 0.0022, + "step": 6413 + }, + { + "epoch": 8.32987012987013, + "grad_norm": 5.079473398099091, + "learning_rate": 1.4297835979246777e-06, + "loss": 0.0072, + "step": 6414 + }, + { + "epoch": 8.331168831168831, + "grad_norm": 3.916258362196294, + "learning_rate": 1.4276169961212294e-06, + "loss": 0.0217, + "step": 6415 + }, + { + "epoch": 8.332467532467533, + "grad_norm": 3.9592718459099623, + "learning_rate": 1.425451910936414e-06, + "loss": 0.0083, + "step": 6416 + }, + { + "epoch": 8.333766233766234, + "grad_norm": 4.669434621346435, + "learning_rate": 1.4232883427532717e-06, + "loss": 0.0082, + "step": 6417 + }, + { + "epoch": 8.335064935064935, + "grad_norm": 3.261042107398802, + "learning_rate": 1.4211262919545865e-06, + "loss": 0.0027, + "step": 6418 + }, + { + "epoch": 8.336363636363636, + "grad_norm": 2.5702809316146076, + "learning_rate": 1.4189657589228568e-06, + "loss": 0.0187, + "step": 6419 + }, + { + "epoch": 8.337662337662337, + "grad_norm": 3.62644632810569, + "learning_rate": 1.4168067440403298e-06, + "loss": 0.0054, + "step": 6420 + }, + { + "epoch": 8.338961038961038, + "grad_norm": 1.4282869848953634, + "learning_rate": 1.4146492476889707e-06, + "loss": 0.0019, + "step": 6421 + }, + { + "epoch": 8.34025974025974, + "grad_norm": 2.3692790019033025, + "learning_rate": 1.41249327025049e-06, + "loss": 0.0026, + "step": 6422 + }, + { + "epoch": 8.341558441558442, + "grad_norm": 1.1942626598911847, + "learning_rate": 1.4103388121063122e-06, + "loss": 0.0165, + "step": 6423 + }, + { + "epoch": 8.342857142857143, + "grad_norm": 1.701011307811885, + "learning_rate": 1.4081858736376085e-06, + "loss": 0.0012, + "step": 6424 + }, + { + "epoch": 8.344155844155845, + "grad_norm": 1.9595111594025756, + "learning_rate": 1.4060344552252702e-06, + "loss": 0.0016, + "step": 6425 + }, + { + "epoch": 8.345454545454546, + "grad_norm": 1.168527162475915, + "learning_rate": 1.4038845572499304e-06, + "loss": 0.0046, + "step": 6426 + }, + { + "epoch": 8.346753246753247, + "grad_norm": 3.2451928491495834, + "learning_rate": 1.4017361800919427e-06, + "loss": 0.009, + "step": 6427 + }, + { + "epoch": 8.348051948051948, + "grad_norm": 4.679424637787728, + "learning_rate": 1.3995893241313974e-06, + "loss": 0.0191, + "step": 6428 + }, + { + "epoch": 8.349350649350649, + "grad_norm": 1.8402918938867927, + "learning_rate": 1.3974439897481151e-06, + "loss": 0.002, + "step": 6429 + }, + { + "epoch": 8.35064935064935, + "grad_norm": 4.749289179004831, + "learning_rate": 1.3953001773216425e-06, + "loss": 0.0287, + "step": 6430 + }, + { + "epoch": 8.351948051948051, + "grad_norm": 3.1789734062390282, + "learning_rate": 1.3931578872312668e-06, + "loss": 0.0016, + "step": 6431 + }, + { + "epoch": 8.353246753246752, + "grad_norm": 2.0593722376294825, + "learning_rate": 1.3910171198559974e-06, + "loss": 0.0029, + "step": 6432 + }, + { + "epoch": 8.354545454545455, + "grad_norm": 4.688105957805071, + "learning_rate": 1.3888778755745768e-06, + "loss": 0.0111, + "step": 6433 + }, + { + "epoch": 8.355844155844157, + "grad_norm": 1.921709779052265, + "learning_rate": 1.386740154765478e-06, + "loss": 0.0186, + "step": 6434 + }, + { + "epoch": 8.357142857142858, + "grad_norm": 4.923881575730746, + "learning_rate": 1.384603957806906e-06, + "loss": 0.0095, + "step": 6435 + }, + { + "epoch": 8.358441558441559, + "grad_norm": 3.5940904854084406, + "learning_rate": 1.3824692850767928e-06, + "loss": 0.0012, + "step": 6436 + }, + { + "epoch": 8.35974025974026, + "grad_norm": 2.326722785289336, + "learning_rate": 1.380336136952809e-06, + "loss": 0.0014, + "step": 6437 + }, + { + "epoch": 8.361038961038961, + "grad_norm": 1.3938012270944644, + "learning_rate": 1.378204513812339e-06, + "loss": 0.0088, + "step": 6438 + }, + { + "epoch": 8.362337662337662, + "grad_norm": 2.696195627199077, + "learning_rate": 1.3760744160325167e-06, + "loss": 0.0011, + "step": 6439 + }, + { + "epoch": 8.363636363636363, + "grad_norm": 5.539146921521937, + "learning_rate": 1.373945843990192e-06, + "loss": 0.0065, + "step": 6440 + }, + { + "epoch": 8.364935064935064, + "grad_norm": 6.60039982375773, + "learning_rate": 1.3718187980619557e-06, + "loss": 0.0167, + "step": 6441 + }, + { + "epoch": 8.366233766233766, + "grad_norm": 5.602623731700719, + "learning_rate": 1.3696932786241146e-06, + "loss": 0.0163, + "step": 6442 + }, + { + "epoch": 8.367532467532467, + "grad_norm": 0.7368271490785772, + "learning_rate": 1.3675692860527213e-06, + "loss": 0.0015, + "step": 6443 + }, + { + "epoch": 8.36883116883117, + "grad_norm": 3.7798123927013805, + "learning_rate": 1.365446820723547e-06, + "loss": 0.0019, + "step": 6444 + }, + { + "epoch": 8.37012987012987, + "grad_norm": 3.6672425126449166, + "learning_rate": 1.3633258830120977e-06, + "loss": 0.0012, + "step": 6445 + }, + { + "epoch": 8.371428571428572, + "grad_norm": 4.834499242643977, + "learning_rate": 1.3612064732936048e-06, + "loss": 0.0269, + "step": 6446 + }, + { + "epoch": 8.372727272727273, + "grad_norm": 0.6619352384125019, + "learning_rate": 1.3590885919430385e-06, + "loss": 0.001, + "step": 6447 + }, + { + "epoch": 8.374025974025974, + "grad_norm": 1.6236017942938814, + "learning_rate": 1.3569722393350882e-06, + "loss": 0.009, + "step": 6448 + }, + { + "epoch": 8.375324675324675, + "grad_norm": 5.499786624448246, + "learning_rate": 1.3548574158441773e-06, + "loss": 0.0019, + "step": 6449 + }, + { + "epoch": 8.376623376623376, + "grad_norm": 3.8517832009789696, + "learning_rate": 1.3527441218444582e-06, + "loss": 0.0014, + "step": 6450 + }, + { + "epoch": 8.377922077922078, + "grad_norm": 0.7524424022381218, + "learning_rate": 1.3506323577098157e-06, + "loss": 0.015, + "step": 6451 + }, + { + "epoch": 8.379220779220779, + "grad_norm": 5.093241670439388, + "learning_rate": 1.3485221238138602e-06, + "loss": 0.0186, + "step": 6452 + }, + { + "epoch": 8.38051948051948, + "grad_norm": 2.279402800173672, + "learning_rate": 1.3464134205299317e-06, + "loss": 0.0026, + "step": 6453 + }, + { + "epoch": 8.381818181818183, + "grad_norm": 6.376010100312236, + "learning_rate": 1.3443062482310987e-06, + "loss": 0.0115, + "step": 6454 + }, + { + "epoch": 8.383116883116884, + "grad_norm": 2.8577204295685177, + "learning_rate": 1.342200607290165e-06, + "loss": 0.0012, + "step": 6455 + }, + { + "epoch": 8.384415584415585, + "grad_norm": 2.66983489899499, + "learning_rate": 1.3400964980796549e-06, + "loss": 0.0015, + "step": 6456 + }, + { + "epoch": 8.385714285714286, + "grad_norm": 2.1833584146313325, + "learning_rate": 1.3379939209718262e-06, + "loss": 0.0013, + "step": 6457 + }, + { + "epoch": 8.387012987012987, + "grad_norm": 1.892671789343376, + "learning_rate": 1.3358928763386647e-06, + "loss": 0.0257, + "step": 6458 + }, + { + "epoch": 8.388311688311688, + "grad_norm": 0.6800136765198398, + "learning_rate": 1.3337933645518874e-06, + "loss": 0.0014, + "step": 6459 + }, + { + "epoch": 8.38961038961039, + "grad_norm": 2.0251678266728392, + "learning_rate": 1.3316953859829372e-06, + "loss": 0.0015, + "step": 6460 + }, + { + "epoch": 8.39090909090909, + "grad_norm": 1.5242436385223024, + "learning_rate": 1.3295989410029864e-06, + "loss": 0.0021, + "step": 6461 + }, + { + "epoch": 8.392207792207792, + "grad_norm": 0.9210072071556246, + "learning_rate": 1.3275040299829333e-06, + "loss": 0.0092, + "step": 6462 + }, + { + "epoch": 8.393506493506493, + "grad_norm": 1.0849576365192637, + "learning_rate": 1.3254106532934131e-06, + "loss": 0.0104, + "step": 6463 + }, + { + "epoch": 8.394805194805194, + "grad_norm": 0.9602882858339858, + "learning_rate": 1.3233188113047824e-06, + "loss": 0.0013, + "step": 6464 + }, + { + "epoch": 8.396103896103897, + "grad_norm": 0.5513192787226643, + "learning_rate": 1.3212285043871243e-06, + "loss": 0.0032, + "step": 6465 + }, + { + "epoch": 8.397402597402598, + "grad_norm": 1.205680365312448, + "learning_rate": 1.3191397329102596e-06, + "loss": 0.0065, + "step": 6466 + }, + { + "epoch": 8.3987012987013, + "grad_norm": 3.410347936182966, + "learning_rate": 1.3170524972437304e-06, + "loss": 0.0063, + "step": 6467 + }, + { + "epoch": 8.4, + "grad_norm": 1.522913823188738, + "learning_rate": 1.314966797756807e-06, + "loss": 0.0086, + "step": 6468 + }, + { + "epoch": 8.401298701298701, + "grad_norm": 0.7787455371526942, + "learning_rate": 1.3128826348184886e-06, + "loss": 0.0015, + "step": 6469 + }, + { + "epoch": 8.402597402597403, + "grad_norm": 1.068281731458151, + "learning_rate": 1.3108000087975093e-06, + "loss": 0.0021, + "step": 6470 + }, + { + "epoch": 8.403896103896104, + "grad_norm": 0.8955516317727982, + "learning_rate": 1.3087189200623174e-06, + "loss": 0.0012, + "step": 6471 + }, + { + "epoch": 8.405194805194805, + "grad_norm": 1.3340288012621764, + "learning_rate": 1.306639368981104e-06, + "loss": 0.0011, + "step": 6472 + }, + { + "epoch": 8.406493506493506, + "grad_norm": 0.7563158131385523, + "learning_rate": 1.3045613559217763e-06, + "loss": 0.002, + "step": 6473 + }, + { + "epoch": 8.407792207792207, + "grad_norm": 0.7846571132859809, + "learning_rate": 1.3024848812519818e-06, + "loss": 0.0009, + "step": 6474 + }, + { + "epoch": 8.409090909090908, + "grad_norm": 1.317347260228361, + "learning_rate": 1.3004099453390794e-06, + "loss": 0.0018, + "step": 6475 + }, + { + "epoch": 8.410389610389611, + "grad_norm": 1.0124987902405898, + "learning_rate": 1.298336548550172e-06, + "loss": 0.0012, + "step": 6476 + }, + { + "epoch": 8.411688311688312, + "grad_norm": 0.3938107992736514, + "learning_rate": 1.2962646912520782e-06, + "loss": 0.0012, + "step": 6477 + }, + { + "epoch": 8.412987012987013, + "grad_norm": 3.069045613130241, + "learning_rate": 1.294194373811356e-06, + "loss": 0.0076, + "step": 6478 + }, + { + "epoch": 8.414285714285715, + "grad_norm": 1.4777114897913912, + "learning_rate": 1.292125596594277e-06, + "loss": 0.0026, + "step": 6479 + }, + { + "epoch": 8.415584415584416, + "grad_norm": 1.612594030270518, + "learning_rate": 1.2900583599668515e-06, + "loss": 0.0015, + "step": 6480 + }, + { + "epoch": 8.416883116883117, + "grad_norm": 1.1480344786488474, + "learning_rate": 1.2879926642948103e-06, + "loss": 0.0013, + "step": 6481 + }, + { + "epoch": 8.418181818181818, + "grad_norm": 0.38870741288433525, + "learning_rate": 1.2859285099436203e-06, + "loss": 0.001, + "step": 6482 + }, + { + "epoch": 8.41948051948052, + "grad_norm": 5.01775111528699, + "learning_rate": 1.2838658972784623e-06, + "loss": 0.0251, + "step": 6483 + }, + { + "epoch": 8.42077922077922, + "grad_norm": 0.550805497386761, + "learning_rate": 1.2818048266642569e-06, + "loss": 0.0014, + "step": 6484 + }, + { + "epoch": 8.422077922077921, + "grad_norm": 1.9745371993649106, + "learning_rate": 1.2797452984656466e-06, + "loss": 0.0098, + "step": 6485 + }, + { + "epoch": 8.423376623376623, + "grad_norm": 1.8503897115431676, + "learning_rate": 1.2776873130470003e-06, + "loss": 0.0091, + "step": 6486 + }, + { + "epoch": 8.424675324675325, + "grad_norm": 3.2841714832833895, + "learning_rate": 1.275630870772413e-06, + "loss": 0.0018, + "step": 6487 + }, + { + "epoch": 8.425974025974027, + "grad_norm": 1.513854401227321, + "learning_rate": 1.2735759720057129e-06, + "loss": 0.0013, + "step": 6488 + }, + { + "epoch": 8.427272727272728, + "grad_norm": 0.5517784661408517, + "learning_rate": 1.2715226171104478e-06, + "loss": 0.0008, + "step": 6489 + }, + { + "epoch": 8.428571428571429, + "grad_norm": 3.697627959448199, + "learning_rate": 1.2694708064498984e-06, + "loss": 0.0153, + "step": 6490 + }, + { + "epoch": 8.42987012987013, + "grad_norm": 1.4291003910615032, + "learning_rate": 1.2674205403870643e-06, + "loss": 0.0014, + "step": 6491 + }, + { + "epoch": 8.431168831168831, + "grad_norm": 2.942251909750061, + "learning_rate": 1.265371819284682e-06, + "loss": 0.0107, + "step": 6492 + }, + { + "epoch": 8.432467532467532, + "grad_norm": 2.7696518090234417, + "learning_rate": 1.2633246435052083e-06, + "loss": 0.0231, + "step": 6493 + }, + { + "epoch": 8.433766233766233, + "grad_norm": 1.8191737329241824, + "learning_rate": 1.2612790134108265e-06, + "loss": 0.0011, + "step": 6494 + }, + { + "epoch": 8.435064935064934, + "grad_norm": 1.1446920102003957, + "learning_rate": 1.2592349293634454e-06, + "loss": 0.0012, + "step": 6495 + }, + { + "epoch": 8.436363636363636, + "grad_norm": 1.9412903372718, + "learning_rate": 1.2571923917247075e-06, + "loss": 0.0094, + "step": 6496 + }, + { + "epoch": 8.437662337662339, + "grad_norm": 2.325550352898212, + "learning_rate": 1.255151400855975e-06, + "loss": 0.0143, + "step": 6497 + }, + { + "epoch": 8.43896103896104, + "grad_norm": 0.7636581367174983, + "learning_rate": 1.253111957118336e-06, + "loss": 0.0077, + "step": 6498 + }, + { + "epoch": 8.44025974025974, + "grad_norm": 0.99533857216044, + "learning_rate": 1.2510740608726113e-06, + "loss": 0.0082, + "step": 6499 + }, + { + "epoch": 8.441558441558442, + "grad_norm": 2.1847666440932856, + "learning_rate": 1.2490377124793406e-06, + "loss": 0.0013, + "step": 6500 + }, + { + "epoch": 8.442857142857143, + "grad_norm": 6.130540875770173, + "learning_rate": 1.2470029122987947e-06, + "loss": 0.0184, + "step": 6501 + }, + { + "epoch": 8.444155844155844, + "grad_norm": 5.1719984479366365, + "learning_rate": 1.2449696606909656e-06, + "loss": 0.008, + "step": 6502 + }, + { + "epoch": 8.445454545454545, + "grad_norm": 3.999647957042696, + "learning_rate": 1.2429379580155776e-06, + "loss": 0.0134, + "step": 6503 + }, + { + "epoch": 8.446753246753246, + "grad_norm": 1.5106808575155664, + "learning_rate": 1.2409078046320778e-06, + "loss": 0.0013, + "step": 6504 + }, + { + "epoch": 8.448051948051948, + "grad_norm": 3.2419985553301403, + "learning_rate": 1.2388792008996376e-06, + "loss": 0.0014, + "step": 6505 + }, + { + "epoch": 8.449350649350649, + "grad_norm": 3.3846125535822686, + "learning_rate": 1.2368521471771556e-06, + "loss": 0.0278, + "step": 6506 + }, + { + "epoch": 8.45064935064935, + "grad_norm": 10.726015140884922, + "learning_rate": 1.2348266438232582e-06, + "loss": 0.0125, + "step": 6507 + }, + { + "epoch": 8.451948051948053, + "grad_norm": 3.2375209516083894, + "learning_rate": 1.2328026911962942e-06, + "loss": 0.0206, + "step": 6508 + }, + { + "epoch": 8.453246753246754, + "grad_norm": 0.8870297982432741, + "learning_rate": 1.230780289654341e-06, + "loss": 0.008, + "step": 6509 + }, + { + "epoch": 8.454545454545455, + "grad_norm": 3.146474430417453, + "learning_rate": 1.2287594395551961e-06, + "loss": 0.0141, + "step": 6510 + }, + { + "epoch": 8.455844155844156, + "grad_norm": 5.806117554775649, + "learning_rate": 1.226740141256395e-06, + "loss": 0.0019, + "step": 6511 + }, + { + "epoch": 8.457142857142857, + "grad_norm": 4.4046870610936635, + "learning_rate": 1.2247223951151799e-06, + "loss": 0.0021, + "step": 6512 + }, + { + "epoch": 8.458441558441558, + "grad_norm": 2.6685128681474475, + "learning_rate": 1.222706201488536e-06, + "loss": 0.0015, + "step": 6513 + }, + { + "epoch": 8.45974025974026, + "grad_norm": 2.7010812522437333, + "learning_rate": 1.2206915607331626e-06, + "loss": 0.0065, + "step": 6514 + }, + { + "epoch": 8.46103896103896, + "grad_norm": 2.7862350751873257, + "learning_rate": 1.2186784732054934e-06, + "loss": 0.002, + "step": 6515 + }, + { + "epoch": 8.462337662337662, + "grad_norm": 2.6758501465855877, + "learning_rate": 1.2166669392616747e-06, + "loss": 0.0021, + "step": 6516 + }, + { + "epoch": 8.463636363636363, + "grad_norm": 4.1591864375599465, + "learning_rate": 1.2146569592575919e-06, + "loss": 0.0017, + "step": 6517 + }, + { + "epoch": 8.464935064935064, + "grad_norm": 3.8074827929934525, + "learning_rate": 1.2126485335488437e-06, + "loss": 0.0018, + "step": 6518 + }, + { + "epoch": 8.466233766233767, + "grad_norm": 3.4193528571209066, + "learning_rate": 1.2106416624907658e-06, + "loss": 0.0228, + "step": 6519 + }, + { + "epoch": 8.467532467532468, + "grad_norm": 1.099121183911347, + "learning_rate": 1.2086363464384032e-06, + "loss": 0.015, + "step": 6520 + }, + { + "epoch": 8.46883116883117, + "grad_norm": 4.622811014618921, + "learning_rate": 1.2066325857465422e-06, + "loss": 0.0032, + "step": 6521 + }, + { + "epoch": 8.47012987012987, + "grad_norm": 6.414157613150749, + "learning_rate": 1.2046303807696835e-06, + "loss": 0.0088, + "step": 6522 + }, + { + "epoch": 8.471428571428572, + "grad_norm": 2.6339292299641968, + "learning_rate": 1.202629731862056e-06, + "loss": 0.0017, + "step": 6523 + }, + { + "epoch": 8.472727272727273, + "grad_norm": 0.9573544362169655, + "learning_rate": 1.200630639377609e-06, + "loss": 0.0084, + "step": 6524 + }, + { + "epoch": 8.474025974025974, + "grad_norm": 0.4057112232740261, + "learning_rate": 1.1986331036700272e-06, + "loss": 0.001, + "step": 6525 + }, + { + "epoch": 8.475324675324675, + "grad_norm": 0.6809404783201706, + "learning_rate": 1.1966371250927088e-06, + "loss": 0.001, + "step": 6526 + }, + { + "epoch": 8.476623376623376, + "grad_norm": 1.4877273286158106, + "learning_rate": 1.1946427039987784e-06, + "loss": 0.001, + "step": 6527 + }, + { + "epoch": 8.477922077922077, + "grad_norm": 1.76926103198961, + "learning_rate": 1.1926498407410924e-06, + "loss": 0.0021, + "step": 6528 + }, + { + "epoch": 8.479220779220778, + "grad_norm": 6.501963878372843, + "learning_rate": 1.1906585356722245e-06, + "loss": 0.0211, + "step": 6529 + }, + { + "epoch": 8.480519480519481, + "grad_norm": 1.2410954540225192, + "learning_rate": 1.1886687891444737e-06, + "loss": 0.0015, + "step": 6530 + }, + { + "epoch": 8.481818181818182, + "grad_norm": 1.1066115262192326, + "learning_rate": 1.186680601509863e-06, + "loss": 0.0011, + "step": 6531 + }, + { + "epoch": 8.483116883116883, + "grad_norm": 0.5337695562386401, + "learning_rate": 1.1846939731201456e-06, + "loss": 0.0079, + "step": 6532 + }, + { + "epoch": 8.484415584415585, + "grad_norm": 3.5546483271212583, + "learning_rate": 1.1827089043267902e-06, + "loss": 0.0162, + "step": 6533 + }, + { + "epoch": 8.485714285714286, + "grad_norm": 1.5985438382560067, + "learning_rate": 1.1807253954809939e-06, + "loss": 0.0029, + "step": 6534 + }, + { + "epoch": 8.487012987012987, + "grad_norm": 1.0447141643050835, + "learning_rate": 1.1787434469336767e-06, + "loss": 0.0013, + "step": 6535 + }, + { + "epoch": 8.488311688311688, + "grad_norm": 1.9565242011816597, + "learning_rate": 1.176763059035485e-06, + "loss": 0.0018, + "step": 6536 + }, + { + "epoch": 8.48961038961039, + "grad_norm": 1.3026321437128217, + "learning_rate": 1.1747842321367886e-06, + "loss": 0.0018, + "step": 6537 + }, + { + "epoch": 8.49090909090909, + "grad_norm": 0.3701927909796618, + "learning_rate": 1.1728069665876773e-06, + "loss": 0.0012, + "step": 6538 + }, + { + "epoch": 8.492207792207791, + "grad_norm": 0.9135477050270499, + "learning_rate": 1.1708312627379647e-06, + "loss": 0.0035, + "step": 6539 + }, + { + "epoch": 8.493506493506494, + "grad_norm": 2.0466822863412464, + "learning_rate": 1.1688571209371969e-06, + "loss": 0.0076, + "step": 6540 + }, + { + "epoch": 8.494805194805195, + "grad_norm": 2.55861418184047, + "learning_rate": 1.1668845415346342e-06, + "loss": 0.0029, + "step": 6541 + }, + { + "epoch": 8.496103896103897, + "grad_norm": 1.2370327174892846, + "learning_rate": 1.1649135248792642e-06, + "loss": 0.0025, + "step": 6542 + }, + { + "epoch": 8.497402597402598, + "grad_norm": 1.8902718606880504, + "learning_rate": 1.1629440713197937e-06, + "loss": 0.0019, + "step": 6543 + }, + { + "epoch": 8.498701298701299, + "grad_norm": 0.3132712038537562, + "learning_rate": 1.1609761812046638e-06, + "loss": 0.001, + "step": 6544 + }, + { + "epoch": 8.5, + "grad_norm": 0.34970931468979694, + "learning_rate": 1.1590098548820273e-06, + "loss": 0.0011, + "step": 6545 + }, + { + "epoch": 8.501298701298701, + "grad_norm": 1.0716666399693031, + "learning_rate": 1.1570450926997657e-06, + "loss": 0.0077, + "step": 6546 + }, + { + "epoch": 8.502597402597402, + "grad_norm": 0.41491794565935325, + "learning_rate": 1.155081895005481e-06, + "loss": 0.0013, + "step": 6547 + }, + { + "epoch": 8.503896103896103, + "grad_norm": 0.7393403298020872, + "learning_rate": 1.1531202621465043e-06, + "loss": 0.001, + "step": 6548 + }, + { + "epoch": 8.505194805194805, + "grad_norm": 0.3998091174750013, + "learning_rate": 1.1511601944698847e-06, + "loss": 0.0014, + "step": 6549 + }, + { + "epoch": 8.506493506493506, + "grad_norm": 1.693513911535196, + "learning_rate": 1.149201692322396e-06, + "loss": 0.0123, + "step": 6550 + }, + { + "epoch": 8.507792207792209, + "grad_norm": 2.2780142965993138, + "learning_rate": 1.1472447560505307e-06, + "loss": 0.0099, + "step": 6551 + }, + { + "epoch": 8.50909090909091, + "grad_norm": 3.374290021276805, + "learning_rate": 1.1452893860005166e-06, + "loss": 0.0084, + "step": 6552 + }, + { + "epoch": 8.51038961038961, + "grad_norm": 2.7465170485082053, + "learning_rate": 1.1433355825182868e-06, + "loss": 0.0023, + "step": 6553 + }, + { + "epoch": 8.511688311688312, + "grad_norm": 1.0832178262271501, + "learning_rate": 1.1413833459495127e-06, + "loss": 0.0074, + "step": 6554 + }, + { + "epoch": 8.512987012987013, + "grad_norm": 1.5568590777443208, + "learning_rate": 1.1394326766395791e-06, + "loss": 0.0091, + "step": 6555 + }, + { + "epoch": 8.514285714285714, + "grad_norm": 2.7389705644703626, + "learning_rate": 1.137483574933601e-06, + "loss": 0.0014, + "step": 6556 + }, + { + "epoch": 8.515584415584415, + "grad_norm": 5.518694434110813, + "learning_rate": 1.135536041176406e-06, + "loss": 0.0108, + "step": 6557 + }, + { + "epoch": 8.516883116883117, + "grad_norm": 8.25034676043785, + "learning_rate": 1.133590075712554e-06, + "loss": 0.0061, + "step": 6558 + }, + { + "epoch": 8.518181818181818, + "grad_norm": 2.5967232928739685, + "learning_rate": 1.1316456788863207e-06, + "loss": 0.0104, + "step": 6559 + }, + { + "epoch": 8.519480519480519, + "grad_norm": 0.49235316714118693, + "learning_rate": 1.1297028510417107e-06, + "loss": 0.0009, + "step": 6560 + }, + { + "epoch": 8.520779220779222, + "grad_norm": 1.2348017158938718, + "learning_rate": 1.1277615925224461e-06, + "loss": 0.009, + "step": 6561 + }, + { + "epoch": 8.522077922077923, + "grad_norm": 1.0047671293666187, + "learning_rate": 1.125821903671973e-06, + "loss": 0.0088, + "step": 6562 + }, + { + "epoch": 8.523376623376624, + "grad_norm": 6.287002938527389, + "learning_rate": 1.1238837848334571e-06, + "loss": 0.0056, + "step": 6563 + }, + { + "epoch": 8.524675324675325, + "grad_norm": 2.0210969567467014, + "learning_rate": 1.1219472363497897e-06, + "loss": 0.0013, + "step": 6564 + }, + { + "epoch": 8.525974025974026, + "grad_norm": 3.7549033069817193, + "learning_rate": 1.1200122585635853e-06, + "loss": 0.0064, + "step": 6565 + }, + { + "epoch": 8.527272727272727, + "grad_norm": 1.8276333961261506, + "learning_rate": 1.118078851817177e-06, + "loss": 0.0075, + "step": 6566 + }, + { + "epoch": 8.528571428571428, + "grad_norm": 2.322853852754734, + "learning_rate": 1.1161470164526223e-06, + "loss": 0.0013, + "step": 6567 + }, + { + "epoch": 8.52987012987013, + "grad_norm": 1.5593766734139103, + "learning_rate": 1.1142167528116965e-06, + "loss": 0.0062, + "step": 6568 + }, + { + "epoch": 8.53116883116883, + "grad_norm": 4.5067778993132706, + "learning_rate": 1.1122880612359043e-06, + "loss": 0.0122, + "step": 6569 + }, + { + "epoch": 8.532467532467532, + "grad_norm": 2.388205268444977, + "learning_rate": 1.1103609420664652e-06, + "loss": 0.0165, + "step": 6570 + }, + { + "epoch": 8.533766233766233, + "grad_norm": 3.4072792206974913, + "learning_rate": 1.1084353956443295e-06, + "loss": 0.0063, + "step": 6571 + }, + { + "epoch": 8.535064935064934, + "grad_norm": 2.8687470189059123, + "learning_rate": 1.1065114223101547e-06, + "loss": 0.0013, + "step": 6572 + }, + { + "epoch": 8.536363636363637, + "grad_norm": 5.138658615570322, + "learning_rate": 1.1045890224043343e-06, + "loss": 0.0134, + "step": 6573 + }, + { + "epoch": 8.537662337662338, + "grad_norm": 2.338798129988818, + "learning_rate": 1.1026681962669728e-06, + "loss": 0.0011, + "step": 6574 + }, + { + "epoch": 8.53896103896104, + "grad_norm": 0.5026635329817312, + "learning_rate": 1.1007489442379093e-06, + "loss": 0.0083, + "step": 6575 + }, + { + "epoch": 8.54025974025974, + "grad_norm": 2.8636703149856526, + "learning_rate": 1.098831266656687e-06, + "loss": 0.0082, + "step": 6576 + }, + { + "epoch": 8.541558441558442, + "grad_norm": 1.1852266370850237, + "learning_rate": 1.0969151638625852e-06, + "loss": 0.0018, + "step": 6577 + }, + { + "epoch": 8.542857142857143, + "grad_norm": 5.237833276159303, + "learning_rate": 1.0950006361945974e-06, + "loss": 0.0109, + "step": 6578 + }, + { + "epoch": 8.544155844155844, + "grad_norm": 5.878819251507727, + "learning_rate": 1.0930876839914418e-06, + "loss": 0.0157, + "step": 6579 + }, + { + "epoch": 8.545454545454545, + "grad_norm": 0.8306259141050125, + "learning_rate": 1.0911763075915527e-06, + "loss": 0.0015, + "step": 6580 + }, + { + "epoch": 8.546753246753246, + "grad_norm": 1.1241189822276598, + "learning_rate": 1.0892665073330932e-06, + "loss": 0.0018, + "step": 6581 + }, + { + "epoch": 8.548051948051947, + "grad_norm": 1.526826690069796, + "learning_rate": 1.0873582835539408e-06, + "loss": 0.001, + "step": 6582 + }, + { + "epoch": 8.54935064935065, + "grad_norm": 5.304558338305907, + "learning_rate": 1.085451636591699e-06, + "loss": 0.006, + "step": 6583 + }, + { + "epoch": 8.550649350649351, + "grad_norm": 5.670698148877469, + "learning_rate": 1.0835465667836876e-06, + "loss": 0.0027, + "step": 6584 + }, + { + "epoch": 8.551948051948052, + "grad_norm": 2.16128240094735, + "learning_rate": 1.0816430744669527e-06, + "loss": 0.0049, + "step": 6585 + }, + { + "epoch": 8.553246753246754, + "grad_norm": 3.2411569707683454, + "learning_rate": 1.0797411599782558e-06, + "loss": 0.0159, + "step": 6586 + }, + { + "epoch": 8.554545454545455, + "grad_norm": 1.2910733859326855, + "learning_rate": 1.077840823654085e-06, + "loss": 0.0023, + "step": 6587 + }, + { + "epoch": 8.555844155844156, + "grad_norm": 0.25838248275044323, + "learning_rate": 1.075942065830642e-06, + "loss": 0.0009, + "step": 6588 + }, + { + "epoch": 8.557142857142857, + "grad_norm": 0.7810086772247422, + "learning_rate": 1.074044886843857e-06, + "loss": 0.0088, + "step": 6589 + }, + { + "epoch": 8.558441558441558, + "grad_norm": 4.314191679103135, + "learning_rate": 1.072149287029377e-06, + "loss": 0.0054, + "step": 6590 + }, + { + "epoch": 8.55974025974026, + "grad_norm": 1.7825911789740483, + "learning_rate": 1.0702552667225685e-06, + "loss": 0.0012, + "step": 6591 + }, + { + "epoch": 8.56103896103896, + "grad_norm": 3.036750773346681, + "learning_rate": 1.0683628262585187e-06, + "loss": 0.0015, + "step": 6592 + }, + { + "epoch": 8.562337662337661, + "grad_norm": 2.3079029875498596, + "learning_rate": 1.0664719659720401e-06, + "loss": 0.0012, + "step": 6593 + }, + { + "epoch": 8.563636363636364, + "grad_norm": 0.5182131832988816, + "learning_rate": 1.0645826861976615e-06, + "loss": 0.008, + "step": 6594 + }, + { + "epoch": 8.564935064935066, + "grad_norm": 1.97227529982949, + "learning_rate": 1.0626949872696302e-06, + "loss": 0.0132, + "step": 6595 + }, + { + "epoch": 8.566233766233767, + "grad_norm": 1.3752894295100495, + "learning_rate": 1.0608088695219154e-06, + "loss": 0.0091, + "step": 6596 + }, + { + "epoch": 8.567532467532468, + "grad_norm": 3.565865301484552, + "learning_rate": 1.0589243332882114e-06, + "loss": 0.0084, + "step": 6597 + }, + { + "epoch": 8.568831168831169, + "grad_norm": 0.626901962737372, + "learning_rate": 1.0570413789019274e-06, + "loss": 0.0015, + "step": 6598 + }, + { + "epoch": 8.57012987012987, + "grad_norm": 0.4185141631385277, + "learning_rate": 1.055160006696192e-06, + "loss": 0.0013, + "step": 6599 + }, + { + "epoch": 8.571428571428571, + "grad_norm": 3.1564847009389236, + "learning_rate": 1.0532802170038591e-06, + "loss": 0.0054, + "step": 6600 + }, + { + "epoch": 8.572727272727272, + "grad_norm": 0.7221499440581137, + "learning_rate": 1.0514020101574974e-06, + "loss": 0.001, + "step": 6601 + }, + { + "epoch": 8.574025974025973, + "grad_norm": 0.5681292622391176, + "learning_rate": 1.0495253864893983e-06, + "loss": 0.0079, + "step": 6602 + }, + { + "epoch": 8.575324675324675, + "grad_norm": 0.5247794659146546, + "learning_rate": 1.0476503463315702e-06, + "loss": 0.0014, + "step": 6603 + }, + { + "epoch": 8.576623376623377, + "grad_norm": 0.5873315897860771, + "learning_rate": 1.0457768900157496e-06, + "loss": 0.0009, + "step": 6604 + }, + { + "epoch": 8.577922077922079, + "grad_norm": 4.545706379648674, + "learning_rate": 1.043905017873379e-06, + "loss": 0.0168, + "step": 6605 + }, + { + "epoch": 8.57922077922078, + "grad_norm": 1.8789949152066194, + "learning_rate": 1.0420347302356337e-06, + "loss": 0.0019, + "step": 6606 + }, + { + "epoch": 8.58051948051948, + "grad_norm": 2.1859071958486083, + "learning_rate": 1.0401660274334013e-06, + "loss": 0.0024, + "step": 6607 + }, + { + "epoch": 8.581818181818182, + "grad_norm": 5.6958922525649935, + "learning_rate": 1.038298909797294e-06, + "loss": 0.0134, + "step": 6608 + }, + { + "epoch": 8.583116883116883, + "grad_norm": 0.5009314443572981, + "learning_rate": 1.036433377657635e-06, + "loss": 0.0012, + "step": 6609 + }, + { + "epoch": 8.584415584415584, + "grad_norm": 1.9260418413869973, + "learning_rate": 1.0345694313444788e-06, + "loss": 0.0076, + "step": 6610 + }, + { + "epoch": 8.585714285714285, + "grad_norm": 1.7019603587241257, + "learning_rate": 1.0327070711875875e-06, + "loss": 0.0028, + "step": 6611 + }, + { + "epoch": 8.587012987012987, + "grad_norm": 1.806817564705184, + "learning_rate": 1.0308462975164547e-06, + "loss": 0.0014, + "step": 6612 + }, + { + "epoch": 8.588311688311688, + "grad_norm": 2.221446759698117, + "learning_rate": 1.02898711066028e-06, + "loss": 0.0088, + "step": 6613 + }, + { + "epoch": 8.589610389610389, + "grad_norm": 0.6367744922717781, + "learning_rate": 1.027129510947995e-06, + "loss": 0.0015, + "step": 6614 + }, + { + "epoch": 8.590909090909092, + "grad_norm": 1.9153159422722548, + "learning_rate": 1.0252734987082391e-06, + "loss": 0.0013, + "step": 6615 + }, + { + "epoch": 8.592207792207793, + "grad_norm": 0.7993384247937267, + "learning_rate": 1.023419074269384e-06, + "loss": 0.0079, + "step": 6616 + }, + { + "epoch": 8.593506493506494, + "grad_norm": 1.8524445203322544, + "learning_rate": 1.0215662379595038e-06, + "loss": 0.0017, + "step": 6617 + }, + { + "epoch": 8.594805194805195, + "grad_norm": 1.3248071029388142, + "learning_rate": 1.0197149901064074e-06, + "loss": 0.0011, + "step": 6618 + }, + { + "epoch": 8.596103896103896, + "grad_norm": 0.883286342198722, + "learning_rate": 1.017865331037613e-06, + "loss": 0.0083, + "step": 6619 + }, + { + "epoch": 8.597402597402597, + "grad_norm": 5.100962177687283, + "learning_rate": 1.016017261080362e-06, + "loss": 0.0123, + "step": 6620 + }, + { + "epoch": 8.598701298701299, + "grad_norm": 4.350952876600733, + "learning_rate": 1.0141707805616109e-06, + "loss": 0.0016, + "step": 6621 + }, + { + "epoch": 8.6, + "grad_norm": 1.3431911247644597, + "learning_rate": 1.0123258898080413e-06, + "loss": 0.0016, + "step": 6622 + }, + { + "epoch": 8.6012987012987, + "grad_norm": 1.4223040218070906, + "learning_rate": 1.010482589146048e-06, + "loss": 0.0086, + "step": 6623 + }, + { + "epoch": 8.602597402597402, + "grad_norm": 2.2429095502542262, + "learning_rate": 1.0086408789017454e-06, + "loss": 0.0011, + "step": 6624 + }, + { + "epoch": 8.603896103896105, + "grad_norm": 1.5894525079124813, + "learning_rate": 1.0068007594009654e-06, + "loss": 0.0018, + "step": 6625 + }, + { + "epoch": 8.605194805194806, + "grad_norm": 4.878569803893205, + "learning_rate": 1.0049622309692652e-06, + "loss": 0.0043, + "step": 6626 + }, + { + "epoch": 8.606493506493507, + "grad_norm": 2.0397771963813174, + "learning_rate": 1.0031252939319136e-06, + "loss": 0.011, + "step": 6627 + }, + { + "epoch": 8.607792207792208, + "grad_norm": 1.2016353968180107, + "learning_rate": 1.0012899486138994e-06, + "loss": 0.0011, + "step": 6628 + }, + { + "epoch": 8.60909090909091, + "grad_norm": 2.222326048860378, + "learning_rate": 9.994561953399284e-07, + "loss": 0.0023, + "step": 6629 + }, + { + "epoch": 8.61038961038961, + "grad_norm": 0.9395324428350165, + "learning_rate": 9.976240344344302e-07, + "loss": 0.0016, + "step": 6630 + }, + { + "epoch": 8.611688311688312, + "grad_norm": 7.054519803824935, + "learning_rate": 9.957934662215485e-07, + "loss": 0.0034, + "step": 6631 + }, + { + "epoch": 8.612987012987013, + "grad_norm": 0.32086951770924343, + "learning_rate": 9.939644910251422e-07, + "loss": 0.0013, + "step": 6632 + }, + { + "epoch": 8.614285714285714, + "grad_norm": 3.348133789683201, + "learning_rate": 9.92137109168797e-07, + "loss": 0.016, + "step": 6633 + }, + { + "epoch": 8.615584415584415, + "grad_norm": 0.963214343711734, + "learning_rate": 9.903113209758098e-07, + "loss": 0.0017, + "step": 6634 + }, + { + "epoch": 8.616883116883116, + "grad_norm": 0.9494149133221969, + "learning_rate": 9.884871267691964e-07, + "loss": 0.0087, + "step": 6635 + }, + { + "epoch": 8.618181818181817, + "grad_norm": 0.949603353638582, + "learning_rate": 9.866645268716912e-07, + "loss": 0.0069, + "step": 6636 + }, + { + "epoch": 8.61948051948052, + "grad_norm": 0.6645195982894816, + "learning_rate": 9.848435216057496e-07, + "loss": 0.0015, + "step": 6637 + }, + { + "epoch": 8.620779220779221, + "grad_norm": 1.3268823647499215, + "learning_rate": 9.830241112935412e-07, + "loss": 0.0009, + "step": 6638 + }, + { + "epoch": 8.622077922077922, + "grad_norm": 1.0759141513457844, + "learning_rate": 9.812062962569546e-07, + "loss": 0.0014, + "step": 6639 + }, + { + "epoch": 8.623376623376624, + "grad_norm": 0.6741973260233222, + "learning_rate": 9.793900768175946e-07, + "loss": 0.0009, + "step": 6640 + }, + { + "epoch": 8.624675324675325, + "grad_norm": 0.7728679106172865, + "learning_rate": 9.775754532967863e-07, + "loss": 0.0012, + "step": 6641 + }, + { + "epoch": 8.625974025974026, + "grad_norm": 3.33093354867827, + "learning_rate": 9.757624260155728e-07, + "loss": 0.0082, + "step": 6642 + }, + { + "epoch": 8.627272727272727, + "grad_norm": 2.8257609278304985, + "learning_rate": 9.739509952947113e-07, + "loss": 0.0098, + "step": 6643 + }, + { + "epoch": 8.628571428571428, + "grad_norm": 0.9192315518952682, + "learning_rate": 9.721411614546772e-07, + "loss": 0.0014, + "step": 6644 + }, + { + "epoch": 8.62987012987013, + "grad_norm": 1.8720607391443027, + "learning_rate": 9.703329248156712e-07, + "loss": 0.0041, + "step": 6645 + }, + { + "epoch": 8.63116883116883, + "grad_norm": 4.851924054835003, + "learning_rate": 9.685262856975963e-07, + "loss": 0.0029, + "step": 6646 + }, + { + "epoch": 8.632467532467533, + "grad_norm": 0.747699492811453, + "learning_rate": 9.667212444200879e-07, + "loss": 0.0032, + "step": 6647 + }, + { + "epoch": 8.633766233766234, + "grad_norm": 1.0684965252696303, + "learning_rate": 9.649178013024873e-07, + "loss": 0.0013, + "step": 6648 + }, + { + "epoch": 8.635064935064936, + "grad_norm": 2.019876054322069, + "learning_rate": 9.631159566638659e-07, + "loss": 0.0052, + "step": 6649 + }, + { + "epoch": 8.636363636363637, + "grad_norm": 0.9859447143888679, + "learning_rate": 9.613157108229952e-07, + "loss": 0.0081, + "step": 6650 + }, + { + "epoch": 8.637662337662338, + "grad_norm": 0.5961860678660174, + "learning_rate": 9.595170640983786e-07, + "loss": 0.0011, + "step": 6651 + }, + { + "epoch": 8.638961038961039, + "grad_norm": 7.366226318047512, + "learning_rate": 9.57720016808229e-07, + "loss": 0.0049, + "step": 6652 + }, + { + "epoch": 8.64025974025974, + "grad_norm": 1.4315293116351107, + "learning_rate": 9.559245692704834e-07, + "loss": 0.0077, + "step": 6653 + }, + { + "epoch": 8.641558441558441, + "grad_norm": 1.012078230144823, + "learning_rate": 9.54130721802784e-07, + "loss": 0.0081, + "step": 6654 + }, + { + "epoch": 8.642857142857142, + "grad_norm": 0.2835633278144155, + "learning_rate": 9.523384747225007e-07, + "loss": 0.001, + "step": 6655 + }, + { + "epoch": 8.644155844155843, + "grad_norm": 3.3891340012239772, + "learning_rate": 9.505478283467129e-07, + "loss": 0.0028, + "step": 6656 + }, + { + "epoch": 8.645454545454545, + "grad_norm": 0.7357934861757071, + "learning_rate": 9.487587829922263e-07, + "loss": 0.0083, + "step": 6657 + }, + { + "epoch": 8.646753246753248, + "grad_norm": 1.9501024410661043, + "learning_rate": 9.469713389755508e-07, + "loss": 0.0013, + "step": 6658 + }, + { + "epoch": 8.648051948051949, + "grad_norm": 1.9032885793375018, + "learning_rate": 9.451854966129237e-07, + "loss": 0.0017, + "step": 6659 + }, + { + "epoch": 8.64935064935065, + "grad_norm": 1.0298008116075057, + "learning_rate": 9.434012562202932e-07, + "loss": 0.0024, + "step": 6660 + }, + { + "epoch": 8.650649350649351, + "grad_norm": 0.5068416574238291, + "learning_rate": 9.416186181133236e-07, + "loss": 0.0013, + "step": 6661 + }, + { + "epoch": 8.651948051948052, + "grad_norm": 5.405043070885156, + "learning_rate": 9.398375826074024e-07, + "loss": 0.0061, + "step": 6662 + }, + { + "epoch": 8.653246753246753, + "grad_norm": 1.28133907942058, + "learning_rate": 9.380581500176256e-07, + "loss": 0.01, + "step": 6663 + }, + { + "epoch": 8.654545454545454, + "grad_norm": 1.253259421365018, + "learning_rate": 9.362803206588089e-07, + "loss": 0.0015, + "step": 6664 + }, + { + "epoch": 8.655844155844155, + "grad_norm": 1.3857912848864473, + "learning_rate": 9.345040948454842e-07, + "loss": 0.0102, + "step": 6665 + }, + { + "epoch": 8.657142857142857, + "grad_norm": 0.6204536593259712, + "learning_rate": 9.327294728919012e-07, + "loss": 0.0024, + "step": 6666 + }, + { + "epoch": 8.658441558441558, + "grad_norm": 2.3535389738065904, + "learning_rate": 9.309564551120254e-07, + "loss": 0.003, + "step": 6667 + }, + { + "epoch": 8.65974025974026, + "grad_norm": 2.9365333555526307, + "learning_rate": 9.291850418195347e-07, + "loss": 0.0212, + "step": 6668 + }, + { + "epoch": 8.661038961038962, + "grad_norm": 1.576781665863075, + "learning_rate": 9.274152333278263e-07, + "loss": 0.0009, + "step": 6669 + }, + { + "epoch": 8.662337662337663, + "grad_norm": 1.0208689248918004, + "learning_rate": 9.256470299500153e-07, + "loss": 0.0101, + "step": 6670 + }, + { + "epoch": 8.663636363636364, + "grad_norm": 5.002616723274495, + "learning_rate": 9.238804319989303e-07, + "loss": 0.013, + "step": 6671 + }, + { + "epoch": 8.664935064935065, + "grad_norm": 2.296322851960507, + "learning_rate": 9.221154397871157e-07, + "loss": 0.0012, + "step": 6672 + }, + { + "epoch": 8.666233766233766, + "grad_norm": 2.1287443245927693, + "learning_rate": 9.203520536268295e-07, + "loss": 0.002, + "step": 6673 + }, + { + "epoch": 8.667532467532467, + "grad_norm": 3.4744560318383306, + "learning_rate": 9.185902738300545e-07, + "loss": 0.0029, + "step": 6674 + }, + { + "epoch": 8.668831168831169, + "grad_norm": 2.136500001627992, + "learning_rate": 9.16830100708479e-07, + "loss": 0.0015, + "step": 6675 + }, + { + "epoch": 8.67012987012987, + "grad_norm": 4.413475577032968, + "learning_rate": 9.150715345735139e-07, + "loss": 0.0152, + "step": 6676 + }, + { + "epoch": 8.67142857142857, + "grad_norm": 1.1222831519387673, + "learning_rate": 9.133145757362782e-07, + "loss": 0.0086, + "step": 6677 + }, + { + "epoch": 8.672727272727272, + "grad_norm": 2.1791265164715896, + "learning_rate": 9.115592245076177e-07, + "loss": 0.0013, + "step": 6678 + }, + { + "epoch": 8.674025974025973, + "grad_norm": 1.7008685833686237, + "learning_rate": 9.098054811980849e-07, + "loss": 0.0018, + "step": 6679 + }, + { + "epoch": 8.675324675324676, + "grad_norm": 5.656934491226225, + "learning_rate": 9.080533461179508e-07, + "loss": 0.0051, + "step": 6680 + }, + { + "epoch": 8.676623376623377, + "grad_norm": 3.5889040894808857, + "learning_rate": 9.063028195771994e-07, + "loss": 0.0263, + "step": 6681 + }, + { + "epoch": 8.677922077922078, + "grad_norm": 1.4311659807517865, + "learning_rate": 9.045539018855365e-07, + "loss": 0.0013, + "step": 6682 + }, + { + "epoch": 8.67922077922078, + "grad_norm": 3.3534151316779077, + "learning_rate": 9.028065933523777e-07, + "loss": 0.0076, + "step": 6683 + }, + { + "epoch": 8.68051948051948, + "grad_norm": 5.861646142058915, + "learning_rate": 9.010608942868537e-07, + "loss": 0.0063, + "step": 6684 + }, + { + "epoch": 8.681818181818182, + "grad_norm": 0.36601289795950076, + "learning_rate": 8.993168049978118e-07, + "loss": 0.0011, + "step": 6685 + }, + { + "epoch": 8.683116883116883, + "grad_norm": 1.034526335240476, + "learning_rate": 8.975743257938186e-07, + "loss": 0.0013, + "step": 6686 + }, + { + "epoch": 8.684415584415584, + "grad_norm": 1.2678223410427802, + "learning_rate": 8.958334569831473e-07, + "loss": 0.0015, + "step": 6687 + }, + { + "epoch": 8.685714285714285, + "grad_norm": 1.6318941528723163, + "learning_rate": 8.940941988737939e-07, + "loss": 0.0013, + "step": 6688 + }, + { + "epoch": 8.687012987012986, + "grad_norm": 0.3434112765631074, + "learning_rate": 8.923565517734633e-07, + "loss": 0.0011, + "step": 6689 + }, + { + "epoch": 8.688311688311689, + "grad_norm": 2.0740853872818827, + "learning_rate": 8.906205159895853e-07, + "loss": 0.0039, + "step": 6690 + }, + { + "epoch": 8.68961038961039, + "grad_norm": 0.2954880570253002, + "learning_rate": 8.888860918292908e-07, + "loss": 0.0009, + "step": 6691 + }, + { + "epoch": 8.690909090909091, + "grad_norm": 0.47312341567012856, + "learning_rate": 8.871532795994354e-07, + "loss": 0.0012, + "step": 6692 + }, + { + "epoch": 8.692207792207792, + "grad_norm": 2.1613949392559273, + "learning_rate": 8.854220796065849e-07, + "loss": 0.0096, + "step": 6693 + }, + { + "epoch": 8.693506493506494, + "grad_norm": 1.3018518581324405, + "learning_rate": 8.836924921570256e-07, + "loss": 0.0016, + "step": 6694 + }, + { + "epoch": 8.694805194805195, + "grad_norm": 3.4330123998924513, + "learning_rate": 8.819645175567527e-07, + "loss": 0.0029, + "step": 6695 + }, + { + "epoch": 8.696103896103896, + "grad_norm": 0.584708685409116, + "learning_rate": 8.802381561114781e-07, + "loss": 0.0015, + "step": 6696 + }, + { + "epoch": 8.697402597402597, + "grad_norm": 0.5964923696835259, + "learning_rate": 8.785134081266244e-07, + "loss": 0.0014, + "step": 6697 + }, + { + "epoch": 8.698701298701298, + "grad_norm": 1.3344631616954177, + "learning_rate": 8.767902739073398e-07, + "loss": 0.0016, + "step": 6698 + }, + { + "epoch": 8.7, + "grad_norm": 1.9270016460691768, + "learning_rate": 8.750687537584745e-07, + "loss": 0.0167, + "step": 6699 + }, + { + "epoch": 8.7012987012987, + "grad_norm": 1.515172603111829, + "learning_rate": 8.733488479845997e-07, + "loss": 0.0034, + "step": 6700 + }, + { + "epoch": 8.702597402597403, + "grad_norm": 3.4354045887902926, + "learning_rate": 8.716305568899997e-07, + "loss": 0.0065, + "step": 6701 + }, + { + "epoch": 8.703896103896104, + "grad_norm": 0.7775260133571893, + "learning_rate": 8.699138807786711e-07, + "loss": 0.0017, + "step": 6702 + }, + { + "epoch": 8.705194805194806, + "grad_norm": 3.7296576949313507, + "learning_rate": 8.681988199543312e-07, + "loss": 0.019, + "step": 6703 + }, + { + "epoch": 8.706493506493507, + "grad_norm": 0.6945716098490822, + "learning_rate": 8.664853747204028e-07, + "loss": 0.0086, + "step": 6704 + }, + { + "epoch": 8.707792207792208, + "grad_norm": 1.1597020407506025, + "learning_rate": 8.647735453800322e-07, + "loss": 0.0011, + "step": 6705 + }, + { + "epoch": 8.709090909090909, + "grad_norm": 4.8421337870059675, + "learning_rate": 8.63063332236067e-07, + "loss": 0.0069, + "step": 6706 + }, + { + "epoch": 8.71038961038961, + "grad_norm": 0.7895013098551567, + "learning_rate": 8.613547355910845e-07, + "loss": 0.0078, + "step": 6707 + }, + { + "epoch": 8.711688311688311, + "grad_norm": 1.7014290369241771, + "learning_rate": 8.596477557473615e-07, + "loss": 0.0014, + "step": 6708 + }, + { + "epoch": 8.712987012987012, + "grad_norm": 1.484687945990477, + "learning_rate": 8.579423930069042e-07, + "loss": 0.0105, + "step": 6709 + }, + { + "epoch": 8.714285714285714, + "grad_norm": 1.5785968928988652, + "learning_rate": 8.562386476714135e-07, + "loss": 0.0013, + "step": 6710 + }, + { + "epoch": 8.715584415584416, + "grad_norm": 3.3462838552582097, + "learning_rate": 8.545365200423216e-07, + "loss": 0.0027, + "step": 6711 + }, + { + "epoch": 8.716883116883118, + "grad_norm": 0.4840729585228598, + "learning_rate": 8.528360104207644e-07, + "loss": 0.0013, + "step": 6712 + }, + { + "epoch": 8.718181818181819, + "grad_norm": 1.1378480739131234, + "learning_rate": 8.511371191075968e-07, + "loss": 0.0013, + "step": 6713 + }, + { + "epoch": 8.71948051948052, + "grad_norm": 3.2419914010547943, + "learning_rate": 8.494398464033815e-07, + "loss": 0.0028, + "step": 6714 + }, + { + "epoch": 8.720779220779221, + "grad_norm": 2.085552529180524, + "learning_rate": 8.47744192608404e-07, + "loss": 0.0145, + "step": 6715 + }, + { + "epoch": 8.722077922077922, + "grad_norm": 1.1312575303419705, + "learning_rate": 8.460501580226532e-07, + "loss": 0.0014, + "step": 6716 + }, + { + "epoch": 8.723376623376623, + "grad_norm": 0.9445509698079555, + "learning_rate": 8.443577429458394e-07, + "loss": 0.0015, + "step": 6717 + }, + { + "epoch": 8.724675324675324, + "grad_norm": 1.4984568784955337, + "learning_rate": 8.426669476773786e-07, + "loss": 0.0022, + "step": 6718 + }, + { + "epoch": 8.725974025974025, + "grad_norm": 0.4443161488961288, + "learning_rate": 8.409777725164103e-07, + "loss": 0.0012, + "step": 6719 + }, + { + "epoch": 8.727272727272727, + "grad_norm": 0.3020819026804785, + "learning_rate": 8.392902177617801e-07, + "loss": 0.0011, + "step": 6720 + }, + { + "epoch": 8.728571428571428, + "grad_norm": 1.2082896656649396, + "learning_rate": 8.37604283712048e-07, + "loss": 0.0088, + "step": 6721 + }, + { + "epoch": 8.72987012987013, + "grad_norm": 4.034246338161194, + "learning_rate": 8.359199706654852e-07, + "loss": 0.0113, + "step": 6722 + }, + { + "epoch": 8.731168831168832, + "grad_norm": 4.527496023543423, + "learning_rate": 8.342372789200848e-07, + "loss": 0.0075, + "step": 6723 + }, + { + "epoch": 8.732467532467533, + "grad_norm": 3.6698671652364494, + "learning_rate": 8.32556208773545e-07, + "loss": 0.0111, + "step": 6724 + }, + { + "epoch": 8.733766233766234, + "grad_norm": 0.8415069763175481, + "learning_rate": 8.308767605232781e-07, + "loss": 0.0146, + "step": 6725 + }, + { + "epoch": 8.735064935064935, + "grad_norm": 4.357147682864052, + "learning_rate": 8.291989344664086e-07, + "loss": 0.0142, + "step": 6726 + }, + { + "epoch": 8.736363636363636, + "grad_norm": 2.9937415717924605, + "learning_rate": 8.275227308997813e-07, + "loss": 0.0049, + "step": 6727 + }, + { + "epoch": 8.737662337662337, + "grad_norm": 0.5388739165872288, + "learning_rate": 8.258481501199455e-07, + "loss": 0.003, + "step": 6728 + }, + { + "epoch": 8.738961038961039, + "grad_norm": 0.9991091165819421, + "learning_rate": 8.241751924231666e-07, + "loss": 0.0018, + "step": 6729 + }, + { + "epoch": 8.74025974025974, + "grad_norm": 1.1087826952993574, + "learning_rate": 8.225038581054223e-07, + "loss": 0.008, + "step": 6730 + }, + { + "epoch": 8.74155844155844, + "grad_norm": 0.8989392548710223, + "learning_rate": 8.208341474624071e-07, + "loss": 0.0013, + "step": 6731 + }, + { + "epoch": 8.742857142857144, + "grad_norm": 2.1586742284886817, + "learning_rate": 8.191660607895214e-07, + "loss": 0.0132, + "step": 6732 + }, + { + "epoch": 8.744155844155845, + "grad_norm": 0.8414371754320649, + "learning_rate": 8.174995983818801e-07, + "loss": 0.0145, + "step": 6733 + }, + { + "epoch": 8.745454545454546, + "grad_norm": 0.30803784289915676, + "learning_rate": 8.158347605343176e-07, + "loss": 0.0013, + "step": 6734 + }, + { + "epoch": 8.746753246753247, + "grad_norm": 5.29016937899635, + "learning_rate": 8.141715475413725e-07, + "loss": 0.006, + "step": 6735 + }, + { + "epoch": 8.748051948051948, + "grad_norm": 13.14511574244913, + "learning_rate": 8.125099596972996e-07, + "loss": 0.0116, + "step": 6736 + }, + { + "epoch": 8.74935064935065, + "grad_norm": 1.5628174278413316, + "learning_rate": 8.108499972960626e-07, + "loss": 0.0019, + "step": 6737 + }, + { + "epoch": 8.75064935064935, + "grad_norm": 0.6013422457166118, + "learning_rate": 8.091916606313465e-07, + "loss": 0.0011, + "step": 6738 + }, + { + "epoch": 8.751948051948052, + "grad_norm": 4.298283709439762, + "learning_rate": 8.075349499965379e-07, + "loss": 0.0087, + "step": 6739 + }, + { + "epoch": 8.753246753246753, + "grad_norm": 6.112279815226919, + "learning_rate": 8.058798656847433e-07, + "loss": 0.0094, + "step": 6740 + }, + { + "epoch": 8.754545454545454, + "grad_norm": 9.281745282154763, + "learning_rate": 8.042264079887762e-07, + "loss": 0.0095, + "step": 6741 + }, + { + "epoch": 8.755844155844155, + "grad_norm": 3.288726413810338, + "learning_rate": 8.025745772011695e-07, + "loss": 0.0084, + "step": 6742 + }, + { + "epoch": 8.757142857142856, + "grad_norm": 1.1101185263354165, + "learning_rate": 8.009243736141581e-07, + "loss": 0.0012, + "step": 6743 + }, + { + "epoch": 8.75844155844156, + "grad_norm": 4.2449147925456945, + "learning_rate": 7.992757975196974e-07, + "loss": 0.0024, + "step": 6744 + }, + { + "epoch": 8.75974025974026, + "grad_norm": 1.9796439563600075, + "learning_rate": 7.976288492094508e-07, + "loss": 0.0235, + "step": 6745 + }, + { + "epoch": 8.761038961038961, + "grad_norm": 0.8256448554385979, + "learning_rate": 7.959835289747997e-07, + "loss": 0.0021, + "step": 6746 + }, + { + "epoch": 8.762337662337663, + "grad_norm": 2.1518945514839585, + "learning_rate": 7.943398371068256e-07, + "loss": 0.0014, + "step": 6747 + }, + { + "epoch": 8.763636363636364, + "grad_norm": 2.817469084456827, + "learning_rate": 7.926977738963348e-07, + "loss": 0.0232, + "step": 6748 + }, + { + "epoch": 8.764935064935065, + "grad_norm": 10.068403741634672, + "learning_rate": 7.91057339633835e-07, + "loss": 0.01, + "step": 6749 + }, + { + "epoch": 8.766233766233766, + "grad_norm": 1.6600930255575335, + "learning_rate": 7.894185346095562e-07, + "loss": 0.0014, + "step": 6750 + }, + { + "epoch": 8.767532467532467, + "grad_norm": 0.44821002251226727, + "learning_rate": 7.877813591134276e-07, + "loss": 0.0011, + "step": 6751 + }, + { + "epoch": 8.768831168831168, + "grad_norm": 0.5388371227312501, + "learning_rate": 7.86145813435103e-07, + "loss": 0.0012, + "step": 6752 + }, + { + "epoch": 8.77012987012987, + "grad_norm": 1.3777191031314195, + "learning_rate": 7.845118978639376e-07, + "loss": 0.0075, + "step": 6753 + }, + { + "epoch": 8.771428571428572, + "grad_norm": 4.177837292097478, + "learning_rate": 7.828796126890048e-07, + "loss": 0.0151, + "step": 6754 + }, + { + "epoch": 8.772727272727273, + "grad_norm": 4.4423772866349625, + "learning_rate": 7.812489581990846e-07, + "loss": 0.0056, + "step": 6755 + }, + { + "epoch": 8.774025974025975, + "grad_norm": 1.0497424269186064, + "learning_rate": 7.796199346826727e-07, + "loss": 0.0101, + "step": 6756 + }, + { + "epoch": 8.775324675324676, + "grad_norm": 4.046620336037784, + "learning_rate": 7.779925424279755e-07, + "loss": 0.0174, + "step": 6757 + }, + { + "epoch": 8.776623376623377, + "grad_norm": 0.7798012138630154, + "learning_rate": 7.763667817229092e-07, + "loss": 0.0079, + "step": 6758 + }, + { + "epoch": 8.777922077922078, + "grad_norm": 3.0661546726304216, + "learning_rate": 7.747426528550983e-07, + "loss": 0.0016, + "step": 6759 + }, + { + "epoch": 8.779220779220779, + "grad_norm": 1.7628859804801558, + "learning_rate": 7.731201561118884e-07, + "loss": 0.0021, + "step": 6760 + }, + { + "epoch": 8.78051948051948, + "grad_norm": 3.7340126271192107, + "learning_rate": 7.714992917803266e-07, + "loss": 0.0063, + "step": 6761 + }, + { + "epoch": 8.781818181818181, + "grad_norm": 2.51026682195014, + "learning_rate": 7.698800601471745e-07, + "loss": 0.0036, + "step": 6762 + }, + { + "epoch": 8.783116883116882, + "grad_norm": 2.0345276803820918, + "learning_rate": 7.682624614989064e-07, + "loss": 0.0029, + "step": 6763 + }, + { + "epoch": 8.784415584415584, + "grad_norm": 1.1710996969115075, + "learning_rate": 7.666464961217069e-07, + "loss": 0.0011, + "step": 6764 + }, + { + "epoch": 8.785714285714286, + "grad_norm": 1.5914691044394995, + "learning_rate": 7.650321643014702e-07, + "loss": 0.002, + "step": 6765 + }, + { + "epoch": 8.787012987012988, + "grad_norm": 3.110299614650767, + "learning_rate": 7.634194663238015e-07, + "loss": 0.0061, + "step": 6766 + }, + { + "epoch": 8.788311688311689, + "grad_norm": 1.0380624577352826, + "learning_rate": 7.618084024740213e-07, + "loss": 0.0089, + "step": 6767 + }, + { + "epoch": 8.78961038961039, + "grad_norm": 2.7215054994922756, + "learning_rate": 7.60198973037155e-07, + "loss": 0.0078, + "step": 6768 + }, + { + "epoch": 8.790909090909091, + "grad_norm": 4.034440721110816, + "learning_rate": 7.585911782979416e-07, + "loss": 0.0026, + "step": 6769 + }, + { + "epoch": 8.792207792207792, + "grad_norm": 4.616744070356855, + "learning_rate": 7.569850185408301e-07, + "loss": 0.0102, + "step": 6770 + }, + { + "epoch": 8.793506493506493, + "grad_norm": 1.2431466509683375, + "learning_rate": 7.553804940499842e-07, + "loss": 0.0071, + "step": 6771 + }, + { + "epoch": 8.794805194805194, + "grad_norm": 1.353929312107366, + "learning_rate": 7.537776051092727e-07, + "loss": 0.0072, + "step": 6772 + }, + { + "epoch": 8.796103896103896, + "grad_norm": 3.5887432379270185, + "learning_rate": 7.521763520022762e-07, + "loss": 0.0012, + "step": 6773 + }, + { + "epoch": 8.797402597402597, + "grad_norm": 2.3872974872973214, + "learning_rate": 7.505767350122884e-07, + "loss": 0.0012, + "step": 6774 + }, + { + "epoch": 8.7987012987013, + "grad_norm": 1.6842544455247863, + "learning_rate": 7.489787544223126e-07, + "loss": 0.0088, + "step": 6775 + }, + { + "epoch": 8.8, + "grad_norm": 1.124044798689858, + "learning_rate": 7.473824105150617e-07, + "loss": 0.0083, + "step": 6776 + }, + { + "epoch": 8.801298701298702, + "grad_norm": 3.9755414422971365, + "learning_rate": 7.457877035729588e-07, + "loss": 0.0108, + "step": 6777 + }, + { + "epoch": 8.802597402597403, + "grad_norm": 1.4288755275643148, + "learning_rate": 7.44194633878138e-07, + "loss": 0.007, + "step": 6778 + }, + { + "epoch": 8.803896103896104, + "grad_norm": 0.4163516857717473, + "learning_rate": 7.426032017124452e-07, + "loss": 0.0013, + "step": 6779 + }, + { + "epoch": 8.805194805194805, + "grad_norm": 1.6667937837031404, + "learning_rate": 7.410134073574349e-07, + "loss": 0.0026, + "step": 6780 + }, + { + "epoch": 8.806493506493506, + "grad_norm": 1.280654679030473, + "learning_rate": 7.394252510943723e-07, + "loss": 0.0023, + "step": 6781 + }, + { + "epoch": 8.807792207792208, + "grad_norm": 1.6898798069417766, + "learning_rate": 7.378387332042303e-07, + "loss": 0.0104, + "step": 6782 + }, + { + "epoch": 8.809090909090909, + "grad_norm": 3.5895037175483515, + "learning_rate": 7.362538539676989e-07, + "loss": 0.0061, + "step": 6783 + }, + { + "epoch": 8.81038961038961, + "grad_norm": 1.4914130580830864, + "learning_rate": 7.346706136651671e-07, + "loss": 0.0017, + "step": 6784 + }, + { + "epoch": 8.811688311688311, + "grad_norm": 0.4098834040670976, + "learning_rate": 7.330890125767465e-07, + "loss": 0.0011, + "step": 6785 + }, + { + "epoch": 8.812987012987014, + "grad_norm": 0.8941006806575253, + "learning_rate": 7.315090509822476e-07, + "loss": 0.0011, + "step": 6786 + }, + { + "epoch": 8.814285714285715, + "grad_norm": 1.836831708612436, + "learning_rate": 7.299307291612023e-07, + "loss": 0.0017, + "step": 6787 + }, + { + "epoch": 8.815584415584416, + "grad_norm": 0.6832183608121759, + "learning_rate": 7.283540473928386e-07, + "loss": 0.0024, + "step": 6788 + }, + { + "epoch": 8.816883116883117, + "grad_norm": 0.30206173238398376, + "learning_rate": 7.267790059561075e-07, + "loss": 0.0011, + "step": 6789 + }, + { + "epoch": 8.818181818181818, + "grad_norm": 0.5955097911193508, + "learning_rate": 7.252056051296586e-07, + "loss": 0.0015, + "step": 6790 + }, + { + "epoch": 8.81948051948052, + "grad_norm": 0.866957113412993, + "learning_rate": 7.236338451918634e-07, + "loss": 0.0009, + "step": 6791 + }, + { + "epoch": 8.82077922077922, + "grad_norm": 2.078551535836769, + "learning_rate": 7.220637264207897e-07, + "loss": 0.0013, + "step": 6792 + }, + { + "epoch": 8.822077922077922, + "grad_norm": 0.3191862166755056, + "learning_rate": 7.204952490942263e-07, + "loss": 0.0011, + "step": 6793 + }, + { + "epoch": 8.823376623376623, + "grad_norm": 1.372722300052341, + "learning_rate": 7.189284134896635e-07, + "loss": 0.008, + "step": 6794 + }, + { + "epoch": 8.824675324675324, + "grad_norm": 2.127381219789317, + "learning_rate": 7.17363219884305e-07, + "loss": 0.0247, + "step": 6795 + }, + { + "epoch": 8.825974025974027, + "grad_norm": 0.8461027815605024, + "learning_rate": 7.157996685550672e-07, + "loss": 0.0076, + "step": 6796 + }, + { + "epoch": 8.827272727272728, + "grad_norm": 0.625931208245725, + "learning_rate": 7.142377597785688e-07, + "loss": 0.008, + "step": 6797 + }, + { + "epoch": 8.82857142857143, + "grad_norm": 2.3857437141535582, + "learning_rate": 7.12677493831141e-07, + "loss": 0.0012, + "step": 6798 + }, + { + "epoch": 8.82987012987013, + "grad_norm": 1.7475485014581313, + "learning_rate": 7.11118870988825e-07, + "loss": 0.0021, + "step": 6799 + }, + { + "epoch": 8.831168831168831, + "grad_norm": 1.80565713716009, + "learning_rate": 7.095618915273738e-07, + "loss": 0.0038, + "step": 6800 + }, + { + "epoch": 8.832467532467533, + "grad_norm": 1.6952620165753596, + "learning_rate": 7.080065557222449e-07, + "loss": 0.0077, + "step": 6801 + }, + { + "epoch": 8.833766233766234, + "grad_norm": 3.914367102576316, + "learning_rate": 7.064528638486068e-07, + "loss": 0.0096, + "step": 6802 + }, + { + "epoch": 8.835064935064935, + "grad_norm": 0.9472505105363145, + "learning_rate": 7.049008161813354e-07, + "loss": 0.0012, + "step": 6803 + }, + { + "epoch": 8.836363636363636, + "grad_norm": 0.580813196004897, + "learning_rate": 7.033504129950208e-07, + "loss": 0.0011, + "step": 6804 + }, + { + "epoch": 8.837662337662337, + "grad_norm": 0.5505113662742241, + "learning_rate": 7.018016545639573e-07, + "loss": 0.001, + "step": 6805 + }, + { + "epoch": 8.838961038961038, + "grad_norm": 0.3449989671281334, + "learning_rate": 7.002545411621542e-07, + "loss": 0.0011, + "step": 6806 + }, + { + "epoch": 8.84025974025974, + "grad_norm": 3.711582825379605, + "learning_rate": 6.987090730633173e-07, + "loss": 0.0092, + "step": 6807 + }, + { + "epoch": 8.841558441558442, + "grad_norm": 0.5746722745289449, + "learning_rate": 6.971652505408766e-07, + "loss": 0.0011, + "step": 6808 + }, + { + "epoch": 8.842857142857143, + "grad_norm": 1.0984466178461134, + "learning_rate": 6.956230738679604e-07, + "loss": 0.0012, + "step": 6809 + }, + { + "epoch": 8.844155844155845, + "grad_norm": 5.436877380170486, + "learning_rate": 6.940825433174103e-07, + "loss": 0.0078, + "step": 6810 + }, + { + "epoch": 8.845454545454546, + "grad_norm": 3.8630526424035483, + "learning_rate": 6.92543659161774e-07, + "loss": 0.0047, + "step": 6811 + }, + { + "epoch": 8.846753246753247, + "grad_norm": 1.9577172011922808, + "learning_rate": 6.910064216733136e-07, + "loss": 0.0081, + "step": 6812 + }, + { + "epoch": 8.848051948051948, + "grad_norm": 3.066229345971026, + "learning_rate": 6.894708311239917e-07, + "loss": 0.0039, + "step": 6813 + }, + { + "epoch": 8.849350649350649, + "grad_norm": 2.733351972824566, + "learning_rate": 6.879368877854864e-07, + "loss": 0.0084, + "step": 6814 + }, + { + "epoch": 8.85064935064935, + "grad_norm": 1.0244175987341804, + "learning_rate": 6.864045919291785e-07, + "loss": 0.0029, + "step": 6815 + }, + { + "epoch": 8.851948051948051, + "grad_norm": 0.7023559394396864, + "learning_rate": 6.848739438261631e-07, + "loss": 0.001, + "step": 6816 + }, + { + "epoch": 8.853246753246752, + "grad_norm": 2.081801730114124, + "learning_rate": 6.833449437472417e-07, + "loss": 0.0021, + "step": 6817 + }, + { + "epoch": 8.854545454545455, + "grad_norm": 3.042015070564821, + "learning_rate": 6.818175919629211e-07, + "loss": 0.01, + "step": 6818 + }, + { + "epoch": 8.855844155844157, + "grad_norm": 6.163986277365507, + "learning_rate": 6.802918887434195e-07, + "loss": 0.0076, + "step": 6819 + }, + { + "epoch": 8.857142857142858, + "grad_norm": 1.824755699410406, + "learning_rate": 6.787678343586679e-07, + "loss": 0.0076, + "step": 6820 + }, + { + "epoch": 8.858441558441559, + "grad_norm": 1.1885385534763628, + "learning_rate": 6.772454290782926e-07, + "loss": 0.0148, + "step": 6821 + }, + { + "epoch": 8.85974025974026, + "grad_norm": 0.31348570602928116, + "learning_rate": 6.757246731716416e-07, + "loss": 0.0008, + "step": 6822 + }, + { + "epoch": 8.861038961038961, + "grad_norm": 5.005974262907015, + "learning_rate": 6.74205566907763e-07, + "loss": 0.0064, + "step": 6823 + }, + { + "epoch": 8.862337662337662, + "grad_norm": 1.0193963277785518, + "learning_rate": 6.726881105554206e-07, + "loss": 0.0009, + "step": 6824 + }, + { + "epoch": 8.863636363636363, + "grad_norm": 3.7367356724237255, + "learning_rate": 6.711723043830753e-07, + "loss": 0.0115, + "step": 6825 + }, + { + "epoch": 8.864935064935064, + "grad_norm": 0.999701376764667, + "learning_rate": 6.696581486589071e-07, + "loss": 0.0025, + "step": 6826 + }, + { + "epoch": 8.866233766233766, + "grad_norm": 2.456764400201545, + "learning_rate": 6.681456436507938e-07, + "loss": 0.0033, + "step": 6827 + }, + { + "epoch": 8.867532467532467, + "grad_norm": 0.8727859413064866, + "learning_rate": 6.666347896263326e-07, + "loss": 0.003, + "step": 6828 + }, + { + "epoch": 8.86883116883117, + "grad_norm": 12.232115047829465, + "learning_rate": 6.651255868528195e-07, + "loss": 0.0142, + "step": 6829 + }, + { + "epoch": 8.87012987012987, + "grad_norm": 0.7823947817280598, + "learning_rate": 6.63618035597261e-07, + "loss": 0.0025, + "step": 6830 + }, + { + "epoch": 8.871428571428572, + "grad_norm": 1.4855682412569413, + "learning_rate": 6.621121361263705e-07, + "loss": 0.0026, + "step": 6831 + }, + { + "epoch": 8.872727272727273, + "grad_norm": 2.0956996128893772, + "learning_rate": 6.606078887065748e-07, + "loss": 0.0191, + "step": 6832 + }, + { + "epoch": 8.874025974025974, + "grad_norm": 2.6351807648347085, + "learning_rate": 6.59105293604001e-07, + "loss": 0.0081, + "step": 6833 + }, + { + "epoch": 8.875324675324675, + "grad_norm": 2.3530801791407945, + "learning_rate": 6.576043510844854e-07, + "loss": 0.0017, + "step": 6834 + }, + { + "epoch": 8.876623376623376, + "grad_norm": 0.40897391944109884, + "learning_rate": 6.561050614135789e-07, + "loss": 0.001, + "step": 6835 + }, + { + "epoch": 8.877922077922078, + "grad_norm": 0.7936478845490119, + "learning_rate": 6.546074248565282e-07, + "loss": 0.0017, + "step": 6836 + }, + { + "epoch": 8.879220779220779, + "grad_norm": 2.449598287965404, + "learning_rate": 6.53111441678298e-07, + "loss": 0.0081, + "step": 6837 + }, + { + "epoch": 8.88051948051948, + "grad_norm": 0.5191093427855383, + "learning_rate": 6.516171121435533e-07, + "loss": 0.0011, + "step": 6838 + }, + { + "epoch": 8.881818181818183, + "grad_norm": 1.7214436494705025, + "learning_rate": 6.501244365166748e-07, + "loss": 0.0133, + "step": 6839 + }, + { + "epoch": 8.883116883116884, + "grad_norm": 0.885443851092412, + "learning_rate": 6.486334150617391e-07, + "loss": 0.0015, + "step": 6840 + }, + { + "epoch": 8.884415584415585, + "grad_norm": 4.2972211694319595, + "learning_rate": 6.471440480425406e-07, + "loss": 0.0066, + "step": 6841 + }, + { + "epoch": 8.885714285714286, + "grad_norm": 8.616834952110686, + "learning_rate": 6.456563357225743e-07, + "loss": 0.012, + "step": 6842 + }, + { + "epoch": 8.887012987012987, + "grad_norm": 0.7002213303671746, + "learning_rate": 6.441702783650494e-07, + "loss": 0.0009, + "step": 6843 + }, + { + "epoch": 8.888311688311688, + "grad_norm": 1.3983242795425976, + "learning_rate": 6.426858762328714e-07, + "loss": 0.001, + "step": 6844 + }, + { + "epoch": 8.88961038961039, + "grad_norm": 1.1071909991142148, + "learning_rate": 6.412031295886645e-07, + "loss": 0.0067, + "step": 6845 + }, + { + "epoch": 8.89090909090909, + "grad_norm": 6.80563936578275, + "learning_rate": 6.397220386947511e-07, + "loss": 0.0052, + "step": 6846 + }, + { + "epoch": 8.892207792207792, + "grad_norm": 1.7706542235905955, + "learning_rate": 6.382426038131706e-07, + "loss": 0.0028, + "step": 6847 + }, + { + "epoch": 8.893506493506493, + "grad_norm": 0.3362813455822271, + "learning_rate": 6.367648252056558e-07, + "loss": 0.0032, + "step": 6848 + }, + { + "epoch": 8.894805194805194, + "grad_norm": 2.9926401330848122, + "learning_rate": 6.352887031336597e-07, + "loss": 0.0093, + "step": 6849 + }, + { + "epoch": 8.896103896103895, + "grad_norm": 0.4163680775292089, + "learning_rate": 6.338142378583345e-07, + "loss": 0.0022, + "step": 6850 + }, + { + "epoch": 8.897402597402598, + "grad_norm": 7.468605064724425, + "learning_rate": 6.323414296405406e-07, + "loss": 0.0198, + "step": 6851 + }, + { + "epoch": 8.8987012987013, + "grad_norm": 0.8272076826587004, + "learning_rate": 6.308702787408461e-07, + "loss": 0.0087, + "step": 6852 + }, + { + "epoch": 8.9, + "grad_norm": 0.6026102250815999, + "learning_rate": 6.294007854195272e-07, + "loss": 0.0083, + "step": 6853 + }, + { + "epoch": 8.901298701298701, + "grad_norm": 1.5341121555428465, + "learning_rate": 6.279329499365649e-07, + "loss": 0.0015, + "step": 6854 + }, + { + "epoch": 8.902597402597403, + "grad_norm": 2.254428675815623, + "learning_rate": 6.26466772551646e-07, + "loss": 0.0025, + "step": 6855 + }, + { + "epoch": 8.903896103896104, + "grad_norm": 1.1525534726855862, + "learning_rate": 6.250022535241662e-07, + "loss": 0.0094, + "step": 6856 + }, + { + "epoch": 8.905194805194805, + "grad_norm": 0.39363111697848735, + "learning_rate": 6.235393931132283e-07, + "loss": 0.0012, + "step": 6857 + }, + { + "epoch": 8.906493506493506, + "grad_norm": 2.779361737567987, + "learning_rate": 6.220781915776386e-07, + "loss": 0.0122, + "step": 6858 + }, + { + "epoch": 8.907792207792207, + "grad_norm": 0.35399348276855375, + "learning_rate": 6.206186491759126e-07, + "loss": 0.0011, + "step": 6859 + }, + { + "epoch": 8.909090909090908, + "grad_norm": 1.4567301481949522, + "learning_rate": 6.191607661662691e-07, + "loss": 0.0093, + "step": 6860 + }, + { + "epoch": 8.910389610389611, + "grad_norm": 0.48005477778219485, + "learning_rate": 6.177045428066397e-07, + "loss": 0.0012, + "step": 6861 + }, + { + "epoch": 8.911688311688312, + "grad_norm": 0.7064214565747592, + "learning_rate": 6.16249979354655e-07, + "loss": 0.0014, + "step": 6862 + }, + { + "epoch": 8.912987012987013, + "grad_norm": 0.2174818623996631, + "learning_rate": 6.147970760676558e-07, + "loss": 0.0009, + "step": 6863 + }, + { + "epoch": 8.914285714285715, + "grad_norm": 0.626470628533472, + "learning_rate": 6.133458332026865e-07, + "loss": 0.0018, + "step": 6864 + }, + { + "epoch": 8.915584415584416, + "grad_norm": 0.5367126282961401, + "learning_rate": 6.118962510165039e-07, + "loss": 0.0013, + "step": 6865 + }, + { + "epoch": 8.916883116883117, + "grad_norm": 1.7949708877602821, + "learning_rate": 6.10448329765565e-07, + "loss": 0.006, + "step": 6866 + }, + { + "epoch": 8.918181818181818, + "grad_norm": 0.4071883322457613, + "learning_rate": 6.090020697060317e-07, + "loss": 0.0007, + "step": 6867 + }, + { + "epoch": 8.91948051948052, + "grad_norm": 2.8782530360579073, + "learning_rate": 6.075574710937804e-07, + "loss": 0.0078, + "step": 6868 + }, + { + "epoch": 8.92077922077922, + "grad_norm": 1.9895201628691537, + "learning_rate": 6.061145341843866e-07, + "loss": 0.0074, + "step": 6869 + }, + { + "epoch": 8.922077922077921, + "grad_norm": 1.1636129856509894, + "learning_rate": 6.046732592331306e-07, + "loss": 0.0019, + "step": 6870 + }, + { + "epoch": 8.923376623376623, + "grad_norm": 4.649754175221732, + "learning_rate": 6.032336464950039e-07, + "loss": 0.0152, + "step": 6871 + }, + { + "epoch": 8.924675324675325, + "grad_norm": 3.3977731908239748, + "learning_rate": 6.017956962247018e-07, + "loss": 0.0026, + "step": 6872 + }, + { + "epoch": 8.925974025974027, + "grad_norm": 4.259912858095725, + "learning_rate": 6.00359408676624e-07, + "loss": 0.0213, + "step": 6873 + }, + { + "epoch": 8.927272727272728, + "grad_norm": 1.146972248659844, + "learning_rate": 5.989247841048785e-07, + "loss": 0.0018, + "step": 6874 + }, + { + "epoch": 8.928571428571429, + "grad_norm": 0.48746978969724053, + "learning_rate": 5.974918227632753e-07, + "loss": 0.008, + "step": 6875 + }, + { + "epoch": 8.92987012987013, + "grad_norm": 2.4507570368388816, + "learning_rate": 5.960605249053386e-07, + "loss": 0.0087, + "step": 6876 + }, + { + "epoch": 8.931168831168831, + "grad_norm": 2.2629420400148827, + "learning_rate": 5.946308907842846e-07, + "loss": 0.0015, + "step": 6877 + }, + { + "epoch": 8.932467532467532, + "grad_norm": 1.1013912963602734, + "learning_rate": 5.932029206530477e-07, + "loss": 0.0016, + "step": 6878 + }, + { + "epoch": 8.933766233766233, + "grad_norm": 5.226525314139875, + "learning_rate": 5.917766147642612e-07, + "loss": 0.0069, + "step": 6879 + }, + { + "epoch": 8.935064935064934, + "grad_norm": 5.169371827447711, + "learning_rate": 5.9035197337027e-07, + "loss": 0.0118, + "step": 6880 + }, + { + "epoch": 8.936363636363636, + "grad_norm": 3.419914940717185, + "learning_rate": 5.889289967231149e-07, + "loss": 0.0094, + "step": 6881 + }, + { + "epoch": 8.937662337662339, + "grad_norm": 4.1128702586051995, + "learning_rate": 5.875076850745509e-07, + "loss": 0.0033, + "step": 6882 + }, + { + "epoch": 8.93896103896104, + "grad_norm": 4.396478092681166, + "learning_rate": 5.860880386760326e-07, + "loss": 0.0075, + "step": 6883 + }, + { + "epoch": 8.94025974025974, + "grad_norm": 1.420181395386949, + "learning_rate": 5.846700577787279e-07, + "loss": 0.0016, + "step": 6884 + }, + { + "epoch": 8.941558441558442, + "grad_norm": 0.8098509362349938, + "learning_rate": 5.832537426334983e-07, + "loss": 0.0023, + "step": 6885 + }, + { + "epoch": 8.942857142857143, + "grad_norm": 1.4022087686600597, + "learning_rate": 5.818390934909213e-07, + "loss": 0.0012, + "step": 6886 + }, + { + "epoch": 8.944155844155844, + "grad_norm": 1.6396955712768349, + "learning_rate": 5.804261106012732e-07, + "loss": 0.0025, + "step": 6887 + }, + { + "epoch": 8.945454545454545, + "grad_norm": 1.1861476242716833, + "learning_rate": 5.790147942145408e-07, + "loss": 0.0081, + "step": 6888 + }, + { + "epoch": 8.946753246753246, + "grad_norm": 1.8552101800704948, + "learning_rate": 5.776051445804087e-07, + "loss": 0.0011, + "step": 6889 + }, + { + "epoch": 8.948051948051948, + "grad_norm": 1.2908266488054376, + "learning_rate": 5.761971619482742e-07, + "loss": 0.005, + "step": 6890 + }, + { + "epoch": 8.949350649350649, + "grad_norm": 2.595376693976779, + "learning_rate": 5.747908465672359e-07, + "loss": 0.0019, + "step": 6891 + }, + { + "epoch": 8.95064935064935, + "grad_norm": 4.971511238195608, + "learning_rate": 5.733861986860967e-07, + "loss": 0.0272, + "step": 6892 + }, + { + "epoch": 8.951948051948053, + "grad_norm": 1.9938451156537353, + "learning_rate": 5.719832185533647e-07, + "loss": 0.0092, + "step": 6893 + }, + { + "epoch": 8.953246753246754, + "grad_norm": 1.6015072172938902, + "learning_rate": 5.705819064172569e-07, + "loss": 0.0011, + "step": 6894 + }, + { + "epoch": 8.954545454545455, + "grad_norm": 6.532491632494852, + "learning_rate": 5.691822625256904e-07, + "loss": 0.0199, + "step": 6895 + }, + { + "epoch": 8.955844155844156, + "grad_norm": 6.358589116898516, + "learning_rate": 5.677842871262895e-07, + "loss": 0.0039, + "step": 6896 + }, + { + "epoch": 8.957142857142857, + "grad_norm": 6.218281026738584, + "learning_rate": 5.66387980466383e-07, + "loss": 0.0108, + "step": 6897 + }, + { + "epoch": 8.958441558441558, + "grad_norm": 0.7512372396823763, + "learning_rate": 5.649933427930043e-07, + "loss": 0.0013, + "step": 6898 + }, + { + "epoch": 8.95974025974026, + "grad_norm": 0.6159884033277246, + "learning_rate": 5.636003743528917e-07, + "loss": 0.0024, + "step": 6899 + }, + { + "epoch": 8.96103896103896, + "grad_norm": 2.590811639246748, + "learning_rate": 5.622090753924847e-07, + "loss": 0.0092, + "step": 6900 + }, + { + "epoch": 8.962337662337662, + "grad_norm": 1.0614966225199758, + "learning_rate": 5.608194461579364e-07, + "loss": 0.0029, + "step": 6901 + }, + { + "epoch": 8.963636363636363, + "grad_norm": 1.130067908569639, + "learning_rate": 5.594314868950957e-07, + "loss": 0.0013, + "step": 6902 + }, + { + "epoch": 8.964935064935066, + "grad_norm": 1.458472014266421, + "learning_rate": 5.580451978495183e-07, + "loss": 0.0091, + "step": 6903 + }, + { + "epoch": 8.966233766233767, + "grad_norm": 2.848588893506172, + "learning_rate": 5.566605792664659e-07, + "loss": 0.004, + "step": 6904 + }, + { + "epoch": 8.967532467532468, + "grad_norm": 2.0701392281000786, + "learning_rate": 5.552776313909059e-07, + "loss": 0.0081, + "step": 6905 + }, + { + "epoch": 8.96883116883117, + "grad_norm": 0.6926073883836449, + "learning_rate": 5.538963544675069e-07, + "loss": 0.0015, + "step": 6906 + }, + { + "epoch": 8.97012987012987, + "grad_norm": 0.47609892550809757, + "learning_rate": 5.525167487406424e-07, + "loss": 0.0014, + "step": 6907 + }, + { + "epoch": 8.971428571428572, + "grad_norm": 1.3343999336201846, + "learning_rate": 5.511388144543906e-07, + "loss": 0.0084, + "step": 6908 + }, + { + "epoch": 8.972727272727273, + "grad_norm": 1.5321621281005007, + "learning_rate": 5.497625518525374e-07, + "loss": 0.0014, + "step": 6909 + }, + { + "epoch": 8.974025974025974, + "grad_norm": 1.0124487355272893, + "learning_rate": 5.483879611785681e-07, + "loss": 0.0014, + "step": 6910 + }, + { + "epoch": 8.975324675324675, + "grad_norm": 0.40272168054389795, + "learning_rate": 5.470150426756738e-07, + "loss": 0.0011, + "step": 6911 + }, + { + "epoch": 8.976623376623376, + "grad_norm": 2.0562165589058377, + "learning_rate": 5.456437965867501e-07, + "loss": 0.0019, + "step": 6912 + }, + { + "epoch": 8.977922077922077, + "grad_norm": 9.428419371027895, + "learning_rate": 5.442742231543974e-07, + "loss": 0.0202, + "step": 6913 + }, + { + "epoch": 8.979220779220778, + "grad_norm": 1.6033143067614386, + "learning_rate": 5.429063226209197e-07, + "loss": 0.0011, + "step": 6914 + }, + { + "epoch": 8.980519480519481, + "grad_norm": 0.9682979713396188, + "learning_rate": 5.415400952283245e-07, + "loss": 0.0009, + "step": 6915 + }, + { + "epoch": 8.981818181818182, + "grad_norm": 0.5381616267377346, + "learning_rate": 5.401755412183218e-07, + "loss": 0.0015, + "step": 6916 + }, + { + "epoch": 8.983116883116883, + "grad_norm": 0.9058555765848623, + "learning_rate": 5.388126608323318e-07, + "loss": 0.0008, + "step": 6917 + }, + { + "epoch": 8.984415584415585, + "grad_norm": 0.8745649135401212, + "learning_rate": 5.374514543114684e-07, + "loss": 0.0016, + "step": 6918 + }, + { + "epoch": 8.985714285714286, + "grad_norm": 1.060719565102875, + "learning_rate": 5.360919218965588e-07, + "loss": 0.0029, + "step": 6919 + }, + { + "epoch": 8.987012987012987, + "grad_norm": 1.5803205065761914, + "learning_rate": 5.347340638281273e-07, + "loss": 0.001, + "step": 6920 + }, + { + "epoch": 8.988311688311688, + "grad_norm": 3.9708372009014394, + "learning_rate": 5.333778803464107e-07, + "loss": 0.0144, + "step": 6921 + }, + { + "epoch": 8.98961038961039, + "grad_norm": 5.975517744375095, + "learning_rate": 5.32023371691337e-07, + "loss": 0.015, + "step": 6922 + }, + { + "epoch": 8.99090909090909, + "grad_norm": 1.0910013821157907, + "learning_rate": 5.306705381025501e-07, + "loss": 0.0012, + "step": 6923 + }, + { + "epoch": 8.992207792207791, + "grad_norm": 0.40607500529327184, + "learning_rate": 5.293193798193874e-07, + "loss": 0.0015, + "step": 6924 + }, + { + "epoch": 8.993506493506494, + "grad_norm": 2.377276750159681, + "learning_rate": 5.279698970809011e-07, + "loss": 0.002, + "step": 6925 + }, + { + "epoch": 8.994805194805195, + "grad_norm": 0.584538686855404, + "learning_rate": 5.266220901258323e-07, + "loss": 0.0011, + "step": 6926 + }, + { + "epoch": 8.996103896103897, + "grad_norm": 0.4763128357907645, + "learning_rate": 5.252759591926404e-07, + "loss": 0.0018, + "step": 6927 + }, + { + "epoch": 8.997402597402598, + "grad_norm": 3.9130029950739544, + "learning_rate": 5.239315045194781e-07, + "loss": 0.0071, + "step": 6928 + }, + { + "epoch": 8.998701298701299, + "grad_norm": 0.3574895536726463, + "learning_rate": 5.225887263442086e-07, + "loss": 0.0011, + "step": 6929 + }, + { + "epoch": 9.0, + "grad_norm": 0.5910958773006902, + "learning_rate": 5.212476249043918e-07, + "loss": 0.0016, + "step": 6930 + }, + { + "epoch": 9.0, + "eval_accuracy": 0.9371139809096013, + "eval_f1": 0.9264480491523116, + "eval_loss": 0.18165066838264465, + "eval_precision": 0.9205724731570057, + "eval_recall": 0.9355043751127309, + "eval_runtime": 12.9875, + "eval_samples_per_second": 137.132, + "eval_steps_per_second": 1.078, + "step": 6930 + }, + { + "epoch": 9.001298701298701, + "grad_norm": 0.6742734070404722, + "learning_rate": 5.199082004372958e-07, + "loss": 0.0013, + "step": 6931 + }, + { + "epoch": 9.002597402597402, + "grad_norm": 4.426726548275311, + "learning_rate": 5.185704531798907e-07, + "loss": 0.0052, + "step": 6932 + }, + { + "epoch": 9.003896103896103, + "grad_norm": 1.528387929320508, + "learning_rate": 5.172343833688464e-07, + "loss": 0.0041, + "step": 6933 + }, + { + "epoch": 9.005194805194805, + "grad_norm": 0.8541553825936717, + "learning_rate": 5.158999912405438e-07, + "loss": 0.001, + "step": 6934 + }, + { + "epoch": 9.006493506493506, + "grad_norm": 0.7498862490525675, + "learning_rate": 5.145672770310606e-07, + "loss": 0.0082, + "step": 6935 + }, + { + "epoch": 9.007792207792209, + "grad_norm": 0.8862905515422869, + "learning_rate": 5.132362409761782e-07, + "loss": 0.0121, + "step": 6936 + }, + { + "epoch": 9.00909090909091, + "grad_norm": 1.34572983514297, + "learning_rate": 5.119068833113828e-07, + "loss": 0.0033, + "step": 6937 + }, + { + "epoch": 9.01038961038961, + "grad_norm": 1.1936507610097071, + "learning_rate": 5.105792042718649e-07, + "loss": 0.0097, + "step": 6938 + }, + { + "epoch": 9.011688311688312, + "grad_norm": 0.8237768465697033, + "learning_rate": 5.092532040925146e-07, + "loss": 0.0008, + "step": 6939 + }, + { + "epoch": 9.012987012987013, + "grad_norm": 0.4546569441517469, + "learning_rate": 5.079288830079288e-07, + "loss": 0.0012, + "step": 6940 + }, + { + "epoch": 9.014285714285714, + "grad_norm": 0.4754802796538065, + "learning_rate": 5.06606241252402e-07, + "loss": 0.001, + "step": 6941 + }, + { + "epoch": 9.015584415584415, + "grad_norm": 0.7236657856925953, + "learning_rate": 5.052852790599383e-07, + "loss": 0.0078, + "step": 6942 + }, + { + "epoch": 9.016883116883117, + "grad_norm": 0.7615299376745172, + "learning_rate": 5.039659966642385e-07, + "loss": 0.0065, + "step": 6943 + }, + { + "epoch": 9.018181818181818, + "grad_norm": 1.4241510952583583, + "learning_rate": 5.026483942987104e-07, + "loss": 0.0018, + "step": 6944 + }, + { + "epoch": 9.019480519480519, + "grad_norm": 1.109952088154184, + "learning_rate": 5.013324721964607e-07, + "loss": 0.0011, + "step": 6945 + }, + { + "epoch": 9.02077922077922, + "grad_norm": 1.190816059706084, + "learning_rate": 5.000182305903045e-07, + "loss": 0.0093, + "step": 6946 + }, + { + "epoch": 9.022077922077923, + "grad_norm": 0.28362575131982176, + "learning_rate": 4.987056697127557e-07, + "loss": 0.0008, + "step": 6947 + }, + { + "epoch": 9.023376623376624, + "grad_norm": 0.5973088705816169, + "learning_rate": 4.973947897960307e-07, + "loss": 0.0017, + "step": 6948 + }, + { + "epoch": 9.024675324675325, + "grad_norm": 4.162392332544689, + "learning_rate": 4.960855910720475e-07, + "loss": 0.0031, + "step": 6949 + }, + { + "epoch": 9.025974025974026, + "grad_norm": 1.9529364307401642, + "learning_rate": 4.947780737724306e-07, + "loss": 0.0072, + "step": 6950 + }, + { + "epoch": 9.027272727272727, + "grad_norm": 1.1250261929057626, + "learning_rate": 4.934722381285051e-07, + "loss": 0.0011, + "step": 6951 + }, + { + "epoch": 9.028571428571428, + "grad_norm": 4.0043717953008535, + "learning_rate": 4.921680843712973e-07, + "loss": 0.01, + "step": 6952 + }, + { + "epoch": 9.02987012987013, + "grad_norm": 6.114642912975425, + "learning_rate": 4.908656127315359e-07, + "loss": 0.0056, + "step": 6953 + }, + { + "epoch": 9.03116883116883, + "grad_norm": 2.8665071258090595, + "learning_rate": 4.895648234396566e-07, + "loss": 0.0027, + "step": 6954 + }, + { + "epoch": 9.032467532467532, + "grad_norm": 1.798221609948964, + "learning_rate": 4.88265716725791e-07, + "loss": 0.0023, + "step": 6955 + }, + { + "epoch": 9.033766233766233, + "grad_norm": 0.4752296485037849, + "learning_rate": 4.869682928197761e-07, + "loss": 0.0016, + "step": 6956 + }, + { + "epoch": 9.035064935064936, + "grad_norm": 1.1144853376019617, + "learning_rate": 4.856725519511507e-07, + "loss": 0.0017, + "step": 6957 + }, + { + "epoch": 9.036363636363637, + "grad_norm": 1.7887655556773405, + "learning_rate": 4.843784943491591e-07, + "loss": 0.0035, + "step": 6958 + }, + { + "epoch": 9.037662337662338, + "grad_norm": 1.3634855486003716, + "learning_rate": 4.830861202427417e-07, + "loss": 0.0012, + "step": 6959 + }, + { + "epoch": 9.03896103896104, + "grad_norm": 3.6399961258953835, + "learning_rate": 4.817954298605454e-07, + "loss": 0.004, + "step": 6960 + }, + { + "epoch": 9.04025974025974, + "grad_norm": 0.8486856934341888, + "learning_rate": 4.805064234309175e-07, + "loss": 0.0069, + "step": 6961 + }, + { + "epoch": 9.041558441558442, + "grad_norm": 1.2915154461184106, + "learning_rate": 4.792191011819092e-07, + "loss": 0.0098, + "step": 6962 + }, + { + "epoch": 9.042857142857143, + "grad_norm": 1.344214909640011, + "learning_rate": 4.779334633412713e-07, + "loss": 0.0009, + "step": 6963 + }, + { + "epoch": 9.044155844155844, + "grad_norm": 2.033914974307812, + "learning_rate": 4.7664951013645875e-07, + "loss": 0.001, + "step": 6964 + }, + { + "epoch": 9.045454545454545, + "grad_norm": 1.323670414774306, + "learning_rate": 4.753672417946242e-07, + "loss": 0.0143, + "step": 6965 + }, + { + "epoch": 9.046753246753246, + "grad_norm": 0.9172284788550076, + "learning_rate": 4.7408665854263067e-07, + "loss": 0.002, + "step": 6966 + }, + { + "epoch": 9.048051948051947, + "grad_norm": 0.6550626987752501, + "learning_rate": 4.728077606070336e-07, + "loss": 0.0009, + "step": 6967 + }, + { + "epoch": 9.04935064935065, + "grad_norm": 1.8231023192293423, + "learning_rate": 4.715305482140964e-07, + "loss": 0.0082, + "step": 6968 + }, + { + "epoch": 9.050649350649351, + "grad_norm": 1.5141773100638982, + "learning_rate": 4.7025502158978276e-07, + "loss": 0.0036, + "step": 6969 + }, + { + "epoch": 9.051948051948052, + "grad_norm": 0.9507115284034077, + "learning_rate": 4.689811809597566e-07, + "loss": 0.0015, + "step": 6970 + }, + { + "epoch": 9.053246753246754, + "grad_norm": 5.860178779609073, + "learning_rate": 4.6770902654938553e-07, + "loss": 0.0055, + "step": 6971 + }, + { + "epoch": 9.054545454545455, + "grad_norm": 2.901269979394849, + "learning_rate": 4.664385585837361e-07, + "loss": 0.0161, + "step": 6972 + }, + { + "epoch": 9.055844155844156, + "grad_norm": 4.83473772897509, + "learning_rate": 4.6516977728758295e-07, + "loss": 0.0051, + "step": 6973 + }, + { + "epoch": 9.057142857142857, + "grad_norm": 1.0524581234410302, + "learning_rate": 4.639026828853921e-07, + "loss": 0.0012, + "step": 6974 + }, + { + "epoch": 9.058441558441558, + "grad_norm": 1.7037956137966552, + "learning_rate": 4.6263727560134086e-07, + "loss": 0.0075, + "step": 6975 + }, + { + "epoch": 9.05974025974026, + "grad_norm": 0.6913244192333964, + "learning_rate": 4.613735556593013e-07, + "loss": 0.001, + "step": 6976 + }, + { + "epoch": 9.06103896103896, + "grad_norm": 0.6531275438838989, + "learning_rate": 4.6011152328285345e-07, + "loss": 0.0071, + "step": 6977 + }, + { + "epoch": 9.062337662337661, + "grad_norm": 0.40894382432158927, + "learning_rate": 4.5885117869526984e-07, + "loss": 0.0011, + "step": 6978 + }, + { + "epoch": 9.063636363636364, + "grad_norm": 1.882907677325237, + "learning_rate": 4.575925221195343e-07, + "loss": 0.0017, + "step": 6979 + }, + { + "epoch": 9.064935064935066, + "grad_norm": 1.0387886577862753, + "learning_rate": 4.5633555377832315e-07, + "loss": 0.0012, + "step": 6980 + }, + { + "epoch": 9.066233766233767, + "grad_norm": 1.438543471133609, + "learning_rate": 4.55080273894023e-07, + "loss": 0.022, + "step": 6981 + }, + { + "epoch": 9.067532467532468, + "grad_norm": 0.33877873786459023, + "learning_rate": 4.5382668268871165e-07, + "loss": 0.0009, + "step": 6982 + }, + { + "epoch": 9.068831168831169, + "grad_norm": 2.8978932187100734, + "learning_rate": 4.5257478038417736e-07, + "loss": 0.01, + "step": 6983 + }, + { + "epoch": 9.07012987012987, + "grad_norm": 1.669856605492557, + "learning_rate": 4.513245672019029e-07, + "loss": 0.0019, + "step": 6984 + }, + { + "epoch": 9.071428571428571, + "grad_norm": 4.902338575694928, + "learning_rate": 4.50076043363078e-07, + "loss": 0.0042, + "step": 6985 + }, + { + "epoch": 9.072727272727272, + "grad_norm": 0.4574050480213024, + "learning_rate": 4.4882920908858595e-07, + "loss": 0.0012, + "step": 6986 + }, + { + "epoch": 9.074025974025973, + "grad_norm": 0.9803633409519419, + "learning_rate": 4.475840645990193e-07, + "loss": 0.0013, + "step": 6987 + }, + { + "epoch": 9.075324675324675, + "grad_norm": 0.6998262637423649, + "learning_rate": 4.4634061011466725e-07, + "loss": 0.0011, + "step": 6988 + }, + { + "epoch": 9.076623376623377, + "grad_norm": 0.43927519415753236, + "learning_rate": 4.4509884585552056e-07, + "loss": 0.0012, + "step": 6989 + }, + { + "epoch": 9.077922077922079, + "grad_norm": 0.3387096399692672, + "learning_rate": 4.4385877204126903e-07, + "loss": 0.001, + "step": 6990 + }, + { + "epoch": 9.07922077922078, + "grad_norm": 0.5362575261825134, + "learning_rate": 4.426203888913083e-07, + "loss": 0.0011, + "step": 6991 + }, + { + "epoch": 9.08051948051948, + "grad_norm": 1.6725724063095815, + "learning_rate": 4.41383696624732e-07, + "loss": 0.003, + "step": 6992 + }, + { + "epoch": 9.081818181818182, + "grad_norm": 0.49943183727409063, + "learning_rate": 4.401486954603329e-07, + "loss": 0.001, + "step": 6993 + }, + { + "epoch": 9.083116883116883, + "grad_norm": 0.48372456357526755, + "learning_rate": 4.3891538561660505e-07, + "loss": 0.0009, + "step": 6994 + }, + { + "epoch": 9.084415584415584, + "grad_norm": 2.4692040276953198, + "learning_rate": 4.3768376731174957e-07, + "loss": 0.0079, + "step": 6995 + }, + { + "epoch": 9.085714285714285, + "grad_norm": 5.675133298526852, + "learning_rate": 4.3645384076365873e-07, + "loss": 0.0047, + "step": 6996 + }, + { + "epoch": 9.087012987012987, + "grad_norm": 1.3381094975989303, + "learning_rate": 4.352256061899329e-07, + "loss": 0.0222, + "step": 6997 + }, + { + "epoch": 9.088311688311688, + "grad_norm": 0.7869509380230537, + "learning_rate": 4.339990638078673e-07, + "loss": 0.0025, + "step": 6998 + }, + { + "epoch": 9.089610389610389, + "grad_norm": 1.514680540954715, + "learning_rate": 4.3277421383446373e-07, + "loss": 0.0017, + "step": 6999 + }, + { + "epoch": 9.090909090909092, + "grad_norm": 0.5426159400427099, + "learning_rate": 4.3155105648642004e-07, + "loss": 0.0066, + "step": 7000 + }, + { + "epoch": 9.092207792207793, + "grad_norm": 0.7641731524904462, + "learning_rate": 4.3032959198013646e-07, + "loss": 0.0078, + "step": 7001 + }, + { + "epoch": 9.093506493506494, + "grad_norm": 1.88095915829341, + "learning_rate": 4.2910982053171344e-07, + "loss": 0.0016, + "step": 7002 + }, + { + "epoch": 9.094805194805195, + "grad_norm": 1.2837477671802244, + "learning_rate": 4.2789174235695283e-07, + "loss": 0.0082, + "step": 7003 + }, + { + "epoch": 9.096103896103896, + "grad_norm": 1.1392287177063467, + "learning_rate": 4.266753576713545e-07, + "loss": 0.0021, + "step": 7004 + }, + { + "epoch": 9.097402597402597, + "grad_norm": 0.5944440167899969, + "learning_rate": 4.2546066669011954e-07, + "loss": 0.0009, + "step": 7005 + }, + { + "epoch": 9.098701298701299, + "grad_norm": 0.3552604984850573, + "learning_rate": 4.2424766962815277e-07, + "loss": 0.0011, + "step": 7006 + }, + { + "epoch": 9.1, + "grad_norm": 9.570220531427921, + "learning_rate": 4.230363667000559e-07, + "loss": 0.0088, + "step": 7007 + }, + { + "epoch": 9.1012987012987, + "grad_norm": 2.2401064511261866, + "learning_rate": 4.218267581201296e-07, + "loss": 0.009, + "step": 7008 + }, + { + "epoch": 9.102597402597402, + "grad_norm": 6.846304595002268, + "learning_rate": 4.206188441023773e-07, + "loss": 0.0073, + "step": 7009 + }, + { + "epoch": 9.103896103896103, + "grad_norm": 1.4239953356367439, + "learning_rate": 4.194126248605046e-07, + "loss": 0.0009, + "step": 7010 + }, + { + "epoch": 9.105194805194806, + "grad_norm": 1.364877394930028, + "learning_rate": 4.1820810060791086e-07, + "loss": 0.0015, + "step": 7011 + }, + { + "epoch": 9.106493506493507, + "grad_norm": 1.0544625969035955, + "learning_rate": 4.170052715577033e-07, + "loss": 0.0033, + "step": 7012 + }, + { + "epoch": 9.107792207792208, + "grad_norm": 1.0614470981201027, + "learning_rate": 4.158041379226818e-07, + "loss": 0.0081, + "step": 7013 + }, + { + "epoch": 9.10909090909091, + "grad_norm": 0.8381588011469009, + "learning_rate": 4.146046999153552e-07, + "loss": 0.0009, + "step": 7014 + }, + { + "epoch": 9.11038961038961, + "grad_norm": 0.2494216141978821, + "learning_rate": 4.134069577479205e-07, + "loss": 0.001, + "step": 7015 + }, + { + "epoch": 9.111688311688312, + "grad_norm": 1.6441374337417072, + "learning_rate": 4.1221091163228477e-07, + "loss": 0.0018, + "step": 7016 + }, + { + "epoch": 9.112987012987013, + "grad_norm": 2.4968262150552527, + "learning_rate": 4.11016561780051e-07, + "loss": 0.0145, + "step": 7017 + }, + { + "epoch": 9.114285714285714, + "grad_norm": 1.6375147574441178, + "learning_rate": 4.098239084025246e-07, + "loss": 0.0016, + "step": 7018 + }, + { + "epoch": 9.115584415584415, + "grad_norm": 3.955187071969439, + "learning_rate": 4.086329517107046e-07, + "loss": 0.0033, + "step": 7019 + }, + { + "epoch": 9.116883116883116, + "grad_norm": 1.6024332916640451, + "learning_rate": 4.0744369191529686e-07, + "loss": 0.0014, + "step": 7020 + }, + { + "epoch": 9.118181818181819, + "grad_norm": 2.4459995479312178, + "learning_rate": 4.06256129226702e-07, + "loss": 0.0011, + "step": 7021 + }, + { + "epoch": 9.11948051948052, + "grad_norm": 1.1764047613908923, + "learning_rate": 4.0507026385502747e-07, + "loss": 0.001, + "step": 7022 + }, + { + "epoch": 9.120779220779221, + "grad_norm": 0.5078354356467474, + "learning_rate": 4.0388609601006876e-07, + "loss": 0.0029, + "step": 7023 + }, + { + "epoch": 9.122077922077922, + "grad_norm": 0.4311756513590117, + "learning_rate": 4.0270362590133284e-07, + "loss": 0.001, + "step": 7024 + }, + { + "epoch": 9.123376623376624, + "grad_norm": 1.56570673036285, + "learning_rate": 4.0152285373802e-07, + "loss": 0.0085, + "step": 7025 + }, + { + "epoch": 9.124675324675325, + "grad_norm": 2.3112065199025653, + "learning_rate": 4.0034377972903103e-07, + "loss": 0.0031, + "step": 7026 + }, + { + "epoch": 9.125974025974026, + "grad_norm": 1.8362564923573668, + "learning_rate": 3.991664040829657e-07, + "loss": 0.0013, + "step": 7027 + }, + { + "epoch": 9.127272727272727, + "grad_norm": 3.582938327188228, + "learning_rate": 3.9799072700812624e-07, + "loss": 0.0089, + "step": 7028 + }, + { + "epoch": 9.128571428571428, + "grad_norm": 4.829885692107608, + "learning_rate": 3.9681674871251187e-07, + "loss": 0.0077, + "step": 7029 + }, + { + "epoch": 9.12987012987013, + "grad_norm": 0.5679564574333188, + "learning_rate": 3.9564446940382087e-07, + "loss": 0.0011, + "step": 7030 + }, + { + "epoch": 9.13116883116883, + "grad_norm": 0.9839092407366185, + "learning_rate": 3.9447388928945285e-07, + "loss": 0.009, + "step": 7031 + }, + { + "epoch": 9.132467532467533, + "grad_norm": 1.825783048248784, + "learning_rate": 3.9330500857650557e-07, + "loss": 0.0079, + "step": 7032 + }, + { + "epoch": 9.133766233766234, + "grad_norm": 2.1088496398327936, + "learning_rate": 3.921378274717769e-07, + "loss": 0.0088, + "step": 7033 + }, + { + "epoch": 9.135064935064936, + "grad_norm": 7.477636550860508, + "learning_rate": 3.909723461817616e-07, + "loss": 0.0168, + "step": 7034 + }, + { + "epoch": 9.136363636363637, + "grad_norm": 0.9818973751850725, + "learning_rate": 3.8980856491265815e-07, + "loss": 0.001, + "step": 7035 + }, + { + "epoch": 9.137662337662338, + "grad_norm": 0.46116939160312465, + "learning_rate": 3.8864648387036074e-07, + "loss": 0.0008, + "step": 7036 + }, + { + "epoch": 9.138961038961039, + "grad_norm": 2.462470378310665, + "learning_rate": 3.8748610326046267e-07, + "loss": 0.0153, + "step": 7037 + }, + { + "epoch": 9.14025974025974, + "grad_norm": 1.9249959048859089, + "learning_rate": 3.8632742328825856e-07, + "loss": 0.0026, + "step": 7038 + }, + { + "epoch": 9.141558441558441, + "grad_norm": 0.40256545673295624, + "learning_rate": 3.8517044415874114e-07, + "loss": 0.0012, + "step": 7039 + }, + { + "epoch": 9.142857142857142, + "grad_norm": 3.882598155191011, + "learning_rate": 3.840151660766023e-07, + "loss": 0.0038, + "step": 7040 + }, + { + "epoch": 9.144155844155843, + "grad_norm": 0.7637173720577951, + "learning_rate": 3.8286158924623287e-07, + "loss": 0.0011, + "step": 7041 + }, + { + "epoch": 9.145454545454545, + "grad_norm": 0.8021806953470572, + "learning_rate": 3.8170971387172074e-07, + "loss": 0.0015, + "step": 7042 + }, + { + "epoch": 9.146753246753248, + "grad_norm": 0.39843180253837873, + "learning_rate": 3.8055954015685736e-07, + "loss": 0.0013, + "step": 7043 + }, + { + "epoch": 9.148051948051949, + "grad_norm": 7.914907629563621, + "learning_rate": 3.7941106830512995e-07, + "loss": 0.0096, + "step": 7044 + }, + { + "epoch": 9.14935064935065, + "grad_norm": 3.7681408885488445, + "learning_rate": 3.7826429851972377e-07, + "loss": 0.0027, + "step": 7045 + }, + { + "epoch": 9.150649350649351, + "grad_norm": 0.6124477089568422, + "learning_rate": 3.7711923100352434e-07, + "loss": 0.0014, + "step": 7046 + }, + { + "epoch": 9.151948051948052, + "grad_norm": 1.5395774404476203, + "learning_rate": 3.7597586595911837e-07, + "loss": 0.001, + "step": 7047 + }, + { + "epoch": 9.153246753246753, + "grad_norm": 0.459325633702013, + "learning_rate": 3.748342035887875e-07, + "loss": 0.0012, + "step": 7048 + }, + { + "epoch": 9.154545454545454, + "grad_norm": 0.9942803782693175, + "learning_rate": 3.7369424409451347e-07, + "loss": 0.0011, + "step": 7049 + }, + { + "epoch": 9.155844155844155, + "grad_norm": 0.5995267602987949, + "learning_rate": 3.7255598767797496e-07, + "loss": 0.0019, + "step": 7050 + }, + { + "epoch": 9.157142857142857, + "grad_norm": 1.1643739402821318, + "learning_rate": 3.714194345405575e-07, + "loss": 0.0074, + "step": 7051 + }, + { + "epoch": 9.158441558441558, + "grad_norm": 0.8446082032296451, + "learning_rate": 3.702845848833314e-07, + "loss": 0.0009, + "step": 7052 + }, + { + "epoch": 9.15974025974026, + "grad_norm": 3.6154195950642074, + "learning_rate": 3.691514389070783e-07, + "loss": 0.0283, + "step": 7053 + }, + { + "epoch": 9.161038961038962, + "grad_norm": 0.3443445641855423, + "learning_rate": 3.6801999681227105e-07, + "loss": 0.0007, + "step": 7054 + }, + { + "epoch": 9.162337662337663, + "grad_norm": 0.6806400735565935, + "learning_rate": 3.6689025879908745e-07, + "loss": 0.0012, + "step": 7055 + }, + { + "epoch": 9.163636363636364, + "grad_norm": 15.821212435631994, + "learning_rate": 3.657622250673942e-07, + "loss": 0.0184, + "step": 7056 + }, + { + "epoch": 9.164935064935065, + "grad_norm": 1.6429834831128443, + "learning_rate": 3.6463589581676507e-07, + "loss": 0.0082, + "step": 7057 + }, + { + "epoch": 9.166233766233766, + "grad_norm": 1.5872576632302933, + "learning_rate": 3.6351127124646833e-07, + "loss": 0.0011, + "step": 7058 + }, + { + "epoch": 9.167532467532467, + "grad_norm": 1.965367396161061, + "learning_rate": 3.623883515554738e-07, + "loss": 0.0014, + "step": 7059 + }, + { + "epoch": 9.168831168831169, + "grad_norm": 1.033888303828671, + "learning_rate": 3.6126713694244363e-07, + "loss": 0.0009, + "step": 7060 + }, + { + "epoch": 9.17012987012987, + "grad_norm": 2.252428634170132, + "learning_rate": 3.6014762760574585e-07, + "loss": 0.002, + "step": 7061 + }, + { + "epoch": 9.17142857142857, + "grad_norm": 0.7353108529847674, + "learning_rate": 3.5902982374343976e-07, + "loss": 0.001, + "step": 7062 + }, + { + "epoch": 9.172727272727272, + "grad_norm": 1.356488781545421, + "learning_rate": 3.579137255532894e-07, + "loss": 0.0073, + "step": 7063 + }, + { + "epoch": 9.174025974025975, + "grad_norm": 1.8785042992873344, + "learning_rate": 3.567993332327524e-07, + "loss": 0.0012, + "step": 7064 + }, + { + "epoch": 9.175324675324676, + "grad_norm": 1.4858734320686595, + "learning_rate": 3.556866469789866e-07, + "loss": 0.0013, + "step": 7065 + }, + { + "epoch": 9.176623376623377, + "grad_norm": 1.2905564425800855, + "learning_rate": 3.545756669888467e-07, + "loss": 0.0013, + "step": 7066 + }, + { + "epoch": 9.177922077922078, + "grad_norm": 0.9250906881600344, + "learning_rate": 3.5346639345888554e-07, + "loss": 0.0069, + "step": 7067 + }, + { + "epoch": 9.17922077922078, + "grad_norm": 0.6961808773642965, + "learning_rate": 3.5235882658535725e-07, + "loss": 0.0013, + "step": 7068 + }, + { + "epoch": 9.18051948051948, + "grad_norm": 7.302935809207607, + "learning_rate": 3.512529665642106e-07, + "loss": 0.0135, + "step": 7069 + }, + { + "epoch": 9.181818181818182, + "grad_norm": 1.666547792283117, + "learning_rate": 3.5014881359109244e-07, + "loss": 0.0092, + "step": 7070 + }, + { + "epoch": 9.183116883116883, + "grad_norm": 0.9983646974704413, + "learning_rate": 3.490463678613487e-07, + "loss": 0.0101, + "step": 7071 + }, + { + "epoch": 9.184415584415584, + "grad_norm": 2.448471871728372, + "learning_rate": 3.4794562957002455e-07, + "loss": 0.0092, + "step": 7072 + }, + { + "epoch": 9.185714285714285, + "grad_norm": 0.23012707341366356, + "learning_rate": 3.4684659891185967e-07, + "loss": 0.0011, + "step": 7073 + }, + { + "epoch": 9.187012987012986, + "grad_norm": 3.4395128934578416, + "learning_rate": 3.457492760812975e-07, + "loss": 0.0109, + "step": 7074 + }, + { + "epoch": 9.188311688311689, + "grad_norm": 0.37082261741631595, + "learning_rate": 3.446536612724694e-07, + "loss": 0.0012, + "step": 7075 + }, + { + "epoch": 9.18961038961039, + "grad_norm": 1.324885555679826, + "learning_rate": 3.4355975467921576e-07, + "loss": 0.0015, + "step": 7076 + }, + { + "epoch": 9.190909090909091, + "grad_norm": 0.8912140239897071, + "learning_rate": 3.4246755649506746e-07, + "loss": 0.001, + "step": 7077 + }, + { + "epoch": 9.192207792207792, + "grad_norm": 1.3527854531470214, + "learning_rate": 3.413770669132566e-07, + "loss": 0.0017, + "step": 7078 + }, + { + "epoch": 9.193506493506494, + "grad_norm": 3.649370851250802, + "learning_rate": 3.4028828612670873e-07, + "loss": 0.0017, + "step": 7079 + }, + { + "epoch": 9.194805194805195, + "grad_norm": 4.054969581899702, + "learning_rate": 3.392012143280532e-07, + "loss": 0.0067, + "step": 7080 + }, + { + "epoch": 9.196103896103896, + "grad_norm": 1.1961695922703122, + "learning_rate": 3.381158517096128e-07, + "loss": 0.0039, + "step": 7081 + }, + { + "epoch": 9.197402597402597, + "grad_norm": 3.251949756485924, + "learning_rate": 3.370321984634084e-07, + "loss": 0.02, + "step": 7082 + }, + { + "epoch": 9.198701298701298, + "grad_norm": 0.8749972841347758, + "learning_rate": 3.359502547811577e-07, + "loss": 0.0029, + "step": 7083 + }, + { + "epoch": 9.2, + "grad_norm": 1.5000458042043103, + "learning_rate": 3.348700208542799e-07, + "loss": 0.0093, + "step": 7084 + }, + { + "epoch": 9.2012987012987, + "grad_norm": 0.2945910982988263, + "learning_rate": 3.3379149687388866e-07, + "loss": 0.0012, + "step": 7085 + }, + { + "epoch": 9.202597402597403, + "grad_norm": 1.4460522705197427, + "learning_rate": 3.327146830307937e-07, + "loss": 0.0013, + "step": 7086 + }, + { + "epoch": 9.203896103896104, + "grad_norm": 1.0683053843634143, + "learning_rate": 3.3163957951550476e-07, + "loss": 0.0158, + "step": 7087 + }, + { + "epoch": 9.205194805194806, + "grad_norm": 2.6662359508052327, + "learning_rate": 3.305661865182308e-07, + "loss": 0.002, + "step": 7088 + }, + { + "epoch": 9.206493506493507, + "grad_norm": 1.0435990960454755, + "learning_rate": 3.2949450422887217e-07, + "loss": 0.0074, + "step": 7089 + }, + { + "epoch": 9.207792207792208, + "grad_norm": 1.0436740250268342, + "learning_rate": 3.2842453283703055e-07, + "loss": 0.0077, + "step": 7090 + }, + { + "epoch": 9.209090909090909, + "grad_norm": 0.6720907079715094, + "learning_rate": 3.273562725320045e-07, + "loss": 0.0015, + "step": 7091 + }, + { + "epoch": 9.21038961038961, + "grad_norm": 0.7024903872396399, + "learning_rate": 3.2628972350279176e-07, + "loss": 0.001, + "step": 7092 + }, + { + "epoch": 9.211688311688311, + "grad_norm": 0.37598311860069256, + "learning_rate": 3.2522488593808357e-07, + "loss": 0.0015, + "step": 7093 + }, + { + "epoch": 9.212987012987012, + "grad_norm": 3.3216729551086344, + "learning_rate": 3.241617600262703e-07, + "loss": 0.0255, + "step": 7094 + }, + { + "epoch": 9.214285714285714, + "grad_norm": 1.5913859578321699, + "learning_rate": 3.2310034595543824e-07, + "loss": 0.0075, + "step": 7095 + }, + { + "epoch": 9.215584415584416, + "grad_norm": 3.817261376974092, + "learning_rate": 3.2204064391337277e-07, + "loss": 0.0028, + "step": 7096 + }, + { + "epoch": 9.216883116883118, + "grad_norm": 0.9966640560747944, + "learning_rate": 3.2098265408755716e-07, + "loss": 0.0014, + "step": 7097 + }, + { + "epoch": 9.218181818181819, + "grad_norm": 0.5131636889831397, + "learning_rate": 3.1992637666516726e-07, + "loss": 0.0084, + "step": 7098 + }, + { + "epoch": 9.21948051948052, + "grad_norm": 0.9200710857064608, + "learning_rate": 3.1887181183307914e-07, + "loss": 0.0084, + "step": 7099 + }, + { + "epoch": 9.220779220779221, + "grad_norm": 0.49236083534566194, + "learning_rate": 3.1781895977786694e-07, + "loss": 0.0008, + "step": 7100 + }, + { + "epoch": 9.222077922077922, + "grad_norm": 5.191427791891197, + "learning_rate": 3.1676782068580047e-07, + "loss": 0.01, + "step": 7101 + }, + { + "epoch": 9.223376623376623, + "grad_norm": 1.730435341360834, + "learning_rate": 3.1571839474284437e-07, + "loss": 0.002, + "step": 7102 + }, + { + "epoch": 9.224675324675324, + "grad_norm": 1.4620192213798628, + "learning_rate": 3.1467068213466345e-07, + "loss": 0.0016, + "step": 7103 + }, + { + "epoch": 9.225974025974025, + "grad_norm": 1.3030769979724102, + "learning_rate": 3.1362468304661944e-07, + "loss": 0.0015, + "step": 7104 + }, + { + "epoch": 9.227272727272727, + "grad_norm": 0.5061533561871399, + "learning_rate": 3.125803976637676e-07, + "loss": 0.0009, + "step": 7105 + }, + { + "epoch": 9.228571428571428, + "grad_norm": 1.1718677226966967, + "learning_rate": 3.1153782617086126e-07, + "loss": 0.0013, + "step": 7106 + }, + { + "epoch": 9.22987012987013, + "grad_norm": 3.0989755871753277, + "learning_rate": 3.1049696875235515e-07, + "loss": 0.0221, + "step": 7107 + }, + { + "epoch": 9.231168831168832, + "grad_norm": 1.216825009402099, + "learning_rate": 3.094578255923919e-07, + "loss": 0.0012, + "step": 7108 + }, + { + "epoch": 9.232467532467533, + "grad_norm": 2.0219463891496225, + "learning_rate": 3.08420396874819e-07, + "loss": 0.0017, + "step": 7109 + }, + { + "epoch": 9.233766233766234, + "grad_norm": 5.065809999915443, + "learning_rate": 3.073846827831761e-07, + "loss": 0.0045, + "step": 7110 + }, + { + "epoch": 9.235064935064935, + "grad_norm": 0.8327574662381317, + "learning_rate": 3.063506835007035e-07, + "loss": 0.0074, + "step": 7111 + }, + { + "epoch": 9.236363636363636, + "grad_norm": 1.7342067047570757, + "learning_rate": 3.0531839921033146e-07, + "loss": 0.006, + "step": 7112 + }, + { + "epoch": 9.237662337662337, + "grad_norm": 2.213798059708507, + "learning_rate": 3.042878300946939e-07, + "loss": 0.002, + "step": 7113 + }, + { + "epoch": 9.238961038961039, + "grad_norm": 6.202919464255222, + "learning_rate": 3.03258976336116e-07, + "loss": 0.0124, + "step": 7114 + }, + { + "epoch": 9.24025974025974, + "grad_norm": 1.4212209683412214, + "learning_rate": 3.022318381166245e-07, + "loss": 0.0011, + "step": 7115 + }, + { + "epoch": 9.24155844155844, + "grad_norm": 0.5749006901053054, + "learning_rate": 3.0120641561793616e-07, + "loss": 0.0011, + "step": 7116 + }, + { + "epoch": 9.242857142857142, + "grad_norm": 0.801188926580731, + "learning_rate": 3.001827090214715e-07, + "loss": 0.0013, + "step": 7117 + }, + { + "epoch": 9.244155844155845, + "grad_norm": 4.499297517887477, + "learning_rate": 2.9916071850834004e-07, + "loss": 0.0204, + "step": 7118 + }, + { + "epoch": 9.245454545454546, + "grad_norm": 1.233366726686066, + "learning_rate": 2.9814044425935605e-07, + "loss": 0.0076, + "step": 7119 + }, + { + "epoch": 9.246753246753247, + "grad_norm": 2.389271133125393, + "learning_rate": 2.971218864550207e-07, + "loss": 0.0081, + "step": 7120 + }, + { + "epoch": 9.248051948051948, + "grad_norm": 1.7168871776500985, + "learning_rate": 2.9610504527553987e-07, + "loss": 0.0016, + "step": 7121 + }, + { + "epoch": 9.24935064935065, + "grad_norm": 0.8258670157455065, + "learning_rate": 2.9508992090081176e-07, + "loss": 0.0031, + "step": 7122 + }, + { + "epoch": 9.25064935064935, + "grad_norm": 3.1925322469730144, + "learning_rate": 2.940765135104295e-07, + "loss": 0.0086, + "step": 7123 + }, + { + "epoch": 9.251948051948052, + "grad_norm": 3.9760364719864816, + "learning_rate": 2.9306482328368523e-07, + "loss": 0.0203, + "step": 7124 + }, + { + "epoch": 9.253246753246753, + "grad_norm": 2.0742377358249326, + "learning_rate": 2.920548503995668e-07, + "loss": 0.0082, + "step": 7125 + }, + { + "epoch": 9.254545454545454, + "grad_norm": 0.5422055483737269, + "learning_rate": 2.9104659503675804e-07, + "loss": 0.0088, + "step": 7126 + }, + { + "epoch": 9.255844155844155, + "grad_norm": 4.46522166864626, + "learning_rate": 2.9004005737363726e-07, + "loss": 0.0121, + "step": 7127 + }, + { + "epoch": 9.257142857142856, + "grad_norm": 1.7709326210352736, + "learning_rate": 2.89035237588281e-07, + "loss": 0.001, + "step": 7128 + }, + { + "epoch": 9.25844155844156, + "grad_norm": 2.352092072461833, + "learning_rate": 2.8803213585846036e-07, + "loss": 0.0026, + "step": 7129 + }, + { + "epoch": 9.25974025974026, + "grad_norm": 4.239313437730481, + "learning_rate": 2.8703075236164556e-07, + "loss": 0.0116, + "step": 7130 + }, + { + "epoch": 9.261038961038961, + "grad_norm": 3.401639476557899, + "learning_rate": 2.860310872749983e-07, + "loss": 0.002, + "step": 7131 + }, + { + "epoch": 9.262337662337663, + "grad_norm": 1.390714080836409, + "learning_rate": 2.8503314077537816e-07, + "loss": 0.0197, + "step": 7132 + }, + { + "epoch": 9.263636363636364, + "grad_norm": 4.241842457371831, + "learning_rate": 2.840369130393428e-07, + "loss": 0.002, + "step": 7133 + }, + { + "epoch": 9.264935064935065, + "grad_norm": 1.5361600613297381, + "learning_rate": 2.8304240424314235e-07, + "loss": 0.0013, + "step": 7134 + }, + { + "epoch": 9.266233766233766, + "grad_norm": 2.7556997871869204, + "learning_rate": 2.8204961456272495e-07, + "loss": 0.0081, + "step": 7135 + }, + { + "epoch": 9.267532467532467, + "grad_norm": 2.3609820397092434, + "learning_rate": 2.8105854417373457e-07, + "loss": 0.0022, + "step": 7136 + }, + { + "epoch": 9.268831168831168, + "grad_norm": 1.2615804392466157, + "learning_rate": 2.8006919325151095e-07, + "loss": 0.0023, + "step": 7137 + }, + { + "epoch": 9.27012987012987, + "grad_norm": 1.4648056000011405, + "learning_rate": 2.7908156197108847e-07, + "loss": 0.0028, + "step": 7138 + }, + { + "epoch": 9.271428571428572, + "grad_norm": 3.407512132054432, + "learning_rate": 2.7809565050719743e-07, + "loss": 0.0082, + "step": 7139 + }, + { + "epoch": 9.272727272727273, + "grad_norm": 1.8853669310096237, + "learning_rate": 2.7711145903426494e-07, + "loss": 0.0016, + "step": 7140 + }, + { + "epoch": 9.274025974025975, + "grad_norm": 1.4532023318580751, + "learning_rate": 2.761289877264139e-07, + "loss": 0.0022, + "step": 7141 + }, + { + "epoch": 9.275324675324676, + "grad_norm": 4.931450915971397, + "learning_rate": 2.7514823675746205e-07, + "loss": 0.0154, + "step": 7142 + }, + { + "epoch": 9.276623376623377, + "grad_norm": 1.6468053646950511, + "learning_rate": 2.741692063009216e-07, + "loss": 0.0113, + "step": 7143 + }, + { + "epoch": 9.277922077922078, + "grad_norm": 1.4780293381958398, + "learning_rate": 2.73191896530004e-07, + "loss": 0.0086, + "step": 7144 + }, + { + "epoch": 9.279220779220779, + "grad_norm": 0.8382416958489365, + "learning_rate": 2.722163076176132e-07, + "loss": 0.0014, + "step": 7145 + }, + { + "epoch": 9.28051948051948, + "grad_norm": 1.398765923656471, + "learning_rate": 2.712424397363489e-07, + "loss": 0.0022, + "step": 7146 + }, + { + "epoch": 9.281818181818181, + "grad_norm": 1.395230806967035, + "learning_rate": 2.702702930585066e-07, + "loss": 0.0105, + "step": 7147 + }, + { + "epoch": 9.283116883116882, + "grad_norm": 2.6968938229750257, + "learning_rate": 2.692998677560799e-07, + "loss": 0.017, + "step": 7148 + }, + { + "epoch": 9.284415584415584, + "grad_norm": 1.519003163047448, + "learning_rate": 2.6833116400075355e-07, + "loss": 0.001, + "step": 7149 + }, + { + "epoch": 9.285714285714286, + "grad_norm": 9.866815298511392, + "learning_rate": 2.673641819639106e-07, + "loss": 0.0069, + "step": 7150 + }, + { + "epoch": 9.287012987012988, + "grad_norm": 7.770445750060313, + "learning_rate": 2.6639892181662744e-07, + "loss": 0.0058, + "step": 7151 + }, + { + "epoch": 9.288311688311689, + "grad_norm": 1.0346914939821703, + "learning_rate": 2.654353837296797e-07, + "loss": 0.0069, + "step": 7152 + }, + { + "epoch": 9.28961038961039, + "grad_norm": 0.6200704015809042, + "learning_rate": 2.6447356787353217e-07, + "loss": 0.002, + "step": 7153 + }, + { + "epoch": 9.290909090909091, + "grad_norm": 4.056283747192585, + "learning_rate": 2.6351347441835207e-07, + "loss": 0.0057, + "step": 7154 + }, + { + "epoch": 9.292207792207792, + "grad_norm": 0.4380793247105217, + "learning_rate": 2.6255510353399574e-07, + "loss": 0.001, + "step": 7155 + }, + { + "epoch": 9.293506493506493, + "grad_norm": 3.051398752431963, + "learning_rate": 2.6159845539001973e-07, + "loss": 0.0053, + "step": 7156 + }, + { + "epoch": 9.294805194805194, + "grad_norm": 2.719778378790745, + "learning_rate": 2.606435301556709e-07, + "loss": 0.008, + "step": 7157 + }, + { + "epoch": 9.296103896103896, + "grad_norm": 2.8205748219136586, + "learning_rate": 2.596903279998952e-07, + "loss": 0.0155, + "step": 7158 + }, + { + "epoch": 9.297402597402597, + "grad_norm": 1.6295221102506776, + "learning_rate": 2.587388490913323e-07, + "loss": 0.0085, + "step": 7159 + }, + { + "epoch": 9.2987012987013, + "grad_norm": 4.351071862716202, + "learning_rate": 2.5778909359831625e-07, + "loss": 0.0035, + "step": 7160 + }, + { + "epoch": 9.3, + "grad_norm": 0.7907317027358056, + "learning_rate": 2.568410616888761e-07, + "loss": 0.0016, + "step": 7161 + }, + { + "epoch": 9.301298701298702, + "grad_norm": 5.58170826842395, + "learning_rate": 2.5589475353073987e-07, + "loss": 0.0169, + "step": 7162 + }, + { + "epoch": 9.302597402597403, + "grad_norm": 0.6638022637142083, + "learning_rate": 2.549501692913259e-07, + "loss": 0.0009, + "step": 7163 + }, + { + "epoch": 9.303896103896104, + "grad_norm": 0.8362044874606477, + "learning_rate": 2.5400730913774817e-07, + "loss": 0.0016, + "step": 7164 + }, + { + "epoch": 9.305194805194805, + "grad_norm": 1.5150298636229815, + "learning_rate": 2.5306617323681894e-07, + "loss": 0.0101, + "step": 7165 + }, + { + "epoch": 9.306493506493506, + "grad_norm": 0.8212019769559403, + "learning_rate": 2.521267617550416e-07, + "loss": 0.0012, + "step": 7166 + }, + { + "epoch": 9.307792207792208, + "grad_norm": 1.1793420429037953, + "learning_rate": 2.5118907485861654e-07, + "loss": 0.0077, + "step": 7167 + }, + { + "epoch": 9.309090909090909, + "grad_norm": 0.6101502875528758, + "learning_rate": 2.5025311271343665e-07, + "loss": 0.0008, + "step": 7168 + }, + { + "epoch": 9.31038961038961, + "grad_norm": 2.5166795831569084, + "learning_rate": 2.4931887548509503e-07, + "loss": 0.008, + "step": 7169 + }, + { + "epoch": 9.311688311688311, + "grad_norm": 0.527353878579148, + "learning_rate": 2.483863633388739e-07, + "loss": 0.0013, + "step": 7170 + }, + { + "epoch": 9.312987012987014, + "grad_norm": 0.41071986282833, + "learning_rate": 2.4745557643975347e-07, + "loss": 0.001, + "step": 7171 + }, + { + "epoch": 9.314285714285715, + "grad_norm": 5.435119999721423, + "learning_rate": 2.465265149524065e-07, + "loss": 0.0215, + "step": 7172 + }, + { + "epoch": 9.315584415584416, + "grad_norm": 4.558109965470126, + "learning_rate": 2.4559917904120266e-07, + "loss": 0.0103, + "step": 7173 + }, + { + "epoch": 9.316883116883117, + "grad_norm": 2.8557548111153634, + "learning_rate": 2.4467356887020624e-07, + "loss": 0.0022, + "step": 7174 + }, + { + "epoch": 9.318181818181818, + "grad_norm": 5.454004479079293, + "learning_rate": 2.437496846031739e-07, + "loss": 0.003, + "step": 7175 + }, + { + "epoch": 9.31948051948052, + "grad_norm": 2.1262969187180527, + "learning_rate": 2.4282752640355846e-07, + "loss": 0.0027, + "step": 7176 + }, + { + "epoch": 9.32077922077922, + "grad_norm": 6.292685148799, + "learning_rate": 2.419070944345081e-07, + "loss": 0.011, + "step": 7177 + }, + { + "epoch": 9.322077922077922, + "grad_norm": 1.6059752120357271, + "learning_rate": 2.409883888588649e-07, + "loss": 0.0063, + "step": 7178 + }, + { + "epoch": 9.323376623376623, + "grad_norm": 1.089766808076682, + "learning_rate": 2.400714098391654e-07, + "loss": 0.0019, + "step": 7179 + }, + { + "epoch": 9.324675324675324, + "grad_norm": 2.1923605964629833, + "learning_rate": 2.391561575376389e-07, + "loss": 0.0074, + "step": 7180 + }, + { + "epoch": 9.325974025974025, + "grad_norm": 1.1893123649415172, + "learning_rate": 2.3824263211621457e-07, + "loss": 0.0029, + "step": 7181 + }, + { + "epoch": 9.327272727272728, + "grad_norm": 1.744584166621901, + "learning_rate": 2.373308337365099e-07, + "loss": 0.016, + "step": 7182 + }, + { + "epoch": 9.32857142857143, + "grad_norm": 6.290245014808479, + "learning_rate": 2.3642076255984027e-07, + "loss": 0.0106, + "step": 7183 + }, + { + "epoch": 9.32987012987013, + "grad_norm": 1.0754328066828607, + "learning_rate": 2.3551241874721353e-07, + "loss": 0.0021, + "step": 7184 + }, + { + "epoch": 9.331168831168831, + "grad_norm": 1.026131911905493, + "learning_rate": 2.3460580245933562e-07, + "loss": 0.0108, + "step": 7185 + }, + { + "epoch": 9.332467532467533, + "grad_norm": 0.7140142353554344, + "learning_rate": 2.3370091385660377e-07, + "loss": 0.0079, + "step": 7186 + }, + { + "epoch": 9.333766233766234, + "grad_norm": 1.8080551179671742, + "learning_rate": 2.327977530991077e-07, + "loss": 0.0016, + "step": 7187 + }, + { + "epoch": 9.335064935064935, + "grad_norm": 1.5548095974928078, + "learning_rate": 2.3189632034663513e-07, + "loss": 0.0011, + "step": 7188 + }, + { + "epoch": 9.336363636363636, + "grad_norm": 2.021879080518838, + "learning_rate": 2.309966157586685e-07, + "loss": 0.0021, + "step": 7189 + }, + { + "epoch": 9.337662337662337, + "grad_norm": 3.8705315253355983, + "learning_rate": 2.3009863949437937e-07, + "loss": 0.0215, + "step": 7190 + }, + { + "epoch": 9.338961038961038, + "grad_norm": 2.497585465440136, + "learning_rate": 2.292023917126407e-07, + "loss": 0.0157, + "step": 7191 + }, + { + "epoch": 9.34025974025974, + "grad_norm": 1.319307962967952, + "learning_rate": 2.2830787257201225e-07, + "loss": 0.0011, + "step": 7192 + }, + { + "epoch": 9.341558441558442, + "grad_norm": 1.8598930334577741, + "learning_rate": 2.2741508223075526e-07, + "loss": 0.0083, + "step": 7193 + }, + { + "epoch": 9.342857142857143, + "grad_norm": 3.44472620728884, + "learning_rate": 2.265240208468178e-07, + "loss": 0.0026, + "step": 7194 + }, + { + "epoch": 9.344155844155845, + "grad_norm": 2.5338708192191066, + "learning_rate": 2.256346885778482e-07, + "loss": 0.0132, + "step": 7195 + }, + { + "epoch": 9.345454545454546, + "grad_norm": 0.3382699193121015, + "learning_rate": 2.2474708558118397e-07, + "loss": 0.001, + "step": 7196 + }, + { + "epoch": 9.346753246753247, + "grad_norm": 3.8171517393503, + "learning_rate": 2.238612120138617e-07, + "loss": 0.0097, + "step": 7197 + }, + { + "epoch": 9.348051948051948, + "grad_norm": 6.858504412263834, + "learning_rate": 2.2297706803260932e-07, + "loss": 0.0196, + "step": 7198 + }, + { + "epoch": 9.349350649350649, + "grad_norm": 3.9646695768001865, + "learning_rate": 2.2209465379384731e-07, + "loss": 0.0097, + "step": 7199 + }, + { + "epoch": 9.35064935064935, + "grad_norm": 3.153854660403098, + "learning_rate": 2.212139694536919e-07, + "loss": 0.0153, + "step": 7200 + }, + { + "epoch": 9.351948051948051, + "grad_norm": 0.6688967297466166, + "learning_rate": 2.2033501516795286e-07, + "loss": 0.0021, + "step": 7201 + }, + { + "epoch": 9.353246753246752, + "grad_norm": 1.4920282362349748, + "learning_rate": 2.1945779109213473e-07, + "loss": 0.0102, + "step": 7202 + }, + { + "epoch": 9.354545454545455, + "grad_norm": 1.6376305836166167, + "learning_rate": 2.1858229738143554e-07, + "loss": 0.0011, + "step": 7203 + }, + { + "epoch": 9.355844155844157, + "grad_norm": 1.0060242107316582, + "learning_rate": 2.177085341907459e-07, + "loss": 0.0012, + "step": 7204 + }, + { + "epoch": 9.357142857142858, + "grad_norm": 1.9309528604562674, + "learning_rate": 2.1683650167465098e-07, + "loss": 0.0011, + "step": 7205 + }, + { + "epoch": 9.358441558441559, + "grad_norm": 0.5948070662256864, + "learning_rate": 2.1596619998743075e-07, + "loss": 0.0073, + "step": 7206 + }, + { + "epoch": 9.35974025974026, + "grad_norm": 0.8190651802256369, + "learning_rate": 2.1509762928305754e-07, + "loss": 0.0011, + "step": 7207 + }, + { + "epoch": 9.361038961038961, + "grad_norm": 2.31171964409837, + "learning_rate": 2.1423078971520183e-07, + "loss": 0.002, + "step": 7208 + }, + { + "epoch": 9.362337662337662, + "grad_norm": 1.0852866843269968, + "learning_rate": 2.133656814372176e-07, + "loss": 0.0011, + "step": 7209 + }, + { + "epoch": 9.363636363636363, + "grad_norm": 0.7981265822137421, + "learning_rate": 2.1250230460216459e-07, + "loss": 0.0089, + "step": 7210 + }, + { + "epoch": 9.364935064935064, + "grad_norm": 0.3732788228009341, + "learning_rate": 2.1164065936278732e-07, + "loss": 0.0008, + "step": 7211 + }, + { + "epoch": 9.366233766233766, + "grad_norm": 0.41857962590665615, + "learning_rate": 2.1078074587152942e-07, + "loss": 0.0009, + "step": 7212 + }, + { + "epoch": 9.367532467532467, + "grad_norm": 1.9755980537793707, + "learning_rate": 2.099225642805247e-07, + "loss": 0.0084, + "step": 7213 + }, + { + "epoch": 9.36883116883117, + "grad_norm": 0.9505098026521261, + "learning_rate": 2.0906611474160288e-07, + "loss": 0.0128, + "step": 7214 + }, + { + "epoch": 9.37012987012987, + "grad_norm": 0.8197086789897519, + "learning_rate": 2.0821139740628604e-07, + "loss": 0.0012, + "step": 7215 + }, + { + "epoch": 9.371428571428572, + "grad_norm": 1.5150284366144324, + "learning_rate": 2.0735841242578992e-07, + "loss": 0.0011, + "step": 7216 + }, + { + "epoch": 9.372727272727273, + "grad_norm": 4.182316208311269, + "learning_rate": 2.0650715995102266e-07, + "loss": 0.0145, + "step": 7217 + }, + { + "epoch": 9.374025974025974, + "grad_norm": 3.632974181773877, + "learning_rate": 2.0565764013258826e-07, + "loss": 0.0085, + "step": 7218 + }, + { + "epoch": 9.375324675324675, + "grad_norm": 0.3347453425348726, + "learning_rate": 2.0480985312078427e-07, + "loss": 0.0008, + "step": 7219 + }, + { + "epoch": 9.376623376623376, + "grad_norm": 1.3168833738143415, + "learning_rate": 2.039637990655985e-07, + "loss": 0.001, + "step": 7220 + }, + { + "epoch": 9.377922077922078, + "grad_norm": 1.7955854806782916, + "learning_rate": 2.0311947811671452e-07, + "loss": 0.0018, + "step": 7221 + }, + { + "epoch": 9.379220779220779, + "grad_norm": 1.3576434567711868, + "learning_rate": 2.0227689042350952e-07, + "loss": 0.0014, + "step": 7222 + }, + { + "epoch": 9.38051948051948, + "grad_norm": 3.2178454031168156, + "learning_rate": 2.0143603613505313e-07, + "loss": 0.009, + "step": 7223 + }, + { + "epoch": 9.381818181818183, + "grad_norm": 0.6143103355151985, + "learning_rate": 2.0059691540010863e-07, + "loss": 0.0033, + "step": 7224 + }, + { + "epoch": 9.383116883116884, + "grad_norm": 2.669694717751029, + "learning_rate": 1.9975952836713163e-07, + "loss": 0.0109, + "step": 7225 + }, + { + "epoch": 9.384415584415585, + "grad_norm": 0.6468479875047185, + "learning_rate": 1.989238751842737e-07, + "loss": 0.0022, + "step": 7226 + }, + { + "epoch": 9.385714285714286, + "grad_norm": 0.5840638133701114, + "learning_rate": 1.9808995599937653e-07, + "loss": 0.0013, + "step": 7227 + }, + { + "epoch": 9.387012987012987, + "grad_norm": 1.0508703609198788, + "learning_rate": 1.9725777095997655e-07, + "loss": 0.0068, + "step": 7228 + }, + { + "epoch": 9.388311688311688, + "grad_norm": 3.1468139795199486, + "learning_rate": 1.9642732021330268e-07, + "loss": 0.0109, + "step": 7229 + }, + { + "epoch": 9.38961038961039, + "grad_norm": 0.5719266334445704, + "learning_rate": 1.955986039062785e-07, + "loss": 0.0031, + "step": 7230 + }, + { + "epoch": 9.39090909090909, + "grad_norm": 3.126790736086955, + "learning_rate": 1.9477162218552115e-07, + "loss": 0.0059, + "step": 7231 + }, + { + "epoch": 9.392207792207792, + "grad_norm": 6.966312397409845, + "learning_rate": 1.9394637519733583e-07, + "loss": 0.0108, + "step": 7232 + }, + { + "epoch": 9.393506493506493, + "grad_norm": 0.7529406531884486, + "learning_rate": 1.9312286308772687e-07, + "loss": 0.0068, + "step": 7233 + }, + { + "epoch": 9.394805194805194, + "grad_norm": 1.9153608534892825, + "learning_rate": 1.9230108600238995e-07, + "loss": 0.0082, + "step": 7234 + }, + { + "epoch": 9.396103896103897, + "grad_norm": 1.0962211904175763, + "learning_rate": 1.9148104408671208e-07, + "loss": 0.0011, + "step": 7235 + }, + { + "epoch": 9.397402597402598, + "grad_norm": 1.0298805186523232, + "learning_rate": 1.9066273748577391e-07, + "loss": 0.0012, + "step": 7236 + }, + { + "epoch": 9.3987012987013, + "grad_norm": 1.728808001598506, + "learning_rate": 1.8984616634435072e-07, + "loss": 0.0014, + "step": 7237 + }, + { + "epoch": 9.4, + "grad_norm": 2.7059296765341814, + "learning_rate": 1.8903133080691027e-07, + "loss": 0.016, + "step": 7238 + }, + { + "epoch": 9.401298701298701, + "grad_norm": 0.863463084624122, + "learning_rate": 1.8821823101760949e-07, + "loss": 0.001, + "step": 7239 + }, + { + "epoch": 9.402597402597403, + "grad_norm": 0.48202898083566537, + "learning_rate": 1.8740686712030332e-07, + "loss": 0.0011, + "step": 7240 + }, + { + "epoch": 9.403896103896104, + "grad_norm": 0.8182746357286638, + "learning_rate": 1.8659723925853913e-07, + "loss": 0.0012, + "step": 7241 + }, + { + "epoch": 9.405194805194805, + "grad_norm": 0.37581852153058215, + "learning_rate": 1.8578934757555233e-07, + "loss": 0.001, + "step": 7242 + }, + { + "epoch": 9.406493506493506, + "grad_norm": 2.3893362572403003, + "learning_rate": 1.849831922142764e-07, + "loss": 0.0021, + "step": 7243 + }, + { + "epoch": 9.407792207792207, + "grad_norm": 2.636942327860397, + "learning_rate": 1.8417877331733393e-07, + "loss": 0.0026, + "step": 7244 + }, + { + "epoch": 9.409090909090908, + "grad_norm": 4.477980739786025, + "learning_rate": 1.8337609102704544e-07, + "loss": 0.0084, + "step": 7245 + }, + { + "epoch": 9.410389610389611, + "grad_norm": 3.143690459921717, + "learning_rate": 1.8257514548541632e-07, + "loss": 0.0042, + "step": 7246 + }, + { + "epoch": 9.411688311688312, + "grad_norm": 0.8864470971916073, + "learning_rate": 1.8177593683415318e-07, + "loss": 0.0014, + "step": 7247 + }, + { + "epoch": 9.412987012987013, + "grad_norm": 3.1331486615036033, + "learning_rate": 1.8097846521464846e-07, + "loss": 0.0192, + "step": 7248 + }, + { + "epoch": 9.414285714285715, + "grad_norm": 0.9201933131644962, + "learning_rate": 1.8018273076799264e-07, + "loss": 0.0085, + "step": 7249 + }, + { + "epoch": 9.415584415584416, + "grad_norm": 1.0712981784077302, + "learning_rate": 1.7938873363496422e-07, + "loss": 0.0012, + "step": 7250 + }, + { + "epoch": 9.416883116883117, + "grad_norm": 1.1370221610585238, + "learning_rate": 1.7859647395603753e-07, + "loss": 0.0012, + "step": 7251 + }, + { + "epoch": 9.418181818181818, + "grad_norm": 0.49206957251040817, + "learning_rate": 1.7780595187137816e-07, + "loss": 0.0011, + "step": 7252 + }, + { + "epoch": 9.41948051948052, + "grad_norm": 5.248004386187279, + "learning_rate": 1.770171675208454e-07, + "loss": 0.0108, + "step": 7253 + }, + { + "epoch": 9.42077922077922, + "grad_norm": 1.1913860876217481, + "learning_rate": 1.7623012104398872e-07, + "loss": 0.001, + "step": 7254 + }, + { + "epoch": 9.422077922077921, + "grad_norm": 2.8960045452876666, + "learning_rate": 1.754448125800523e-07, + "loss": 0.0098, + "step": 7255 + }, + { + "epoch": 9.423376623376623, + "grad_norm": 1.3574430774455581, + "learning_rate": 1.7466124226797386e-07, + "loss": 0.001, + "step": 7256 + }, + { + "epoch": 9.424675324675325, + "grad_norm": 0.5005255464686963, + "learning_rate": 1.7387941024638033e-07, + "loss": 0.0078, + "step": 7257 + }, + { + "epoch": 9.425974025974027, + "grad_norm": 1.046293123587476, + "learning_rate": 1.7309931665359213e-07, + "loss": 0.0015, + "step": 7258 + }, + { + "epoch": 9.427272727272728, + "grad_norm": 0.28725233696650654, + "learning_rate": 1.7232096162762335e-07, + "loss": 0.0027, + "step": 7259 + }, + { + "epoch": 9.428571428571429, + "grad_norm": 0.35410712101313646, + "learning_rate": 1.7154434530618158e-07, + "loss": 0.003, + "step": 7260 + }, + { + "epoch": 9.42987012987013, + "grad_norm": 2.502935457979415, + "learning_rate": 1.707694678266636e-07, + "loss": 0.0071, + "step": 7261 + }, + { + "epoch": 9.431168831168831, + "grad_norm": 0.23560408431906865, + "learning_rate": 1.6999632932615974e-07, + "loss": 0.0011, + "step": 7262 + }, + { + "epoch": 9.432467532467532, + "grad_norm": 1.2816259328322133, + "learning_rate": 1.692249299414539e-07, + "loss": 0.0101, + "step": 7263 + }, + { + "epoch": 9.433766233766233, + "grad_norm": 1.499653188605821, + "learning_rate": 1.6845526980902028e-07, + "loss": 0.0086, + "step": 7264 + }, + { + "epoch": 9.435064935064934, + "grad_norm": 0.3560262999558271, + "learning_rate": 1.676873490650277e-07, + "loss": 0.0011, + "step": 7265 + }, + { + "epoch": 9.436363636363636, + "grad_norm": 0.6181254158854833, + "learning_rate": 1.669211678453353e-07, + "loss": 0.0009, + "step": 7266 + }, + { + "epoch": 9.437662337662339, + "grad_norm": 0.5129211921970737, + "learning_rate": 1.6615672628549572e-07, + "loss": 0.001, + "step": 7267 + }, + { + "epoch": 9.43896103896104, + "grad_norm": 0.6253608247583345, + "learning_rate": 1.6539402452075304e-07, + "loss": 0.009, + "step": 7268 + }, + { + "epoch": 9.44025974025974, + "grad_norm": 5.56353410726274, + "learning_rate": 1.6463306268604372e-07, + "loss": 0.0074, + "step": 7269 + }, + { + "epoch": 9.441558441558442, + "grad_norm": 0.21018194781315905, + "learning_rate": 1.6387384091599679e-07, + "loss": 0.001, + "step": 7270 + }, + { + "epoch": 9.442857142857143, + "grad_norm": 1.8755310472649047, + "learning_rate": 1.6311635934493254e-07, + "loss": 0.01, + "step": 7271 + }, + { + "epoch": 9.444155844155844, + "grad_norm": 2.039765477596474, + "learning_rate": 1.6236061810686488e-07, + "loss": 0.0144, + "step": 7272 + }, + { + "epoch": 9.445454545454545, + "grad_norm": 2.808606139042277, + "learning_rate": 1.6160661733549687e-07, + "loss": 0.0054, + "step": 7273 + }, + { + "epoch": 9.446753246753246, + "grad_norm": 1.0548410818599445, + "learning_rate": 1.6085435716422848e-07, + "loss": 0.001, + "step": 7274 + }, + { + "epoch": 9.448051948051948, + "grad_norm": 5.898943511880636, + "learning_rate": 1.6010383772614658e-07, + "loss": 0.0106, + "step": 7275 + }, + { + "epoch": 9.449350649350649, + "grad_norm": 1.9676133066130315, + "learning_rate": 1.5935505915403494e-07, + "loss": 0.0019, + "step": 7276 + }, + { + "epoch": 9.45064935064935, + "grad_norm": 1.0749651730097496, + "learning_rate": 1.5860802158036314e-07, + "loss": 0.001, + "step": 7277 + }, + { + "epoch": 9.451948051948053, + "grad_norm": 1.7761219224339555, + "learning_rate": 1.5786272513729994e-07, + "loss": 0.0087, + "step": 7278 + }, + { + "epoch": 9.453246753246754, + "grad_norm": 10.323936275479745, + "learning_rate": 1.5711916995670095e-07, + "loss": 0.0163, + "step": 7279 + }, + { + "epoch": 9.454545454545455, + "grad_norm": 0.7308844412371612, + "learning_rate": 1.563773561701165e-07, + "loss": 0.0022, + "step": 7280 + }, + { + "epoch": 9.455844155844156, + "grad_norm": 4.800039719575001, + "learning_rate": 1.5563728390878496e-07, + "loss": 0.0042, + "step": 7281 + }, + { + "epoch": 9.457142857142857, + "grad_norm": 0.47966893069060346, + "learning_rate": 1.5489895330364268e-07, + "loss": 0.0008, + "step": 7282 + }, + { + "epoch": 9.458441558441558, + "grad_norm": 1.3400067951482388, + "learning_rate": 1.5416236448531075e-07, + "loss": 0.0017, + "step": 7283 + }, + { + "epoch": 9.45974025974026, + "grad_norm": 3.751893735962381, + "learning_rate": 1.5342751758410935e-07, + "loss": 0.0061, + "step": 7284 + }, + { + "epoch": 9.46103896103896, + "grad_norm": 2.6674680175932113, + "learning_rate": 1.5269441273004448e-07, + "loss": 0.0071, + "step": 7285 + }, + { + "epoch": 9.462337662337662, + "grad_norm": 0.7355929777273058, + "learning_rate": 1.519630500528191e-07, + "loss": 0.0024, + "step": 7286 + }, + { + "epoch": 9.463636363636363, + "grad_norm": 1.2017275303953128, + "learning_rate": 1.5123342968182297e-07, + "loss": 0.0111, + "step": 7287 + }, + { + "epoch": 9.464935064935064, + "grad_norm": 0.43569521958101876, + "learning_rate": 1.5050555174613957e-07, + "loss": 0.0009, + "step": 7288 + }, + { + "epoch": 9.466233766233767, + "grad_norm": 0.7058202962622219, + "learning_rate": 1.4977941637454585e-07, + "loss": 0.0032, + "step": 7289 + }, + { + "epoch": 9.467532467532468, + "grad_norm": 0.6441807580800017, + "learning_rate": 1.4905502369551017e-07, + "loss": 0.0009, + "step": 7290 + }, + { + "epoch": 9.46883116883117, + "grad_norm": 1.4118195726265903, + "learning_rate": 1.483323738371878e-07, + "loss": 0.008, + "step": 7291 + }, + { + "epoch": 9.47012987012987, + "grad_norm": 1.0337855529327766, + "learning_rate": 1.47611466927432e-07, + "loss": 0.0149, + "step": 7292 + }, + { + "epoch": 9.471428571428572, + "grad_norm": 1.2417003992774598, + "learning_rate": 1.468923030937841e-07, + "loss": 0.0094, + "step": 7293 + }, + { + "epoch": 9.472727272727273, + "grad_norm": 1.0077125911236346, + "learning_rate": 1.4617488246348012e-07, + "loss": 0.0011, + "step": 7294 + }, + { + "epoch": 9.474025974025974, + "grad_norm": 0.36148744566341484, + "learning_rate": 1.454592051634407e-07, + "loss": 0.0015, + "step": 7295 + }, + { + "epoch": 9.475324675324675, + "grad_norm": 0.8125232679338027, + "learning_rate": 1.4474527132028793e-07, + "loss": 0.0073, + "step": 7296 + }, + { + "epoch": 9.476623376623376, + "grad_norm": 0.8152572709890465, + "learning_rate": 1.4403308106032855e-07, + "loss": 0.0012, + "step": 7297 + }, + { + "epoch": 9.477922077922077, + "grad_norm": 3.0453474276508543, + "learning_rate": 1.4332263450956064e-07, + "loss": 0.005, + "step": 7298 + }, + { + "epoch": 9.479220779220778, + "grad_norm": 1.1431289135064668, + "learning_rate": 1.426139317936781e-07, + "loss": 0.0011, + "step": 7299 + }, + { + "epoch": 9.480519480519481, + "grad_norm": 2.4230221422000198, + "learning_rate": 1.41906973038064e-07, + "loss": 0.0126, + "step": 7300 + }, + { + "epoch": 9.481818181818182, + "grad_norm": 5.989546732111273, + "learning_rate": 1.412017583677927e-07, + "loss": 0.0094, + "step": 7301 + }, + { + "epoch": 9.483116883116883, + "grad_norm": 1.1185767993260156, + "learning_rate": 1.4049828790762775e-07, + "loss": 0.0082, + "step": 7302 + }, + { + "epoch": 9.484415584415585, + "grad_norm": 1.9551576715497914, + "learning_rate": 1.3979656178203073e-07, + "loss": 0.0024, + "step": 7303 + }, + { + "epoch": 9.485714285714286, + "grad_norm": 1.3340922401200044, + "learning_rate": 1.390965801151467e-07, + "loss": 0.0012, + "step": 7304 + }, + { + "epoch": 9.487012987012987, + "grad_norm": 1.5030862396933278, + "learning_rate": 1.383983430308189e-07, + "loss": 0.0056, + "step": 7305 + }, + { + "epoch": 9.488311688311688, + "grad_norm": 0.5644890508772228, + "learning_rate": 1.3770185065257513e-07, + "loss": 0.0011, + "step": 7306 + }, + { + "epoch": 9.48961038961039, + "grad_norm": 1.0106277284876024, + "learning_rate": 1.370071031036424e-07, + "loss": 0.0073, + "step": 7307 + }, + { + "epoch": 9.49090909090909, + "grad_norm": 1.1891681570878394, + "learning_rate": 1.363141005069324e-07, + "loss": 0.0013, + "step": 7308 + }, + { + "epoch": 9.492207792207791, + "grad_norm": 2.556237925824387, + "learning_rate": 1.3562284298505035e-07, + "loss": 0.0169, + "step": 7309 + }, + { + "epoch": 9.493506493506494, + "grad_norm": 1.7806038393534254, + "learning_rate": 1.349333306602929e-07, + "loss": 0.0034, + "step": 7310 + }, + { + "epoch": 9.494805194805195, + "grad_norm": 1.567137100982009, + "learning_rate": 1.3424556365465024e-07, + "loss": 0.0074, + "step": 7311 + }, + { + "epoch": 9.496103896103897, + "grad_norm": 0.9308698244198451, + "learning_rate": 1.335595420897995e-07, + "loss": 0.0069, + "step": 7312 + }, + { + "epoch": 9.497402597402598, + "grad_norm": 1.244633574208129, + "learning_rate": 1.3287526608711132e-07, + "loss": 0.0013, + "step": 7313 + }, + { + "epoch": 9.498701298701299, + "grad_norm": 0.6041318035344863, + "learning_rate": 1.3219273576764668e-07, + "loss": 0.0011, + "step": 7314 + }, + { + "epoch": 9.5, + "grad_norm": 0.36005559544726284, + "learning_rate": 1.315119512521601e-07, + "loss": 0.0031, + "step": 7315 + }, + { + "epoch": 9.501298701298701, + "grad_norm": 0.6733937801498276, + "learning_rate": 1.30832912661093e-07, + "loss": 0.0013, + "step": 7316 + }, + { + "epoch": 9.502597402597402, + "grad_norm": 0.6293152717874231, + "learning_rate": 1.301556201145826e-07, + "loss": 0.0026, + "step": 7317 + }, + { + "epoch": 9.503896103896103, + "grad_norm": 0.2477078530004973, + "learning_rate": 1.2948007373245307e-07, + "loss": 0.0009, + "step": 7318 + }, + { + "epoch": 9.505194805194805, + "grad_norm": 1.4198696172805738, + "learning_rate": 1.2880627363422215e-07, + "loss": 0.0027, + "step": 7319 + }, + { + "epoch": 9.506493506493506, + "grad_norm": 1.1418924692854513, + "learning_rate": 1.2813421993909892e-07, + "loss": 0.0019, + "step": 7320 + }, + { + "epoch": 9.507792207792209, + "grad_norm": 9.510455779695116, + "learning_rate": 1.274639127659816e-07, + "loss": 0.0114, + "step": 7321 + }, + { + "epoch": 9.50909090909091, + "grad_norm": 1.3059291441871128, + "learning_rate": 1.2679535223345972e-07, + "loss": 0.0014, + "step": 7322 + }, + { + "epoch": 9.51038961038961, + "grad_norm": 0.4754009332413694, + "learning_rate": 1.2612853845981654e-07, + "loss": 0.0009, + "step": 7323 + }, + { + "epoch": 9.511688311688312, + "grad_norm": 0.3187972545163556, + "learning_rate": 1.2546347156302098e-07, + "loss": 0.001, + "step": 7324 + }, + { + "epoch": 9.512987012987013, + "grad_norm": 2.2713025082241685, + "learning_rate": 1.248001516607389e-07, + "loss": 0.0025, + "step": 7325 + }, + { + "epoch": 9.514285714285714, + "grad_norm": 2.9271264751541772, + "learning_rate": 1.241385788703231e-07, + "loss": 0.0087, + "step": 7326 + }, + { + "epoch": 9.515584415584415, + "grad_norm": 0.7886561411018564, + "learning_rate": 1.2347875330881886e-07, + "loss": 0.0029, + "step": 7327 + }, + { + "epoch": 9.516883116883117, + "grad_norm": 0.35425501207014665, + "learning_rate": 1.2282067509296282e-07, + "loss": 0.0013, + "step": 7328 + }, + { + "epoch": 9.518181818181818, + "grad_norm": 1.4279239978812717, + "learning_rate": 1.2216434433917956e-07, + "loss": 0.0076, + "step": 7329 + }, + { + "epoch": 9.519480519480519, + "grad_norm": 0.3496125905858675, + "learning_rate": 1.2150976116358738e-07, + "loss": 0.0009, + "step": 7330 + }, + { + "epoch": 9.520779220779222, + "grad_norm": 1.4920792413189903, + "learning_rate": 1.208569256819947e-07, + "loss": 0.0076, + "step": 7331 + }, + { + "epoch": 9.522077922077923, + "grad_norm": 3.638611955345609, + "learning_rate": 1.2020583800990137e-07, + "loss": 0.0061, + "step": 7332 + }, + { + "epoch": 9.523376623376624, + "grad_norm": 2.177614254467264, + "learning_rate": 1.1955649826249638e-07, + "loss": 0.0127, + "step": 7333 + }, + { + "epoch": 9.524675324675325, + "grad_norm": 2.1695593320842876, + "learning_rate": 1.1890890655466003e-07, + "loss": 0.0143, + "step": 7334 + }, + { + "epoch": 9.525974025974026, + "grad_norm": 1.0207569974548154, + "learning_rate": 1.1826306300096401e-07, + "loss": 0.0008, + "step": 7335 + }, + { + "epoch": 9.527272727272727, + "grad_norm": 0.4059021800774691, + "learning_rate": 1.1761896771567139e-07, + "loss": 0.0012, + "step": 7336 + }, + { + "epoch": 9.528571428571428, + "grad_norm": 1.4709900029873693, + "learning_rate": 1.169766208127332e-07, + "loss": 0.002, + "step": 7337 + }, + { + "epoch": 9.52987012987013, + "grad_norm": 1.2136450736754103, + "learning_rate": 1.1633602240579522e-07, + "loss": 0.0013, + "step": 7338 + }, + { + "epoch": 9.53116883116883, + "grad_norm": 0.2832127820425306, + "learning_rate": 1.1569717260819014e-07, + "loss": 0.0009, + "step": 7339 + }, + { + "epoch": 9.532467532467532, + "grad_norm": 0.5782958883067252, + "learning_rate": 1.1506007153294197e-07, + "loss": 0.0075, + "step": 7340 + }, + { + "epoch": 9.533766233766233, + "grad_norm": 0.9748448997759809, + "learning_rate": 1.144247192927661e-07, + "loss": 0.0022, + "step": 7341 + }, + { + "epoch": 9.535064935064934, + "grad_norm": 3.438412300463025, + "learning_rate": 1.1379111600007153e-07, + "loss": 0.0025, + "step": 7342 + }, + { + "epoch": 9.536363636363637, + "grad_norm": 2.0661313167472493, + "learning_rate": 1.131592617669508e-07, + "loss": 0.0028, + "step": 7343 + }, + { + "epoch": 9.537662337662338, + "grad_norm": 0.7051044869249998, + "learning_rate": 1.1252915670519449e-07, + "loss": 0.0034, + "step": 7344 + }, + { + "epoch": 9.53896103896104, + "grad_norm": 0.5060217361693459, + "learning_rate": 1.1190080092627675e-07, + "loss": 0.0067, + "step": 7345 + }, + { + "epoch": 9.54025974025974, + "grad_norm": 0.6541427802913012, + "learning_rate": 1.1127419454136868e-07, + "loss": 0.0011, + "step": 7346 + }, + { + "epoch": 9.541558441558442, + "grad_norm": 0.6484321569056238, + "learning_rate": 1.1064933766132713e-07, + "loss": 0.001, + "step": 7347 + }, + { + "epoch": 9.542857142857143, + "grad_norm": 0.5528615251968568, + "learning_rate": 1.1002623039670257e-07, + "loss": 0.001, + "step": 7348 + }, + { + "epoch": 9.544155844155844, + "grad_norm": 0.962933562353069, + "learning_rate": 1.094048728577346e-07, + "loss": 0.0026, + "step": 7349 + }, + { + "epoch": 9.545454545454545, + "grad_norm": 3.55876616735181, + "learning_rate": 1.0878526515435195e-07, + "loss": 0.0118, + "step": 7350 + }, + { + "epoch": 9.546753246753246, + "grad_norm": 0.5883582267812725, + "learning_rate": 1.0816740739617471e-07, + "loss": 0.0028, + "step": 7351 + }, + { + "epoch": 9.548051948051947, + "grad_norm": 4.498086957629556, + "learning_rate": 1.0755129969251543e-07, + "loss": 0.0099, + "step": 7352 + }, + { + "epoch": 9.54935064935065, + "grad_norm": 2.312057241879436, + "learning_rate": 1.0693694215237583e-07, + "loss": 0.0088, + "step": 7353 + }, + { + "epoch": 9.550649350649351, + "grad_norm": 2.8512663573677175, + "learning_rate": 1.063243348844456e-07, + "loss": 0.0095, + "step": 7354 + }, + { + "epoch": 9.551948051948052, + "grad_norm": 1.3310001958417872, + "learning_rate": 1.0571347799710696e-07, + "loss": 0.0016, + "step": 7355 + }, + { + "epoch": 9.553246753246754, + "grad_norm": 1.2106196654728525, + "learning_rate": 1.0510437159843346e-07, + "loss": 0.0078, + "step": 7356 + }, + { + "epoch": 9.554545454545455, + "grad_norm": 0.8978434782176523, + "learning_rate": 1.0449701579618665e-07, + "loss": 0.0011, + "step": 7357 + }, + { + "epoch": 9.555844155844156, + "grad_norm": 2.4180893697744734, + "learning_rate": 1.038914106978195e-07, + "loss": 0.0024, + "step": 7358 + }, + { + "epoch": 9.557142857142857, + "grad_norm": 1.7255058399971488, + "learning_rate": 1.0328755641047517e-07, + "loss": 0.0017, + "step": 7359 + }, + { + "epoch": 9.558441558441558, + "grad_norm": 5.639733868828318, + "learning_rate": 1.026854530409882e-07, + "loss": 0.0046, + "step": 7360 + }, + { + "epoch": 9.55974025974026, + "grad_norm": 0.9943962856879099, + "learning_rate": 1.0208510069588007e-07, + "loss": 0.0014, + "step": 7361 + }, + { + "epoch": 9.56103896103896, + "grad_norm": 1.560311581458851, + "learning_rate": 1.014864994813669e-07, + "loss": 0.0017, + "step": 7362 + }, + { + "epoch": 9.562337662337661, + "grad_norm": 0.3227119080629072, + "learning_rate": 1.0088964950335067e-07, + "loss": 0.0031, + "step": 7363 + }, + { + "epoch": 9.563636363636364, + "grad_norm": 6.9922759029829855, + "learning_rate": 1.002945508674269e-07, + "loss": 0.011, + "step": 7364 + }, + { + "epoch": 9.564935064935066, + "grad_norm": 0.5145569431491823, + "learning_rate": 9.970120367887914e-08, + "loss": 0.0012, + "step": 7365 + }, + { + "epoch": 9.566233766233767, + "grad_norm": 0.3371025399873421, + "learning_rate": 9.910960804268232e-08, + "loss": 0.0011, + "step": 7366 + }, + { + "epoch": 9.567532467532468, + "grad_norm": 0.8382940622968573, + "learning_rate": 9.85197640635005e-08, + "loss": 0.0068, + "step": 7367 + }, + { + "epoch": 9.568831168831169, + "grad_norm": 1.442620725536722, + "learning_rate": 9.793167184568908e-08, + "loss": 0.0155, + "step": 7368 + }, + { + "epoch": 9.57012987012987, + "grad_norm": 0.25730807058653204, + "learning_rate": 9.734533149329261e-08, + "loss": 0.001, + "step": 7369 + }, + { + "epoch": 9.571428571428571, + "grad_norm": 0.4308141099993008, + "learning_rate": 9.676074311004591e-08, + "loss": 0.0013, + "step": 7370 + }, + { + "epoch": 9.572727272727272, + "grad_norm": 0.43302416922705345, + "learning_rate": 9.617790679937288e-08, + "loss": 0.0011, + "step": 7371 + }, + { + "epoch": 9.574025974025973, + "grad_norm": 1.609210501530052, + "learning_rate": 9.559682266439108e-08, + "loss": 0.0051, + "step": 7372 + }, + { + "epoch": 9.575324675324675, + "grad_norm": 0.47382768270670805, + "learning_rate": 9.501749080790269e-08, + "loss": 0.0014, + "step": 7373 + }, + { + "epoch": 9.576623376623377, + "grad_norm": 0.8072004928351698, + "learning_rate": 9.44399113324035e-08, + "loss": 0.0014, + "step": 7374 + }, + { + "epoch": 9.577922077922079, + "grad_norm": 3.332783807303196, + "learning_rate": 9.386408434007843e-08, + "loss": 0.0189, + "step": 7375 + }, + { + "epoch": 9.57922077922078, + "grad_norm": 0.45307172798987244, + "learning_rate": 9.329000993280269e-08, + "loss": 0.0008, + "step": 7376 + }, + { + "epoch": 9.58051948051948, + "grad_norm": 0.2959535702965184, + "learning_rate": 9.271768821213944e-08, + "loss": 0.001, + "step": 7377 + }, + { + "epoch": 9.581818181818182, + "grad_norm": 0.8433273806591187, + "learning_rate": 9.21471192793455e-08, + "loss": 0.0019, + "step": 7378 + }, + { + "epoch": 9.583116883116883, + "grad_norm": 0.8104204780119116, + "learning_rate": 9.157830323536566e-08, + "loss": 0.0013, + "step": 7379 + }, + { + "epoch": 9.584415584415584, + "grad_norm": 0.6117514856226864, + "learning_rate": 9.101124018083051e-08, + "loss": 0.0012, + "step": 7380 + }, + { + "epoch": 9.585714285714285, + "grad_norm": 6.928882898770837, + "learning_rate": 9.044593021606873e-08, + "loss": 0.0024, + "step": 7381 + }, + { + "epoch": 9.587012987012987, + "grad_norm": 1.4644725233731446, + "learning_rate": 8.988237344109251e-08, + "loss": 0.0071, + "step": 7382 + }, + { + "epoch": 9.588311688311688, + "grad_norm": 1.963818358527185, + "learning_rate": 8.932056995560656e-08, + "loss": 0.0117, + "step": 7383 + }, + { + "epoch": 9.589610389610389, + "grad_norm": 6.676159299113281, + "learning_rate": 8.876051985900359e-08, + "loss": 0.0139, + "step": 7384 + }, + { + "epoch": 9.590909090909092, + "grad_norm": 4.461286623384958, + "learning_rate": 8.820222325036765e-08, + "loss": 0.0052, + "step": 7385 + }, + { + "epoch": 9.592207792207793, + "grad_norm": 1.1283493750147657, + "learning_rate": 8.764568022847198e-08, + "loss": 0.0082, + "step": 7386 + }, + { + "epoch": 9.593506493506494, + "grad_norm": 3.1639083976498106, + "learning_rate": 8.709089089178225e-08, + "loss": 0.0025, + "step": 7387 + }, + { + "epoch": 9.594805194805195, + "grad_norm": 0.7227973340127124, + "learning_rate": 8.65378553384466e-08, + "loss": 0.0064, + "step": 7388 + }, + { + "epoch": 9.596103896103896, + "grad_norm": 0.5616434029728734, + "learning_rate": 8.598657366631124e-08, + "loss": 0.001, + "step": 7389 + }, + { + "epoch": 9.597402597402597, + "grad_norm": 6.701532254476568, + "learning_rate": 8.543704597290592e-08, + "loss": 0.0058, + "step": 7390 + }, + { + "epoch": 9.598701298701299, + "grad_norm": 0.3531647594893516, + "learning_rate": 8.488927235545508e-08, + "loss": 0.001, + "step": 7391 + }, + { + "epoch": 9.6, + "grad_norm": 3.6931665995292575, + "learning_rate": 8.434325291086898e-08, + "loss": 0.0146, + "step": 7392 + }, + { + "epoch": 9.6012987012987, + "grad_norm": 2.400462969073021, + "learning_rate": 8.379898773574924e-08, + "loss": 0.009, + "step": 7393 + }, + { + "epoch": 9.602597402597402, + "grad_norm": 0.7924344411327113, + "learning_rate": 8.325647692638661e-08, + "loss": 0.0011, + "step": 7394 + }, + { + "epoch": 9.603896103896105, + "grad_norm": 2.3670284885564006, + "learning_rate": 8.271572057876098e-08, + "loss": 0.0089, + "step": 7395 + }, + { + "epoch": 9.605194805194806, + "grad_norm": 0.9215448923280404, + "learning_rate": 8.21767187885436e-08, + "loss": 0.0012, + "step": 7396 + }, + { + "epoch": 9.606493506493507, + "grad_norm": 1.0675775134094558, + "learning_rate": 8.163947165109487e-08, + "loss": 0.0089, + "step": 7397 + }, + { + "epoch": 9.607792207792208, + "grad_norm": 1.6898231918763345, + "learning_rate": 8.110397926146208e-08, + "loss": 0.0125, + "step": 7398 + }, + { + "epoch": 9.60909090909091, + "grad_norm": 1.5963707579789503, + "learning_rate": 8.0570241714385e-08, + "loss": 0.0015, + "step": 7399 + }, + { + "epoch": 9.61038961038961, + "grad_norm": 4.913722804363909, + "learning_rate": 8.003825910429252e-08, + "loss": 0.0136, + "step": 7400 + }, + { + "epoch": 9.611688311688312, + "grad_norm": 0.9409588274892334, + "learning_rate": 7.950803152530162e-08, + "loss": 0.0075, + "step": 7401 + }, + { + "epoch": 9.612987012987013, + "grad_norm": 0.7604164206696489, + "learning_rate": 7.897955907122057e-08, + "loss": 0.001, + "step": 7402 + }, + { + "epoch": 9.614285714285714, + "grad_norm": 2.487755854914751, + "learning_rate": 7.845284183554569e-08, + "loss": 0.0073, + "step": 7403 + }, + { + "epoch": 9.615584415584415, + "grad_norm": 2.6846444565167076, + "learning_rate": 7.792787991146356e-08, + "loss": 0.0072, + "step": 7404 + }, + { + "epoch": 9.616883116883116, + "grad_norm": 1.4352225576509277, + "learning_rate": 7.740467339184987e-08, + "loss": 0.0085, + "step": 7405 + }, + { + "epoch": 9.618181818181817, + "grad_norm": 3.4336728890983435, + "learning_rate": 7.688322236927059e-08, + "loss": 0.0026, + "step": 7406 + }, + { + "epoch": 9.61948051948052, + "grad_norm": 0.33551546800254384, + "learning_rate": 7.636352693597859e-08, + "loss": 0.0009, + "step": 7407 + }, + { + "epoch": 9.620779220779221, + "grad_norm": 1.1294214979479786, + "learning_rate": 7.584558718391922e-08, + "loss": 0.0017, + "step": 7408 + }, + { + "epoch": 9.622077922077922, + "grad_norm": 0.3017746728605583, + "learning_rate": 7.532940320472693e-08, + "loss": 0.0008, + "step": 7409 + }, + { + "epoch": 9.623376623376624, + "grad_norm": 3.4004839886143112, + "learning_rate": 7.481497508972313e-08, + "loss": 0.0034, + "step": 7410 + }, + { + "epoch": 9.624675324675325, + "grad_norm": 0.9282443167371983, + "learning_rate": 7.430230292991947e-08, + "loss": 0.0014, + "step": 7411 + }, + { + "epoch": 9.625974025974026, + "grad_norm": 3.8155769625236804, + "learning_rate": 7.379138681601894e-08, + "loss": 0.003, + "step": 7412 + }, + { + "epoch": 9.627272727272727, + "grad_norm": 11.713839824016697, + "learning_rate": 7.328222683841147e-08, + "loss": 0.0127, + "step": 7413 + }, + { + "epoch": 9.628571428571428, + "grad_norm": 0.23636249687932542, + "learning_rate": 7.277482308717832e-08, + "loss": 0.0011, + "step": 7414 + }, + { + "epoch": 9.62987012987013, + "grad_norm": 0.4848673361529798, + "learning_rate": 7.226917565208658e-08, + "loss": 0.0014, + "step": 7415 + }, + { + "epoch": 9.63116883116883, + "grad_norm": 2.921520389206663, + "learning_rate": 7.176528462259802e-08, + "loss": 0.0118, + "step": 7416 + }, + { + "epoch": 9.632467532467533, + "grad_norm": 3.773570825336545, + "learning_rate": 7.1263150087858e-08, + "loss": 0.0028, + "step": 7417 + }, + { + "epoch": 9.633766233766234, + "grad_norm": 3.9335043308509685, + "learning_rate": 7.076277213670546e-08, + "loss": 0.0098, + "step": 7418 + }, + { + "epoch": 9.635064935064936, + "grad_norm": 0.5387633629732395, + "learning_rate": 7.026415085766513e-08, + "loss": 0.0011, + "step": 7419 + }, + { + "epoch": 9.636363636363637, + "grad_norm": 0.7265494983973654, + "learning_rate": 6.976728633895536e-08, + "loss": 0.0018, + "step": 7420 + }, + { + "epoch": 9.637662337662338, + "grad_norm": 0.8081076571548022, + "learning_rate": 6.92721786684769e-08, + "loss": 0.0082, + "step": 7421 + }, + { + "epoch": 9.638961038961039, + "grad_norm": 5.526618116111052, + "learning_rate": 6.877882793382862e-08, + "loss": 0.0266, + "step": 7422 + }, + { + "epoch": 9.64025974025974, + "grad_norm": 3.7033506520069297, + "learning_rate": 6.828723422228955e-08, + "loss": 0.0101, + "step": 7423 + }, + { + "epoch": 9.641558441558441, + "grad_norm": 0.5228578264012287, + "learning_rate": 6.779739762083459e-08, + "loss": 0.0018, + "step": 7424 + }, + { + "epoch": 9.642857142857142, + "grad_norm": 0.2162615238766758, + "learning_rate": 6.730931821612441e-08, + "loss": 0.001, + "step": 7425 + }, + { + "epoch": 9.644155844155843, + "grad_norm": 3.1508781684312357, + "learning_rate": 6.682299609450993e-08, + "loss": 0.0157, + "step": 7426 + }, + { + "epoch": 9.645454545454545, + "grad_norm": 0.7506885446749001, + "learning_rate": 6.6338431342029e-08, + "loss": 0.0014, + "step": 7427 + }, + { + "epoch": 9.646753246753248, + "grad_norm": 1.979515814860476, + "learning_rate": 6.585562404441304e-08, + "loss": 0.0162, + "step": 7428 + }, + { + "epoch": 9.648051948051949, + "grad_norm": 2.7108271741069756, + "learning_rate": 6.537457428707816e-08, + "loss": 0.015, + "step": 7429 + }, + { + "epoch": 9.64935064935065, + "grad_norm": 1.1426278904037765, + "learning_rate": 6.489528215513296e-08, + "loss": 0.0022, + "step": 7430 + }, + { + "epoch": 9.650649350649351, + "grad_norm": 4.737751071858586, + "learning_rate": 6.441774773337184e-08, + "loss": 0.0041, + "step": 7431 + }, + { + "epoch": 9.651948051948052, + "grad_norm": 0.7311424719538008, + "learning_rate": 6.394197110627943e-08, + "loss": 0.0016, + "step": 7432 + }, + { + "epoch": 9.653246753246753, + "grad_norm": 3.0219215306429383, + "learning_rate": 6.34679523580306e-08, + "loss": 0.0054, + "step": 7433 + }, + { + "epoch": 9.654545454545454, + "grad_norm": 1.1776643492749068, + "learning_rate": 6.299569157248942e-08, + "loss": 0.0026, + "step": 7434 + }, + { + "epoch": 9.655844155844155, + "grad_norm": 1.3001025368173542, + "learning_rate": 6.252518883320568e-08, + "loss": 0.0014, + "step": 7435 + }, + { + "epoch": 9.657142857142857, + "grad_norm": 1.995192789448526, + "learning_rate": 6.20564442234206e-08, + "loss": 0.0022, + "step": 7436 + }, + { + "epoch": 9.658441558441558, + "grad_norm": 0.3797167266852697, + "learning_rate": 6.158945782606674e-08, + "loss": 0.0014, + "step": 7437 + }, + { + "epoch": 9.65974025974026, + "grad_norm": 0.8292382805421411, + "learning_rate": 6.11242297237602e-08, + "loss": 0.0011, + "step": 7438 + }, + { + "epoch": 9.661038961038962, + "grad_norm": 1.550908884665677, + "learning_rate": 6.06607599988085e-08, + "loss": 0.0072, + "step": 7439 + }, + { + "epoch": 9.662337662337663, + "grad_norm": 0.8431279095022013, + "learning_rate": 6.019904873321047e-08, + "loss": 0.0011, + "step": 7440 + }, + { + "epoch": 9.663636363636364, + "grad_norm": 0.9741455986328289, + "learning_rate": 5.973909600864968e-08, + "loss": 0.0011, + "step": 7441 + }, + { + "epoch": 9.664935064935065, + "grad_norm": 1.441469980143382, + "learning_rate": 5.9280901906502156e-08, + "loss": 0.003, + "step": 7442 + }, + { + "epoch": 9.666233766233766, + "grad_norm": 1.8554843201058961, + "learning_rate": 5.88244665078308e-08, + "loss": 0.0021, + "step": 7443 + }, + { + "epoch": 9.667532467532467, + "grad_norm": 0.7509805318288919, + "learning_rate": 5.8369789893386595e-08, + "loss": 0.0073, + "step": 7444 + }, + { + "epoch": 9.668831168831169, + "grad_norm": 1.6435961448732679, + "learning_rate": 5.7916872143611856e-08, + "loss": 0.0081, + "step": 7445 + }, + { + "epoch": 9.67012987012987, + "grad_norm": 2.231769698190018, + "learning_rate": 5.746571333863582e-08, + "loss": 0.0022, + "step": 7446 + }, + { + "epoch": 9.67142857142857, + "grad_norm": 1.6172857343639653, + "learning_rate": 5.701631355827797e-08, + "loss": 0.0142, + "step": 7447 + }, + { + "epoch": 9.672727272727272, + "grad_norm": 2.5065118685399677, + "learning_rate": 5.6568672882043595e-08, + "loss": 0.0114, + "step": 7448 + }, + { + "epoch": 9.674025974025973, + "grad_norm": 0.8711973554529616, + "learning_rate": 5.612279138913157e-08, + "loss": 0.0047, + "step": 7449 + }, + { + "epoch": 9.675324675324676, + "grad_norm": 6.109636355907629, + "learning_rate": 5.5678669158424346e-08, + "loss": 0.0129, + "step": 7450 + }, + { + "epoch": 9.676623376623377, + "grad_norm": 0.4869893237797513, + "learning_rate": 5.523630626849796e-08, + "loss": 0.0009, + "step": 7451 + }, + { + "epoch": 9.677922077922078, + "grad_norm": 0.331401125559993, + "learning_rate": 5.479570279761315e-08, + "loss": 0.001, + "step": 7452 + }, + { + "epoch": 9.67922077922078, + "grad_norm": 5.457489219811562, + "learning_rate": 5.4356858823721994e-08, + "loss": 0.0086, + "step": 7453 + }, + { + "epoch": 9.68051948051948, + "grad_norm": 0.9668781258787675, + "learning_rate": 5.391977442446461e-08, + "loss": 0.0013, + "step": 7454 + }, + { + "epoch": 9.681818181818182, + "grad_norm": 1.2976342173629398, + "learning_rate": 5.348444967716804e-08, + "loss": 0.0019, + "step": 7455 + }, + { + "epoch": 9.683116883116883, + "grad_norm": 0.8640028367743259, + "learning_rate": 5.305088465885067e-08, + "loss": 0.008, + "step": 7456 + }, + { + "epoch": 9.684415584415584, + "grad_norm": 1.9222847177079427, + "learning_rate": 5.2619079446220025e-08, + "loss": 0.0072, + "step": 7457 + }, + { + "epoch": 9.685714285714285, + "grad_norm": 3.1197085872433274, + "learning_rate": 5.2189034115668334e-08, + "loss": 0.0029, + "step": 7458 + }, + { + "epoch": 9.687012987012986, + "grad_norm": 1.8777650694765942, + "learning_rate": 5.176074874327919e-08, + "loss": 0.0091, + "step": 7459 + }, + { + "epoch": 9.688311688311689, + "grad_norm": 1.6898184248177377, + "learning_rate": 5.1334223404824193e-08, + "loss": 0.0082, + "step": 7460 + }, + { + "epoch": 9.68961038961039, + "grad_norm": 0.28319881346962894, + "learning_rate": 5.0909458175766315e-08, + "loss": 0.001, + "step": 7461 + }, + { + "epoch": 9.690909090909091, + "grad_norm": 1.607491907523123, + "learning_rate": 5.048645313125211e-08, + "loss": 0.0012, + "step": 7462 + }, + { + "epoch": 9.692207792207792, + "grad_norm": 3.94851796048626, + "learning_rate": 5.006520834612061e-08, + "loss": 0.0027, + "step": 7463 + }, + { + "epoch": 9.693506493506494, + "grad_norm": 0.5488277116682871, + "learning_rate": 4.9645723894896636e-08, + "loss": 0.0011, + "step": 7464 + }, + { + "epoch": 9.694805194805195, + "grad_norm": 1.7364320296704572, + "learning_rate": 4.922799985179638e-08, + "loss": 0.0065, + "step": 7465 + }, + { + "epoch": 9.696103896103896, + "grad_norm": 2.0900143949965044, + "learning_rate": 4.8812036290722955e-08, + "loss": 0.016, + "step": 7466 + }, + { + "epoch": 9.697402597402597, + "grad_norm": 5.720355937124311, + "learning_rate": 4.839783328526748e-08, + "loss": 0.0052, + "step": 7467 + }, + { + "epoch": 9.698701298701298, + "grad_norm": 2.1653653939621247, + "learning_rate": 4.7985390908711346e-08, + "loss": 0.0013, + "step": 7468 + }, + { + "epoch": 9.7, + "grad_norm": 0.2775098616341244, + "learning_rate": 4.757470923402285e-08, + "loss": 0.001, + "step": 7469 + }, + { + "epoch": 9.7012987012987, + "grad_norm": 3.888938396304779, + "learning_rate": 4.716578833386054e-08, + "loss": 0.012, + "step": 7470 + }, + { + "epoch": 9.702597402597403, + "grad_norm": 1.2043352648801047, + "learning_rate": 4.675862828056876e-08, + "loss": 0.0012, + "step": 7471 + }, + { + "epoch": 9.703896103896104, + "grad_norm": 12.145910356247851, + "learning_rate": 4.635322914618323e-08, + "loss": 0.0084, + "step": 7472 + }, + { + "epoch": 9.705194805194806, + "grad_norm": 2.0018701745262617, + "learning_rate": 4.594959100242546e-08, + "loss": 0.0017, + "step": 7473 + }, + { + "epoch": 9.706493506493507, + "grad_norm": 7.503525844662557, + "learning_rate": 4.554771392070834e-08, + "loss": 0.0066, + "step": 7474 + }, + { + "epoch": 9.707792207792208, + "grad_norm": 1.212157207520829, + "learning_rate": 4.514759797213053e-08, + "loss": 0.0077, + "step": 7475 + }, + { + "epoch": 9.709090909090909, + "grad_norm": 0.21445798339906091, + "learning_rate": 4.4749243227480975e-08, + "loss": 0.0009, + "step": 7476 + }, + { + "epoch": 9.71038961038961, + "grad_norm": 1.198330948969595, + "learning_rate": 4.435264975723552e-08, + "loss": 0.0079, + "step": 7477 + }, + { + "epoch": 9.711688311688311, + "grad_norm": 0.6543545110800103, + "learning_rate": 4.395781763156026e-08, + "loss": 0.0014, + "step": 7478 + }, + { + "epoch": 9.712987012987012, + "grad_norm": 0.7923819601901944, + "learning_rate": 4.356474692030599e-08, + "loss": 0.0012, + "step": 7479 + }, + { + "epoch": 9.714285714285714, + "grad_norm": 1.6516571039900865, + "learning_rate": 4.317343769301818e-08, + "loss": 0.0266, + "step": 7480 + }, + { + "epoch": 9.715584415584416, + "grad_norm": 0.7144540587428299, + "learning_rate": 4.278389001892369e-08, + "loss": 0.0019, + "step": 7481 + }, + { + "epoch": 9.716883116883118, + "grad_norm": 1.6818725632646834, + "learning_rate": 4.2396103966941824e-08, + "loss": 0.009, + "step": 7482 + }, + { + "epoch": 9.718181818181819, + "grad_norm": 1.0848985333476397, + "learning_rate": 4.2010079605679934e-08, + "loss": 0.0086, + "step": 7483 + }, + { + "epoch": 9.71948051948052, + "grad_norm": 0.23957856536850525, + "learning_rate": 4.162581700343449e-08, + "loss": 0.0011, + "step": 7484 + }, + { + "epoch": 9.720779220779221, + "grad_norm": 0.5973875548712735, + "learning_rate": 4.1243316228185556e-08, + "loss": 0.0012, + "step": 7485 + }, + { + "epoch": 9.722077922077922, + "grad_norm": 8.041938252056552, + "learning_rate": 4.08625773476079e-08, + "loss": 0.0121, + "step": 7486 + }, + { + "epoch": 9.723376623376623, + "grad_norm": 0.9177632977517695, + "learning_rate": 4.048360042905985e-08, + "loss": 0.007, + "step": 7487 + }, + { + "epoch": 9.724675324675324, + "grad_norm": 1.490488766109946, + "learning_rate": 4.0106385539590006e-08, + "loss": 0.0089, + "step": 7488 + }, + { + "epoch": 9.725974025974025, + "grad_norm": 0.4376270818729866, + "learning_rate": 3.973093274593387e-08, + "loss": 0.0011, + "step": 7489 + }, + { + "epoch": 9.727272727272727, + "grad_norm": 1.2724334790017868, + "learning_rate": 3.93572421145183e-08, + "loss": 0.0015, + "step": 7490 + }, + { + "epoch": 9.728571428571428, + "grad_norm": 0.7186889082172387, + "learning_rate": 3.898531371145597e-08, + "loss": 0.0018, + "step": 7491 + }, + { + "epoch": 9.72987012987013, + "grad_norm": 2.728044210276041, + "learning_rate": 3.861514760254648e-08, + "loss": 0.0043, + "step": 7492 + }, + { + "epoch": 9.731168831168832, + "grad_norm": 5.118822890089458, + "learning_rate": 3.8246743853280754e-08, + "loss": 0.0174, + "step": 7493 + }, + { + "epoch": 9.732467532467533, + "grad_norm": 0.38523335733734787, + "learning_rate": 3.788010252883556e-08, + "loss": 0.0028, + "step": 7494 + }, + { + "epoch": 9.733766233766234, + "grad_norm": 1.5050451699789669, + "learning_rate": 3.75152236940779e-08, + "loss": 0.0017, + "step": 7495 + }, + { + "epoch": 9.735064935064935, + "grad_norm": 15.165450028304793, + "learning_rate": 3.7152107413560566e-08, + "loss": 0.0084, + "step": 7496 + }, + { + "epoch": 9.736363636363636, + "grad_norm": 0.833160933778754, + "learning_rate": 3.679075375152663e-08, + "loss": 0.0016, + "step": 7497 + }, + { + "epoch": 9.737662337662337, + "grad_norm": 0.2640181103304927, + "learning_rate": 3.643116277190606e-08, + "loss": 0.0009, + "step": 7498 + }, + { + "epoch": 9.738961038961039, + "grad_norm": 3.9517216029897093, + "learning_rate": 3.607333453831796e-08, + "loss": 0.0118, + "step": 7499 + }, + { + "epoch": 9.74025974025974, + "grad_norm": 1.342246271482849, + "learning_rate": 3.571726911406836e-08, + "loss": 0.0071, + "step": 7500 + }, + { + "epoch": 9.74155844155844, + "grad_norm": 0.7342933470246346, + "learning_rate": 3.536296656215132e-08, + "loss": 0.0017, + "step": 7501 + }, + { + "epoch": 9.742857142857144, + "grad_norm": 0.691168040208204, + "learning_rate": 3.501042694525225e-08, + "loss": 0.0012, + "step": 7502 + }, + { + "epoch": 9.744155844155845, + "grad_norm": 12.416128436899399, + "learning_rate": 3.465965032574015e-08, + "loss": 0.0234, + "step": 7503 + }, + { + "epoch": 9.745454545454546, + "grad_norm": 1.163889746595954, + "learning_rate": 3.431063676567425e-08, + "loss": 0.0015, + "step": 7504 + }, + { + "epoch": 9.746753246753247, + "grad_norm": 8.697199484755126, + "learning_rate": 3.396338632680185e-08, + "loss": 0.0059, + "step": 7505 + }, + { + "epoch": 9.748051948051948, + "grad_norm": 2.262180292769642, + "learning_rate": 3.3617899070558236e-08, + "loss": 0.0037, + "step": 7506 + }, + { + "epoch": 9.74935064935065, + "grad_norm": 0.3804105871175758, + "learning_rate": 3.327417505806785e-08, + "loss": 0.0013, + "step": 7507 + }, + { + "epoch": 9.75064935064935, + "grad_norm": 0.9942284722346013, + "learning_rate": 3.2932214350139825e-08, + "loss": 0.0014, + "step": 7508 + }, + { + "epoch": 9.751948051948052, + "grad_norm": 0.34590783208046294, + "learning_rate": 3.259201700727577e-08, + "loss": 0.0009, + "step": 7509 + }, + { + "epoch": 9.753246753246753, + "grad_norm": 1.3601007723035672, + "learning_rate": 3.225358308966198e-08, + "loss": 0.0018, + "step": 7510 + }, + { + "epoch": 9.754545454545454, + "grad_norm": 2.6133086929195084, + "learning_rate": 3.191691265717389e-08, + "loss": 0.0039, + "step": 7511 + }, + { + "epoch": 9.755844155844155, + "grad_norm": 1.0981946572534567, + "learning_rate": 3.158200576937498e-08, + "loss": 0.0012, + "step": 7512 + }, + { + "epoch": 9.757142857142856, + "grad_norm": 0.49584455369195035, + "learning_rate": 3.124886248551673e-08, + "loss": 0.0009, + "step": 7513 + }, + { + "epoch": 9.75844155844156, + "grad_norm": 1.168178965810171, + "learning_rate": 3.091748286453866e-08, + "loss": 0.0059, + "step": 7514 + }, + { + "epoch": 9.75974025974026, + "grad_norm": 4.10486264469073, + "learning_rate": 3.0587866965068324e-08, + "loss": 0.0131, + "step": 7515 + }, + { + "epoch": 9.761038961038961, + "grad_norm": 5.846119330562018, + "learning_rate": 3.0260014845421294e-08, + "loss": 0.011, + "step": 7516 + }, + { + "epoch": 9.762337662337663, + "grad_norm": 0.4785055937227914, + "learning_rate": 2.993392656360006e-08, + "loss": 0.0011, + "step": 7517 + }, + { + "epoch": 9.763636363636364, + "grad_norm": 0.624617253892039, + "learning_rate": 2.960960217729625e-08, + "loss": 0.0019, + "step": 7518 + }, + { + "epoch": 9.764935064935065, + "grad_norm": 7.0764496881092365, + "learning_rate": 2.9287041743889523e-08, + "loss": 0.0102, + "step": 7519 + }, + { + "epoch": 9.766233766233766, + "grad_norm": 0.7316848129160244, + "learning_rate": 2.896624532044534e-08, + "loss": 0.0011, + "step": 7520 + }, + { + "epoch": 9.767532467532467, + "grad_norm": 1.569808204905777, + "learning_rate": 2.864721296372164e-08, + "loss": 0.0018, + "step": 7521 + }, + { + "epoch": 9.768831168831168, + "grad_norm": 5.046704733109517, + "learning_rate": 2.832994473015882e-08, + "loss": 0.0089, + "step": 7522 + }, + { + "epoch": 9.77012987012987, + "grad_norm": 8.857501198241211, + "learning_rate": 2.8014440675888653e-08, + "loss": 0.0147, + "step": 7523 + }, + { + "epoch": 9.771428571428572, + "grad_norm": 0.8958436698849783, + "learning_rate": 2.770070085672982e-08, + "loss": 0.0072, + "step": 7524 + }, + { + "epoch": 9.772727272727273, + "grad_norm": 1.2973345160579617, + "learning_rate": 2.7388725328189036e-08, + "loss": 0.0077, + "step": 7525 + }, + { + "epoch": 9.774025974025975, + "grad_norm": 0.8615155358338779, + "learning_rate": 2.7078514145459924e-08, + "loss": 0.007, + "step": 7526 + }, + { + "epoch": 9.775324675324676, + "grad_norm": 1.3958130745274384, + "learning_rate": 2.6770067363426354e-08, + "loss": 0.0081, + "step": 7527 + }, + { + "epoch": 9.776623376623377, + "grad_norm": 1.8990877728600728, + "learning_rate": 2.6463385036656906e-08, + "loss": 0.0081, + "step": 7528 + }, + { + "epoch": 9.777922077922078, + "grad_norm": 1.5903420212374382, + "learning_rate": 2.61584672194104e-08, + "loss": 0.0049, + "step": 7529 + }, + { + "epoch": 9.779220779220779, + "grad_norm": 1.391440415477073, + "learning_rate": 2.5855313965631456e-08, + "loss": 0.01, + "step": 7530 + }, + { + "epoch": 9.78051948051948, + "grad_norm": 2.681622999186809, + "learning_rate": 2.5553925328956066e-08, + "loss": 0.0338, + "step": 7531 + }, + { + "epoch": 9.781818181818181, + "grad_norm": 6.996041132273538, + "learning_rate": 2.5254301362703792e-08, + "loss": 0.0118, + "step": 7532 + }, + { + "epoch": 9.783116883116882, + "grad_norm": 7.19580466839928, + "learning_rate": 2.495644211988335e-08, + "loss": 0.0044, + "step": 7533 + }, + { + "epoch": 9.784415584415584, + "grad_norm": 1.0461886249123744, + "learning_rate": 2.4660347653193694e-08, + "loss": 0.0073, + "step": 7534 + }, + { + "epoch": 9.785714285714286, + "grad_norm": 0.6982556064515092, + "learning_rate": 2.4366018015018477e-08, + "loss": 0.0026, + "step": 7535 + }, + { + "epoch": 9.787012987012988, + "grad_norm": 1.3477029711428496, + "learning_rate": 2.407345325743049e-08, + "loss": 0.0097, + "step": 7536 + }, + { + "epoch": 9.788311688311689, + "grad_norm": 0.49560057142753505, + "learning_rate": 2.378265343218944e-08, + "loss": 0.0011, + "step": 7537 + }, + { + "epoch": 9.78961038961039, + "grad_norm": 1.006390855353987, + "learning_rate": 2.3493618590744173e-08, + "loss": 0.0013, + "step": 7538 + }, + { + "epoch": 9.790909090909091, + "grad_norm": 2.4913055903987984, + "learning_rate": 2.3206348784229337e-08, + "loss": 0.0087, + "step": 7539 + }, + { + "epoch": 9.792207792207792, + "grad_norm": 6.545229472680061, + "learning_rate": 2.292084406347095e-08, + "loss": 0.0155, + "step": 7540 + }, + { + "epoch": 9.793506493506493, + "grad_norm": 0.8229863087060024, + "learning_rate": 2.263710447897638e-08, + "loss": 0.0015, + "step": 7541 + }, + { + "epoch": 9.794805194805194, + "grad_norm": 0.35315125059295815, + "learning_rate": 2.2355130080948807e-08, + "loss": 0.0008, + "step": 7542 + }, + { + "epoch": 9.796103896103896, + "grad_norm": 0.30841827958704, + "learning_rate": 2.2074920919271658e-08, + "loss": 0.0011, + "step": 7543 + }, + { + "epoch": 9.797402597402597, + "grad_norm": 6.416942851137385, + "learning_rate": 2.1796477043520836e-08, + "loss": 0.0065, + "step": 7544 + }, + { + "epoch": 9.7987012987013, + "grad_norm": 1.728220712013137, + "learning_rate": 2.1519798502958045e-08, + "loss": 0.0114, + "step": 7545 + }, + { + "epoch": 9.8, + "grad_norm": 8.468304091313755, + "learning_rate": 2.1244885346533016e-08, + "loss": 0.021, + "step": 7546 + }, + { + "epoch": 9.801298701298702, + "grad_norm": 1.508416884735701, + "learning_rate": 2.0971737622883515e-08, + "loss": 0.002, + "step": 7547 + }, + { + "epoch": 9.802597402597403, + "grad_norm": 2.4165402141231547, + "learning_rate": 2.0700355380334215e-08, + "loss": 0.0081, + "step": 7548 + }, + { + "epoch": 9.803896103896104, + "grad_norm": 1.3040421705635514, + "learning_rate": 2.043073866689782e-08, + "loss": 0.0081, + "step": 7549 + }, + { + "epoch": 9.805194805194805, + "grad_norm": 3.0932788111912757, + "learning_rate": 2.0162887530273956e-08, + "loss": 0.0186, + "step": 7550 + }, + { + "epoch": 9.806493506493506, + "grad_norm": 1.417730784941316, + "learning_rate": 1.989680201785249e-08, + "loss": 0.0034, + "step": 7551 + }, + { + "epoch": 9.807792207792208, + "grad_norm": 1.1345090379250724, + "learning_rate": 1.963248217670688e-08, + "loss": 0.0012, + "step": 7552 + }, + { + "epoch": 9.809090909090909, + "grad_norm": 0.42923542157373196, + "learning_rate": 1.936992805360194e-08, + "loss": 0.0012, + "step": 7553 + }, + { + "epoch": 9.81038961038961, + "grad_norm": 0.4726100370041406, + "learning_rate": 1.91091396949894e-08, + "loss": 0.0028, + "step": 7554 + }, + { + "epoch": 9.811688311688311, + "grad_norm": 1.072854011855302, + "learning_rate": 1.8850117147004577e-08, + "loss": 0.0009, + "step": 7555 + }, + { + "epoch": 9.812987012987014, + "grad_norm": 2.5702916844753125, + "learning_rate": 1.8592860455476368e-08, + "loss": 0.0136, + "step": 7556 + }, + { + "epoch": 9.814285714285715, + "grad_norm": 2.6543680745964195, + "learning_rate": 1.833736966591615e-08, + "loss": 0.0218, + "step": 7557 + }, + { + "epoch": 9.815584415584416, + "grad_norm": 1.0669294231129114, + "learning_rate": 1.8083644823526645e-08, + "loss": 0.0081, + "step": 7558 + }, + { + "epoch": 9.816883116883117, + "grad_norm": 6.720950790633935, + "learning_rate": 1.78316859731964e-08, + "loss": 0.0046, + "step": 7559 + }, + { + "epoch": 9.818181818181818, + "grad_norm": 0.7929238627595117, + "learning_rate": 1.7581493159501974e-08, + "loss": 0.0015, + "step": 7560 + }, + { + "epoch": 9.81948051948052, + "grad_norm": 5.6719587238420965, + "learning_rate": 1.7333066426706845e-08, + "loss": 0.004, + "step": 7561 + }, + { + "epoch": 9.82077922077922, + "grad_norm": 0.7400467494512154, + "learning_rate": 1.708640581876253e-08, + "loss": 0.0083, + "step": 7562 + }, + { + "epoch": 9.822077922077922, + "grad_norm": 8.835198954968993, + "learning_rate": 1.6841511379308563e-08, + "loss": 0.0062, + "step": 7563 + }, + { + "epoch": 9.823376623376623, + "grad_norm": 1.3234694425806957, + "learning_rate": 1.6598383151670282e-08, + "loss": 0.0079, + "step": 7564 + }, + { + "epoch": 9.824675324675324, + "grad_norm": 3.255271562691422, + "learning_rate": 1.6357021178862176e-08, + "loss": 0.0056, + "step": 7565 + }, + { + "epoch": 9.825974025974027, + "grad_norm": 1.723382659370355, + "learning_rate": 1.611742550358564e-08, + "loss": 0.008, + "step": 7566 + }, + { + "epoch": 9.827272727272728, + "grad_norm": 1.1721548093159366, + "learning_rate": 1.5879596168231203e-08, + "loss": 0.003, + "step": 7567 + }, + { + "epoch": 9.82857142857143, + "grad_norm": 0.8724421317973289, + "learning_rate": 1.5643533214874108e-08, + "loss": 0.001, + "step": 7568 + }, + { + "epoch": 9.82987012987013, + "grad_norm": 0.6578139033689339, + "learning_rate": 1.5409236685277608e-08, + "loss": 0.0014, + "step": 7569 + }, + { + "epoch": 9.831168831168831, + "grad_norm": 0.9566499918893631, + "learning_rate": 1.517670662089521e-08, + "loss": 0.0071, + "step": 7570 + }, + { + "epoch": 9.832467532467533, + "grad_norm": 0.4776036534400758, + "learning_rate": 1.494594306286401e-08, + "loss": 0.001, + "step": 7571 + }, + { + "epoch": 9.833766233766234, + "grad_norm": 1.1357612635379903, + "learning_rate": 1.471694605201246e-08, + "loss": 0.0161, + "step": 7572 + }, + { + "epoch": 9.835064935064935, + "grad_norm": 0.7852698509021684, + "learning_rate": 1.4489715628853707e-08, + "loss": 0.0094, + "step": 7573 + }, + { + "epoch": 9.836363636363636, + "grad_norm": 1.7661753450479927, + "learning_rate": 1.4264251833587816e-08, + "loss": 0.0018, + "step": 7574 + }, + { + "epoch": 9.837662337662337, + "grad_norm": 0.9985319071397226, + "learning_rate": 1.4040554706105103e-08, + "loss": 0.002, + "step": 7575 + }, + { + "epoch": 9.838961038961038, + "grad_norm": 1.8872081251352961, + "learning_rate": 1.3818624285981685e-08, + "loss": 0.0019, + "step": 7576 + }, + { + "epoch": 9.84025974025974, + "grad_norm": 0.5112695686999953, + "learning_rate": 1.3598460612481712e-08, + "loss": 0.0014, + "step": 7577 + }, + { + "epoch": 9.841558441558442, + "grad_norm": 0.9924888334630978, + "learning_rate": 1.3380063724555136e-08, + "loss": 0.0011, + "step": 7578 + }, + { + "epoch": 9.842857142857143, + "grad_norm": 0.9481805149512188, + "learning_rate": 1.3163433660842163e-08, + "loss": 0.0009, + "step": 7579 + }, + { + "epoch": 9.844155844155845, + "grad_norm": 1.506882025923929, + "learning_rate": 1.2948570459667687e-08, + "loss": 0.0087, + "step": 7580 + }, + { + "epoch": 9.845454545454546, + "grad_norm": 0.5783992003460504, + "learning_rate": 1.2735474159044637e-08, + "loss": 0.0074, + "step": 7581 + }, + { + "epoch": 9.846753246753247, + "grad_norm": 1.088795987085842, + "learning_rate": 1.2524144796673965e-08, + "loss": 0.0016, + "step": 7582 + }, + { + "epoch": 9.848051948051948, + "grad_norm": 1.2664749332917522, + "learning_rate": 1.2314582409945764e-08, + "loss": 0.0149, + "step": 7583 + }, + { + "epoch": 9.849350649350649, + "grad_norm": 0.19184005460136935, + "learning_rate": 1.2106787035933709e-08, + "loss": 0.0007, + "step": 7584 + }, + { + "epoch": 9.85064935064935, + "grad_norm": 0.764545549627212, + "learning_rate": 1.1900758711401728e-08, + "loss": 0.0008, + "step": 7585 + }, + { + "epoch": 9.851948051948051, + "grad_norm": 3.1285998087490485, + "learning_rate": 1.1696497472798441e-08, + "loss": 0.0025, + "step": 7586 + }, + { + "epoch": 9.853246753246752, + "grad_norm": 11.333216337016767, + "learning_rate": 1.1494003356263827e-08, + "loss": 0.0123, + "step": 7587 + }, + { + "epoch": 9.854545454545455, + "grad_norm": 1.0649647078039255, + "learning_rate": 1.1293276397622566e-08, + "loss": 0.0071, + "step": 7588 + }, + { + "epoch": 9.855844155844157, + "grad_norm": 0.7796994782225403, + "learning_rate": 1.1094316632386248e-08, + "loss": 0.0023, + "step": 7589 + }, + { + "epoch": 9.857142857142858, + "grad_norm": 2.1708883327400783, + "learning_rate": 1.0897124095754497e-08, + "loss": 0.0087, + "step": 7590 + }, + { + "epoch": 9.858441558441559, + "grad_norm": 1.1240524055095742, + "learning_rate": 1.070169882261496e-08, + "loss": 0.0019, + "step": 7591 + }, + { + "epoch": 9.85974025974026, + "grad_norm": 0.484142555771593, + "learning_rate": 1.0508040847543311e-08, + "loss": 0.001, + "step": 7592 + }, + { + "epoch": 9.861038961038961, + "grad_norm": 2.8949080874475603, + "learning_rate": 1.0316150204798813e-08, + "loss": 0.0057, + "step": 7593 + }, + { + "epoch": 9.862337662337662, + "grad_norm": 8.173044952184803, + "learning_rate": 1.0126026928332089e-08, + "loss": 0.0057, + "step": 7594 + }, + { + "epoch": 9.863636363636363, + "grad_norm": 1.1927941021261592, + "learning_rate": 9.937671051780673e-09, + "loss": 0.0078, + "step": 7595 + }, + { + "epoch": 9.864935064935064, + "grad_norm": 0.6109903279091758, + "learning_rate": 9.75108260846569e-09, + "loss": 0.0083, + "step": 7596 + }, + { + "epoch": 9.866233766233766, + "grad_norm": 0.7294181304949942, + "learning_rate": 9.56626163140073e-09, + "loss": 0.0013, + "step": 7597 + }, + { + "epoch": 9.867532467532467, + "grad_norm": 2.8047884456046592, + "learning_rate": 9.383208153281864e-09, + "loss": 0.0147, + "step": 7598 + }, + { + "epoch": 9.86883116883117, + "grad_norm": 0.7124679037300737, + "learning_rate": 9.201922206496516e-09, + "loss": 0.0014, + "step": 7599 + }, + { + "epoch": 9.87012987012987, + "grad_norm": 0.29170383600252076, + "learning_rate": 9.02240382311681e-09, + "loss": 0.0013, + "step": 7600 + }, + { + "epoch": 9.871428571428572, + "grad_norm": 0.5309329604296875, + "learning_rate": 8.844653034902895e-09, + "loss": 0.0012, + "step": 7601 + }, + { + "epoch": 9.872727272727273, + "grad_norm": 0.4736465664418744, + "learning_rate": 8.668669873304058e-09, + "loss": 0.0008, + "step": 7602 + }, + { + "epoch": 9.874025974025974, + "grad_norm": 0.9640218982209368, + "learning_rate": 8.494454369452065e-09, + "loss": 0.0013, + "step": 7603 + }, + { + "epoch": 9.875324675324675, + "grad_norm": 0.3249249714980823, + "learning_rate": 8.322006554171147e-09, + "loss": 0.0009, + "step": 7604 + }, + { + "epoch": 9.876623376623376, + "grad_norm": 3.033079428820859, + "learning_rate": 8.151326457970233e-09, + "loss": 0.0018, + "step": 7605 + }, + { + "epoch": 9.877922077922078, + "grad_norm": 2.0572629661773307, + "learning_rate": 7.982414111045167e-09, + "loss": 0.0194, + "step": 7606 + }, + { + "epoch": 9.879220779220779, + "grad_norm": 2.05889352467776, + "learning_rate": 7.815269543280934e-09, + "loss": 0.0052, + "step": 7607 + }, + { + "epoch": 9.88051948051948, + "grad_norm": 0.5365166859042194, + "learning_rate": 7.649892784247215e-09, + "loss": 0.001, + "step": 7608 + }, + { + "epoch": 9.881818181818183, + "grad_norm": 0.8380504316583147, + "learning_rate": 7.486283863203935e-09, + "loss": 0.001, + "step": 7609 + }, + { + "epoch": 9.883116883116884, + "grad_norm": 0.3675856911799692, + "learning_rate": 7.324442809095722e-09, + "loss": 0.001, + "step": 7610 + }, + { + "epoch": 9.884415584415585, + "grad_norm": 0.7094132538847099, + "learning_rate": 7.164369650555225e-09, + "loss": 0.008, + "step": 7611 + }, + { + "epoch": 9.885714285714286, + "grad_norm": 0.9579850694388623, + "learning_rate": 7.0060644159020144e-09, + "loss": 0.0012, + "step": 7612 + }, + { + "epoch": 9.887012987012987, + "grad_norm": 0.7344910664367065, + "learning_rate": 6.849527133144795e-09, + "loss": 0.0014, + "step": 7613 + }, + { + "epoch": 9.888311688311688, + "grad_norm": 5.5676186223186255, + "learning_rate": 6.694757829976972e-09, + "loss": 0.0078, + "step": 7614 + }, + { + "epoch": 9.88961038961039, + "grad_norm": 2.073558497216596, + "learning_rate": 6.541756533779975e-09, + "loss": 0.0018, + "step": 7615 + }, + { + "epoch": 9.89090909090909, + "grad_norm": 1.1007014863602356, + "learning_rate": 6.3905232716232616e-09, + "loss": 0.0045, + "step": 7616 + }, + { + "epoch": 9.892207792207792, + "grad_norm": 0.9897720616948059, + "learning_rate": 6.241058070263206e-09, + "loss": 0.0072, + "step": 7617 + }, + { + "epoch": 9.893506493506493, + "grad_norm": 0.8266799688438546, + "learning_rate": 6.093360956141992e-09, + "loss": 0.0009, + "step": 7618 + }, + { + "epoch": 9.894805194805194, + "grad_norm": 0.40814956651103096, + "learning_rate": 5.947431955390937e-09, + "loss": 0.001, + "step": 7619 + }, + { + "epoch": 9.896103896103895, + "grad_norm": 1.2779950716173236, + "learning_rate": 5.8032710938271675e-09, + "loss": 0.0093, + "step": 7620 + }, + { + "epoch": 9.897402597402598, + "grad_norm": 0.5536306908805549, + "learning_rate": 5.6608783969558375e-09, + "loss": 0.0009, + "step": 7621 + }, + { + "epoch": 9.8987012987013, + "grad_norm": 0.34538973404079676, + "learning_rate": 5.520253889969018e-09, + "loss": 0.0011, + "step": 7622 + }, + { + "epoch": 9.9, + "grad_norm": 0.6907457402136398, + "learning_rate": 5.381397597744587e-09, + "loss": 0.0024, + "step": 7623 + }, + { + "epoch": 9.901298701298701, + "grad_norm": 0.39769557885974366, + "learning_rate": 5.2443095448506674e-09, + "loss": 0.001, + "step": 7624 + }, + { + "epoch": 9.902597402597403, + "grad_norm": 2.51207074048038, + "learning_rate": 5.108989755540083e-09, + "loss": 0.002, + "step": 7625 + }, + { + "epoch": 9.903896103896104, + "grad_norm": 1.0132775605198068, + "learning_rate": 4.975438253753684e-09, + "loss": 0.0015, + "step": 7626 + }, + { + "epoch": 9.905194805194805, + "grad_norm": 4.769809277118773, + "learning_rate": 4.8436550631192345e-09, + "loss": 0.0037, + "step": 7627 + }, + { + "epoch": 9.906493506493506, + "grad_norm": 2.195673876970017, + "learning_rate": 4.7136402069514194e-09, + "loss": 0.0196, + "step": 7628 + }, + { + "epoch": 9.907792207792207, + "grad_norm": 7.809902275701846, + "learning_rate": 4.58539370825184e-09, + "loss": 0.0048, + "step": 7629 + }, + { + "epoch": 9.909090909090908, + "grad_norm": 7.10348117406294, + "learning_rate": 4.458915589711232e-09, + "loss": 0.0116, + "step": 7630 + }, + { + "epoch": 9.910389610389611, + "grad_norm": 0.6367010153421079, + "learning_rate": 4.334205873705033e-09, + "loss": 0.0014, + "step": 7631 + }, + { + "epoch": 9.911688311688312, + "grad_norm": 3.0523939341365205, + "learning_rate": 4.211264582295593e-09, + "loss": 0.0197, + "step": 7632 + }, + { + "epoch": 9.912987012987013, + "grad_norm": 3.2939907338850642, + "learning_rate": 4.090091737236623e-09, + "loss": 0.0029, + "step": 7633 + }, + { + "epoch": 9.914285714285715, + "grad_norm": 0.3842361825261678, + "learning_rate": 3.970687359963199e-09, + "loss": 0.0025, + "step": 7634 + }, + { + "epoch": 9.915584415584416, + "grad_norm": 0.680609902532925, + "learning_rate": 3.853051471601754e-09, + "loss": 0.0009, + "step": 7635 + }, + { + "epoch": 9.916883116883117, + "grad_norm": 0.3054907117938859, + "learning_rate": 3.737184092963419e-09, + "loss": 0.0009, + "step": 7636 + }, + { + "epoch": 9.918181818181818, + "grad_norm": 0.6032948692551355, + "learning_rate": 3.623085244547353e-09, + "loss": 0.001, + "step": 7637 + }, + { + "epoch": 9.91948051948052, + "grad_norm": 1.06158365674446, + "learning_rate": 3.51075494654074e-09, + "loss": 0.0013, + "step": 7638 + }, + { + "epoch": 9.92077922077922, + "grad_norm": 4.790736408775846, + "learning_rate": 3.400193218816572e-09, + "loss": 0.0121, + "step": 7639 + }, + { + "epoch": 9.922077922077921, + "grad_norm": 0.9290347656832907, + "learning_rate": 3.291400080934759e-09, + "loss": 0.001, + "step": 7640 + }, + { + "epoch": 9.923376623376623, + "grad_norm": 1.4101679854617626, + "learning_rate": 3.1843755521443474e-09, + "loss": 0.0106, + "step": 7641 + }, + { + "epoch": 9.924675324675325, + "grad_norm": 6.760920363187259, + "learning_rate": 3.0791196513779707e-09, + "loss": 0.0079, + "step": 7642 + }, + { + "epoch": 9.925974025974027, + "grad_norm": 3.1635960960257585, + "learning_rate": 2.97563239725851e-09, + "loss": 0.0032, + "step": 7643 + }, + { + "epoch": 9.927272727272728, + "grad_norm": 0.7919127494305525, + "learning_rate": 2.8739138080946526e-09, + "loss": 0.0009, + "step": 7644 + }, + { + "epoch": 9.928571428571429, + "grad_norm": 1.7948010060124484, + "learning_rate": 2.7739639018831145e-09, + "loss": 0.0023, + "step": 7645 + }, + { + "epoch": 9.92987012987013, + "grad_norm": 1.332848088810919, + "learning_rate": 2.675782696305307e-09, + "loss": 0.0015, + "step": 7646 + }, + { + "epoch": 9.931168831168831, + "grad_norm": 0.4345598401887244, + "learning_rate": 2.579370208734e-09, + "loss": 0.001, + "step": 7647 + }, + { + "epoch": 9.932467532467532, + "grad_norm": 0.9255486475352925, + "learning_rate": 2.48472645622333e-09, + "loss": 0.0009, + "step": 7648 + }, + { + "epoch": 9.933766233766233, + "grad_norm": 1.2609432322085932, + "learning_rate": 2.39185145551879e-09, + "loss": 0.0089, + "step": 7649 + }, + { + "epoch": 9.935064935064934, + "grad_norm": 0.6195006143501326, + "learning_rate": 2.3007452230527917e-09, + "loss": 0.0075, + "step": 7650 + }, + { + "epoch": 9.936363636363636, + "grad_norm": 0.8760329273535077, + "learning_rate": 2.2114077749435524e-09, + "loss": 0.0079, + "step": 7651 + }, + { + "epoch": 9.937662337662339, + "grad_norm": 0.31689861642675754, + "learning_rate": 2.1238391269950977e-09, + "loss": 0.0012, + "step": 7652 + }, + { + "epoch": 9.93896103896104, + "grad_norm": 1.205899628810383, + "learning_rate": 2.0380392947005887e-09, + "loss": 0.0073, + "step": 7653 + }, + { + "epoch": 9.94025974025974, + "grad_norm": 0.6832900440680333, + "learning_rate": 1.9540082932412162e-09, + "loss": 0.0081, + "step": 7654 + }, + { + "epoch": 9.941558441558442, + "grad_norm": 1.350448144846318, + "learning_rate": 1.8717461374817558e-09, + "loss": 0.0012, + "step": 7655 + }, + { + "epoch": 9.942857142857143, + "grad_norm": 2.3540840354246377, + "learning_rate": 1.791252841976121e-09, + "loss": 0.0162, + "step": 7656 + }, + { + "epoch": 9.944155844155844, + "grad_norm": 0.3118832350467507, + "learning_rate": 1.712528420966253e-09, + "loss": 0.0011, + "step": 7657 + }, + { + "epoch": 9.945454545454545, + "grad_norm": 1.6806006935522841, + "learning_rate": 1.635572888378789e-09, + "loss": 0.0161, + "step": 7658 + }, + { + "epoch": 9.946753246753246, + "grad_norm": 1.3610308205605208, + "learning_rate": 1.5603862578295049e-09, + "loss": 0.0053, + "step": 7659 + }, + { + "epoch": 9.948051948051948, + "grad_norm": 1.1113353136539441, + "learning_rate": 1.486968542619982e-09, + "loss": 0.0013, + "step": 7660 + }, + { + "epoch": 9.949350649350649, + "grad_norm": 2.8923675060079903, + "learning_rate": 1.4153197557398302e-09, + "loss": 0.0086, + "step": 7661 + }, + { + "epoch": 9.95064935064935, + "grad_norm": 0.6941261220653735, + "learning_rate": 1.3454399098633552e-09, + "loss": 0.0083, + "step": 7662 + }, + { + "epoch": 9.951948051948053, + "grad_norm": 0.41908688492870505, + "learning_rate": 1.2773290173562213e-09, + "loss": 0.0014, + "step": 7663 + }, + { + "epoch": 9.953246753246754, + "grad_norm": 6.460363435872477, + "learning_rate": 1.2109870902665687e-09, + "loss": 0.0263, + "step": 7664 + }, + { + "epoch": 9.954545454545455, + "grad_norm": 2.7242682656319523, + "learning_rate": 1.1464141403316752e-09, + "loss": 0.016, + "step": 7665 + }, + { + "epoch": 9.955844155844156, + "grad_norm": 1.6091832193729305, + "learning_rate": 1.0836101789768462e-09, + "loss": 0.0077, + "step": 7666 + }, + { + "epoch": 9.957142857142857, + "grad_norm": 0.5583130869608747, + "learning_rate": 1.022575217313193e-09, + "loss": 0.0013, + "step": 7667 + }, + { + "epoch": 9.958441558441558, + "grad_norm": 1.0792774413891995, + "learning_rate": 9.633092661376353e-10, + "loss": 0.0014, + "step": 7668 + }, + { + "epoch": 9.95974025974026, + "grad_norm": 1.5839464773800236, + "learning_rate": 9.058123359362292e-10, + "loss": 0.0017, + "step": 7669 + }, + { + "epoch": 9.96103896103896, + "grad_norm": 0.3276447377519524, + "learning_rate": 8.500844368819483e-10, + "loss": 0.0008, + "step": 7670 + }, + { + "epoch": 9.962337662337662, + "grad_norm": 8.795831665799183, + "learning_rate": 7.961255788324629e-10, + "loss": 0.0091, + "step": 7671 + }, + { + "epoch": 9.963636363636363, + "grad_norm": 1.720472488607327, + "learning_rate": 7.439357713356909e-10, + "loss": 0.0019, + "step": 7672 + }, + { + "epoch": 9.964935064935066, + "grad_norm": 0.9678153253174788, + "learning_rate": 6.935150236253574e-10, + "loss": 0.0016, + "step": 7673 + }, + { + "epoch": 9.966233766233767, + "grad_norm": 5.445340688860022, + "learning_rate": 6.448633446198837e-10, + "loss": 0.0037, + "step": 7674 + }, + { + "epoch": 9.967532467532468, + "grad_norm": 0.6608028366466133, + "learning_rate": 5.979807429279394e-10, + "loss": 0.0009, + "step": 7675 + }, + { + "epoch": 9.96883116883117, + "grad_norm": 2.1138157934148665, + "learning_rate": 5.528672268440005e-10, + "loss": 0.0041, + "step": 7676 + }, + { + "epoch": 9.97012987012987, + "grad_norm": 0.29269262049994477, + "learning_rate": 5.095228043494604e-10, + "loss": 0.0009, + "step": 7677 + }, + { + "epoch": 9.971428571428572, + "grad_norm": 0.4573947323256857, + "learning_rate": 4.679474831126296e-10, + "loss": 0.0008, + "step": 7678 + }, + { + "epoch": 9.972727272727273, + "grad_norm": 2.0179171730522367, + "learning_rate": 4.2814127048873553e-10, + "loss": 0.0024, + "step": 7679 + }, + { + "epoch": 9.974025974025974, + "grad_norm": 0.16356878223455998, + "learning_rate": 3.9010417352103313e-10, + "loss": 0.0009, + "step": 7680 + }, + { + "epoch": 9.975324675324675, + "grad_norm": 0.5957252053823271, + "learning_rate": 3.538361989374739e-10, + "loss": 0.0014, + "step": 7681 + }, + { + "epoch": 9.976623376623376, + "grad_norm": 0.7639402261634324, + "learning_rate": 3.1933735315625713e-10, + "loss": 0.0025, + "step": 7682 + }, + { + "epoch": 9.977922077922077, + "grad_norm": 3.7442627197970335, + "learning_rate": 2.866076422802788e-10, + "loss": 0.0036, + "step": 7683 + }, + { + "epoch": 9.979220779220778, + "grad_norm": 3.0608548257624095, + "learning_rate": 2.55647072099352e-10, + "loss": 0.0027, + "step": 7684 + }, + { + "epoch": 9.980519480519481, + "grad_norm": 1.3561788052456867, + "learning_rate": 2.2645564809242736e-10, + "loss": 0.0017, + "step": 7685 + }, + { + "epoch": 9.981818181818182, + "grad_norm": 0.8047166226560061, + "learning_rate": 1.9903337542204192e-10, + "loss": 0.001, + "step": 7686 + }, + { + "epoch": 9.983116883116883, + "grad_norm": 4.062116320784228, + "learning_rate": 1.7338025894209077e-10, + "loss": 0.0075, + "step": 7687 + }, + { + "epoch": 9.984415584415585, + "grad_norm": 2.7360621842121207, + "learning_rate": 1.4949630318894516e-10, + "loss": 0.0048, + "step": 7688 + }, + { + "epoch": 9.985714285714286, + "grad_norm": 11.28034759633753, + "learning_rate": 1.2738151239033435e-10, + "loss": 0.006, + "step": 7689 + }, + { + "epoch": 9.987012987012987, + "grad_norm": 0.35240003873681475, + "learning_rate": 1.070358904564639e-10, + "loss": 0.001, + "step": 7690 + }, + { + "epoch": 9.988311688311688, + "grad_norm": 0.46533092484565536, + "learning_rate": 8.845944098778703e-11, + "loss": 0.0015, + "step": 7691 + }, + { + "epoch": 9.98961038961039, + "grad_norm": 0.7787266690473452, + "learning_rate": 7.165216727167412e-11, + "loss": 0.0013, + "step": 7692 + }, + { + "epoch": 9.99090909090909, + "grad_norm": 3.764032529704328, + "learning_rate": 5.6614072280192224e-11, + "loss": 0.0046, + "step": 7693 + }, + { + "epoch": 9.992207792207791, + "grad_norm": 2.35737179980628, + "learning_rate": 4.3345158675656144e-11, + "loss": 0.015, + "step": 7694 + }, + { + "epoch": 9.993506493506494, + "grad_norm": 1.1168562302714782, + "learning_rate": 3.184542880396713e-11, + "loss": 0.0017, + "step": 7695 + }, + { + "epoch": 9.994805194805195, + "grad_norm": 4.934373922360939, + "learning_rate": 2.2114884700163986e-11, + "loss": 0.0075, + "step": 7696 + }, + { + "epoch": 9.996103896103897, + "grad_norm": 1.5588675630199313, + "learning_rate": 1.4153528086202628e-11, + "loss": 0.0085, + "step": 7697 + }, + { + "epoch": 9.997402597402598, + "grad_norm": 1.1263227436907806, + "learning_rate": 7.961360369845849e-12, + "loss": 0.0085, + "step": 7698 + }, + { + "epoch": 9.998701298701299, + "grad_norm": 1.878616027847936, + "learning_rate": 3.5383826479939986e-12, + "loss": 0.0092, + "step": 7699 + }, + { + "epoch": 10.0, + "grad_norm": 0.9146996176062973, + "learning_rate": 8.845957011338613e-13, + "loss": 0.0018, + "step": 7700 + }, + { + "epoch": 10.0, + "eval_accuracy": 0.939921392476137, + "eval_f1": 0.9287781074188871, + "eval_loss": 0.16708879172801971, + "eval_precision": 0.922539255038175, + "eval_recall": 0.9380805638160603, + "eval_runtime": 12.7906, + "eval_samples_per_second": 139.243, + "eval_steps_per_second": 1.095, + "step": 7700 + }, + { + "epoch": 10.0, + "step": 7700, + "total_flos": 9868738426699776.0, + "train_loss": 0.20223668603153971, + "train_runtime": 97142.0937, + "train_samples_per_second": 38.665, + "train_steps_per_second": 0.079 + } + ], + "logging_steps": 1.0, + "max_steps": 7700, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 9868738426699776.0, + "train_batch_size": 61, + "trial_name": null, + "trial_params": null +}