{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 837, "global_step": 837, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0011947431302270011, "grad_norm": 0.6640625, "learning_rate": 1e-05, "loss": 1.4509, "step": 1 }, { "epoch": 0.0023894862604540022, "grad_norm": 0.328125, "learning_rate": 9.98805256869773e-06, "loss": 1.4667, "step": 2 }, { "epoch": 0.0035842293906810036, "grad_norm": 0.265625, "learning_rate": 9.97610513739546e-06, "loss": 1.5168, "step": 3 }, { "epoch": 0.0047789725209080045, "grad_norm": 0.283203125, "learning_rate": 9.96415770609319e-06, "loss": 1.5592, "step": 4 }, { "epoch": 0.005973715651135006, "grad_norm": 0.2412109375, "learning_rate": 9.95221027479092e-06, "loss": 1.485, "step": 5 }, { "epoch": 0.007168458781362007, "grad_norm": 0.23046875, "learning_rate": 9.940262843488651e-06, "loss": 1.5185, "step": 6 }, { "epoch": 0.008363201911589008, "grad_norm": 0.205078125, "learning_rate": 9.928315412186382e-06, "loss": 1.3904, "step": 7 }, { "epoch": 0.009557945041816009, "grad_norm": 0.244140625, "learning_rate": 9.916367980884111e-06, "loss": 1.5337, "step": 8 }, { "epoch": 0.010752688172043012, "grad_norm": 0.19140625, "learning_rate": 9.90442054958184e-06, "loss": 1.4088, "step": 9 }, { "epoch": 0.011947431302270013, "grad_norm": 0.2421875, "learning_rate": 9.89247311827957e-06, "loss": 1.613, "step": 10 }, { "epoch": 0.013142174432497013, "grad_norm": 0.2578125, "learning_rate": 9.8805256869773e-06, "loss": 1.5922, "step": 11 }, { "epoch": 0.014336917562724014, "grad_norm": 0.263671875, "learning_rate": 9.86857825567503e-06, "loss": 1.4745, "step": 12 }, { "epoch": 0.015531660692951015, "grad_norm": 0.201171875, "learning_rate": 9.856630824372761e-06, "loss": 1.3199, "step": 13 }, { "epoch": 0.016726403823178016, "grad_norm": 0.138671875, "learning_rate": 9.84468339307049e-06, "loss": 1.3981, "step": 14 }, { "epoch": 0.017921146953405017, "grad_norm": 0.150390625, "learning_rate": 9.832735961768221e-06, "loss": 1.4429, "step": 15 }, { "epoch": 0.019115890083632018, "grad_norm": 0.1650390625, "learning_rate": 9.820788530465952e-06, "loss": 1.4959, "step": 16 }, { "epoch": 0.02031063321385902, "grad_norm": 0.220703125, "learning_rate": 9.80884109916368e-06, "loss": 1.4593, "step": 17 }, { "epoch": 0.021505376344086023, "grad_norm": 0.1630859375, "learning_rate": 9.79689366786141e-06, "loss": 1.4518, "step": 18 }, { "epoch": 0.022700119474313024, "grad_norm": 0.197265625, "learning_rate": 9.78494623655914e-06, "loss": 1.3832, "step": 19 }, { "epoch": 0.023894862604540025, "grad_norm": 0.220703125, "learning_rate": 9.77299880525687e-06, "loss": 1.428, "step": 20 }, { "epoch": 0.025089605734767026, "grad_norm": 0.189453125, "learning_rate": 9.7610513739546e-06, "loss": 1.3976, "step": 21 }, { "epoch": 0.026284348864994027, "grad_norm": 0.1884765625, "learning_rate": 9.749103942652331e-06, "loss": 1.4803, "step": 22 }, { "epoch": 0.027479091995221028, "grad_norm": 0.177734375, "learning_rate": 9.737156511350062e-06, "loss": 1.5784, "step": 23 }, { "epoch": 0.02867383512544803, "grad_norm": 0.16015625, "learning_rate": 9.725209080047791e-06, "loss": 1.4369, "step": 24 }, { "epoch": 0.02986857825567503, "grad_norm": 0.1669921875, "learning_rate": 9.71326164874552e-06, "loss": 1.4774, "step": 25 }, { "epoch": 0.03106332138590203, "grad_norm": 0.1455078125, "learning_rate": 9.70131421744325e-06, "loss": 1.3351, "step": 26 }, { "epoch": 0.03225806451612903, "grad_norm": 0.220703125, "learning_rate": 9.68936678614098e-06, "loss": 1.4909, "step": 27 }, { "epoch": 0.03345280764635603, "grad_norm": 0.1689453125, "learning_rate": 9.67741935483871e-06, "loss": 1.2378, "step": 28 }, { "epoch": 0.03464755077658303, "grad_norm": 0.16015625, "learning_rate": 9.66547192353644e-06, "loss": 1.4896, "step": 29 }, { "epoch": 0.035842293906810034, "grad_norm": 0.140625, "learning_rate": 9.65352449223417e-06, "loss": 1.5645, "step": 30 }, { "epoch": 0.037037037037037035, "grad_norm": 0.1591796875, "learning_rate": 9.641577060931901e-06, "loss": 1.4164, "step": 31 }, { "epoch": 0.038231780167264036, "grad_norm": 0.181640625, "learning_rate": 9.62962962962963e-06, "loss": 1.5123, "step": 32 }, { "epoch": 0.03942652329749104, "grad_norm": 0.138671875, "learning_rate": 9.61768219832736e-06, "loss": 1.3468, "step": 33 }, { "epoch": 0.04062126642771804, "grad_norm": 0.14453125, "learning_rate": 9.60573476702509e-06, "loss": 1.3082, "step": 34 }, { "epoch": 0.04181600955794504, "grad_norm": 0.150390625, "learning_rate": 9.59378733572282e-06, "loss": 1.4152, "step": 35 }, { "epoch": 0.043010752688172046, "grad_norm": 0.283203125, "learning_rate": 9.58183990442055e-06, "loss": 1.4631, "step": 36 }, { "epoch": 0.04420549581839905, "grad_norm": 0.1201171875, "learning_rate": 9.56989247311828e-06, "loss": 1.3573, "step": 37 }, { "epoch": 0.04540023894862605, "grad_norm": 0.12255859375, "learning_rate": 9.557945041816011e-06, "loss": 1.3414, "step": 38 }, { "epoch": 0.04659498207885305, "grad_norm": 0.265625, "learning_rate": 9.54599761051374e-06, "loss": 1.4339, "step": 39 }, { "epoch": 0.04778972520908005, "grad_norm": 0.16015625, "learning_rate": 9.53405017921147e-06, "loss": 1.3224, "step": 40 }, { "epoch": 0.04898446833930705, "grad_norm": 0.1796875, "learning_rate": 9.5221027479092e-06, "loss": 1.6164, "step": 41 }, { "epoch": 0.05017921146953405, "grad_norm": 0.185546875, "learning_rate": 9.51015531660693e-06, "loss": 1.5909, "step": 42 }, { "epoch": 0.05137395459976105, "grad_norm": 0.13671875, "learning_rate": 9.49820788530466e-06, "loss": 1.2923, "step": 43 }, { "epoch": 0.052568697729988054, "grad_norm": 0.1484375, "learning_rate": 9.48626045400239e-06, "loss": 1.3446, "step": 44 }, { "epoch": 0.053763440860215055, "grad_norm": 0.154296875, "learning_rate": 9.47431302270012e-06, "loss": 1.2121, "step": 45 }, { "epoch": 0.054958183990442055, "grad_norm": 0.1484375, "learning_rate": 9.46236559139785e-06, "loss": 1.3305, "step": 46 }, { "epoch": 0.056152927120669056, "grad_norm": 0.1328125, "learning_rate": 9.450418160095581e-06, "loss": 1.1723, "step": 47 }, { "epoch": 0.05734767025089606, "grad_norm": 0.1416015625, "learning_rate": 9.43847072879331e-06, "loss": 1.3501, "step": 48 }, { "epoch": 0.05854241338112306, "grad_norm": 0.1474609375, "learning_rate": 9.42652329749104e-06, "loss": 1.4093, "step": 49 }, { "epoch": 0.05973715651135006, "grad_norm": 0.1357421875, "learning_rate": 9.41457586618877e-06, "loss": 1.3779, "step": 50 }, { "epoch": 0.06093189964157706, "grad_norm": 0.16796875, "learning_rate": 9.4026284348865e-06, "loss": 1.6416, "step": 51 }, { "epoch": 0.06212664277180406, "grad_norm": 0.158203125, "learning_rate": 9.39068100358423e-06, "loss": 1.5283, "step": 52 }, { "epoch": 0.06332138590203107, "grad_norm": 0.123046875, "learning_rate": 9.37873357228196e-06, "loss": 1.3239, "step": 53 }, { "epoch": 0.06451612903225806, "grad_norm": 0.1435546875, "learning_rate": 9.36678614097969e-06, "loss": 1.4087, "step": 54 }, { "epoch": 0.06571087216248507, "grad_norm": 0.1240234375, "learning_rate": 9.35483870967742e-06, "loss": 1.4155, "step": 55 }, { "epoch": 0.06690561529271206, "grad_norm": 0.1474609375, "learning_rate": 9.34289127837515e-06, "loss": 1.2758, "step": 56 }, { "epoch": 0.06810035842293907, "grad_norm": 0.126953125, "learning_rate": 9.33094384707288e-06, "loss": 1.0777, "step": 57 }, { "epoch": 0.06929510155316607, "grad_norm": 0.1455078125, "learning_rate": 9.31899641577061e-06, "loss": 1.1965, "step": 58 }, { "epoch": 0.07048984468339307, "grad_norm": 0.146484375, "learning_rate": 9.30704898446834e-06, "loss": 1.2827, "step": 59 }, { "epoch": 0.07168458781362007, "grad_norm": 0.1552734375, "learning_rate": 9.29510155316607e-06, "loss": 1.3057, "step": 60 }, { "epoch": 0.07287933094384708, "grad_norm": 0.2294921875, "learning_rate": 9.2831541218638e-06, "loss": 1.3622, "step": 61 }, { "epoch": 0.07407407407407407, "grad_norm": 0.1171875, "learning_rate": 9.27120669056153e-06, "loss": 1.3517, "step": 62 }, { "epoch": 0.07526881720430108, "grad_norm": 0.318359375, "learning_rate": 9.25925925925926e-06, "loss": 1.4463, "step": 63 }, { "epoch": 0.07646356033452807, "grad_norm": 1.2890625, "learning_rate": 9.24731182795699e-06, "loss": 1.3783, "step": 64 }, { "epoch": 0.07765830346475508, "grad_norm": 0.1474609375, "learning_rate": 9.23536439665472e-06, "loss": 1.2306, "step": 65 }, { "epoch": 0.07885304659498207, "grad_norm": 0.15625, "learning_rate": 9.22341696535245e-06, "loss": 1.3297, "step": 66 }, { "epoch": 0.08004778972520908, "grad_norm": 0.1533203125, "learning_rate": 9.21146953405018e-06, "loss": 1.4157, "step": 67 }, { "epoch": 0.08124253285543608, "grad_norm": 0.1298828125, "learning_rate": 9.19952210274791e-06, "loss": 1.386, "step": 68 }, { "epoch": 0.08243727598566308, "grad_norm": 0.1103515625, "learning_rate": 9.18757467144564e-06, "loss": 1.2735, "step": 69 }, { "epoch": 0.08363201911589008, "grad_norm": 0.1376953125, "learning_rate": 9.17562724014337e-06, "loss": 1.3003, "step": 70 }, { "epoch": 0.08482676224611709, "grad_norm": 0.162109375, "learning_rate": 9.1636798088411e-06, "loss": 1.3123, "step": 71 }, { "epoch": 0.08602150537634409, "grad_norm": 0.1689453125, "learning_rate": 9.15173237753883e-06, "loss": 1.3509, "step": 72 }, { "epoch": 0.08721624850657109, "grad_norm": 0.130859375, "learning_rate": 9.13978494623656e-06, "loss": 1.2555, "step": 73 }, { "epoch": 0.0884109916367981, "grad_norm": 0.208984375, "learning_rate": 9.12783751493429e-06, "loss": 1.3674, "step": 74 }, { "epoch": 0.08960573476702509, "grad_norm": 0.1494140625, "learning_rate": 9.11589008363202e-06, "loss": 1.3415, "step": 75 }, { "epoch": 0.0908004778972521, "grad_norm": 0.140625, "learning_rate": 9.10394265232975e-06, "loss": 1.555, "step": 76 }, { "epoch": 0.09199522102747909, "grad_norm": 0.1611328125, "learning_rate": 9.09199522102748e-06, "loss": 1.2856, "step": 77 }, { "epoch": 0.0931899641577061, "grad_norm": 0.1533203125, "learning_rate": 9.08004778972521e-06, "loss": 1.2737, "step": 78 }, { "epoch": 0.09438470728793309, "grad_norm": 0.13671875, "learning_rate": 9.068100358422939e-06, "loss": 1.286, "step": 79 }, { "epoch": 0.0955794504181601, "grad_norm": 0.12060546875, "learning_rate": 9.05615292712067e-06, "loss": 1.2542, "step": 80 }, { "epoch": 0.0967741935483871, "grad_norm": 0.1904296875, "learning_rate": 9.0442054958184e-06, "loss": 1.3861, "step": 81 }, { "epoch": 0.0979689366786141, "grad_norm": 0.16015625, "learning_rate": 9.03225806451613e-06, "loss": 1.3931, "step": 82 }, { "epoch": 0.0991636798088411, "grad_norm": 0.18359375, "learning_rate": 9.02031063321386e-06, "loss": 1.3573, "step": 83 }, { "epoch": 0.1003584229390681, "grad_norm": 0.1259765625, "learning_rate": 9.00836320191159e-06, "loss": 1.4077, "step": 84 }, { "epoch": 0.1015531660692951, "grad_norm": 0.11474609375, "learning_rate": 8.99641577060932e-06, "loss": 1.2472, "step": 85 }, { "epoch": 0.1027479091995221, "grad_norm": 0.1572265625, "learning_rate": 8.98446833930705e-06, "loss": 1.3902, "step": 86 }, { "epoch": 0.1039426523297491, "grad_norm": 0.1572265625, "learning_rate": 8.97252090800478e-06, "loss": 1.3607, "step": 87 }, { "epoch": 0.10513739545997611, "grad_norm": 0.1376953125, "learning_rate": 8.96057347670251e-06, "loss": 1.3551, "step": 88 }, { "epoch": 0.1063321385902031, "grad_norm": 0.1552734375, "learning_rate": 8.94862604540024e-06, "loss": 1.356, "step": 89 }, { "epoch": 0.10752688172043011, "grad_norm": 0.19921875, "learning_rate": 8.93667861409797e-06, "loss": 1.5084, "step": 90 }, { "epoch": 0.1087216248506571, "grad_norm": 0.1201171875, "learning_rate": 8.9247311827957e-06, "loss": 1.2582, "step": 91 }, { "epoch": 0.10991636798088411, "grad_norm": 0.125, "learning_rate": 8.91278375149343e-06, "loss": 1.2923, "step": 92 }, { "epoch": 0.1111111111111111, "grad_norm": 0.1484375, "learning_rate": 8.90083632019116e-06, "loss": 1.3095, "step": 93 }, { "epoch": 0.11230585424133811, "grad_norm": 0.1298828125, "learning_rate": 8.888888888888888e-06, "loss": 1.1731, "step": 94 }, { "epoch": 0.1135005973715651, "grad_norm": 0.1767578125, "learning_rate": 8.876941457586619e-06, "loss": 1.3654, "step": 95 }, { "epoch": 0.11469534050179211, "grad_norm": 0.1494140625, "learning_rate": 8.86499402628435e-06, "loss": 1.3491, "step": 96 }, { "epoch": 0.11589008363201912, "grad_norm": 0.1455078125, "learning_rate": 8.85304659498208e-06, "loss": 1.3979, "step": 97 }, { "epoch": 0.11708482676224612, "grad_norm": 0.140625, "learning_rate": 8.84109916367981e-06, "loss": 1.2811, "step": 98 }, { "epoch": 0.11827956989247312, "grad_norm": 0.1640625, "learning_rate": 8.82915173237754e-06, "loss": 1.4037, "step": 99 }, { "epoch": 0.11947431302270012, "grad_norm": 0.1337890625, "learning_rate": 8.81720430107527e-06, "loss": 1.2943, "step": 100 }, { "epoch": 0.12066905615292713, "grad_norm": 0.1494140625, "learning_rate": 8.805256869772998e-06, "loss": 1.2369, "step": 101 }, { "epoch": 0.12186379928315412, "grad_norm": 0.1220703125, "learning_rate": 8.793309438470729e-06, "loss": 1.4284, "step": 102 }, { "epoch": 0.12305854241338113, "grad_norm": 0.14453125, "learning_rate": 8.78136200716846e-06, "loss": 1.2186, "step": 103 }, { "epoch": 0.12425328554360812, "grad_norm": 0.1552734375, "learning_rate": 8.769414575866189e-06, "loss": 1.4268, "step": 104 }, { "epoch": 0.12544802867383512, "grad_norm": 0.177734375, "learning_rate": 8.75746714456392e-06, "loss": 1.5379, "step": 105 }, { "epoch": 0.12664277180406214, "grad_norm": 0.13671875, "learning_rate": 8.74551971326165e-06, "loss": 1.2474, "step": 106 }, { "epoch": 0.12783751493428913, "grad_norm": 0.1572265625, "learning_rate": 8.73357228195938e-06, "loss": 1.2705, "step": 107 }, { "epoch": 0.12903225806451613, "grad_norm": 0.1748046875, "learning_rate": 8.72162485065711e-06, "loss": 1.4317, "step": 108 }, { "epoch": 0.13022700119474312, "grad_norm": 0.1328125, "learning_rate": 8.70967741935484e-06, "loss": 1.4022, "step": 109 }, { "epoch": 0.13142174432497014, "grad_norm": 0.140625, "learning_rate": 8.697729988052568e-06, "loss": 1.3346, "step": 110 }, { "epoch": 0.13261648745519714, "grad_norm": 0.13671875, "learning_rate": 8.685782556750299e-06, "loss": 1.3898, "step": 111 }, { "epoch": 0.13381123058542413, "grad_norm": 0.1259765625, "learning_rate": 8.67383512544803e-06, "loss": 1.3461, "step": 112 }, { "epoch": 0.13500597371565112, "grad_norm": 0.15234375, "learning_rate": 8.66188769414576e-06, "loss": 1.1186, "step": 113 }, { "epoch": 0.13620071684587814, "grad_norm": 0.1337890625, "learning_rate": 8.64994026284349e-06, "loss": 1.4236, "step": 114 }, { "epoch": 0.13739545997610514, "grad_norm": 0.1591796875, "learning_rate": 8.63799283154122e-06, "loss": 1.4034, "step": 115 }, { "epoch": 0.13859020310633213, "grad_norm": 0.140625, "learning_rate": 8.62604540023895e-06, "loss": 1.1879, "step": 116 }, { "epoch": 0.13978494623655913, "grad_norm": 0.1259765625, "learning_rate": 8.614097968936678e-06, "loss": 1.1556, "step": 117 }, { "epoch": 0.14097968936678615, "grad_norm": 0.138671875, "learning_rate": 8.602150537634409e-06, "loss": 1.3156, "step": 118 }, { "epoch": 0.14217443249701314, "grad_norm": 0.150390625, "learning_rate": 8.590203106332138e-06, "loss": 1.2808, "step": 119 }, { "epoch": 0.14336917562724014, "grad_norm": 0.16796875, "learning_rate": 8.578255675029869e-06, "loss": 1.5488, "step": 120 }, { "epoch": 0.14456391875746716, "grad_norm": 0.1611328125, "learning_rate": 8.5663082437276e-06, "loss": 1.2217, "step": 121 }, { "epoch": 0.14575866188769415, "grad_norm": 0.13671875, "learning_rate": 8.55436081242533e-06, "loss": 1.3641, "step": 122 }, { "epoch": 0.14695340501792115, "grad_norm": 0.181640625, "learning_rate": 8.54241338112306e-06, "loss": 1.2743, "step": 123 }, { "epoch": 0.14814814814814814, "grad_norm": 0.1494140625, "learning_rate": 8.530465949820788e-06, "loss": 1.4144, "step": 124 }, { "epoch": 0.14934289127837516, "grad_norm": 0.12890625, "learning_rate": 8.518518518518519e-06, "loss": 1.3276, "step": 125 }, { "epoch": 0.15053763440860216, "grad_norm": 0.203125, "learning_rate": 8.506571087216248e-06, "loss": 1.3218, "step": 126 }, { "epoch": 0.15173237753882915, "grad_norm": 0.142578125, "learning_rate": 8.494623655913979e-06, "loss": 1.3604, "step": 127 }, { "epoch": 0.15292712066905614, "grad_norm": 0.11962890625, "learning_rate": 8.48267622461171e-06, "loss": 1.4547, "step": 128 }, { "epoch": 0.15412186379928317, "grad_norm": 0.1416015625, "learning_rate": 8.470728793309439e-06, "loss": 1.2508, "step": 129 }, { "epoch": 0.15531660692951016, "grad_norm": 0.11376953125, "learning_rate": 8.45878136200717e-06, "loss": 1.2601, "step": 130 }, { "epoch": 0.15651135005973715, "grad_norm": 0.171875, "learning_rate": 8.4468339307049e-06, "loss": 1.3102, "step": 131 }, { "epoch": 0.15770609318996415, "grad_norm": 0.14453125, "learning_rate": 8.43488649940263e-06, "loss": 1.3148, "step": 132 }, { "epoch": 0.15890083632019117, "grad_norm": 0.1572265625, "learning_rate": 8.422939068100358e-06, "loss": 1.3949, "step": 133 }, { "epoch": 0.16009557945041816, "grad_norm": 0.1728515625, "learning_rate": 8.410991636798089e-06, "loss": 1.3004, "step": 134 }, { "epoch": 0.16129032258064516, "grad_norm": 0.1376953125, "learning_rate": 8.399044205495818e-06, "loss": 1.3714, "step": 135 }, { "epoch": 0.16248506571087215, "grad_norm": 0.1962890625, "learning_rate": 8.387096774193549e-06, "loss": 1.5332, "step": 136 }, { "epoch": 0.16367980884109917, "grad_norm": 0.1630859375, "learning_rate": 8.37514934289128e-06, "loss": 1.4755, "step": 137 }, { "epoch": 0.16487455197132617, "grad_norm": 0.142578125, "learning_rate": 8.36320191158901e-06, "loss": 1.2918, "step": 138 }, { "epoch": 0.16606929510155316, "grad_norm": 0.25, "learning_rate": 8.35125448028674e-06, "loss": 1.4565, "step": 139 }, { "epoch": 0.16726403823178015, "grad_norm": 0.1806640625, "learning_rate": 8.339307048984468e-06, "loss": 1.3773, "step": 140 }, { "epoch": 0.16845878136200718, "grad_norm": 0.1328125, "learning_rate": 8.327359617682199e-06, "loss": 1.2499, "step": 141 }, { "epoch": 0.16965352449223417, "grad_norm": 0.1826171875, "learning_rate": 8.315412186379928e-06, "loss": 1.6546, "step": 142 }, { "epoch": 0.17084826762246116, "grad_norm": 0.1298828125, "learning_rate": 8.303464755077659e-06, "loss": 1.1191, "step": 143 }, { "epoch": 0.17204301075268819, "grad_norm": 0.12890625, "learning_rate": 8.291517323775388e-06, "loss": 1.1895, "step": 144 }, { "epoch": 0.17323775388291518, "grad_norm": 0.1455078125, "learning_rate": 8.279569892473119e-06, "loss": 1.2718, "step": 145 }, { "epoch": 0.17443249701314217, "grad_norm": 0.150390625, "learning_rate": 8.26762246117085e-06, "loss": 1.2813, "step": 146 }, { "epoch": 0.17562724014336917, "grad_norm": 0.16796875, "learning_rate": 8.255675029868578e-06, "loss": 1.1815, "step": 147 }, { "epoch": 0.1768219832735962, "grad_norm": 0.1416015625, "learning_rate": 8.24372759856631e-06, "loss": 1.4834, "step": 148 }, { "epoch": 0.17801672640382318, "grad_norm": 0.1357421875, "learning_rate": 8.231780167264038e-06, "loss": 1.2885, "step": 149 }, { "epoch": 0.17921146953405018, "grad_norm": 0.181640625, "learning_rate": 8.219832735961769e-06, "loss": 1.4216, "step": 150 }, { "epoch": 0.18040621266427717, "grad_norm": 0.166015625, "learning_rate": 8.207885304659498e-06, "loss": 1.4332, "step": 151 }, { "epoch": 0.1816009557945042, "grad_norm": 0.142578125, "learning_rate": 8.195937873357229e-06, "loss": 1.3, "step": 152 }, { "epoch": 0.1827956989247312, "grad_norm": 0.140625, "learning_rate": 8.18399044205496e-06, "loss": 1.3436, "step": 153 }, { "epoch": 0.18399044205495818, "grad_norm": 0.171875, "learning_rate": 8.172043010752689e-06, "loss": 1.2863, "step": 154 }, { "epoch": 0.18518518518518517, "grad_norm": 0.1728515625, "learning_rate": 8.16009557945042e-06, "loss": 1.2297, "step": 155 }, { "epoch": 0.1863799283154122, "grad_norm": 0.1611328125, "learning_rate": 8.148148148148148e-06, "loss": 1.2696, "step": 156 }, { "epoch": 0.1875746714456392, "grad_norm": 0.169921875, "learning_rate": 8.136200716845879e-06, "loss": 1.34, "step": 157 }, { "epoch": 0.18876941457586618, "grad_norm": 0.126953125, "learning_rate": 8.124253285543608e-06, "loss": 1.479, "step": 158 }, { "epoch": 0.18996415770609318, "grad_norm": 0.1689453125, "learning_rate": 8.112305854241339e-06, "loss": 1.5052, "step": 159 }, { "epoch": 0.1911589008363202, "grad_norm": 0.130859375, "learning_rate": 8.100358422939068e-06, "loss": 1.2681, "step": 160 }, { "epoch": 0.1923536439665472, "grad_norm": 0.12890625, "learning_rate": 8.088410991636799e-06, "loss": 1.1317, "step": 161 }, { "epoch": 0.1935483870967742, "grad_norm": 0.17578125, "learning_rate": 8.07646356033453e-06, "loss": 1.1166, "step": 162 }, { "epoch": 0.19474313022700118, "grad_norm": 0.16796875, "learning_rate": 8.064516129032258e-06, "loss": 1.2446, "step": 163 }, { "epoch": 0.1959378733572282, "grad_norm": 0.14453125, "learning_rate": 8.052568697729989e-06, "loss": 1.3049, "step": 164 }, { "epoch": 0.1971326164874552, "grad_norm": 0.17578125, "learning_rate": 8.040621266427718e-06, "loss": 1.2346, "step": 165 }, { "epoch": 0.1983273596176822, "grad_norm": 0.162109375, "learning_rate": 8.028673835125449e-06, "loss": 1.4159, "step": 166 }, { "epoch": 0.19952210274790919, "grad_norm": 0.1201171875, "learning_rate": 8.016726403823178e-06, "loss": 1.1978, "step": 167 }, { "epoch": 0.2007168458781362, "grad_norm": 0.15234375, "learning_rate": 8.004778972520909e-06, "loss": 1.2876, "step": 168 }, { "epoch": 0.2019115890083632, "grad_norm": 0.1279296875, "learning_rate": 7.992831541218638e-06, "loss": 1.2175, "step": 169 }, { "epoch": 0.2031063321385902, "grad_norm": 0.171875, "learning_rate": 7.980884109916368e-06, "loss": 1.4345, "step": 170 }, { "epoch": 0.20430107526881722, "grad_norm": 0.18359375, "learning_rate": 7.9689366786141e-06, "loss": 1.2738, "step": 171 }, { "epoch": 0.2054958183990442, "grad_norm": 0.115234375, "learning_rate": 7.956989247311828e-06, "loss": 1.0421, "step": 172 }, { "epoch": 0.2066905615292712, "grad_norm": 0.1533203125, "learning_rate": 7.945041816009559e-06, "loss": 1.2782, "step": 173 }, { "epoch": 0.2078853046594982, "grad_norm": 0.2119140625, "learning_rate": 7.933094384707288e-06, "loss": 1.4168, "step": 174 }, { "epoch": 0.20908004778972522, "grad_norm": 0.17578125, "learning_rate": 7.921146953405019e-06, "loss": 1.3737, "step": 175 }, { "epoch": 0.21027479091995221, "grad_norm": 0.1513671875, "learning_rate": 7.909199522102748e-06, "loss": 1.3603, "step": 176 }, { "epoch": 0.2114695340501792, "grad_norm": 0.1611328125, "learning_rate": 7.897252090800479e-06, "loss": 1.2961, "step": 177 }, { "epoch": 0.2126642771804062, "grad_norm": 0.1728515625, "learning_rate": 7.88530465949821e-06, "loss": 1.1647, "step": 178 }, { "epoch": 0.21385902031063322, "grad_norm": 0.142578125, "learning_rate": 7.873357228195938e-06, "loss": 1.1951, "step": 179 }, { "epoch": 0.21505376344086022, "grad_norm": 0.1708984375, "learning_rate": 7.861409796893669e-06, "loss": 1.3549, "step": 180 }, { "epoch": 0.2162485065710872, "grad_norm": 0.1162109375, "learning_rate": 7.849462365591398e-06, "loss": 1.3461, "step": 181 }, { "epoch": 0.2174432497013142, "grad_norm": 0.1484375, "learning_rate": 7.837514934289129e-06, "loss": 1.3462, "step": 182 }, { "epoch": 0.21863799283154123, "grad_norm": 0.1533203125, "learning_rate": 7.825567502986858e-06, "loss": 1.1902, "step": 183 }, { "epoch": 0.21983273596176822, "grad_norm": 0.1748046875, "learning_rate": 7.813620071684589e-06, "loss": 1.3002, "step": 184 }, { "epoch": 0.22102747909199522, "grad_norm": 0.1513671875, "learning_rate": 7.801672640382318e-06, "loss": 1.2242, "step": 185 }, { "epoch": 0.2222222222222222, "grad_norm": 0.1689453125, "learning_rate": 7.789725209080048e-06, "loss": 1.2705, "step": 186 }, { "epoch": 0.22341696535244923, "grad_norm": 0.146484375, "learning_rate": 7.77777777777778e-06, "loss": 1.2903, "step": 187 }, { "epoch": 0.22461170848267623, "grad_norm": 0.189453125, "learning_rate": 7.765830346475508e-06, "loss": 1.2576, "step": 188 }, { "epoch": 0.22580645161290322, "grad_norm": 0.171875, "learning_rate": 7.753882915173239e-06, "loss": 1.3013, "step": 189 }, { "epoch": 0.2270011947431302, "grad_norm": 0.123046875, "learning_rate": 7.741935483870968e-06, "loss": 1.0837, "step": 190 }, { "epoch": 0.22819593787335724, "grad_norm": 0.1435546875, "learning_rate": 7.729988052568699e-06, "loss": 1.1949, "step": 191 }, { "epoch": 0.22939068100358423, "grad_norm": 0.1279296875, "learning_rate": 7.718040621266428e-06, "loss": 1.4139, "step": 192 }, { "epoch": 0.23058542413381122, "grad_norm": 0.185546875, "learning_rate": 7.706093189964159e-06, "loss": 1.534, "step": 193 }, { "epoch": 0.23178016726403824, "grad_norm": 0.15625, "learning_rate": 7.694145758661888e-06, "loss": 1.1639, "step": 194 }, { "epoch": 0.23297491039426524, "grad_norm": 0.1533203125, "learning_rate": 7.682198327359618e-06, "loss": 1.4376, "step": 195 }, { "epoch": 0.23416965352449223, "grad_norm": 0.212890625, "learning_rate": 7.670250896057349e-06, "loss": 1.2473, "step": 196 }, { "epoch": 0.23536439665471923, "grad_norm": 0.15234375, "learning_rate": 7.658303464755078e-06, "loss": 1.3324, "step": 197 }, { "epoch": 0.23655913978494625, "grad_norm": 0.169921875, "learning_rate": 7.646356033452809e-06, "loss": 1.3865, "step": 198 }, { "epoch": 0.23775388291517324, "grad_norm": 0.2099609375, "learning_rate": 7.634408602150538e-06, "loss": 1.1739, "step": 199 }, { "epoch": 0.23894862604540024, "grad_norm": 0.1708984375, "learning_rate": 7.622461170848269e-06, "loss": 1.3748, "step": 200 }, { "epoch": 0.24014336917562723, "grad_norm": 0.16015625, "learning_rate": 7.6105137395459985e-06, "loss": 1.3559, "step": 201 }, { "epoch": 0.24133811230585425, "grad_norm": 0.154296875, "learning_rate": 7.5985663082437275e-06, "loss": 1.3006, "step": 202 }, { "epoch": 0.24253285543608125, "grad_norm": 0.1318359375, "learning_rate": 7.586618876941458e-06, "loss": 1.3309, "step": 203 }, { "epoch": 0.24372759856630824, "grad_norm": 0.1435546875, "learning_rate": 7.574671445639188e-06, "loss": 1.2972, "step": 204 }, { "epoch": 0.24492234169653523, "grad_norm": 0.216796875, "learning_rate": 7.562724014336919e-06, "loss": 1.3567, "step": 205 }, { "epoch": 0.24611708482676226, "grad_norm": 0.1279296875, "learning_rate": 7.550776583034648e-06, "loss": 1.3811, "step": 206 }, { "epoch": 0.24731182795698925, "grad_norm": 0.142578125, "learning_rate": 7.538829151732379e-06, "loss": 1.4817, "step": 207 }, { "epoch": 0.24850657108721624, "grad_norm": 0.1865234375, "learning_rate": 7.526881720430108e-06, "loss": 1.2887, "step": 208 }, { "epoch": 0.24970131421744324, "grad_norm": 0.140625, "learning_rate": 7.514934289127838e-06, "loss": 1.3778, "step": 209 }, { "epoch": 0.25089605734767023, "grad_norm": 0.1318359375, "learning_rate": 7.502986857825568e-06, "loss": 1.2292, "step": 210 }, { "epoch": 0.2520908004778972, "grad_norm": 0.140625, "learning_rate": 7.491039426523297e-06, "loss": 1.2065, "step": 211 }, { "epoch": 0.2532855436081243, "grad_norm": 0.1953125, "learning_rate": 7.479091995221028e-06, "loss": 1.2945, "step": 212 }, { "epoch": 0.25448028673835127, "grad_norm": 0.1572265625, "learning_rate": 7.467144563918758e-06, "loss": 1.3488, "step": 213 }, { "epoch": 0.25567502986857826, "grad_norm": 0.1416015625, "learning_rate": 7.455197132616489e-06, "loss": 1.1996, "step": 214 }, { "epoch": 0.25686977299880526, "grad_norm": 0.126953125, "learning_rate": 7.443249701314218e-06, "loss": 1.3442, "step": 215 }, { "epoch": 0.25806451612903225, "grad_norm": 0.138671875, "learning_rate": 7.4313022700119485e-06, "loss": 1.4076, "step": 216 }, { "epoch": 0.25925925925925924, "grad_norm": 0.11669921875, "learning_rate": 7.4193548387096784e-06, "loss": 1.2832, "step": 217 }, { "epoch": 0.26045400238948624, "grad_norm": 0.181640625, "learning_rate": 7.4074074074074075e-06, "loss": 1.2893, "step": 218 }, { "epoch": 0.2616487455197133, "grad_norm": 0.173828125, "learning_rate": 7.395459976105138e-06, "loss": 1.3172, "step": 219 }, { "epoch": 0.2628434886499403, "grad_norm": 0.1494140625, "learning_rate": 7.383512544802868e-06, "loss": 1.1289, "step": 220 }, { "epoch": 0.2640382317801673, "grad_norm": 0.130859375, "learning_rate": 7.371565113500598e-06, "loss": 1.2023, "step": 221 }, { "epoch": 0.26523297491039427, "grad_norm": 0.1552734375, "learning_rate": 7.359617682198328e-06, "loss": 1.5701, "step": 222 }, { "epoch": 0.26642771804062126, "grad_norm": 0.1640625, "learning_rate": 7.347670250896059e-06, "loss": 1.3336, "step": 223 }, { "epoch": 0.26762246117084826, "grad_norm": 0.142578125, "learning_rate": 7.335722819593788e-06, "loss": 1.3593, "step": 224 }, { "epoch": 0.26881720430107525, "grad_norm": 0.1494140625, "learning_rate": 7.3237753882915176e-06, "loss": 1.331, "step": 225 }, { "epoch": 0.27001194743130225, "grad_norm": 0.1572265625, "learning_rate": 7.311827956989248e-06, "loss": 1.202, "step": 226 }, { "epoch": 0.2712066905615293, "grad_norm": 0.162109375, "learning_rate": 7.299880525686977e-06, "loss": 1.2147, "step": 227 }, { "epoch": 0.2724014336917563, "grad_norm": 0.138671875, "learning_rate": 7.287933094384708e-06, "loss": 1.3168, "step": 228 }, { "epoch": 0.2735961768219833, "grad_norm": 0.1298828125, "learning_rate": 7.275985663082438e-06, "loss": 1.1998, "step": 229 }, { "epoch": 0.2747909199522103, "grad_norm": 0.18359375, "learning_rate": 7.264038231780169e-06, "loss": 1.3082, "step": 230 }, { "epoch": 0.27598566308243727, "grad_norm": 0.1826171875, "learning_rate": 7.252090800477898e-06, "loss": 1.2928, "step": 231 }, { "epoch": 0.27718040621266427, "grad_norm": 0.1767578125, "learning_rate": 7.240143369175628e-06, "loss": 1.3262, "step": 232 }, { "epoch": 0.27837514934289126, "grad_norm": 0.169921875, "learning_rate": 7.2281959378733575e-06, "loss": 1.2831, "step": 233 }, { "epoch": 0.27956989247311825, "grad_norm": 0.15234375, "learning_rate": 7.2162485065710874e-06, "loss": 1.3247, "step": 234 }, { "epoch": 0.2807646356033453, "grad_norm": 0.14453125, "learning_rate": 7.204301075268818e-06, "loss": 1.3009, "step": 235 }, { "epoch": 0.2819593787335723, "grad_norm": 0.1728515625, "learning_rate": 7.192353643966547e-06, "loss": 1.2877, "step": 236 }, { "epoch": 0.2831541218637993, "grad_norm": 0.181640625, "learning_rate": 7.180406212664278e-06, "loss": 1.2823, "step": 237 }, { "epoch": 0.2843488649940263, "grad_norm": 0.1650390625, "learning_rate": 7.168458781362008e-06, "loss": 1.2959, "step": 238 }, { "epoch": 0.2855436081242533, "grad_norm": 0.17578125, "learning_rate": 7.156511350059739e-06, "loss": 1.3079, "step": 239 }, { "epoch": 0.2867383512544803, "grad_norm": 0.1923828125, "learning_rate": 7.144563918757468e-06, "loss": 1.2627, "step": 240 }, { "epoch": 0.28793309438470727, "grad_norm": 0.1376953125, "learning_rate": 7.1326164874551975e-06, "loss": 1.2634, "step": 241 }, { "epoch": 0.2891278375149343, "grad_norm": 0.1416015625, "learning_rate": 7.120669056152928e-06, "loss": 1.2925, "step": 242 }, { "epoch": 0.2903225806451613, "grad_norm": 0.1396484375, "learning_rate": 7.108721624850657e-06, "loss": 1.2411, "step": 243 }, { "epoch": 0.2915173237753883, "grad_norm": 0.19921875, "learning_rate": 7.096774193548388e-06, "loss": 1.614, "step": 244 }, { "epoch": 0.2927120669056153, "grad_norm": 0.146484375, "learning_rate": 7.084826762246118e-06, "loss": 1.1332, "step": 245 }, { "epoch": 0.2939068100358423, "grad_norm": 0.16796875, "learning_rate": 7.072879330943848e-06, "loss": 1.4039, "step": 246 }, { "epoch": 0.2951015531660693, "grad_norm": 0.1572265625, "learning_rate": 7.060931899641578e-06, "loss": 1.1856, "step": 247 }, { "epoch": 0.2962962962962963, "grad_norm": 0.1923828125, "learning_rate": 7.048984468339307e-06, "loss": 1.2932, "step": 248 }, { "epoch": 0.2974910394265233, "grad_norm": 0.11767578125, "learning_rate": 7.0370370370370375e-06, "loss": 1.3091, "step": 249 }, { "epoch": 0.2986857825567503, "grad_norm": 0.181640625, "learning_rate": 7.025089605734767e-06, "loss": 1.2378, "step": 250 }, { "epoch": 0.2998805256869773, "grad_norm": 0.1357421875, "learning_rate": 7.013142174432498e-06, "loss": 1.5809, "step": 251 }, { "epoch": 0.3010752688172043, "grad_norm": 0.201171875, "learning_rate": 7.001194743130227e-06, "loss": 1.244, "step": 252 }, { "epoch": 0.3022700119474313, "grad_norm": 0.1689453125, "learning_rate": 6.989247311827958e-06, "loss": 1.6929, "step": 253 }, { "epoch": 0.3034647550776583, "grad_norm": 0.1435546875, "learning_rate": 6.977299880525688e-06, "loss": 1.2964, "step": 254 }, { "epoch": 0.3046594982078853, "grad_norm": 0.166015625, "learning_rate": 6.965352449223417e-06, "loss": 1.2962, "step": 255 }, { "epoch": 0.3058542413381123, "grad_norm": 0.2138671875, "learning_rate": 6.9534050179211476e-06, "loss": 1.2878, "step": 256 }, { "epoch": 0.3070489844683393, "grad_norm": 0.142578125, "learning_rate": 6.9414575866188775e-06, "loss": 1.2781, "step": 257 }, { "epoch": 0.30824372759856633, "grad_norm": 0.1630859375, "learning_rate": 6.929510155316607e-06, "loss": 1.2164, "step": 258 }, { "epoch": 0.3094384707287933, "grad_norm": 0.1689453125, "learning_rate": 6.917562724014337e-06, "loss": 1.2729, "step": 259 }, { "epoch": 0.3106332138590203, "grad_norm": 0.16015625, "learning_rate": 6.905615292712068e-06, "loss": 1.3797, "step": 260 }, { "epoch": 0.3118279569892473, "grad_norm": 0.1474609375, "learning_rate": 6.893667861409797e-06, "loss": 1.348, "step": 261 }, { "epoch": 0.3130227001194743, "grad_norm": 0.11865234375, "learning_rate": 6.881720430107528e-06, "loss": 1.2675, "step": 262 }, { "epoch": 0.3142174432497013, "grad_norm": 0.150390625, "learning_rate": 6.869772998805258e-06, "loss": 1.567, "step": 263 }, { "epoch": 0.3154121863799283, "grad_norm": 0.1494140625, "learning_rate": 6.857825567502987e-06, "loss": 1.2333, "step": 264 }, { "epoch": 0.31660692951015534, "grad_norm": 0.1591796875, "learning_rate": 6.8458781362007174e-06, "loss": 1.5743, "step": 265 }, { "epoch": 0.31780167264038234, "grad_norm": 0.1689453125, "learning_rate": 6.833930704898447e-06, "loss": 1.2533, "step": 266 }, { "epoch": 0.31899641577060933, "grad_norm": 0.1572265625, "learning_rate": 6.821983273596178e-06, "loss": 1.3673, "step": 267 }, { "epoch": 0.3201911589008363, "grad_norm": 0.14453125, "learning_rate": 6.810035842293907e-06, "loss": 1.283, "step": 268 }, { "epoch": 0.3213859020310633, "grad_norm": 0.16015625, "learning_rate": 6.798088410991638e-06, "loss": 1.2981, "step": 269 }, { "epoch": 0.3225806451612903, "grad_norm": 0.181640625, "learning_rate": 6.786140979689368e-06, "loss": 1.4267, "step": 270 }, { "epoch": 0.3237753882915173, "grad_norm": 0.173828125, "learning_rate": 6.774193548387097e-06, "loss": 1.2832, "step": 271 }, { "epoch": 0.3249701314217443, "grad_norm": 0.150390625, "learning_rate": 6.7622461170848275e-06, "loss": 1.2868, "step": 272 }, { "epoch": 0.32616487455197135, "grad_norm": 0.1240234375, "learning_rate": 6.7502986857825566e-06, "loss": 1.2439, "step": 273 }, { "epoch": 0.32735961768219835, "grad_norm": 0.142578125, "learning_rate": 6.738351254480287e-06, "loss": 1.2764, "step": 274 }, { "epoch": 0.32855436081242534, "grad_norm": 0.162109375, "learning_rate": 6.726403823178017e-06, "loss": 1.3295, "step": 275 }, { "epoch": 0.32974910394265233, "grad_norm": 0.1552734375, "learning_rate": 6.714456391875748e-06, "loss": 1.3259, "step": 276 }, { "epoch": 0.3309438470728793, "grad_norm": 0.1748046875, "learning_rate": 6.702508960573477e-06, "loss": 1.3171, "step": 277 }, { "epoch": 0.3321385902031063, "grad_norm": 0.181640625, "learning_rate": 6.690561529271207e-06, "loss": 1.3371, "step": 278 }, { "epoch": 0.3333333333333333, "grad_norm": 0.12890625, "learning_rate": 6.678614097968938e-06, "loss": 1.2428, "step": 279 }, { "epoch": 0.3345280764635603, "grad_norm": 0.166015625, "learning_rate": 6.666666666666667e-06, "loss": 1.487, "step": 280 }, { "epoch": 0.33572281959378736, "grad_norm": 0.169921875, "learning_rate": 6.654719235364397e-06, "loss": 1.2833, "step": 281 }, { "epoch": 0.33691756272401435, "grad_norm": 0.1640625, "learning_rate": 6.642771804062127e-06, "loss": 1.3123, "step": 282 }, { "epoch": 0.33811230585424135, "grad_norm": 0.1455078125, "learning_rate": 6.630824372759857e-06, "loss": 1.2953, "step": 283 }, { "epoch": 0.33930704898446834, "grad_norm": 0.173828125, "learning_rate": 6.618876941457587e-06, "loss": 1.3526, "step": 284 }, { "epoch": 0.34050179211469533, "grad_norm": 0.1455078125, "learning_rate": 6.606929510155318e-06, "loss": 1.3243, "step": 285 }, { "epoch": 0.34169653524492233, "grad_norm": 0.14453125, "learning_rate": 6.594982078853047e-06, "loss": 1.4546, "step": 286 }, { "epoch": 0.3428912783751493, "grad_norm": 0.1845703125, "learning_rate": 6.583034647550777e-06, "loss": 1.4906, "step": 287 }, { "epoch": 0.34408602150537637, "grad_norm": 0.1826171875, "learning_rate": 6.5710872162485075e-06, "loss": 1.1975, "step": 288 }, { "epoch": 0.34528076463560337, "grad_norm": 0.1533203125, "learning_rate": 6.5591397849462365e-06, "loss": 1.1861, "step": 289 }, { "epoch": 0.34647550776583036, "grad_norm": 0.1689453125, "learning_rate": 6.547192353643967e-06, "loss": 1.1677, "step": 290 }, { "epoch": 0.34767025089605735, "grad_norm": 0.150390625, "learning_rate": 6.535244922341697e-06, "loss": 1.1349, "step": 291 }, { "epoch": 0.34886499402628435, "grad_norm": 0.1767578125, "learning_rate": 6.523297491039428e-06, "loss": 1.2388, "step": 292 }, { "epoch": 0.35005973715651134, "grad_norm": 0.15234375, "learning_rate": 6.511350059737157e-06, "loss": 1.2476, "step": 293 }, { "epoch": 0.35125448028673834, "grad_norm": 0.1708984375, "learning_rate": 6.499402628434887e-06, "loss": 1.4268, "step": 294 }, { "epoch": 0.35244922341696533, "grad_norm": 0.1650390625, "learning_rate": 6.4874551971326176e-06, "loss": 1.258, "step": 295 }, { "epoch": 0.3536439665471924, "grad_norm": 0.1806640625, "learning_rate": 6.475507765830347e-06, "loss": 1.2468, "step": 296 }, { "epoch": 0.3548387096774194, "grad_norm": 0.146484375, "learning_rate": 6.463560334528077e-06, "loss": 1.1867, "step": 297 }, { "epoch": 0.35603345280764637, "grad_norm": 0.1962890625, "learning_rate": 6.451612903225806e-06, "loss": 1.299, "step": 298 }, { "epoch": 0.35722819593787336, "grad_norm": 0.2177734375, "learning_rate": 6.439665471923537e-06, "loss": 1.2773, "step": 299 }, { "epoch": 0.35842293906810035, "grad_norm": 0.177734375, "learning_rate": 6.427718040621267e-06, "loss": 1.3814, "step": 300 }, { "epoch": 0.35961768219832735, "grad_norm": 0.1669921875, "learning_rate": 6.415770609318996e-06, "loss": 1.1391, "step": 301 }, { "epoch": 0.36081242532855434, "grad_norm": 0.1962890625, "learning_rate": 6.403823178016727e-06, "loss": 1.2752, "step": 302 }, { "epoch": 0.36200716845878134, "grad_norm": 0.1474609375, "learning_rate": 6.391875746714457e-06, "loss": 1.295, "step": 303 }, { "epoch": 0.3632019115890084, "grad_norm": 0.1474609375, "learning_rate": 6.379928315412187e-06, "loss": 1.238, "step": 304 }, { "epoch": 0.3643966547192354, "grad_norm": 0.1767578125, "learning_rate": 6.3679808841099165e-06, "loss": 1.2734, "step": 305 }, { "epoch": 0.3655913978494624, "grad_norm": 0.1494140625, "learning_rate": 6.356033452807647e-06, "loss": 1.1974, "step": 306 }, { "epoch": 0.36678614097968937, "grad_norm": 0.1474609375, "learning_rate": 6.344086021505377e-06, "loss": 1.0923, "step": 307 }, { "epoch": 0.36798088410991636, "grad_norm": 0.15234375, "learning_rate": 6.332138590203107e-06, "loss": 1.1811, "step": 308 }, { "epoch": 0.36917562724014336, "grad_norm": 0.154296875, "learning_rate": 6.320191158900837e-06, "loss": 1.0611, "step": 309 }, { "epoch": 0.37037037037037035, "grad_norm": 0.1630859375, "learning_rate": 6.308243727598567e-06, "loss": 1.2785, "step": 310 }, { "epoch": 0.3715651135005974, "grad_norm": 0.1552734375, "learning_rate": 6.296296296296297e-06, "loss": 1.2812, "step": 311 }, { "epoch": 0.3727598566308244, "grad_norm": 0.26953125, "learning_rate": 6.2843488649940265e-06, "loss": 1.4601, "step": 312 }, { "epoch": 0.3739545997610514, "grad_norm": 0.1455078125, "learning_rate": 6.272401433691757e-06, "loss": 1.2725, "step": 313 }, { "epoch": 0.3751493428912784, "grad_norm": 0.158203125, "learning_rate": 6.260454002389486e-06, "loss": 1.1564, "step": 314 }, { "epoch": 0.3763440860215054, "grad_norm": 0.130859375, "learning_rate": 6.248506571087217e-06, "loss": 1.2358, "step": 315 }, { "epoch": 0.37753882915173237, "grad_norm": 0.181640625, "learning_rate": 6.236559139784947e-06, "loss": 1.2514, "step": 316 }, { "epoch": 0.37873357228195936, "grad_norm": 0.1650390625, "learning_rate": 6.224611708482676e-06, "loss": 1.2373, "step": 317 }, { "epoch": 0.37992831541218636, "grad_norm": 0.1591796875, "learning_rate": 6.212664277180407e-06, "loss": 1.1799, "step": 318 }, { "epoch": 0.3811230585424134, "grad_norm": 0.169921875, "learning_rate": 6.200716845878137e-06, "loss": 1.3099, "step": 319 }, { "epoch": 0.3823178016726404, "grad_norm": 0.1376953125, "learning_rate": 6.188769414575867e-06, "loss": 1.3336, "step": 320 }, { "epoch": 0.3835125448028674, "grad_norm": 0.1533203125, "learning_rate": 6.176821983273596e-06, "loss": 1.2955, "step": 321 }, { "epoch": 0.3847072879330944, "grad_norm": 0.146484375, "learning_rate": 6.164874551971327e-06, "loss": 1.2624, "step": 322 }, { "epoch": 0.3859020310633214, "grad_norm": 0.1806640625, "learning_rate": 6.152927120669057e-06, "loss": 1.2467, "step": 323 }, { "epoch": 0.3870967741935484, "grad_norm": 0.1591796875, "learning_rate": 6.140979689366786e-06, "loss": 1.2755, "step": 324 }, { "epoch": 0.38829151732377537, "grad_norm": 0.1435546875, "learning_rate": 6.129032258064517e-06, "loss": 1.3667, "step": 325 }, { "epoch": 0.38948626045400236, "grad_norm": 0.1337890625, "learning_rate": 6.117084826762246e-06, "loss": 1.2913, "step": 326 }, { "epoch": 0.3906810035842294, "grad_norm": 0.15234375, "learning_rate": 6.105137395459977e-06, "loss": 1.2037, "step": 327 }, { "epoch": 0.3918757467144564, "grad_norm": 0.2177734375, "learning_rate": 6.0931899641577065e-06, "loss": 1.3325, "step": 328 }, { "epoch": 0.3930704898446834, "grad_norm": 0.162109375, "learning_rate": 6.081242532855437e-06, "loss": 1.2265, "step": 329 }, { "epoch": 0.3942652329749104, "grad_norm": 0.1533203125, "learning_rate": 6.069295101553166e-06, "loss": 1.1603, "step": 330 }, { "epoch": 0.3954599761051374, "grad_norm": 0.173828125, "learning_rate": 6.057347670250897e-06, "loss": 1.2027, "step": 331 }, { "epoch": 0.3966547192353644, "grad_norm": 0.1748046875, "learning_rate": 6.045400238948627e-06, "loss": 1.3296, "step": 332 }, { "epoch": 0.3978494623655914, "grad_norm": 0.169921875, "learning_rate": 6.033452807646356e-06, "loss": 1.3252, "step": 333 }, { "epoch": 0.39904420549581837, "grad_norm": 0.1494140625, "learning_rate": 6.021505376344087e-06, "loss": 1.3507, "step": 334 }, { "epoch": 0.4002389486260454, "grad_norm": 0.14453125, "learning_rate": 6.009557945041817e-06, "loss": 1.5733, "step": 335 }, { "epoch": 0.4014336917562724, "grad_norm": 0.1513671875, "learning_rate": 5.9976105137395465e-06, "loss": 1.1916, "step": 336 }, { "epoch": 0.4026284348864994, "grad_norm": 1.484375, "learning_rate": 5.985663082437276e-06, "loss": 1.3269, "step": 337 }, { "epoch": 0.4038231780167264, "grad_norm": 0.1552734375, "learning_rate": 5.973715651135007e-06, "loss": 1.3344, "step": 338 }, { "epoch": 0.4050179211469534, "grad_norm": 0.11474609375, "learning_rate": 5.961768219832736e-06, "loss": 1.2201, "step": 339 }, { "epoch": 0.4062126642771804, "grad_norm": 0.1572265625, "learning_rate": 5.949820788530466e-06, "loss": 1.2989, "step": 340 }, { "epoch": 0.4074074074074074, "grad_norm": 0.162109375, "learning_rate": 5.937873357228197e-06, "loss": 1.3398, "step": 341 }, { "epoch": 0.40860215053763443, "grad_norm": 0.150390625, "learning_rate": 5.925925925925926e-06, "loss": 1.2762, "step": 342 }, { "epoch": 0.40979689366786143, "grad_norm": 0.171875, "learning_rate": 5.9139784946236566e-06, "loss": 1.4253, "step": 343 }, { "epoch": 0.4109916367980884, "grad_norm": 0.138671875, "learning_rate": 5.9020310633213864e-06, "loss": 1.4116, "step": 344 }, { "epoch": 0.4121863799283154, "grad_norm": 0.1533203125, "learning_rate": 5.890083632019117e-06, "loss": 1.2108, "step": 345 }, { "epoch": 0.4133811230585424, "grad_norm": 0.11328125, "learning_rate": 5.878136200716846e-06, "loss": 1.2885, "step": 346 }, { "epoch": 0.4145758661887694, "grad_norm": 0.154296875, "learning_rate": 5.866188769414576e-06, "loss": 1.3023, "step": 347 }, { "epoch": 0.4157706093189964, "grad_norm": 0.1435546875, "learning_rate": 5.854241338112307e-06, "loss": 1.2373, "step": 348 }, { "epoch": 0.4169653524492234, "grad_norm": 0.181640625, "learning_rate": 5.842293906810036e-06, "loss": 1.2922, "step": 349 }, { "epoch": 0.41816009557945044, "grad_norm": 0.201171875, "learning_rate": 5.830346475507767e-06, "loss": 1.4338, "step": 350 }, { "epoch": 0.41935483870967744, "grad_norm": 0.1845703125, "learning_rate": 5.818399044205496e-06, "loss": 1.5826, "step": 351 }, { "epoch": 0.42054958183990443, "grad_norm": 0.189453125, "learning_rate": 5.806451612903226e-06, "loss": 1.2975, "step": 352 }, { "epoch": 0.4217443249701314, "grad_norm": 0.1533203125, "learning_rate": 5.794504181600956e-06, "loss": 1.3193, "step": 353 }, { "epoch": 0.4229390681003584, "grad_norm": 0.1943359375, "learning_rate": 5.782556750298687e-06, "loss": 1.3176, "step": 354 }, { "epoch": 0.4241338112305854, "grad_norm": 0.16015625, "learning_rate": 5.770609318996416e-06, "loss": 1.2539, "step": 355 }, { "epoch": 0.4253285543608124, "grad_norm": 0.16015625, "learning_rate": 5.758661887694146e-06, "loss": 1.1879, "step": 356 }, { "epoch": 0.4265232974910394, "grad_norm": 0.166015625, "learning_rate": 5.746714456391877e-06, "loss": 1.361, "step": 357 }, { "epoch": 0.42771804062126645, "grad_norm": 0.1650390625, "learning_rate": 5.734767025089606e-06, "loss": 1.0906, "step": 358 }, { "epoch": 0.42891278375149344, "grad_norm": 0.1845703125, "learning_rate": 5.7228195937873365e-06, "loss": 1.2086, "step": 359 }, { "epoch": 0.43010752688172044, "grad_norm": 0.1669921875, "learning_rate": 5.710872162485066e-06, "loss": 1.2157, "step": 360 }, { "epoch": 0.43130227001194743, "grad_norm": 0.1455078125, "learning_rate": 5.698924731182796e-06, "loss": 1.2718, "step": 361 }, { "epoch": 0.4324970131421744, "grad_norm": 0.2041015625, "learning_rate": 5.686977299880526e-06, "loss": 1.329, "step": 362 }, { "epoch": 0.4336917562724014, "grad_norm": 0.1728515625, "learning_rate": 5.675029868578256e-06, "loss": 1.3045, "step": 363 }, { "epoch": 0.4348864994026284, "grad_norm": 0.15234375, "learning_rate": 5.663082437275986e-06, "loss": 1.2349, "step": 364 }, { "epoch": 0.43608124253285546, "grad_norm": 0.1552734375, "learning_rate": 5.651135005973716e-06, "loss": 1.2536, "step": 365 }, { "epoch": 0.43727598566308246, "grad_norm": 0.1416015625, "learning_rate": 5.639187574671447e-06, "loss": 1.1278, "step": 366 }, { "epoch": 0.43847072879330945, "grad_norm": 0.1845703125, "learning_rate": 5.627240143369176e-06, "loss": 1.4194, "step": 367 }, { "epoch": 0.43966547192353644, "grad_norm": 0.1865234375, "learning_rate": 5.615292712066906e-06, "loss": 1.4226, "step": 368 }, { "epoch": 0.44086021505376344, "grad_norm": 0.1484375, "learning_rate": 5.603345280764636e-06, "loss": 1.368, "step": 369 }, { "epoch": 0.44205495818399043, "grad_norm": 0.1962890625, "learning_rate": 5.591397849462365e-06, "loss": 1.3358, "step": 370 }, { "epoch": 0.4432497013142174, "grad_norm": 0.1962890625, "learning_rate": 5.579450418160096e-06, "loss": 1.2234, "step": 371 }, { "epoch": 0.4444444444444444, "grad_norm": 0.1669921875, "learning_rate": 5.567502986857826e-06, "loss": 1.1269, "step": 372 }, { "epoch": 0.44563918757467147, "grad_norm": 0.138671875, "learning_rate": 5.555555555555557e-06, "loss": 1.3428, "step": 373 }, { "epoch": 0.44683393070489846, "grad_norm": 0.140625, "learning_rate": 5.543608124253286e-06, "loss": 1.3882, "step": 374 }, { "epoch": 0.44802867383512546, "grad_norm": 0.1923828125, "learning_rate": 5.5316606929510165e-06, "loss": 1.2331, "step": 375 }, { "epoch": 0.44922341696535245, "grad_norm": 0.2138671875, "learning_rate": 5.5197132616487455e-06, "loss": 1.2352, "step": 376 }, { "epoch": 0.45041816009557945, "grad_norm": 0.189453125, "learning_rate": 5.507765830346476e-06, "loss": 1.4254, "step": 377 }, { "epoch": 0.45161290322580644, "grad_norm": 0.185546875, "learning_rate": 5.495818399044206e-06, "loss": 1.3285, "step": 378 }, { "epoch": 0.45280764635603343, "grad_norm": 0.1474609375, "learning_rate": 5.483870967741935e-06, "loss": 1.2513, "step": 379 }, { "epoch": 0.4540023894862604, "grad_norm": 0.158203125, "learning_rate": 5.471923536439666e-06, "loss": 1.3586, "step": 380 }, { "epoch": 0.4551971326164875, "grad_norm": 0.1640625, "learning_rate": 5.459976105137396e-06, "loss": 1.3535, "step": 381 }, { "epoch": 0.45639187574671447, "grad_norm": 0.158203125, "learning_rate": 5.4480286738351265e-06, "loss": 1.3762, "step": 382 }, { "epoch": 0.45758661887694146, "grad_norm": 0.1728515625, "learning_rate": 5.436081242532856e-06, "loss": 1.1904, "step": 383 }, { "epoch": 0.45878136200716846, "grad_norm": 0.154296875, "learning_rate": 5.424133811230586e-06, "loss": 1.4002, "step": 384 }, { "epoch": 0.45997610513739545, "grad_norm": 0.15625, "learning_rate": 5.412186379928316e-06, "loss": 1.187, "step": 385 }, { "epoch": 0.46117084826762245, "grad_norm": 0.1630859375, "learning_rate": 5.400238948626045e-06, "loss": 1.2398, "step": 386 }, { "epoch": 0.46236559139784944, "grad_norm": 0.171875, "learning_rate": 5.388291517323776e-06, "loss": 1.2184, "step": 387 }, { "epoch": 0.4635603345280765, "grad_norm": 0.1982421875, "learning_rate": 5.376344086021506e-06, "loss": 1.438, "step": 388 }, { "epoch": 0.4647550776583035, "grad_norm": 0.1572265625, "learning_rate": 5.364396654719236e-06, "loss": 1.318, "step": 389 }, { "epoch": 0.4659498207885305, "grad_norm": 0.1904296875, "learning_rate": 5.352449223416966e-06, "loss": 1.2785, "step": 390 }, { "epoch": 0.46714456391875747, "grad_norm": 0.1455078125, "learning_rate": 5.340501792114696e-06, "loss": 1.1807, "step": 391 }, { "epoch": 0.46833930704898447, "grad_norm": 0.18359375, "learning_rate": 5.3285543608124254e-06, "loss": 1.3087, "step": 392 }, { "epoch": 0.46953405017921146, "grad_norm": 0.138671875, "learning_rate": 5.316606929510155e-06, "loss": 1.2896, "step": 393 }, { "epoch": 0.47072879330943845, "grad_norm": 0.1708984375, "learning_rate": 5.304659498207886e-06, "loss": 1.3416, "step": 394 }, { "epoch": 0.47192353643966545, "grad_norm": 0.2001953125, "learning_rate": 5.292712066905615e-06, "loss": 1.1436, "step": 395 }, { "epoch": 0.4731182795698925, "grad_norm": 0.1279296875, "learning_rate": 5.280764635603346e-06, "loss": 1.2203, "step": 396 }, { "epoch": 0.4743130227001195, "grad_norm": 0.1611328125, "learning_rate": 5.268817204301076e-06, "loss": 1.4103, "step": 397 }, { "epoch": 0.4755077658303465, "grad_norm": 0.2021484375, "learning_rate": 5.2568697729988065e-06, "loss": 1.3071, "step": 398 }, { "epoch": 0.4767025089605735, "grad_norm": 0.1669921875, "learning_rate": 5.2449223416965355e-06, "loss": 1.2683, "step": 399 }, { "epoch": 0.4778972520908005, "grad_norm": 0.1806640625, "learning_rate": 5.232974910394266e-06, "loss": 1.2062, "step": 400 }, { "epoch": 0.47909199522102747, "grad_norm": 0.1171875, "learning_rate": 5.221027479091995e-06, "loss": 1.1485, "step": 401 }, { "epoch": 0.48028673835125446, "grad_norm": 0.134765625, "learning_rate": 5.209080047789725e-06, "loss": 1.1335, "step": 402 }, { "epoch": 0.48148148148148145, "grad_norm": 0.1591796875, "learning_rate": 5.197132616487456e-06, "loss": 1.2124, "step": 403 }, { "epoch": 0.4826762246117085, "grad_norm": 0.142578125, "learning_rate": 5.185185185185185e-06, "loss": 1.2114, "step": 404 }, { "epoch": 0.4838709677419355, "grad_norm": 0.1806640625, "learning_rate": 5.173237753882916e-06, "loss": 1.3917, "step": 405 }, { "epoch": 0.4850657108721625, "grad_norm": 0.1494140625, "learning_rate": 5.161290322580646e-06, "loss": 1.2517, "step": 406 }, { "epoch": 0.4862604540023895, "grad_norm": 0.1552734375, "learning_rate": 5.149342891278376e-06, "loss": 1.3838, "step": 407 }, { "epoch": 0.4874551971326165, "grad_norm": 0.1845703125, "learning_rate": 5.137395459976105e-06, "loss": 1.2299, "step": 408 }, { "epoch": 0.4886499402628435, "grad_norm": 0.1337890625, "learning_rate": 5.125448028673835e-06, "loss": 1.3879, "step": 409 }, { "epoch": 0.48984468339307047, "grad_norm": 0.1552734375, "learning_rate": 5.113500597371566e-06, "loss": 1.335, "step": 410 }, { "epoch": 0.4910394265232975, "grad_norm": 0.173828125, "learning_rate": 5.101553166069295e-06, "loss": 1.3968, "step": 411 }, { "epoch": 0.4922341696535245, "grad_norm": 0.1376953125, "learning_rate": 5.089605734767026e-06, "loss": 1.2414, "step": 412 }, { "epoch": 0.4934289127837515, "grad_norm": 0.1533203125, "learning_rate": 5.077658303464756e-06, "loss": 1.4783, "step": 413 }, { "epoch": 0.4946236559139785, "grad_norm": 0.1904296875, "learning_rate": 5.065710872162486e-06, "loss": 1.4038, "step": 414 }, { "epoch": 0.4958183990442055, "grad_norm": 0.1533203125, "learning_rate": 5.0537634408602155e-06, "loss": 1.3558, "step": 415 }, { "epoch": 0.4970131421744325, "grad_norm": 0.1591796875, "learning_rate": 5.0418160095579445e-06, "loss": 1.2472, "step": 416 }, { "epoch": 0.4982078853046595, "grad_norm": 0.1640625, "learning_rate": 5.029868578255675e-06, "loss": 1.2114, "step": 417 }, { "epoch": 0.4994026284348865, "grad_norm": 0.1806640625, "learning_rate": 5.017921146953405e-06, "loss": 1.2461, "step": 418 }, { "epoch": 0.5005973715651135, "grad_norm": 0.177734375, "learning_rate": 5.005973715651136e-06, "loss": 1.31, "step": 419 }, { "epoch": 0.5017921146953405, "grad_norm": 0.19140625, "learning_rate": 4.994026284348865e-06, "loss": 1.1489, "step": 420 }, { "epoch": 0.5029868578255675, "grad_norm": 0.1611328125, "learning_rate": 4.982078853046595e-06, "loss": 1.3762, "step": 421 }, { "epoch": 0.5041816009557945, "grad_norm": 0.169921875, "learning_rate": 4.9701314217443256e-06, "loss": 1.239, "step": 422 }, { "epoch": 0.5053763440860215, "grad_norm": 0.1640625, "learning_rate": 4.9581839904420555e-06, "loss": 1.3169, "step": 423 }, { "epoch": 0.5065710872162486, "grad_norm": 0.2001953125, "learning_rate": 4.946236559139785e-06, "loss": 1.0841, "step": 424 }, { "epoch": 0.5077658303464755, "grad_norm": 0.1484375, "learning_rate": 4.934289127837515e-06, "loss": 1.1415, "step": 425 }, { "epoch": 0.5089605734767025, "grad_norm": 0.154296875, "learning_rate": 4.922341696535245e-06, "loss": 1.3312, "step": 426 }, { "epoch": 0.5101553166069295, "grad_norm": 0.20703125, "learning_rate": 4.910394265232976e-06, "loss": 1.2682, "step": 427 }, { "epoch": 0.5113500597371565, "grad_norm": 0.1728515625, "learning_rate": 4.898446833930705e-06, "loss": 1.3214, "step": 428 }, { "epoch": 0.5125448028673835, "grad_norm": 0.1748046875, "learning_rate": 4.886499402628435e-06, "loss": 1.3794, "step": 429 }, { "epoch": 0.5137395459976105, "grad_norm": 0.138671875, "learning_rate": 4.8745519713261655e-06, "loss": 1.4087, "step": 430 }, { "epoch": 0.5149342891278376, "grad_norm": 0.16015625, "learning_rate": 4.8626045400238954e-06, "loss": 1.2653, "step": 431 }, { "epoch": 0.5161290322580645, "grad_norm": 0.1865234375, "learning_rate": 4.850657108721625e-06, "loss": 1.3382, "step": 432 }, { "epoch": 0.5173237753882916, "grad_norm": 0.12158203125, "learning_rate": 4.838709677419355e-06, "loss": 1.2219, "step": 433 }, { "epoch": 0.5185185185185185, "grad_norm": 0.177734375, "learning_rate": 4.826762246117085e-06, "loss": 1.4112, "step": 434 }, { "epoch": 0.5197132616487455, "grad_norm": 0.64453125, "learning_rate": 4.814814814814815e-06, "loss": 1.2248, "step": 435 }, { "epoch": 0.5209080047789725, "grad_norm": 0.1494140625, "learning_rate": 4.802867383512545e-06, "loss": 1.125, "step": 436 }, { "epoch": 0.5221027479091995, "grad_norm": 0.166015625, "learning_rate": 4.790919952210275e-06, "loss": 1.2466, "step": 437 }, { "epoch": 0.5232974910394266, "grad_norm": 0.1640625, "learning_rate": 4.7789725209080055e-06, "loss": 1.1772, "step": 438 }, { "epoch": 0.5244922341696535, "grad_norm": 0.1591796875, "learning_rate": 4.767025089605735e-06, "loss": 1.2308, "step": 439 }, { "epoch": 0.5256869772998806, "grad_norm": 0.142578125, "learning_rate": 4.755077658303465e-06, "loss": 1.2679, "step": 440 }, { "epoch": 0.5268817204301075, "grad_norm": 0.1669921875, "learning_rate": 4.743130227001195e-06, "loss": 1.184, "step": 441 }, { "epoch": 0.5280764635603346, "grad_norm": 0.2021484375, "learning_rate": 4.731182795698925e-06, "loss": 1.307, "step": 442 }, { "epoch": 0.5292712066905615, "grad_norm": 0.162109375, "learning_rate": 4.719235364396655e-06, "loss": 1.2384, "step": 443 }, { "epoch": 0.5304659498207885, "grad_norm": 0.171875, "learning_rate": 4.707287933094385e-06, "loss": 1.4159, "step": 444 }, { "epoch": 0.5316606929510155, "grad_norm": 0.1767578125, "learning_rate": 4.695340501792115e-06, "loss": 1.3369, "step": 445 }, { "epoch": 0.5328554360812425, "grad_norm": 0.1689453125, "learning_rate": 4.683393070489845e-06, "loss": 1.1344, "step": 446 }, { "epoch": 0.5340501792114696, "grad_norm": 0.1455078125, "learning_rate": 4.671445639187575e-06, "loss": 1.1954, "step": 447 }, { "epoch": 0.5352449223416965, "grad_norm": 0.1982421875, "learning_rate": 4.659498207885305e-06, "loss": 1.2806, "step": 448 }, { "epoch": 0.5364396654719236, "grad_norm": 0.1416015625, "learning_rate": 4.647550776583035e-06, "loss": 1.4307, "step": 449 }, { "epoch": 0.5376344086021505, "grad_norm": 0.1640625, "learning_rate": 4.635603345280765e-06, "loss": 1.2695, "step": 450 }, { "epoch": 0.5388291517323776, "grad_norm": 0.16796875, "learning_rate": 4.623655913978495e-06, "loss": 1.1549, "step": 451 }, { "epoch": 0.5400238948626045, "grad_norm": 0.203125, "learning_rate": 4.611708482676225e-06, "loss": 1.2119, "step": 452 }, { "epoch": 0.5412186379928315, "grad_norm": 0.1533203125, "learning_rate": 4.599761051373955e-06, "loss": 1.083, "step": 453 }, { "epoch": 0.5424133811230586, "grad_norm": 0.2119140625, "learning_rate": 4.587813620071685e-06, "loss": 1.2492, "step": 454 }, { "epoch": 0.5436081242532855, "grad_norm": 0.166015625, "learning_rate": 4.575866188769415e-06, "loss": 1.1171, "step": 455 }, { "epoch": 0.5448028673835126, "grad_norm": 0.177734375, "learning_rate": 4.563918757467145e-06, "loss": 1.4317, "step": 456 }, { "epoch": 0.5459976105137395, "grad_norm": 0.1376953125, "learning_rate": 4.551971326164875e-06, "loss": 1.2578, "step": 457 }, { "epoch": 0.5471923536439666, "grad_norm": 0.1865234375, "learning_rate": 4.540023894862605e-06, "loss": 1.1973, "step": 458 }, { "epoch": 0.5483870967741935, "grad_norm": 0.1416015625, "learning_rate": 4.528076463560335e-06, "loss": 1.1519, "step": 459 }, { "epoch": 0.5495818399044206, "grad_norm": 0.1845703125, "learning_rate": 4.516129032258065e-06, "loss": 1.21, "step": 460 }, { "epoch": 0.5507765830346476, "grad_norm": 0.189453125, "learning_rate": 4.504181600955795e-06, "loss": 1.2306, "step": 461 }, { "epoch": 0.5519713261648745, "grad_norm": 0.1767578125, "learning_rate": 4.492234169653525e-06, "loss": 1.1895, "step": 462 }, { "epoch": 0.5531660692951016, "grad_norm": 0.123046875, "learning_rate": 4.480286738351255e-06, "loss": 1.319, "step": 463 }, { "epoch": 0.5543608124253285, "grad_norm": 0.15625, "learning_rate": 4.468339307048985e-06, "loss": 1.2122, "step": 464 }, { "epoch": 0.5555555555555556, "grad_norm": 0.154296875, "learning_rate": 4.456391875746715e-06, "loss": 1.3262, "step": 465 }, { "epoch": 0.5567502986857825, "grad_norm": 0.1513671875, "learning_rate": 4.444444444444444e-06, "loss": 1.1151, "step": 466 }, { "epoch": 0.5579450418160096, "grad_norm": 0.1689453125, "learning_rate": 4.432497013142175e-06, "loss": 1.3749, "step": 467 }, { "epoch": 0.5591397849462365, "grad_norm": 0.181640625, "learning_rate": 4.420549581839905e-06, "loss": 1.2827, "step": 468 }, { "epoch": 0.5603345280764636, "grad_norm": 0.126953125, "learning_rate": 4.408602150537635e-06, "loss": 1.2817, "step": 469 }, { "epoch": 0.5615292712066906, "grad_norm": 0.16015625, "learning_rate": 4.3966547192353646e-06, "loss": 1.3158, "step": 470 }, { "epoch": 0.5627240143369175, "grad_norm": 0.1591796875, "learning_rate": 4.3847072879330945e-06, "loss": 1.3302, "step": 471 }, { "epoch": 0.5639187574671446, "grad_norm": 0.1591796875, "learning_rate": 4.372759856630825e-06, "loss": 1.4083, "step": 472 }, { "epoch": 0.5651135005973715, "grad_norm": 0.1630859375, "learning_rate": 4.360812425328555e-06, "loss": 1.2664, "step": 473 }, { "epoch": 0.5663082437275986, "grad_norm": 0.166015625, "learning_rate": 4.348864994026284e-06, "loss": 1.1637, "step": 474 }, { "epoch": 0.5675029868578255, "grad_norm": 0.16796875, "learning_rate": 4.336917562724015e-06, "loss": 1.3524, "step": 475 }, { "epoch": 0.5686977299880526, "grad_norm": 0.1796875, "learning_rate": 4.324970131421745e-06, "loss": 1.4541, "step": 476 }, { "epoch": 0.5698924731182796, "grad_norm": 0.158203125, "learning_rate": 4.313022700119475e-06, "loss": 1.1622, "step": 477 }, { "epoch": 0.5710872162485066, "grad_norm": 0.173828125, "learning_rate": 4.3010752688172045e-06, "loss": 1.1623, "step": 478 }, { "epoch": 0.5722819593787336, "grad_norm": 0.15625, "learning_rate": 4.2891278375149344e-06, "loss": 1.292, "step": 479 }, { "epoch": 0.5734767025089605, "grad_norm": 0.14453125, "learning_rate": 4.277180406212665e-06, "loss": 1.2059, "step": 480 }, { "epoch": 0.5746714456391876, "grad_norm": 0.1982421875, "learning_rate": 4.265232974910394e-06, "loss": 1.3813, "step": 481 }, { "epoch": 0.5758661887694145, "grad_norm": 0.13671875, "learning_rate": 4.253285543608124e-06, "loss": 1.2626, "step": 482 }, { "epoch": 0.5770609318996416, "grad_norm": 0.25, "learning_rate": 4.241338112305855e-06, "loss": 1.155, "step": 483 }, { "epoch": 0.5782556750298686, "grad_norm": 0.1982421875, "learning_rate": 4.229390681003585e-06, "loss": 1.2747, "step": 484 }, { "epoch": 0.5794504181600956, "grad_norm": 0.1787109375, "learning_rate": 4.217443249701315e-06, "loss": 1.3227, "step": 485 }, { "epoch": 0.5806451612903226, "grad_norm": 0.2060546875, "learning_rate": 4.2054958183990445e-06, "loss": 1.1815, "step": 486 }, { "epoch": 0.5818399044205496, "grad_norm": 0.138671875, "learning_rate": 4.193548387096774e-06, "loss": 1.2139, "step": 487 }, { "epoch": 0.5830346475507766, "grad_norm": 0.134765625, "learning_rate": 4.181600955794505e-06, "loss": 1.2053, "step": 488 }, { "epoch": 0.5842293906810035, "grad_norm": 0.1640625, "learning_rate": 4.169653524492234e-06, "loss": 1.1345, "step": 489 }, { "epoch": 0.5854241338112306, "grad_norm": 0.1806640625, "learning_rate": 4.157706093189964e-06, "loss": 1.2724, "step": 490 }, { "epoch": 0.5866188769414575, "grad_norm": 0.1591796875, "learning_rate": 4.145758661887694e-06, "loss": 1.2087, "step": 491 }, { "epoch": 0.5878136200716846, "grad_norm": 0.2255859375, "learning_rate": 4.133811230585425e-06, "loss": 1.4632, "step": 492 }, { "epoch": 0.5890083632019116, "grad_norm": 0.1943359375, "learning_rate": 4.121863799283155e-06, "loss": 1.3671, "step": 493 }, { "epoch": 0.5902031063321386, "grad_norm": 0.1572265625, "learning_rate": 4.1099163679808845e-06, "loss": 1.2642, "step": 494 }, { "epoch": 0.5913978494623656, "grad_norm": 0.1376953125, "learning_rate": 4.097968936678614e-06, "loss": 1.2404, "step": 495 }, { "epoch": 0.5925925925925926, "grad_norm": 0.173828125, "learning_rate": 4.086021505376344e-06, "loss": 1.2781, "step": 496 }, { "epoch": 0.5937873357228196, "grad_norm": 0.16015625, "learning_rate": 4.074074074074074e-06, "loss": 1.1652, "step": 497 }, { "epoch": 0.5949820788530465, "grad_norm": 0.1845703125, "learning_rate": 4.062126642771804e-06, "loss": 1.3346, "step": 498 }, { "epoch": 0.5961768219832736, "grad_norm": 0.1767578125, "learning_rate": 4.050179211469534e-06, "loss": 1.142, "step": 499 }, { "epoch": 0.5973715651135006, "grad_norm": 0.1884765625, "learning_rate": 4.038231780167265e-06, "loss": 1.3386, "step": 500 }, { "epoch": 0.5985663082437276, "grad_norm": 0.2021484375, "learning_rate": 4.0262843488649946e-06, "loss": 1.3045, "step": 501 }, { "epoch": 0.5997610513739546, "grad_norm": 0.158203125, "learning_rate": 4.0143369175627245e-06, "loss": 1.1961, "step": 502 }, { "epoch": 0.6009557945041816, "grad_norm": 0.1630859375, "learning_rate": 4.002389486260454e-06, "loss": 1.3117, "step": 503 }, { "epoch": 0.6021505376344086, "grad_norm": 0.181640625, "learning_rate": 3.990442054958184e-06, "loss": 1.344, "step": 504 }, { "epoch": 0.6033452807646356, "grad_norm": 0.1875, "learning_rate": 3.978494623655914e-06, "loss": 1.321, "step": 505 }, { "epoch": 0.6045400238948626, "grad_norm": 0.1357421875, "learning_rate": 3.966547192353644e-06, "loss": 1.1921, "step": 506 }, { "epoch": 0.6057347670250897, "grad_norm": 0.154296875, "learning_rate": 3.954599761051374e-06, "loss": 1.2438, "step": 507 }, { "epoch": 0.6069295101553166, "grad_norm": 0.169921875, "learning_rate": 3.942652329749105e-06, "loss": 1.273, "step": 508 }, { "epoch": 0.6081242532855436, "grad_norm": 0.15625, "learning_rate": 3.9307048984468345e-06, "loss": 1.4018, "step": 509 }, { "epoch": 0.6093189964157706, "grad_norm": 0.1552734375, "learning_rate": 3.9187574671445644e-06, "loss": 1.1961, "step": 510 }, { "epoch": 0.6105137395459976, "grad_norm": 0.169921875, "learning_rate": 3.906810035842294e-06, "loss": 1.29, "step": 511 }, { "epoch": 0.6117084826762246, "grad_norm": 0.19921875, "learning_rate": 3.894862604540024e-06, "loss": 1.3009, "step": 512 }, { "epoch": 0.6129032258064516, "grad_norm": 0.1806640625, "learning_rate": 3.882915173237754e-06, "loss": 1.2891, "step": 513 }, { "epoch": 0.6140979689366786, "grad_norm": 0.1435546875, "learning_rate": 3.870967741935484e-06, "loss": 1.2072, "step": 514 }, { "epoch": 0.6152927120669056, "grad_norm": 0.177734375, "learning_rate": 3.859020310633214e-06, "loss": 1.3979, "step": 515 }, { "epoch": 0.6164874551971327, "grad_norm": 0.1630859375, "learning_rate": 3.847072879330944e-06, "loss": 1.2032, "step": 516 }, { "epoch": 0.6176821983273596, "grad_norm": 0.173828125, "learning_rate": 3.8351254480286745e-06, "loss": 1.4091, "step": 517 }, { "epoch": 0.6188769414575866, "grad_norm": 0.173828125, "learning_rate": 3.823178016726404e-06, "loss": 1.1258, "step": 518 }, { "epoch": 0.6200716845878136, "grad_norm": 0.1728515625, "learning_rate": 3.8112305854241343e-06, "loss": 1.2935, "step": 519 }, { "epoch": 0.6212664277180406, "grad_norm": 0.1533203125, "learning_rate": 3.7992831541218638e-06, "loss": 1.0984, "step": 520 }, { "epoch": 0.6224611708482676, "grad_norm": 0.158203125, "learning_rate": 3.787335722819594e-06, "loss": 1.2416, "step": 521 }, { "epoch": 0.6236559139784946, "grad_norm": 0.1748046875, "learning_rate": 3.775388291517324e-06, "loss": 1.3958, "step": 522 }, { "epoch": 0.6248506571087217, "grad_norm": 0.12353515625, "learning_rate": 3.763440860215054e-06, "loss": 1.204, "step": 523 }, { "epoch": 0.6260454002389486, "grad_norm": 0.158203125, "learning_rate": 3.751493428912784e-06, "loss": 1.26, "step": 524 }, { "epoch": 0.6272401433691757, "grad_norm": 0.1748046875, "learning_rate": 3.739545997610514e-06, "loss": 1.2274, "step": 525 }, { "epoch": 0.6284348864994026, "grad_norm": 0.146484375, "learning_rate": 3.7275985663082444e-06, "loss": 1.3944, "step": 526 }, { "epoch": 0.6296296296296297, "grad_norm": 0.2041015625, "learning_rate": 3.7156511350059743e-06, "loss": 1.3079, "step": 527 }, { "epoch": 0.6308243727598566, "grad_norm": 0.1494140625, "learning_rate": 3.7037037037037037e-06, "loss": 1.2741, "step": 528 }, { "epoch": 0.6320191158900836, "grad_norm": 0.171875, "learning_rate": 3.691756272401434e-06, "loss": 1.2152, "step": 529 }, { "epoch": 0.6332138590203107, "grad_norm": 0.12451171875, "learning_rate": 3.679808841099164e-06, "loss": 1.1674, "step": 530 }, { "epoch": 0.6344086021505376, "grad_norm": 0.1796875, "learning_rate": 3.667861409796894e-06, "loss": 1.3538, "step": 531 }, { "epoch": 0.6356033452807647, "grad_norm": 0.1474609375, "learning_rate": 3.655913978494624e-06, "loss": 1.2513, "step": 532 }, { "epoch": 0.6367980884109916, "grad_norm": 0.150390625, "learning_rate": 3.643966547192354e-06, "loss": 1.2384, "step": 533 }, { "epoch": 0.6379928315412187, "grad_norm": 0.150390625, "learning_rate": 3.6320191158900844e-06, "loss": 1.1252, "step": 534 }, { "epoch": 0.6391875746714456, "grad_norm": 0.158203125, "learning_rate": 3.620071684587814e-06, "loss": 1.1529, "step": 535 }, { "epoch": 0.6403823178016727, "grad_norm": 0.1962890625, "learning_rate": 3.6081242532855437e-06, "loss": 1.2982, "step": 536 }, { "epoch": 0.6415770609318996, "grad_norm": 0.150390625, "learning_rate": 3.5961768219832736e-06, "loss": 1.3431, "step": 537 }, { "epoch": 0.6427718040621266, "grad_norm": 0.17578125, "learning_rate": 3.584229390681004e-06, "loss": 1.3001, "step": 538 }, { "epoch": 0.6439665471923537, "grad_norm": 0.154296875, "learning_rate": 3.572281959378734e-06, "loss": 1.1943, "step": 539 }, { "epoch": 0.6451612903225806, "grad_norm": 0.1640625, "learning_rate": 3.560334528076464e-06, "loss": 1.2509, "step": 540 }, { "epoch": 0.6463560334528077, "grad_norm": 0.134765625, "learning_rate": 3.548387096774194e-06, "loss": 1.2274, "step": 541 }, { "epoch": 0.6475507765830346, "grad_norm": 0.15625, "learning_rate": 3.536439665471924e-06, "loss": 1.3466, "step": 542 }, { "epoch": 0.6487455197132617, "grad_norm": 0.224609375, "learning_rate": 3.5244922341696534e-06, "loss": 1.245, "step": 543 }, { "epoch": 0.6499402628434886, "grad_norm": 0.1259765625, "learning_rate": 3.5125448028673837e-06, "loss": 1.2976, "step": 544 }, { "epoch": 0.6511350059737157, "grad_norm": 0.1796875, "learning_rate": 3.5005973715651136e-06, "loss": 1.351, "step": 545 }, { "epoch": 0.6523297491039427, "grad_norm": 0.1728515625, "learning_rate": 3.488649940262844e-06, "loss": 1.2547, "step": 546 }, { "epoch": 0.6535244922341696, "grad_norm": 0.1689453125, "learning_rate": 3.4767025089605738e-06, "loss": 1.2494, "step": 547 }, { "epoch": 0.6547192353643967, "grad_norm": 0.173828125, "learning_rate": 3.4647550776583037e-06, "loss": 1.3359, "step": 548 }, { "epoch": 0.6559139784946236, "grad_norm": 0.171875, "learning_rate": 3.452807646356034e-06, "loss": 1.1594, "step": 549 }, { "epoch": 0.6571087216248507, "grad_norm": 0.140625, "learning_rate": 3.440860215053764e-06, "loss": 1.2571, "step": 550 }, { "epoch": 0.6583034647550776, "grad_norm": 0.1591796875, "learning_rate": 3.4289127837514933e-06, "loss": 1.4732, "step": 551 }, { "epoch": 0.6594982078853047, "grad_norm": 0.1533203125, "learning_rate": 3.4169653524492237e-06, "loss": 1.4168, "step": 552 }, { "epoch": 0.6606929510155317, "grad_norm": 0.1796875, "learning_rate": 3.4050179211469536e-06, "loss": 1.2989, "step": 553 }, { "epoch": 0.6618876941457587, "grad_norm": 0.1591796875, "learning_rate": 3.393070489844684e-06, "loss": 1.2183, "step": 554 }, { "epoch": 0.6630824372759857, "grad_norm": 0.154296875, "learning_rate": 3.3811230585424138e-06, "loss": 1.1948, "step": 555 }, { "epoch": 0.6642771804062126, "grad_norm": 0.1494140625, "learning_rate": 3.3691756272401437e-06, "loss": 1.2031, "step": 556 }, { "epoch": 0.6654719235364397, "grad_norm": 0.1259765625, "learning_rate": 3.357228195937874e-06, "loss": 1.2616, "step": 557 }, { "epoch": 0.6666666666666666, "grad_norm": 0.1689453125, "learning_rate": 3.3452807646356034e-06, "loss": 1.2815, "step": 558 }, { "epoch": 0.6678614097968937, "grad_norm": 0.181640625, "learning_rate": 3.3333333333333333e-06, "loss": 1.2649, "step": 559 }, { "epoch": 0.6690561529271206, "grad_norm": 0.1806640625, "learning_rate": 3.3213859020310636e-06, "loss": 1.3091, "step": 560 }, { "epoch": 0.6702508960573477, "grad_norm": 0.177734375, "learning_rate": 3.3094384707287935e-06, "loss": 1.1911, "step": 561 }, { "epoch": 0.6714456391875747, "grad_norm": 0.1748046875, "learning_rate": 3.2974910394265234e-06, "loss": 1.3408, "step": 562 }, { "epoch": 0.6726403823178017, "grad_norm": 0.1923828125, "learning_rate": 3.2855436081242537e-06, "loss": 1.2503, "step": 563 }, { "epoch": 0.6738351254480287, "grad_norm": 0.173828125, "learning_rate": 3.2735961768219836e-06, "loss": 1.406, "step": 564 }, { "epoch": 0.6750298685782556, "grad_norm": 0.1826171875, "learning_rate": 3.261648745519714e-06, "loss": 1.2744, "step": 565 }, { "epoch": 0.6762246117084827, "grad_norm": 0.177734375, "learning_rate": 3.2497013142174434e-06, "loss": 1.2799, "step": 566 }, { "epoch": 0.6774193548387096, "grad_norm": 0.1826171875, "learning_rate": 3.2377538829151733e-06, "loss": 1.2846, "step": 567 }, { "epoch": 0.6786140979689367, "grad_norm": 0.1630859375, "learning_rate": 3.225806451612903e-06, "loss": 1.2514, "step": 568 }, { "epoch": 0.6798088410991637, "grad_norm": 0.1943359375, "learning_rate": 3.2138590203106335e-06, "loss": 1.1945, "step": 569 }, { "epoch": 0.6810035842293907, "grad_norm": 0.1552734375, "learning_rate": 3.2019115890083634e-06, "loss": 1.2622, "step": 570 }, { "epoch": 0.6821983273596177, "grad_norm": 0.1572265625, "learning_rate": 3.1899641577060937e-06, "loss": 1.2037, "step": 571 }, { "epoch": 0.6833930704898447, "grad_norm": 0.193359375, "learning_rate": 3.1780167264038236e-06, "loss": 1.2939, "step": 572 }, { "epoch": 0.6845878136200717, "grad_norm": 0.146484375, "learning_rate": 3.1660692951015535e-06, "loss": 1.2833, "step": 573 }, { "epoch": 0.6857825567502986, "grad_norm": 0.1787109375, "learning_rate": 3.1541218637992834e-06, "loss": 1.285, "step": 574 }, { "epoch": 0.6869772998805257, "grad_norm": 0.1669921875, "learning_rate": 3.1421744324970133e-06, "loss": 1.3868, "step": 575 }, { "epoch": 0.6881720430107527, "grad_norm": 0.1591796875, "learning_rate": 3.130227001194743e-06, "loss": 1.2575, "step": 576 }, { "epoch": 0.6893667861409797, "grad_norm": 0.17578125, "learning_rate": 3.1182795698924735e-06, "loss": 1.286, "step": 577 }, { "epoch": 0.6905615292712067, "grad_norm": 0.1552734375, "learning_rate": 3.1063321385902034e-06, "loss": 1.2714, "step": 578 }, { "epoch": 0.6917562724014337, "grad_norm": 0.16796875, "learning_rate": 3.0943847072879337e-06, "loss": 1.2799, "step": 579 }, { "epoch": 0.6929510155316607, "grad_norm": 0.13671875, "learning_rate": 3.0824372759856636e-06, "loss": 1.3179, "step": 580 }, { "epoch": 0.6941457586618877, "grad_norm": 0.1689453125, "learning_rate": 3.070489844683393e-06, "loss": 1.3062, "step": 581 }, { "epoch": 0.6953405017921147, "grad_norm": 0.181640625, "learning_rate": 3.058542413381123e-06, "loss": 1.3339, "step": 582 }, { "epoch": 0.6965352449223416, "grad_norm": 0.1953125, "learning_rate": 3.0465949820788532e-06, "loss": 1.2865, "step": 583 }, { "epoch": 0.6977299880525687, "grad_norm": 0.150390625, "learning_rate": 3.034647550776583e-06, "loss": 1.1325, "step": 584 }, { "epoch": 0.6989247311827957, "grad_norm": 0.1640625, "learning_rate": 3.0227001194743135e-06, "loss": 1.389, "step": 585 }, { "epoch": 0.7001194743130227, "grad_norm": 0.173828125, "learning_rate": 3.0107526881720433e-06, "loss": 1.2164, "step": 586 }, { "epoch": 0.7013142174432497, "grad_norm": 0.1650390625, "learning_rate": 2.9988052568697732e-06, "loss": 1.2775, "step": 587 }, { "epoch": 0.7025089605734767, "grad_norm": 0.1572265625, "learning_rate": 2.9868578255675035e-06, "loss": 1.1907, "step": 588 }, { "epoch": 0.7037037037037037, "grad_norm": 0.1513671875, "learning_rate": 2.974910394265233e-06, "loss": 1.4315, "step": 589 }, { "epoch": 0.7048984468339307, "grad_norm": 0.345703125, "learning_rate": 2.962962962962963e-06, "loss": 1.2156, "step": 590 }, { "epoch": 0.7060931899641577, "grad_norm": 0.1689453125, "learning_rate": 2.9510155316606932e-06, "loss": 1.419, "step": 591 }, { "epoch": 0.7072879330943848, "grad_norm": 0.1796875, "learning_rate": 2.939068100358423e-06, "loss": 1.4412, "step": 592 }, { "epoch": 0.7084826762246117, "grad_norm": 0.154296875, "learning_rate": 2.9271206690561534e-06, "loss": 1.2077, "step": 593 }, { "epoch": 0.7096774193548387, "grad_norm": 0.185546875, "learning_rate": 2.9151732377538833e-06, "loss": 1.3387, "step": 594 }, { "epoch": 0.7108721624850657, "grad_norm": 0.2041015625, "learning_rate": 2.903225806451613e-06, "loss": 1.3809, "step": 595 }, { "epoch": 0.7120669056152927, "grad_norm": 0.126953125, "learning_rate": 2.8912783751493435e-06, "loss": 1.117, "step": 596 }, { "epoch": 0.7132616487455197, "grad_norm": 0.169921875, "learning_rate": 2.879330943847073e-06, "loss": 1.1617, "step": 597 }, { "epoch": 0.7144563918757467, "grad_norm": 0.2109375, "learning_rate": 2.867383512544803e-06, "loss": 1.2493, "step": 598 }, { "epoch": 0.7156511350059738, "grad_norm": 0.14453125, "learning_rate": 2.855436081242533e-06, "loss": 1.2606, "step": 599 }, { "epoch": 0.7168458781362007, "grad_norm": 0.189453125, "learning_rate": 2.843488649940263e-06, "loss": 1.3819, "step": 600 }, { "epoch": 0.7180406212664278, "grad_norm": 0.1484375, "learning_rate": 2.831541218637993e-06, "loss": 1.226, "step": 601 }, { "epoch": 0.7192353643966547, "grad_norm": 0.1591796875, "learning_rate": 2.8195937873357233e-06, "loss": 1.1768, "step": 602 }, { "epoch": 0.7204301075268817, "grad_norm": 0.177734375, "learning_rate": 2.807646356033453e-06, "loss": 1.1835, "step": 603 }, { "epoch": 0.7216248506571087, "grad_norm": 0.1806640625, "learning_rate": 2.7956989247311827e-06, "loss": 1.1958, "step": 604 }, { "epoch": 0.7228195937873357, "grad_norm": 0.1650390625, "learning_rate": 2.783751493428913e-06, "loss": 1.1961, "step": 605 }, { "epoch": 0.7240143369175627, "grad_norm": 0.185546875, "learning_rate": 2.771804062126643e-06, "loss": 1.2288, "step": 606 }, { "epoch": 0.7252090800477897, "grad_norm": 0.1689453125, "learning_rate": 2.7598566308243727e-06, "loss": 1.1704, "step": 607 }, { "epoch": 0.7264038231780168, "grad_norm": 0.17578125, "learning_rate": 2.747909199522103e-06, "loss": 1.2517, "step": 608 }, { "epoch": 0.7275985663082437, "grad_norm": 0.140625, "learning_rate": 2.735961768219833e-06, "loss": 1.2844, "step": 609 }, { "epoch": 0.7287933094384708, "grad_norm": 0.1767578125, "learning_rate": 2.7240143369175633e-06, "loss": 1.2427, "step": 610 }, { "epoch": 0.7299880525686977, "grad_norm": 0.1533203125, "learning_rate": 2.712066905615293e-06, "loss": 1.3714, "step": 611 }, { "epoch": 0.7311827956989247, "grad_norm": 0.1611328125, "learning_rate": 2.7001194743130226e-06, "loss": 1.4169, "step": 612 }, { "epoch": 0.7323775388291517, "grad_norm": 0.162109375, "learning_rate": 2.688172043010753e-06, "loss": 1.2067, "step": 613 }, { "epoch": 0.7335722819593787, "grad_norm": 0.1865234375, "learning_rate": 2.676224611708483e-06, "loss": 1.2321, "step": 614 }, { "epoch": 0.7347670250896058, "grad_norm": 0.1630859375, "learning_rate": 2.6642771804062127e-06, "loss": 1.1757, "step": 615 }, { "epoch": 0.7359617682198327, "grad_norm": 0.158203125, "learning_rate": 2.652329749103943e-06, "loss": 1.2508, "step": 616 }, { "epoch": 0.7371565113500598, "grad_norm": 0.185546875, "learning_rate": 2.640382317801673e-06, "loss": 1.2102, "step": 617 }, { "epoch": 0.7383512544802867, "grad_norm": 0.1455078125, "learning_rate": 2.6284348864994032e-06, "loss": 1.1703, "step": 618 }, { "epoch": 0.7395459976105138, "grad_norm": 0.181640625, "learning_rate": 2.616487455197133e-06, "loss": 1.3654, "step": 619 }, { "epoch": 0.7407407407407407, "grad_norm": 0.166015625, "learning_rate": 2.6045400238948626e-06, "loss": 1.2677, "step": 620 }, { "epoch": 0.7419354838709677, "grad_norm": 0.1962890625, "learning_rate": 2.5925925925925925e-06, "loss": 1.3553, "step": 621 }, { "epoch": 0.7431302270011948, "grad_norm": 0.1572265625, "learning_rate": 2.580645161290323e-06, "loss": 1.1447, "step": 622 }, { "epoch": 0.7443249701314217, "grad_norm": 0.1494140625, "learning_rate": 2.5686977299880527e-06, "loss": 1.2364, "step": 623 }, { "epoch": 0.7455197132616488, "grad_norm": 0.140625, "learning_rate": 2.556750298685783e-06, "loss": 1.1665, "step": 624 }, { "epoch": 0.7467144563918757, "grad_norm": 0.173828125, "learning_rate": 2.544802867383513e-06, "loss": 1.2342, "step": 625 }, { "epoch": 0.7479091995221028, "grad_norm": 0.1640625, "learning_rate": 2.532855436081243e-06, "loss": 1.3281, "step": 626 }, { "epoch": 0.7491039426523297, "grad_norm": 0.1923828125, "learning_rate": 2.5209080047789723e-06, "loss": 1.2851, "step": 627 }, { "epoch": 0.7502986857825568, "grad_norm": 0.1962890625, "learning_rate": 2.5089605734767026e-06, "loss": 1.257, "step": 628 }, { "epoch": 0.7514934289127837, "grad_norm": 0.1552734375, "learning_rate": 2.4970131421744325e-06, "loss": 1.1764, "step": 629 }, { "epoch": 0.7526881720430108, "grad_norm": 0.16015625, "learning_rate": 2.4850657108721628e-06, "loss": 1.1137, "step": 630 }, { "epoch": 0.7538829151732378, "grad_norm": 0.2158203125, "learning_rate": 2.4731182795698927e-06, "loss": 1.4247, "step": 631 }, { "epoch": 0.7550776583034647, "grad_norm": 0.1640625, "learning_rate": 2.4611708482676226e-06, "loss": 1.344, "step": 632 }, { "epoch": 0.7562724014336918, "grad_norm": 0.1494140625, "learning_rate": 2.4492234169653525e-06, "loss": 1.1723, "step": 633 }, { "epoch": 0.7574671445639187, "grad_norm": 0.169921875, "learning_rate": 2.4372759856630828e-06, "loss": 1.2698, "step": 634 }, { "epoch": 0.7586618876941458, "grad_norm": 0.158203125, "learning_rate": 2.4253285543608127e-06, "loss": 1.3512, "step": 635 }, { "epoch": 0.7598566308243727, "grad_norm": 0.15625, "learning_rate": 2.4133811230585425e-06, "loss": 1.2468, "step": 636 }, { "epoch": 0.7610513739545998, "grad_norm": 0.1513671875, "learning_rate": 2.4014336917562724e-06, "loss": 1.1423, "step": 637 }, { "epoch": 0.7622461170848268, "grad_norm": 0.15625, "learning_rate": 2.3894862604540028e-06, "loss": 1.1911, "step": 638 }, { "epoch": 0.7634408602150538, "grad_norm": 0.140625, "learning_rate": 2.3775388291517326e-06, "loss": 1.2762, "step": 639 }, { "epoch": 0.7646356033452808, "grad_norm": 0.185546875, "learning_rate": 2.3655913978494625e-06, "loss": 1.3602, "step": 640 }, { "epoch": 0.7658303464755077, "grad_norm": 0.1787109375, "learning_rate": 2.3536439665471924e-06, "loss": 1.2277, "step": 641 }, { "epoch": 0.7670250896057348, "grad_norm": 0.1533203125, "learning_rate": 2.3416965352449223e-06, "loss": 1.3354, "step": 642 }, { "epoch": 0.7682198327359617, "grad_norm": 0.169921875, "learning_rate": 2.3297491039426526e-06, "loss": 1.4165, "step": 643 }, { "epoch": 0.7694145758661888, "grad_norm": 0.1796875, "learning_rate": 2.3178016726403825e-06, "loss": 1.3166, "step": 644 }, { "epoch": 0.7706093189964157, "grad_norm": 0.2080078125, "learning_rate": 2.3058542413381124e-06, "loss": 1.4224, "step": 645 }, { "epoch": 0.7718040621266428, "grad_norm": 0.154296875, "learning_rate": 2.2939068100358423e-06, "loss": 1.2132, "step": 646 }, { "epoch": 0.7729988052568698, "grad_norm": 0.1494140625, "learning_rate": 2.2819593787335726e-06, "loss": 1.2996, "step": 647 }, { "epoch": 0.7741935483870968, "grad_norm": 0.19140625, "learning_rate": 2.2700119474313025e-06, "loss": 1.2563, "step": 648 }, { "epoch": 0.7753882915173238, "grad_norm": 0.158203125, "learning_rate": 2.2580645161290324e-06, "loss": 1.2519, "step": 649 }, { "epoch": 0.7765830346475507, "grad_norm": 0.1494140625, "learning_rate": 2.2461170848267623e-06, "loss": 1.1983, "step": 650 }, { "epoch": 0.7777777777777778, "grad_norm": 0.150390625, "learning_rate": 2.2341696535244926e-06, "loss": 1.1635, "step": 651 }, { "epoch": 0.7789725209080047, "grad_norm": 0.1875, "learning_rate": 2.222222222222222e-06, "loss": 1.2158, "step": 652 }, { "epoch": 0.7801672640382318, "grad_norm": 0.14453125, "learning_rate": 2.2102747909199524e-06, "loss": 1.4534, "step": 653 }, { "epoch": 0.7813620071684588, "grad_norm": 0.1298828125, "learning_rate": 2.1983273596176823e-06, "loss": 1.2348, "step": 654 }, { "epoch": 0.7825567502986858, "grad_norm": 0.181640625, "learning_rate": 2.1863799283154126e-06, "loss": 1.353, "step": 655 }, { "epoch": 0.7837514934289128, "grad_norm": 0.1630859375, "learning_rate": 2.174432497013142e-06, "loss": 1.4101, "step": 656 }, { "epoch": 0.7849462365591398, "grad_norm": 0.1435546875, "learning_rate": 2.1624850657108724e-06, "loss": 1.1962, "step": 657 }, { "epoch": 0.7861409796893668, "grad_norm": 0.158203125, "learning_rate": 2.1505376344086023e-06, "loss": 1.3317, "step": 658 }, { "epoch": 0.7873357228195937, "grad_norm": 0.1728515625, "learning_rate": 2.1385902031063326e-06, "loss": 1.3185, "step": 659 }, { "epoch": 0.7885304659498208, "grad_norm": 0.1494140625, "learning_rate": 2.126642771804062e-06, "loss": 1.4309, "step": 660 }, { "epoch": 0.7897252090800478, "grad_norm": 0.177734375, "learning_rate": 2.1146953405017924e-06, "loss": 1.2009, "step": 661 }, { "epoch": 0.7909199522102748, "grad_norm": 0.18359375, "learning_rate": 2.1027479091995223e-06, "loss": 1.1984, "step": 662 }, { "epoch": 0.7921146953405018, "grad_norm": 0.169921875, "learning_rate": 2.0908004778972526e-06, "loss": 1.0954, "step": 663 }, { "epoch": 0.7933094384707288, "grad_norm": 0.1826171875, "learning_rate": 2.078853046594982e-06, "loss": 1.2221, "step": 664 }, { "epoch": 0.7945041816009558, "grad_norm": 0.1650390625, "learning_rate": 2.0669056152927124e-06, "loss": 1.2591, "step": 665 }, { "epoch": 0.7956989247311828, "grad_norm": 0.1630859375, "learning_rate": 2.0549581839904422e-06, "loss": 1.4182, "step": 666 }, { "epoch": 0.7968936678614098, "grad_norm": 0.1875, "learning_rate": 2.043010752688172e-06, "loss": 1.2673, "step": 667 }, { "epoch": 0.7980884109916367, "grad_norm": 0.1513671875, "learning_rate": 2.031063321385902e-06, "loss": 1.1739, "step": 668 }, { "epoch": 0.7992831541218638, "grad_norm": 0.181640625, "learning_rate": 2.0191158900836323e-06, "loss": 1.2787, "step": 669 }, { "epoch": 0.8004778972520908, "grad_norm": 0.1484375, "learning_rate": 2.0071684587813622e-06, "loss": 1.3246, "step": 670 }, { "epoch": 0.8016726403823178, "grad_norm": 0.1376953125, "learning_rate": 1.995221027479092e-06, "loss": 1.1937, "step": 671 }, { "epoch": 0.8028673835125448, "grad_norm": 0.1845703125, "learning_rate": 1.983273596176822e-06, "loss": 1.2353, "step": 672 }, { "epoch": 0.8040621266427718, "grad_norm": 0.1484375, "learning_rate": 1.9713261648745523e-06, "loss": 1.2745, "step": 673 }, { "epoch": 0.8052568697729988, "grad_norm": 0.142578125, "learning_rate": 1.9593787335722822e-06, "loss": 1.2057, "step": 674 }, { "epoch": 0.8064516129032258, "grad_norm": 0.181640625, "learning_rate": 1.947431302270012e-06, "loss": 1.3725, "step": 675 }, { "epoch": 0.8076463560334528, "grad_norm": 0.1689453125, "learning_rate": 1.935483870967742e-06, "loss": 1.2863, "step": 676 }, { "epoch": 0.8088410991636799, "grad_norm": 0.11767578125, "learning_rate": 1.923536439665472e-06, "loss": 1.1518, "step": 677 }, { "epoch": 0.8100358422939068, "grad_norm": 0.1865234375, "learning_rate": 1.911589008363202e-06, "loss": 1.2749, "step": 678 }, { "epoch": 0.8112305854241338, "grad_norm": 0.1806640625, "learning_rate": 1.8996415770609319e-06, "loss": 1.324, "step": 679 }, { "epoch": 0.8124253285543608, "grad_norm": 0.181640625, "learning_rate": 1.887694145758662e-06, "loss": 1.3186, "step": 680 }, { "epoch": 0.8136200716845878, "grad_norm": 0.2001953125, "learning_rate": 1.875746714456392e-06, "loss": 1.4525, "step": 681 }, { "epoch": 0.8148148148148148, "grad_norm": 0.1953125, "learning_rate": 1.8637992831541222e-06, "loss": 1.2035, "step": 682 }, { "epoch": 0.8160095579450418, "grad_norm": 0.1455078125, "learning_rate": 1.8518518518518519e-06, "loss": 1.2042, "step": 683 }, { "epoch": 0.8172043010752689, "grad_norm": 0.208984375, "learning_rate": 1.839904420549582e-06, "loss": 1.2318, "step": 684 }, { "epoch": 0.8183990442054958, "grad_norm": 0.166015625, "learning_rate": 1.827956989247312e-06, "loss": 1.3962, "step": 685 }, { "epoch": 0.8195937873357229, "grad_norm": 0.1337890625, "learning_rate": 1.8160095579450422e-06, "loss": 1.3468, "step": 686 }, { "epoch": 0.8207885304659498, "grad_norm": 0.1201171875, "learning_rate": 1.8040621266427719e-06, "loss": 1.2415, "step": 687 }, { "epoch": 0.8219832735961768, "grad_norm": 0.294921875, "learning_rate": 1.792114695340502e-06, "loss": 1.1968, "step": 688 }, { "epoch": 0.8231780167264038, "grad_norm": 0.1689453125, "learning_rate": 1.780167264038232e-06, "loss": 1.2276, "step": 689 }, { "epoch": 0.8243727598566308, "grad_norm": 0.1435546875, "learning_rate": 1.768219832735962e-06, "loss": 1.2377, "step": 690 }, { "epoch": 0.8255675029868578, "grad_norm": 0.1591796875, "learning_rate": 1.7562724014336918e-06, "loss": 1.1912, "step": 691 }, { "epoch": 0.8267622461170848, "grad_norm": 0.1689453125, "learning_rate": 1.744324970131422e-06, "loss": 1.1974, "step": 692 }, { "epoch": 0.8279569892473119, "grad_norm": 0.1513671875, "learning_rate": 1.7323775388291518e-06, "loss": 1.4764, "step": 693 }, { "epoch": 0.8291517323775388, "grad_norm": 0.1689453125, "learning_rate": 1.720430107526882e-06, "loss": 1.3297, "step": 694 }, { "epoch": 0.8303464755077659, "grad_norm": 0.177734375, "learning_rate": 1.7084826762246118e-06, "loss": 1.278, "step": 695 }, { "epoch": 0.8315412186379928, "grad_norm": 0.1591796875, "learning_rate": 1.696535244922342e-06, "loss": 1.2004, "step": 696 }, { "epoch": 0.8327359617682198, "grad_norm": 0.1474609375, "learning_rate": 1.6845878136200718e-06, "loss": 1.3295, "step": 697 }, { "epoch": 0.8339307048984468, "grad_norm": 0.1455078125, "learning_rate": 1.6726403823178017e-06, "loss": 1.205, "step": 698 }, { "epoch": 0.8351254480286738, "grad_norm": 0.1376953125, "learning_rate": 1.6606929510155318e-06, "loss": 1.34, "step": 699 }, { "epoch": 0.8363201911589009, "grad_norm": 0.169921875, "learning_rate": 1.6487455197132617e-06, "loss": 1.2214, "step": 700 }, { "epoch": 0.8375149342891278, "grad_norm": 0.15234375, "learning_rate": 1.6367980884109918e-06, "loss": 1.4084, "step": 701 }, { "epoch": 0.8387096774193549, "grad_norm": 0.1826171875, "learning_rate": 1.6248506571087217e-06, "loss": 1.3336, "step": 702 }, { "epoch": 0.8399044205495818, "grad_norm": 0.1748046875, "learning_rate": 1.6129032258064516e-06, "loss": 1.1475, "step": 703 }, { "epoch": 0.8410991636798089, "grad_norm": 0.1767578125, "learning_rate": 1.6009557945041817e-06, "loss": 1.3053, "step": 704 }, { "epoch": 0.8422939068100358, "grad_norm": 0.1826171875, "learning_rate": 1.5890083632019118e-06, "loss": 1.2372, "step": 705 }, { "epoch": 0.8434886499402628, "grad_norm": 0.15234375, "learning_rate": 1.5770609318996417e-06, "loss": 1.2075, "step": 706 }, { "epoch": 0.8446833930704899, "grad_norm": 0.16796875, "learning_rate": 1.5651135005973716e-06, "loss": 1.3597, "step": 707 }, { "epoch": 0.8458781362007168, "grad_norm": 0.205078125, "learning_rate": 1.5531660692951017e-06, "loss": 1.3602, "step": 708 }, { "epoch": 0.8470728793309439, "grad_norm": 0.1630859375, "learning_rate": 1.5412186379928318e-06, "loss": 1.2644, "step": 709 }, { "epoch": 0.8482676224611708, "grad_norm": 0.1796875, "learning_rate": 1.5292712066905615e-06, "loss": 1.3819, "step": 710 }, { "epoch": 0.8494623655913979, "grad_norm": 0.1533203125, "learning_rate": 1.5173237753882916e-06, "loss": 1.0992, "step": 711 }, { "epoch": 0.8506571087216248, "grad_norm": 0.1728515625, "learning_rate": 1.5053763440860217e-06, "loss": 1.2713, "step": 712 }, { "epoch": 0.8518518518518519, "grad_norm": 0.1591796875, "learning_rate": 1.4934289127837518e-06, "loss": 1.4945, "step": 713 }, { "epoch": 0.8530465949820788, "grad_norm": 0.173828125, "learning_rate": 1.4814814814814815e-06, "loss": 1.2899, "step": 714 }, { "epoch": 0.8542413381123058, "grad_norm": 0.1826171875, "learning_rate": 1.4695340501792116e-06, "loss": 1.287, "step": 715 }, { "epoch": 0.8554360812425329, "grad_norm": 0.181640625, "learning_rate": 1.4575866188769417e-06, "loss": 1.2304, "step": 716 }, { "epoch": 0.8566308243727598, "grad_norm": 0.169921875, "learning_rate": 1.4456391875746718e-06, "loss": 1.093, "step": 717 }, { "epoch": 0.8578255675029869, "grad_norm": 0.1748046875, "learning_rate": 1.4336917562724014e-06, "loss": 1.1063, "step": 718 }, { "epoch": 0.8590203106332138, "grad_norm": 0.18359375, "learning_rate": 1.4217443249701315e-06, "loss": 1.2275, "step": 719 }, { "epoch": 0.8602150537634409, "grad_norm": 0.166015625, "learning_rate": 1.4097968936678616e-06, "loss": 1.2148, "step": 720 }, { "epoch": 0.8614097968936678, "grad_norm": 0.1708984375, "learning_rate": 1.3978494623655913e-06, "loss": 1.2378, "step": 721 }, { "epoch": 0.8626045400238949, "grad_norm": 0.18359375, "learning_rate": 1.3859020310633214e-06, "loss": 1.2951, "step": 722 }, { "epoch": 0.8637992831541219, "grad_norm": 0.146484375, "learning_rate": 1.3739545997610515e-06, "loss": 1.3365, "step": 723 }, { "epoch": 0.8649940262843488, "grad_norm": 0.1787109375, "learning_rate": 1.3620071684587816e-06, "loss": 1.2063, "step": 724 }, { "epoch": 0.8661887694145759, "grad_norm": 0.1689453125, "learning_rate": 1.3500597371565113e-06, "loss": 1.2738, "step": 725 }, { "epoch": 0.8673835125448028, "grad_norm": 0.169921875, "learning_rate": 1.3381123058542414e-06, "loss": 1.2565, "step": 726 }, { "epoch": 0.8685782556750299, "grad_norm": 0.1611328125, "learning_rate": 1.3261648745519715e-06, "loss": 1.1646, "step": 727 }, { "epoch": 0.8697729988052568, "grad_norm": 0.1396484375, "learning_rate": 1.3142174432497016e-06, "loss": 1.4782, "step": 728 }, { "epoch": 0.8709677419354839, "grad_norm": 0.1552734375, "learning_rate": 1.3022700119474313e-06, "loss": 1.1693, "step": 729 }, { "epoch": 0.8721624850657109, "grad_norm": 0.169921875, "learning_rate": 1.2903225806451614e-06, "loss": 1.2645, "step": 730 }, { "epoch": 0.8733572281959379, "grad_norm": 0.1533203125, "learning_rate": 1.2783751493428915e-06, "loss": 1.3592, "step": 731 }, { "epoch": 0.8745519713261649, "grad_norm": 0.181640625, "learning_rate": 1.2664277180406214e-06, "loss": 1.2366, "step": 732 }, { "epoch": 0.8757467144563919, "grad_norm": 0.1455078125, "learning_rate": 1.2544802867383513e-06, "loss": 1.4252, "step": 733 }, { "epoch": 0.8769414575866189, "grad_norm": 0.1611328125, "learning_rate": 1.2425328554360814e-06, "loss": 1.2501, "step": 734 }, { "epoch": 0.8781362007168458, "grad_norm": 0.1943359375, "learning_rate": 1.2305854241338113e-06, "loss": 1.3838, "step": 735 }, { "epoch": 0.8793309438470729, "grad_norm": 0.1494140625, "learning_rate": 1.2186379928315414e-06, "loss": 1.4124, "step": 736 }, { "epoch": 0.8805256869772998, "grad_norm": 0.1474609375, "learning_rate": 1.2066905615292713e-06, "loss": 1.374, "step": 737 }, { "epoch": 0.8817204301075269, "grad_norm": 0.158203125, "learning_rate": 1.1947431302270014e-06, "loss": 1.3488, "step": 738 }, { "epoch": 0.8829151732377539, "grad_norm": 0.1572265625, "learning_rate": 1.1827956989247313e-06, "loss": 1.3162, "step": 739 }, { "epoch": 0.8841099163679809, "grad_norm": 0.1552734375, "learning_rate": 1.1708482676224612e-06, "loss": 1.5824, "step": 740 }, { "epoch": 0.8853046594982079, "grad_norm": 0.1689453125, "learning_rate": 1.1589008363201913e-06, "loss": 1.3205, "step": 741 }, { "epoch": 0.8864994026284349, "grad_norm": 0.1650390625, "learning_rate": 1.1469534050179212e-06, "loss": 1.2278, "step": 742 }, { "epoch": 0.8876941457586619, "grad_norm": 0.1767578125, "learning_rate": 1.1350059737156513e-06, "loss": 1.2349, "step": 743 }, { "epoch": 0.8888888888888888, "grad_norm": 0.14453125, "learning_rate": 1.1230585424133811e-06, "loss": 1.3307, "step": 744 }, { "epoch": 0.8900836320191159, "grad_norm": 0.189453125, "learning_rate": 1.111111111111111e-06, "loss": 1.3272, "step": 745 }, { "epoch": 0.8912783751493429, "grad_norm": 0.150390625, "learning_rate": 1.0991636798088411e-06, "loss": 1.3992, "step": 746 }, { "epoch": 0.8924731182795699, "grad_norm": 0.14453125, "learning_rate": 1.087216248506571e-06, "loss": 1.3452, "step": 747 }, { "epoch": 0.8936678614097969, "grad_norm": 0.1806640625, "learning_rate": 1.0752688172043011e-06, "loss": 1.2129, "step": 748 }, { "epoch": 0.8948626045400239, "grad_norm": 0.1748046875, "learning_rate": 1.063321385902031e-06, "loss": 1.239, "step": 749 }, { "epoch": 0.8960573476702509, "grad_norm": 0.2138671875, "learning_rate": 1.0513739545997611e-06, "loss": 1.3921, "step": 750 }, { "epoch": 0.8972520908004779, "grad_norm": 0.1435546875, "learning_rate": 1.039426523297491e-06, "loss": 1.25, "step": 751 }, { "epoch": 0.8984468339307049, "grad_norm": 0.193359375, "learning_rate": 1.0274790919952211e-06, "loss": 1.3116, "step": 752 }, { "epoch": 0.899641577060932, "grad_norm": 0.115234375, "learning_rate": 1.015531660692951e-06, "loss": 1.2386, "step": 753 }, { "epoch": 0.9008363201911589, "grad_norm": 0.1533203125, "learning_rate": 1.0035842293906811e-06, "loss": 1.3241, "step": 754 }, { "epoch": 0.9020310633213859, "grad_norm": 0.189453125, "learning_rate": 9.91636798088411e-07, "loss": 1.4751, "step": 755 }, { "epoch": 0.9032258064516129, "grad_norm": 0.1865234375, "learning_rate": 9.796893667861411e-07, "loss": 1.3457, "step": 756 }, { "epoch": 0.9044205495818399, "grad_norm": 0.2060546875, "learning_rate": 9.67741935483871e-07, "loss": 1.3708, "step": 757 }, { "epoch": 0.9056152927120669, "grad_norm": 0.146484375, "learning_rate": 9.55794504181601e-07, "loss": 1.2647, "step": 758 }, { "epoch": 0.9068100358422939, "grad_norm": 0.1494140625, "learning_rate": 9.43847072879331e-07, "loss": 1.2678, "step": 759 }, { "epoch": 0.9080047789725209, "grad_norm": 0.1728515625, "learning_rate": 9.318996415770611e-07, "loss": 1.3375, "step": 760 }, { "epoch": 0.9091995221027479, "grad_norm": 0.1484375, "learning_rate": 9.19952210274791e-07, "loss": 1.2456, "step": 761 }, { "epoch": 0.910394265232975, "grad_norm": 0.1884765625, "learning_rate": 9.080047789725211e-07, "loss": 1.2339, "step": 762 }, { "epoch": 0.9115890083632019, "grad_norm": 0.140625, "learning_rate": 8.96057347670251e-07, "loss": 1.2939, "step": 763 }, { "epoch": 0.9127837514934289, "grad_norm": 0.1884765625, "learning_rate": 8.84109916367981e-07, "loss": 1.3251, "step": 764 }, { "epoch": 0.9139784946236559, "grad_norm": 0.1552734375, "learning_rate": 8.72162485065711e-07, "loss": 1.2398, "step": 765 }, { "epoch": 0.9151732377538829, "grad_norm": 0.1904296875, "learning_rate": 8.60215053763441e-07, "loss": 1.3111, "step": 766 }, { "epoch": 0.9163679808841099, "grad_norm": 0.1767578125, "learning_rate": 8.48267622461171e-07, "loss": 1.3237, "step": 767 }, { "epoch": 0.9175627240143369, "grad_norm": 0.2060546875, "learning_rate": 8.363201911589009e-07, "loss": 1.3477, "step": 768 }, { "epoch": 0.918757467144564, "grad_norm": 0.158203125, "learning_rate": 8.243727598566309e-07, "loss": 1.2309, "step": 769 }, { "epoch": 0.9199522102747909, "grad_norm": 0.1650390625, "learning_rate": 8.124253285543609e-07, "loss": 1.2078, "step": 770 }, { "epoch": 0.921146953405018, "grad_norm": 0.1748046875, "learning_rate": 8.004778972520908e-07, "loss": 1.3097, "step": 771 }, { "epoch": 0.9223416965352449, "grad_norm": 0.142578125, "learning_rate": 7.885304659498208e-07, "loss": 1.1484, "step": 772 }, { "epoch": 0.9235364396654719, "grad_norm": 0.154296875, "learning_rate": 7.765830346475508e-07, "loss": 1.3577, "step": 773 }, { "epoch": 0.9247311827956989, "grad_norm": 0.1533203125, "learning_rate": 7.646356033452807e-07, "loss": 1.3955, "step": 774 }, { "epoch": 0.9259259259259259, "grad_norm": 0.1416015625, "learning_rate": 7.526881720430108e-07, "loss": 1.2418, "step": 775 }, { "epoch": 0.927120669056153, "grad_norm": 0.1767578125, "learning_rate": 7.407407407407407e-07, "loss": 1.2209, "step": 776 }, { "epoch": 0.9283154121863799, "grad_norm": 0.1572265625, "learning_rate": 7.287933094384708e-07, "loss": 1.1761, "step": 777 }, { "epoch": 0.929510155316607, "grad_norm": 0.18359375, "learning_rate": 7.168458781362007e-07, "loss": 1.3824, "step": 778 }, { "epoch": 0.9307048984468339, "grad_norm": 0.16015625, "learning_rate": 7.048984468339308e-07, "loss": 1.2029, "step": 779 }, { "epoch": 0.931899641577061, "grad_norm": 0.16015625, "learning_rate": 6.929510155316607e-07, "loss": 1.1916, "step": 780 }, { "epoch": 0.9330943847072879, "grad_norm": 0.1708984375, "learning_rate": 6.810035842293908e-07, "loss": 1.3474, "step": 781 }, { "epoch": 0.9342891278375149, "grad_norm": 0.2099609375, "learning_rate": 6.690561529271207e-07, "loss": 1.2158, "step": 782 }, { "epoch": 0.9354838709677419, "grad_norm": 0.166015625, "learning_rate": 6.571087216248508e-07, "loss": 1.22, "step": 783 }, { "epoch": 0.9366786140979689, "grad_norm": 0.162109375, "learning_rate": 6.451612903225807e-07, "loss": 1.1201, "step": 784 }, { "epoch": 0.937873357228196, "grad_norm": 0.150390625, "learning_rate": 6.332138590203107e-07, "loss": 1.2379, "step": 785 }, { "epoch": 0.9390681003584229, "grad_norm": 0.1640625, "learning_rate": 6.212664277180407e-07, "loss": 1.2518, "step": 786 }, { "epoch": 0.94026284348865, "grad_norm": 0.15625, "learning_rate": 6.093189964157707e-07, "loss": 1.3023, "step": 787 }, { "epoch": 0.9414575866188769, "grad_norm": 0.1640625, "learning_rate": 5.973715651135007e-07, "loss": 1.226, "step": 788 }, { "epoch": 0.942652329749104, "grad_norm": 0.158203125, "learning_rate": 5.854241338112306e-07, "loss": 1.3075, "step": 789 }, { "epoch": 0.9438470728793309, "grad_norm": 0.1630859375, "learning_rate": 5.734767025089606e-07, "loss": 1.2533, "step": 790 }, { "epoch": 0.945041816009558, "grad_norm": 0.1728515625, "learning_rate": 5.615292712066906e-07, "loss": 1.3121, "step": 791 }, { "epoch": 0.946236559139785, "grad_norm": 0.17578125, "learning_rate": 5.495818399044206e-07, "loss": 1.3095, "step": 792 }, { "epoch": 0.9474313022700119, "grad_norm": 0.169921875, "learning_rate": 5.376344086021506e-07, "loss": 1.1445, "step": 793 }, { "epoch": 0.948626045400239, "grad_norm": 0.158203125, "learning_rate": 5.256869772998806e-07, "loss": 1.1984, "step": 794 }, { "epoch": 0.9498207885304659, "grad_norm": 0.1796875, "learning_rate": 5.137395459976106e-07, "loss": 1.3145, "step": 795 }, { "epoch": 0.951015531660693, "grad_norm": 0.169921875, "learning_rate": 5.017921146953406e-07, "loss": 1.3616, "step": 796 }, { "epoch": 0.9522102747909199, "grad_norm": 0.2578125, "learning_rate": 4.898446833930706e-07, "loss": 1.0834, "step": 797 }, { "epoch": 0.953405017921147, "grad_norm": 0.13671875, "learning_rate": 4.778972520908006e-07, "loss": 1.3056, "step": 798 }, { "epoch": 0.954599761051374, "grad_norm": 0.150390625, "learning_rate": 4.6594982078853055e-07, "loss": 1.0818, "step": 799 }, { "epoch": 0.955794504181601, "grad_norm": 0.1875, "learning_rate": 4.5400238948626054e-07, "loss": 1.5021, "step": 800 }, { "epoch": 0.956989247311828, "grad_norm": 0.15625, "learning_rate": 4.420549581839905e-07, "loss": 1.3175, "step": 801 }, { "epoch": 0.9581839904420549, "grad_norm": 0.1494140625, "learning_rate": 4.301075268817205e-07, "loss": 1.2476, "step": 802 }, { "epoch": 0.959378733572282, "grad_norm": 0.1865234375, "learning_rate": 4.1816009557945043e-07, "loss": 1.3317, "step": 803 }, { "epoch": 0.9605734767025089, "grad_norm": 0.13671875, "learning_rate": 4.062126642771804e-07, "loss": 1.1866, "step": 804 }, { "epoch": 0.961768219832736, "grad_norm": 0.1337890625, "learning_rate": 3.942652329749104e-07, "loss": 1.1995, "step": 805 }, { "epoch": 0.9629629629629629, "grad_norm": 0.13671875, "learning_rate": 3.8231780167264037e-07, "loss": 1.1862, "step": 806 }, { "epoch": 0.96415770609319, "grad_norm": 0.193359375, "learning_rate": 3.7037037037037036e-07, "loss": 1.2252, "step": 807 }, { "epoch": 0.965352449223417, "grad_norm": 0.1474609375, "learning_rate": 3.5842293906810036e-07, "loss": 1.3437, "step": 808 }, { "epoch": 0.966547192353644, "grad_norm": 0.158203125, "learning_rate": 3.4647550776583036e-07, "loss": 1.2078, "step": 809 }, { "epoch": 0.967741935483871, "grad_norm": 0.150390625, "learning_rate": 3.3452807646356035e-07, "loss": 1.1443, "step": 810 }, { "epoch": 0.9689366786140979, "grad_norm": 0.1416015625, "learning_rate": 3.2258064516129035e-07, "loss": 1.1793, "step": 811 }, { "epoch": 0.970131421744325, "grad_norm": 0.1796875, "learning_rate": 3.1063321385902035e-07, "loss": 1.2831, "step": 812 }, { "epoch": 0.9713261648745519, "grad_norm": 0.1845703125, "learning_rate": 2.9868578255675034e-07, "loss": 1.3647, "step": 813 }, { "epoch": 0.972520908004779, "grad_norm": 0.1669921875, "learning_rate": 2.867383512544803e-07, "loss": 1.2307, "step": 814 }, { "epoch": 0.973715651135006, "grad_norm": 0.2001953125, "learning_rate": 2.747909199522103e-07, "loss": 1.1814, "step": 815 }, { "epoch": 0.974910394265233, "grad_norm": 0.130859375, "learning_rate": 2.628434886499403e-07, "loss": 1.149, "step": 816 }, { "epoch": 0.97610513739546, "grad_norm": 0.2216796875, "learning_rate": 2.508960573476703e-07, "loss": 1.1915, "step": 817 }, { "epoch": 0.977299880525687, "grad_norm": 0.2109375, "learning_rate": 2.389486260454003e-07, "loss": 1.2675, "step": 818 }, { "epoch": 0.978494623655914, "grad_norm": 0.447265625, "learning_rate": 2.2700119474313027e-07, "loss": 1.3439, "step": 819 }, { "epoch": 0.9796893667861409, "grad_norm": 0.1748046875, "learning_rate": 2.1505376344086024e-07, "loss": 1.2051, "step": 820 }, { "epoch": 0.980884109916368, "grad_norm": 0.1552734375, "learning_rate": 2.031063321385902e-07, "loss": 1.2599, "step": 821 }, { "epoch": 0.982078853046595, "grad_norm": 0.1708984375, "learning_rate": 1.9115890083632018e-07, "loss": 1.4368, "step": 822 }, { "epoch": 0.983273596176822, "grad_norm": 0.1796875, "learning_rate": 1.7921146953405018e-07, "loss": 1.3741, "step": 823 }, { "epoch": 0.984468339307049, "grad_norm": 0.1787109375, "learning_rate": 1.6726403823178018e-07, "loss": 1.3216, "step": 824 }, { "epoch": 0.985663082437276, "grad_norm": 0.1845703125, "learning_rate": 1.5531660692951017e-07, "loss": 1.6831, "step": 825 }, { "epoch": 0.986857825567503, "grad_norm": 0.17578125, "learning_rate": 1.4336917562724014e-07, "loss": 1.1681, "step": 826 }, { "epoch": 0.98805256869773, "grad_norm": 0.171875, "learning_rate": 1.3142174432497014e-07, "loss": 1.3766, "step": 827 }, { "epoch": 0.989247311827957, "grad_norm": 0.15234375, "learning_rate": 1.1947431302270014e-07, "loss": 1.2671, "step": 828 }, { "epoch": 0.9904420549581839, "grad_norm": 0.17578125, "learning_rate": 1.0752688172043012e-07, "loss": 1.2126, "step": 829 }, { "epoch": 0.991636798088411, "grad_norm": 0.2021484375, "learning_rate": 9.557945041816009e-08, "loss": 1.2978, "step": 830 }, { "epoch": 0.992831541218638, "grad_norm": 0.169921875, "learning_rate": 8.363201911589009e-08, "loss": 1.3779, "step": 831 }, { "epoch": 0.994026284348865, "grad_norm": 0.1728515625, "learning_rate": 7.168458781362007e-08, "loss": 1.1895, "step": 832 }, { "epoch": 0.995221027479092, "grad_norm": 0.16015625, "learning_rate": 5.973715651135007e-08, "loss": 1.4024, "step": 833 }, { "epoch": 0.996415770609319, "grad_norm": 0.1728515625, "learning_rate": 4.7789725209080046e-08, "loss": 1.2619, "step": 834 }, { "epoch": 0.997610513739546, "grad_norm": 0.1669921875, "learning_rate": 3.5842293906810036e-08, "loss": 1.3146, "step": 835 }, { "epoch": 0.998805256869773, "grad_norm": 0.15234375, "learning_rate": 2.3894862604540023e-08, "loss": 1.0386, "step": 836 }, { "epoch": 1.0, "grad_norm": 0.154296875, "learning_rate": 1.1947431302270011e-08, "loss": 1.3097, "step": 837 }, { "epoch": 1.0, "eval_loss": 1.2749302387237549, "eval_runtime": 15.1254, "eval_samples_per_second": 2.843, "eval_steps_per_second": 0.397, "step": 837 } ], "logging_steps": 1.0, "max_steps": 837, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.893193859670475e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }