{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 103.44827586206897, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.034482758620689655, "grad_norm": 8.8125, "learning_rate": 0.0, "loss": 0.7107, "step": 1 }, { "epoch": 0.06896551724137931, "grad_norm": 4.5, "learning_rate": 5.0000000000000004e-08, "loss": 0.7461, "step": 2 }, { "epoch": 0.10344827586206896, "grad_norm": 5.65625, "learning_rate": 1.0000000000000001e-07, "loss": 0.7312, "step": 3 }, { "epoch": 0.13793103448275862, "grad_norm": 12.125, "learning_rate": 1.5000000000000002e-07, "loss": 0.7378, "step": 4 }, { "epoch": 0.1724137931034483, "grad_norm": 3.953125, "learning_rate": 2.0000000000000002e-07, "loss": 0.7383, "step": 5 }, { "epoch": 0.20689655172413793, "grad_norm": 10.5, "learning_rate": 2.5000000000000004e-07, "loss": 0.7482, "step": 6 }, { "epoch": 0.2413793103448276, "grad_norm": 14.75, "learning_rate": 3.0000000000000004e-07, "loss": 0.7542, "step": 7 }, { "epoch": 0.27586206896551724, "grad_norm": 23.0, "learning_rate": 3.5000000000000004e-07, "loss": 0.7327, "step": 8 }, { "epoch": 0.3103448275862069, "grad_norm": 18.25, "learning_rate": 4.0000000000000003e-07, "loss": 0.7592, "step": 9 }, { "epoch": 0.3448275862068966, "grad_norm": 3.84375, "learning_rate": 4.5e-07, "loss": 0.7048, "step": 10 }, { "epoch": 0.3793103448275862, "grad_norm": 16.0, "learning_rate": 5.000000000000001e-07, "loss": 0.7361, "step": 11 }, { "epoch": 0.41379310344827586, "grad_norm": 9.9375, "learning_rate": 5.5e-07, "loss": 0.7368, "step": 12 }, { "epoch": 0.4482758620689655, "grad_norm": 4.75, "learning_rate": 6.000000000000001e-07, "loss": 0.7495, "step": 13 }, { "epoch": 0.4827586206896552, "grad_norm": 6.3125, "learning_rate": 6.5e-07, "loss": 0.7323, "step": 14 }, { "epoch": 0.5172413793103449, "grad_norm": 16.125, "learning_rate": 7.000000000000001e-07, "loss": 0.7342, "step": 15 }, { "epoch": 0.5517241379310345, "grad_norm": 10.6875, "learning_rate": 7.5e-07, "loss": 0.724, "step": 16 }, { "epoch": 0.5862068965517241, "grad_norm": 5.59375, "learning_rate": 8.000000000000001e-07, "loss": 0.7105, "step": 17 }, { "epoch": 0.6206896551724138, "grad_norm": 17.625, "learning_rate": 8.500000000000001e-07, "loss": 0.7394, "step": 18 }, { "epoch": 0.6551724137931034, "grad_norm": 5.21875, "learning_rate": 9e-07, "loss": 0.7115, "step": 19 }, { "epoch": 0.6896551724137931, "grad_norm": 8.1875, "learning_rate": 9.5e-07, "loss": 0.7384, "step": 20 }, { "epoch": 0.7241379310344828, "grad_norm": 4.71875, "learning_rate": 1.0000000000000002e-06, "loss": 0.7343, "step": 21 }, { "epoch": 0.7586206896551724, "grad_norm": 5.0, "learning_rate": 1.0500000000000001e-06, "loss": 0.73, "step": 22 }, { "epoch": 0.7931034482758621, "grad_norm": 4.21875, "learning_rate": 1.1e-06, "loss": 0.7074, "step": 23 }, { "epoch": 0.8275862068965517, "grad_norm": 15.125, "learning_rate": 1.15e-06, "loss": 0.7243, "step": 24 }, { "epoch": 0.8620689655172413, "grad_norm": 6.5, "learning_rate": 1.2000000000000002e-06, "loss": 0.7292, "step": 25 }, { "epoch": 0.896551724137931, "grad_norm": 6.4375, "learning_rate": 1.25e-06, "loss": 0.7178, "step": 26 }, { "epoch": 0.9310344827586207, "grad_norm": 7.84375, "learning_rate": 1.3e-06, "loss": 0.7393, "step": 27 }, { "epoch": 0.9655172413793104, "grad_norm": 8.9375, "learning_rate": 1.35e-06, "loss": 0.7338, "step": 28 }, { "epoch": 1.0, "grad_norm": 17.0, "learning_rate": 1.4000000000000001e-06, "loss": 0.7258, "step": 29 }, { "epoch": 1.0344827586206897, "grad_norm": 4.3125, "learning_rate": 1.45e-06, "loss": 0.7363, "step": 30 }, { "epoch": 1.0689655172413792, "grad_norm": 9.375, "learning_rate": 1.5e-06, "loss": 0.72, "step": 31 }, { "epoch": 1.103448275862069, "grad_norm": 9.9375, "learning_rate": 1.55e-06, "loss": 0.7262, "step": 32 }, { "epoch": 1.1379310344827587, "grad_norm": 5.8125, "learning_rate": 1.6000000000000001e-06, "loss": 0.7216, "step": 33 }, { "epoch": 1.1724137931034484, "grad_norm": 4.84375, "learning_rate": 1.65e-06, "loss": 0.712, "step": 34 }, { "epoch": 1.206896551724138, "grad_norm": 4.65625, "learning_rate": 1.7000000000000002e-06, "loss": 0.7215, "step": 35 }, { "epoch": 1.2413793103448276, "grad_norm": 4.96875, "learning_rate": 1.7500000000000002e-06, "loss": 0.7188, "step": 36 }, { "epoch": 1.2758620689655173, "grad_norm": 3.921875, "learning_rate": 1.8e-06, "loss": 0.7239, "step": 37 }, { "epoch": 1.3103448275862069, "grad_norm": 6.625, "learning_rate": 1.85e-06, "loss": 0.7141, "step": 38 }, { "epoch": 1.3448275862068966, "grad_norm": 10.75, "learning_rate": 1.9e-06, "loss": 0.7113, "step": 39 }, { "epoch": 1.3793103448275863, "grad_norm": 8.625, "learning_rate": 1.95e-06, "loss": 0.7227, "step": 40 }, { "epoch": 1.4137931034482758, "grad_norm": 4.4375, "learning_rate": 2.0000000000000003e-06, "loss": 0.7163, "step": 41 }, { "epoch": 1.4482758620689655, "grad_norm": 4.5, "learning_rate": 2.0500000000000003e-06, "loss": 0.7271, "step": 42 }, { "epoch": 1.4827586206896552, "grad_norm": 4.0, "learning_rate": 2.1000000000000002e-06, "loss": 0.7191, "step": 43 }, { "epoch": 1.5172413793103448, "grad_norm": 4.6875, "learning_rate": 2.1499999999999997e-06, "loss": 0.7146, "step": 44 }, { "epoch": 1.5517241379310345, "grad_norm": 3.53125, "learning_rate": 2.2e-06, "loss": 0.7173, "step": 45 }, { "epoch": 1.5862068965517242, "grad_norm": 5.8125, "learning_rate": 2.25e-06, "loss": 0.7096, "step": 46 }, { "epoch": 1.6206896551724137, "grad_norm": 8.625, "learning_rate": 2.3e-06, "loss": 0.7229, "step": 47 }, { "epoch": 1.6551724137931034, "grad_norm": 4.21875, "learning_rate": 2.35e-06, "loss": 0.7107, "step": 48 }, { "epoch": 1.6896551724137931, "grad_norm": 3.609375, "learning_rate": 2.4000000000000003e-06, "loss": 0.723, "step": 49 }, { "epoch": 1.7241379310344827, "grad_norm": 13.1875, "learning_rate": 2.4500000000000003e-06, "loss": 0.7188, "step": 50 }, { "epoch": 1.7586206896551724, "grad_norm": 12.5, "learning_rate": 2.5e-06, "loss": 0.7112, "step": 51 }, { "epoch": 1.793103448275862, "grad_norm": 4.9375, "learning_rate": 2.55e-06, "loss": 0.711, "step": 52 }, { "epoch": 1.8275862068965516, "grad_norm": 7.0, "learning_rate": 2.6e-06, "loss": 0.7271, "step": 53 }, { "epoch": 1.8620689655172413, "grad_norm": 3.4375, "learning_rate": 2.65e-06, "loss": 0.6981, "step": 54 }, { "epoch": 1.896551724137931, "grad_norm": 7.25, "learning_rate": 2.7e-06, "loss": 0.7138, "step": 55 }, { "epoch": 1.9310344827586206, "grad_norm": 11.9375, "learning_rate": 2.7500000000000004e-06, "loss": 0.7186, "step": 56 }, { "epoch": 1.9655172413793105, "grad_norm": 3.78125, "learning_rate": 2.8000000000000003e-06, "loss": 0.7137, "step": 57 }, { "epoch": 2.0, "grad_norm": 6.1875, "learning_rate": 2.8500000000000002e-06, "loss": 0.7084, "step": 58 }, { "epoch": 2.0344827586206895, "grad_norm": 4.59375, "learning_rate": 2.9e-06, "loss": 0.7096, "step": 59 }, { "epoch": 2.0689655172413794, "grad_norm": 6.09375, "learning_rate": 2.95e-06, "loss": 0.7092, "step": 60 }, { "epoch": 2.103448275862069, "grad_norm": 3.953125, "learning_rate": 3e-06, "loss": 0.7146, "step": 61 }, { "epoch": 2.1379310344827585, "grad_norm": 3.3125, "learning_rate": 3.05e-06, "loss": 0.7, "step": 62 }, { "epoch": 2.1724137931034484, "grad_norm": 4.8125, "learning_rate": 3.1e-06, "loss": 0.705, "step": 63 }, { "epoch": 2.206896551724138, "grad_norm": 11.25, "learning_rate": 3.1500000000000003e-06, "loss": 0.7022, "step": 64 }, { "epoch": 2.2413793103448274, "grad_norm": 6.0, "learning_rate": 3.2000000000000003e-06, "loss": 0.7057, "step": 65 }, { "epoch": 2.2758620689655173, "grad_norm": 5.40625, "learning_rate": 3.2500000000000002e-06, "loss": 0.713, "step": 66 }, { "epoch": 2.310344827586207, "grad_norm": 6.1875, "learning_rate": 3.3e-06, "loss": 0.7208, "step": 67 }, { "epoch": 2.344827586206897, "grad_norm": 6.96875, "learning_rate": 3.3500000000000005e-06, "loss": 0.7093, "step": 68 }, { "epoch": 2.3793103448275863, "grad_norm": 10.5625, "learning_rate": 3.4000000000000005e-06, "loss": 0.7091, "step": 69 }, { "epoch": 2.413793103448276, "grad_norm": 8.4375, "learning_rate": 3.4500000000000004e-06, "loss": 0.7109, "step": 70 }, { "epoch": 2.4482758620689653, "grad_norm": 6.96875, "learning_rate": 3.5000000000000004e-06, "loss": 0.7049, "step": 71 }, { "epoch": 2.4827586206896552, "grad_norm": 9.0625, "learning_rate": 3.55e-06, "loss": 0.6893, "step": 72 }, { "epoch": 2.5172413793103448, "grad_norm": 9.4375, "learning_rate": 3.6e-06, "loss": 0.7085, "step": 73 }, { "epoch": 2.5517241379310347, "grad_norm": 4.4375, "learning_rate": 3.6499999999999998e-06, "loss": 0.6961, "step": 74 }, { "epoch": 2.586206896551724, "grad_norm": 3.765625, "learning_rate": 3.7e-06, "loss": 0.6999, "step": 75 }, { "epoch": 2.6206896551724137, "grad_norm": 11.5625, "learning_rate": 3.75e-06, "loss": 0.7024, "step": 76 }, { "epoch": 2.655172413793103, "grad_norm": 4.625, "learning_rate": 3.8e-06, "loss": 0.6998, "step": 77 }, { "epoch": 2.689655172413793, "grad_norm": 11.4375, "learning_rate": 3.85e-06, "loss": 0.6904, "step": 78 }, { "epoch": 2.7241379310344827, "grad_norm": 29.5, "learning_rate": 3.9e-06, "loss": 0.7219, "step": 79 }, { "epoch": 2.7586206896551726, "grad_norm": 15.0, "learning_rate": 3.95e-06, "loss": 0.7005, "step": 80 }, { "epoch": 2.793103448275862, "grad_norm": 4.875, "learning_rate": 4.000000000000001e-06, "loss": 0.6703, "step": 81 }, { "epoch": 2.8275862068965516, "grad_norm": 6.15625, "learning_rate": 4.05e-06, "loss": 0.6898, "step": 82 }, { "epoch": 2.862068965517241, "grad_norm": 9.0625, "learning_rate": 4.1000000000000006e-06, "loss": 0.7069, "step": 83 }, { "epoch": 2.896551724137931, "grad_norm": 9.625, "learning_rate": 4.15e-06, "loss": 0.6989, "step": 84 }, { "epoch": 2.9310344827586206, "grad_norm": 4.5, "learning_rate": 4.2000000000000004e-06, "loss": 0.6934, "step": 85 }, { "epoch": 2.9655172413793105, "grad_norm": 16.125, "learning_rate": 4.250000000000001e-06, "loss": 0.686, "step": 86 }, { "epoch": 3.0, "grad_norm": 6.4375, "learning_rate": 4.2999999999999995e-06, "loss": 0.6698, "step": 87 }, { "epoch": 3.0344827586206895, "grad_norm": 18.0, "learning_rate": 4.35e-06, "loss": 0.6895, "step": 88 }, { "epoch": 3.0689655172413794, "grad_norm": 6.625, "learning_rate": 4.4e-06, "loss": 0.665, "step": 89 }, { "epoch": 3.103448275862069, "grad_norm": 13.375, "learning_rate": 4.45e-06, "loss": 0.6635, "step": 90 }, { "epoch": 3.1379310344827585, "grad_norm": 10.375, "learning_rate": 4.5e-06, "loss": 0.6505, "step": 91 }, { "epoch": 3.1724137931034484, "grad_norm": 10.3125, "learning_rate": 4.5500000000000005e-06, "loss": 0.6346, "step": 92 }, { "epoch": 3.206896551724138, "grad_norm": 24.375, "learning_rate": 4.6e-06, "loss": 0.6598, "step": 93 }, { "epoch": 3.2413793103448274, "grad_norm": 26.375, "learning_rate": 4.65e-06, "loss": 0.6699, "step": 94 }, { "epoch": 3.2758620689655173, "grad_norm": 16.5, "learning_rate": 4.7e-06, "loss": 0.6123, "step": 95 }, { "epoch": 3.310344827586207, "grad_norm": 9.0625, "learning_rate": 4.75e-06, "loss": 0.5965, "step": 96 }, { "epoch": 3.344827586206897, "grad_norm": 15.125, "learning_rate": 4.800000000000001e-06, "loss": 0.5423, "step": 97 }, { "epoch": 3.3793103448275863, "grad_norm": 20.5, "learning_rate": 4.85e-06, "loss": 0.5838, "step": 98 }, { "epoch": 3.413793103448276, "grad_norm": 47.0, "learning_rate": 4.9000000000000005e-06, "loss": 0.6004, "step": 99 }, { "epoch": 3.4482758620689653, "grad_norm": 44.5, "learning_rate": 4.950000000000001e-06, "loss": 0.5217, "step": 100 }, { "epoch": 3.4827586206896552, "grad_norm": 24.5, "learning_rate": 5e-06, "loss": 0.4938, "step": 101 }, { "epoch": 3.5172413793103448, "grad_norm": 26.125, "learning_rate": 5.050000000000001e-06, "loss": 0.5074, "step": 102 }, { "epoch": 3.5517241379310347, "grad_norm": 30.75, "learning_rate": 5.1e-06, "loss": 0.464, "step": 103 }, { "epoch": 3.586206896551724, "grad_norm": 22.125, "learning_rate": 5.15e-06, "loss": 0.5041, "step": 104 }, { "epoch": 3.6206896551724137, "grad_norm": 22.5, "learning_rate": 5.2e-06, "loss": 0.4965, "step": 105 }, { "epoch": 3.655172413793103, "grad_norm": 19.625, "learning_rate": 5.25e-06, "loss": 0.5263, "step": 106 }, { "epoch": 3.689655172413793, "grad_norm": 31.25, "learning_rate": 5.3e-06, "loss": 0.481, "step": 107 }, { "epoch": 3.7241379310344827, "grad_norm": 54.75, "learning_rate": 5.3500000000000004e-06, "loss": 0.4768, "step": 108 }, { "epoch": 3.7586206896551726, "grad_norm": 45.25, "learning_rate": 5.4e-06, "loss": 0.4056, "step": 109 }, { "epoch": 3.793103448275862, "grad_norm": 70.0, "learning_rate": 5.45e-06, "loss": 0.4997, "step": 110 }, { "epoch": 3.8275862068965516, "grad_norm": 43.0, "learning_rate": 5.500000000000001e-06, "loss": 0.4419, "step": 111 }, { "epoch": 3.862068965517241, "grad_norm": 21.5, "learning_rate": 5.55e-06, "loss": 0.4615, "step": 112 }, { "epoch": 3.896551724137931, "grad_norm": 28.125, "learning_rate": 5.600000000000001e-06, "loss": 0.4312, "step": 113 }, { "epoch": 3.9310344827586206, "grad_norm": 29.75, "learning_rate": 5.65e-06, "loss": 0.4113, "step": 114 }, { "epoch": 3.9655172413793105, "grad_norm": 28.5, "learning_rate": 5.7000000000000005e-06, "loss": 0.3913, "step": 115 }, { "epoch": 4.0, "grad_norm": 32.5, "learning_rate": 5.750000000000001e-06, "loss": 0.437, "step": 116 }, { "epoch": 4.0344827586206895, "grad_norm": 25.125, "learning_rate": 5.8e-06, "loss": 0.3551, "step": 117 }, { "epoch": 4.068965517241379, "grad_norm": 35.25, "learning_rate": 5.850000000000001e-06, "loss": 0.5692, "step": 118 }, { "epoch": 4.103448275862069, "grad_norm": 28.625, "learning_rate": 5.9e-06, "loss": 0.5023, "step": 119 }, { "epoch": 4.137931034482759, "grad_norm": 37.5, "learning_rate": 5.95e-06, "loss": 0.4632, "step": 120 }, { "epoch": 4.172413793103448, "grad_norm": 31.125, "learning_rate": 6e-06, "loss": 0.3953, "step": 121 }, { "epoch": 4.206896551724138, "grad_norm": 28.625, "learning_rate": 6.0500000000000005e-06, "loss": 0.4501, "step": 122 }, { "epoch": 4.241379310344827, "grad_norm": 41.25, "learning_rate": 6.1e-06, "loss": 0.4428, "step": 123 }, { "epoch": 4.275862068965517, "grad_norm": 39.75, "learning_rate": 6.15e-06, "loss": 0.3524, "step": 124 }, { "epoch": 4.310344827586207, "grad_norm": 18.125, "learning_rate": 6.2e-06, "loss": 0.4711, "step": 125 }, { "epoch": 4.344827586206897, "grad_norm": 25.0, "learning_rate": 6.25e-06, "loss": 0.4079, "step": 126 }, { "epoch": 4.379310344827586, "grad_norm": 66.5, "learning_rate": 6.300000000000001e-06, "loss": 0.4782, "step": 127 }, { "epoch": 4.413793103448276, "grad_norm": 79.5, "learning_rate": 6.35e-06, "loss": 0.5137, "step": 128 }, { "epoch": 4.448275862068965, "grad_norm": 48.25, "learning_rate": 6.4000000000000006e-06, "loss": 0.4467, "step": 129 }, { "epoch": 4.482758620689655, "grad_norm": 37.75, "learning_rate": 6.45e-06, "loss": 0.2963, "step": 130 }, { "epoch": 4.517241379310345, "grad_norm": 30.125, "learning_rate": 6.5000000000000004e-06, "loss": 0.2491, "step": 131 }, { "epoch": 4.551724137931035, "grad_norm": 67.0, "learning_rate": 6.550000000000001e-06, "loss": 0.4223, "step": 132 }, { "epoch": 4.586206896551724, "grad_norm": 21.25, "learning_rate": 6.6e-06, "loss": 0.2833, "step": 133 }, { "epoch": 4.620689655172414, "grad_norm": 18.75, "learning_rate": 6.650000000000001e-06, "loss": 0.2986, "step": 134 }, { "epoch": 4.655172413793103, "grad_norm": 17.125, "learning_rate": 6.700000000000001e-06, "loss": 0.3281, "step": 135 }, { "epoch": 4.689655172413794, "grad_norm": 21.875, "learning_rate": 6.750000000000001e-06, "loss": 0.3052, "step": 136 }, { "epoch": 4.724137931034483, "grad_norm": 18.375, "learning_rate": 6.800000000000001e-06, "loss": 0.3726, "step": 137 }, { "epoch": 4.758620689655173, "grad_norm": 23.5, "learning_rate": 6.8500000000000005e-06, "loss": 0.3366, "step": 138 }, { "epoch": 4.793103448275862, "grad_norm": 33.25, "learning_rate": 6.900000000000001e-06, "loss": 0.3349, "step": 139 }, { "epoch": 4.827586206896552, "grad_norm": 60.5, "learning_rate": 6.950000000000001e-06, "loss": 0.3021, "step": 140 }, { "epoch": 4.862068965517241, "grad_norm": 51.5, "learning_rate": 7.000000000000001e-06, "loss": 0.2827, "step": 141 }, { "epoch": 4.896551724137931, "grad_norm": 20.375, "learning_rate": 7.049999999999999e-06, "loss": 0.2363, "step": 142 }, { "epoch": 4.931034482758621, "grad_norm": 22.75, "learning_rate": 7.1e-06, "loss": 0.3175, "step": 143 }, { "epoch": 4.9655172413793105, "grad_norm": 35.5, "learning_rate": 7.15e-06, "loss": 0.3624, "step": 144 }, { "epoch": 5.0, "grad_norm": 35.0, "learning_rate": 7.2e-06, "loss": 0.3818, "step": 145 }, { "epoch": 5.0344827586206895, "grad_norm": 29.875, "learning_rate": 7.25e-06, "loss": 0.3482, "step": 146 }, { "epoch": 5.068965517241379, "grad_norm": 28.625, "learning_rate": 7.2999999999999996e-06, "loss": 0.2633, "step": 147 }, { "epoch": 5.103448275862069, "grad_norm": 25.875, "learning_rate": 7.35e-06, "loss": 0.2787, "step": 148 }, { "epoch": 5.137931034482759, "grad_norm": 23.0, "learning_rate": 7.4e-06, "loss": 0.2883, "step": 149 }, { "epoch": 5.172413793103448, "grad_norm": 37.75, "learning_rate": 7.45e-06, "loss": 0.2635, "step": 150 }, { "epoch": 5.206896551724138, "grad_norm": 32.0, "learning_rate": 7.5e-06, "loss": 0.2844, "step": 151 }, { "epoch": 5.241379310344827, "grad_norm": 26.25, "learning_rate": 7.55e-06, "loss": 0.2835, "step": 152 }, { "epoch": 5.275862068965517, "grad_norm": 52.75, "learning_rate": 7.6e-06, "loss": 0.3202, "step": 153 }, { "epoch": 5.310344827586207, "grad_norm": 29.375, "learning_rate": 7.65e-06, "loss": 0.2282, "step": 154 }, { "epoch": 5.344827586206897, "grad_norm": 50.5, "learning_rate": 7.7e-06, "loss": 0.3416, "step": 155 }, { "epoch": 5.379310344827586, "grad_norm": 22.125, "learning_rate": 7.75e-06, "loss": 0.2205, "step": 156 }, { "epoch": 5.413793103448276, "grad_norm": 35.25, "learning_rate": 7.8e-06, "loss": 0.2582, "step": 157 }, { "epoch": 5.448275862068965, "grad_norm": 53.0, "learning_rate": 7.850000000000001e-06, "loss": 0.3194, "step": 158 }, { "epoch": 5.482758620689655, "grad_norm": 48.75, "learning_rate": 7.9e-06, "loss": 0.3036, "step": 159 }, { "epoch": 5.517241379310345, "grad_norm": 42.25, "learning_rate": 7.95e-06, "loss": 0.3065, "step": 160 }, { "epoch": 5.551724137931035, "grad_norm": 15.875, "learning_rate": 8.000000000000001e-06, "loss": 0.3297, "step": 161 }, { "epoch": 5.586206896551724, "grad_norm": 23.0, "learning_rate": 8.050000000000001e-06, "loss": 0.4744, "step": 162 }, { "epoch": 5.620689655172414, "grad_norm": 21.0, "learning_rate": 8.1e-06, "loss": 0.2616, "step": 163 }, { "epoch": 5.655172413793103, "grad_norm": 18.5, "learning_rate": 8.15e-06, "loss": 0.2868, "step": 164 }, { "epoch": 5.689655172413794, "grad_norm": 25.625, "learning_rate": 8.200000000000001e-06, "loss": 0.2513, "step": 165 }, { "epoch": 5.724137931034483, "grad_norm": 32.25, "learning_rate": 8.25e-06, "loss": 0.4692, "step": 166 }, { "epoch": 5.758620689655173, "grad_norm": 14.5625, "learning_rate": 8.3e-06, "loss": 0.4369, "step": 167 }, { "epoch": 5.793103448275862, "grad_norm": 25.375, "learning_rate": 8.350000000000001e-06, "loss": 0.4037, "step": 168 }, { "epoch": 5.827586206896552, "grad_norm": 26.0, "learning_rate": 8.400000000000001e-06, "loss": 0.3738, "step": 169 }, { "epoch": 5.862068965517241, "grad_norm": 14.1875, "learning_rate": 8.45e-06, "loss": 0.3548, "step": 170 }, { "epoch": 5.896551724137931, "grad_norm": 23.125, "learning_rate": 8.500000000000002e-06, "loss": 0.3732, "step": 171 }, { "epoch": 5.931034482758621, "grad_norm": 21.625, "learning_rate": 8.550000000000001e-06, "loss": 0.3027, "step": 172 }, { "epoch": 5.9655172413793105, "grad_norm": 20.5, "learning_rate": 8.599999999999999e-06, "loss": 0.2649, "step": 173 }, { "epoch": 6.0, "grad_norm": 25.375, "learning_rate": 8.65e-06, "loss": 0.3738, "step": 174 }, { "epoch": 6.0344827586206895, "grad_norm": 19.875, "learning_rate": 8.7e-06, "loss": 0.21, "step": 175 }, { "epoch": 6.068965517241379, "grad_norm": 17.875, "learning_rate": 8.75e-06, "loss": 0.2903, "step": 176 }, { "epoch": 6.103448275862069, "grad_norm": 26.125, "learning_rate": 8.8e-06, "loss": 0.2582, "step": 177 }, { "epoch": 6.137931034482759, "grad_norm": 14.3125, "learning_rate": 8.85e-06, "loss": 0.2323, "step": 178 }, { "epoch": 6.172413793103448, "grad_norm": 25.875, "learning_rate": 8.9e-06, "loss": 0.2669, "step": 179 }, { "epoch": 6.206896551724138, "grad_norm": 22.875, "learning_rate": 8.95e-06, "loss": 0.2732, "step": 180 }, { "epoch": 6.241379310344827, "grad_norm": 21.0, "learning_rate": 9e-06, "loss": 0.285, "step": 181 }, { "epoch": 6.275862068965517, "grad_norm": 36.5, "learning_rate": 9.05e-06, "loss": 0.2759, "step": 182 }, { "epoch": 6.310344827586207, "grad_norm": 14.625, "learning_rate": 9.100000000000001e-06, "loss": 0.2509, "step": 183 }, { "epoch": 6.344827586206897, "grad_norm": 13.5, "learning_rate": 9.15e-06, "loss": 0.2683, "step": 184 }, { "epoch": 6.379310344827586, "grad_norm": 25.625, "learning_rate": 9.2e-06, "loss": 0.2194, "step": 185 }, { "epoch": 6.413793103448276, "grad_norm": 16.0, "learning_rate": 9.25e-06, "loss": 0.2961, "step": 186 }, { "epoch": 6.448275862068965, "grad_norm": 29.5, "learning_rate": 9.3e-06, "loss": 0.2366, "step": 187 }, { "epoch": 6.482758620689655, "grad_norm": 17.0, "learning_rate": 9.35e-06, "loss": 0.2152, "step": 188 }, { "epoch": 6.517241379310345, "grad_norm": 18.0, "learning_rate": 9.4e-06, "loss": 0.1646, "step": 189 }, { "epoch": 6.551724137931035, "grad_norm": 17.75, "learning_rate": 9.450000000000001e-06, "loss": 0.1774, "step": 190 }, { "epoch": 6.586206896551724, "grad_norm": 29.25, "learning_rate": 9.5e-06, "loss": 0.236, "step": 191 }, { "epoch": 6.620689655172414, "grad_norm": 23.625, "learning_rate": 9.55e-06, "loss": 0.2805, "step": 192 }, { "epoch": 6.655172413793103, "grad_norm": 39.0, "learning_rate": 9.600000000000001e-06, "loss": 0.2858, "step": 193 }, { "epoch": 6.689655172413794, "grad_norm": 19.75, "learning_rate": 9.65e-06, "loss": 0.2094, "step": 194 }, { "epoch": 6.724137931034483, "grad_norm": 16.375, "learning_rate": 9.7e-06, "loss": 0.2151, "step": 195 }, { "epoch": 6.758620689655173, "grad_norm": 12.9375, "learning_rate": 9.750000000000002e-06, "loss": 0.2277, "step": 196 }, { "epoch": 6.793103448275862, "grad_norm": 46.5, "learning_rate": 9.800000000000001e-06, "loss": 0.2084, "step": 197 }, { "epoch": 6.827586206896552, "grad_norm": 21.25, "learning_rate": 9.85e-06, "loss": 0.2452, "step": 198 }, { "epoch": 6.862068965517241, "grad_norm": 10.5, "learning_rate": 9.900000000000002e-06, "loss": 0.1438, "step": 199 }, { "epoch": 6.896551724137931, "grad_norm": 53.5, "learning_rate": 9.950000000000001e-06, "loss": 0.2776, "step": 200 }, { "epoch": 6.931034482758621, "grad_norm": 53.0, "learning_rate": 1e-05, "loss": 0.304, "step": 201 }, { "epoch": 6.9655172413793105, "grad_norm": 14.9375, "learning_rate": 1.005e-05, "loss": 0.2347, "step": 202 }, { "epoch": 7.0, "grad_norm": 13.0, "learning_rate": 1.0100000000000002e-05, "loss": 0.1558, "step": 203 }, { "epoch": 7.0344827586206895, "grad_norm": 46.75, "learning_rate": 1.0150000000000001e-05, "loss": 0.2916, "step": 204 }, { "epoch": 7.068965517241379, "grad_norm": 32.0, "learning_rate": 1.02e-05, "loss": 0.16, "step": 205 }, { "epoch": 7.103448275862069, "grad_norm": 34.75, "learning_rate": 1.025e-05, "loss": 0.2463, "step": 206 }, { "epoch": 7.137931034482759, "grad_norm": 17.875, "learning_rate": 1.03e-05, "loss": 0.1858, "step": 207 }, { "epoch": 7.172413793103448, "grad_norm": 16.375, "learning_rate": 1.035e-05, "loss": 0.2358, "step": 208 }, { "epoch": 7.206896551724138, "grad_norm": 17.125, "learning_rate": 1.04e-05, "loss": 0.2538, "step": 209 }, { "epoch": 7.241379310344827, "grad_norm": 17.0, "learning_rate": 1.045e-05, "loss": 0.1536, "step": 210 }, { "epoch": 7.275862068965517, "grad_norm": 24.0, "learning_rate": 1.05e-05, "loss": 0.1953, "step": 211 }, { "epoch": 7.310344827586207, "grad_norm": 19.375, "learning_rate": 1.055e-05, "loss": 0.231, "step": 212 }, { "epoch": 7.344827586206897, "grad_norm": 20.625, "learning_rate": 1.06e-05, "loss": 0.1878, "step": 213 }, { "epoch": 7.379310344827586, "grad_norm": 15.125, "learning_rate": 1.065e-05, "loss": 0.2316, "step": 214 }, { "epoch": 7.413793103448276, "grad_norm": 19.25, "learning_rate": 1.0700000000000001e-05, "loss": 0.1528, "step": 215 }, { "epoch": 7.448275862068965, "grad_norm": 15.8125, "learning_rate": 1.075e-05, "loss": 0.1629, "step": 216 }, { "epoch": 7.482758620689655, "grad_norm": 16.875, "learning_rate": 1.08e-05, "loss": 0.1903, "step": 217 }, { "epoch": 7.517241379310345, "grad_norm": 6.25, "learning_rate": 1.0850000000000001e-05, "loss": 0.1712, "step": 218 }, { "epoch": 7.551724137931035, "grad_norm": 30.75, "learning_rate": 1.09e-05, "loss": 0.2271, "step": 219 }, { "epoch": 7.586206896551724, "grad_norm": 14.5, "learning_rate": 1.095e-05, "loss": 0.1609, "step": 220 }, { "epoch": 7.620689655172414, "grad_norm": 25.5, "learning_rate": 1.1000000000000001e-05, "loss": 0.2665, "step": 221 }, { "epoch": 7.655172413793103, "grad_norm": 15.0625, "learning_rate": 1.1050000000000001e-05, "loss": 0.2531, "step": 222 }, { "epoch": 7.689655172413794, "grad_norm": 14.3125, "learning_rate": 1.11e-05, "loss": 0.1329, "step": 223 }, { "epoch": 7.724137931034483, "grad_norm": 20.5, "learning_rate": 1.115e-05, "loss": 0.1491, "step": 224 }, { "epoch": 7.758620689655173, "grad_norm": 14.375, "learning_rate": 1.1200000000000001e-05, "loss": 0.1803, "step": 225 }, { "epoch": 7.793103448275862, "grad_norm": 22.5, "learning_rate": 1.125e-05, "loss": 0.2701, "step": 226 }, { "epoch": 7.827586206896552, "grad_norm": 32.25, "learning_rate": 1.13e-05, "loss": 0.1857, "step": 227 }, { "epoch": 7.862068965517241, "grad_norm": 23.5, "learning_rate": 1.1350000000000001e-05, "loss": 0.2262, "step": 228 }, { "epoch": 7.896551724137931, "grad_norm": 16.375, "learning_rate": 1.1400000000000001e-05, "loss": 0.1697, "step": 229 }, { "epoch": 7.931034482758621, "grad_norm": 25.75, "learning_rate": 1.145e-05, "loss": 0.1563, "step": 230 }, { "epoch": 7.9655172413793105, "grad_norm": 33.0, "learning_rate": 1.1500000000000002e-05, "loss": 0.1764, "step": 231 }, { "epoch": 8.0, "grad_norm": 23.5, "learning_rate": 1.1550000000000001e-05, "loss": 0.1753, "step": 232 }, { "epoch": 8.03448275862069, "grad_norm": 17.875, "learning_rate": 1.16e-05, "loss": 0.1277, "step": 233 }, { "epoch": 8.068965517241379, "grad_norm": 21.75, "learning_rate": 1.1650000000000002e-05, "loss": 0.165, "step": 234 }, { "epoch": 8.10344827586207, "grad_norm": 25.25, "learning_rate": 1.1700000000000001e-05, "loss": 0.2277, "step": 235 }, { "epoch": 8.137931034482758, "grad_norm": 26.375, "learning_rate": 1.175e-05, "loss": 0.2646, "step": 236 }, { "epoch": 8.172413793103448, "grad_norm": 13.6875, "learning_rate": 1.18e-05, "loss": 0.1515, "step": 237 }, { "epoch": 8.206896551724139, "grad_norm": 13.4375, "learning_rate": 1.185e-05, "loss": 0.1394, "step": 238 }, { "epoch": 8.241379310344827, "grad_norm": 10.5625, "learning_rate": 1.19e-05, "loss": 0.1713, "step": 239 }, { "epoch": 8.275862068965518, "grad_norm": 8.75, "learning_rate": 1.195e-05, "loss": 0.1675, "step": 240 }, { "epoch": 8.310344827586206, "grad_norm": 10.6875, "learning_rate": 1.2e-05, "loss": 0.1638, "step": 241 }, { "epoch": 8.344827586206897, "grad_norm": 28.625, "learning_rate": 1.205e-05, "loss": 0.2874, "step": 242 }, { "epoch": 8.379310344827585, "grad_norm": 12.8125, "learning_rate": 1.2100000000000001e-05, "loss": 0.2588, "step": 243 }, { "epoch": 8.413793103448276, "grad_norm": 17.5, "learning_rate": 1.215e-05, "loss": 0.1419, "step": 244 }, { "epoch": 8.448275862068966, "grad_norm": 22.5, "learning_rate": 1.22e-05, "loss": 0.2143, "step": 245 }, { "epoch": 8.482758620689655, "grad_norm": 15.875, "learning_rate": 1.225e-05, "loss": 0.1645, "step": 246 }, { "epoch": 8.517241379310345, "grad_norm": 16.0, "learning_rate": 1.23e-05, "loss": 0.1993, "step": 247 }, { "epoch": 8.551724137931034, "grad_norm": 15.5625, "learning_rate": 1.235e-05, "loss": 0.1824, "step": 248 }, { "epoch": 8.586206896551724, "grad_norm": 37.5, "learning_rate": 1.24e-05, "loss": 0.2379, "step": 249 }, { "epoch": 8.620689655172415, "grad_norm": 20.375, "learning_rate": 1.2450000000000001e-05, "loss": 0.2083, "step": 250 }, { "epoch": 8.655172413793103, "grad_norm": 52.75, "learning_rate": 1.25e-05, "loss": 0.249, "step": 251 }, { "epoch": 8.689655172413794, "grad_norm": 21.75, "learning_rate": 1.255e-05, "loss": 0.1214, "step": 252 }, { "epoch": 8.724137931034482, "grad_norm": 4.46875, "learning_rate": 1.2600000000000001e-05, "loss": 0.0499, "step": 253 }, { "epoch": 8.758620689655173, "grad_norm": 32.0, "learning_rate": 1.2650000000000001e-05, "loss": 0.342, "step": 254 }, { "epoch": 8.793103448275861, "grad_norm": 16.25, "learning_rate": 1.27e-05, "loss": 0.172, "step": 255 }, { "epoch": 8.827586206896552, "grad_norm": 20.75, "learning_rate": 1.2750000000000002e-05, "loss": 0.1833, "step": 256 }, { "epoch": 8.862068965517242, "grad_norm": 13.8125, "learning_rate": 1.2800000000000001e-05, "loss": 0.1007, "step": 257 }, { "epoch": 8.89655172413793, "grad_norm": 8.375, "learning_rate": 1.285e-05, "loss": 0.1065, "step": 258 }, { "epoch": 8.931034482758621, "grad_norm": 14.4375, "learning_rate": 1.29e-05, "loss": 0.193, "step": 259 }, { "epoch": 8.96551724137931, "grad_norm": 23.75, "learning_rate": 1.2950000000000001e-05, "loss": 0.219, "step": 260 }, { "epoch": 9.0, "grad_norm": 14.875, "learning_rate": 1.3000000000000001e-05, "loss": 0.2064, "step": 261 }, { "epoch": 9.03448275862069, "grad_norm": 20.625, "learning_rate": 1.305e-05, "loss": 0.1799, "step": 262 }, { "epoch": 9.068965517241379, "grad_norm": 10.0625, "learning_rate": 1.3100000000000002e-05, "loss": 0.1371, "step": 263 }, { "epoch": 9.10344827586207, "grad_norm": 12.125, "learning_rate": 1.3150000000000001e-05, "loss": 0.165, "step": 264 }, { "epoch": 9.137931034482758, "grad_norm": 20.875, "learning_rate": 1.32e-05, "loss": 0.218, "step": 265 }, { "epoch": 9.172413793103448, "grad_norm": 8.9375, "learning_rate": 1.3250000000000002e-05, "loss": 0.1014, "step": 266 }, { "epoch": 9.206896551724139, "grad_norm": 21.125, "learning_rate": 1.3300000000000001e-05, "loss": 0.1424, "step": 267 }, { "epoch": 9.241379310344827, "grad_norm": 15.1875, "learning_rate": 1.3350000000000001e-05, "loss": 0.201, "step": 268 }, { "epoch": 9.275862068965518, "grad_norm": 15.5, "learning_rate": 1.3400000000000002e-05, "loss": 0.1195, "step": 269 }, { "epoch": 9.310344827586206, "grad_norm": 19.25, "learning_rate": 1.3450000000000002e-05, "loss": 0.1704, "step": 270 }, { "epoch": 9.344827586206897, "grad_norm": 27.5, "learning_rate": 1.3500000000000001e-05, "loss": 0.2065, "step": 271 }, { "epoch": 9.379310344827585, "grad_norm": 26.375, "learning_rate": 1.3550000000000002e-05, "loss": 0.3222, "step": 272 }, { "epoch": 9.413793103448276, "grad_norm": 18.375, "learning_rate": 1.3600000000000002e-05, "loss": 0.1291, "step": 273 }, { "epoch": 9.448275862068966, "grad_norm": 19.75, "learning_rate": 1.3650000000000001e-05, "loss": 0.1134, "step": 274 }, { "epoch": 9.482758620689655, "grad_norm": 18.0, "learning_rate": 1.3700000000000001e-05, "loss": 0.1577, "step": 275 }, { "epoch": 9.517241379310345, "grad_norm": 12.625, "learning_rate": 1.3750000000000002e-05, "loss": 0.157, "step": 276 }, { "epoch": 9.551724137931034, "grad_norm": 14.5, "learning_rate": 1.3800000000000002e-05, "loss": 0.166, "step": 277 }, { "epoch": 9.586206896551724, "grad_norm": 34.25, "learning_rate": 1.3850000000000001e-05, "loss": 0.2796, "step": 278 }, { "epoch": 9.620689655172415, "grad_norm": 12.0, "learning_rate": 1.3900000000000002e-05, "loss": 0.1708, "step": 279 }, { "epoch": 9.655172413793103, "grad_norm": 11.875, "learning_rate": 1.3950000000000002e-05, "loss": 0.1201, "step": 280 }, { "epoch": 9.689655172413794, "grad_norm": 10.0625, "learning_rate": 1.4000000000000001e-05, "loss": 0.1325, "step": 281 }, { "epoch": 9.724137931034482, "grad_norm": 16.0, "learning_rate": 1.4050000000000003e-05, "loss": 0.1056, "step": 282 }, { "epoch": 9.758620689655173, "grad_norm": 10.6875, "learning_rate": 1.4099999999999999e-05, "loss": 0.1418, "step": 283 }, { "epoch": 9.793103448275861, "grad_norm": 22.75, "learning_rate": 1.415e-05, "loss": 0.2227, "step": 284 }, { "epoch": 9.827586206896552, "grad_norm": 19.625, "learning_rate": 1.42e-05, "loss": 0.1218, "step": 285 }, { "epoch": 9.862068965517242, "grad_norm": 22.0, "learning_rate": 1.4249999999999999e-05, "loss": 0.1074, "step": 286 }, { "epoch": 9.89655172413793, "grad_norm": 18.25, "learning_rate": 1.43e-05, "loss": 0.1391, "step": 287 }, { "epoch": 9.931034482758621, "grad_norm": 37.25, "learning_rate": 1.435e-05, "loss": 0.2287, "step": 288 }, { "epoch": 9.96551724137931, "grad_norm": 35.75, "learning_rate": 1.44e-05, "loss": 0.3175, "step": 289 }, { "epoch": 10.0, "grad_norm": 61.25, "learning_rate": 1.4449999999999999e-05, "loss": 0.2396, "step": 290 }, { "epoch": 10.03448275862069, "grad_norm": 20.5, "learning_rate": 1.45e-05, "loss": 0.1449, "step": 291 }, { "epoch": 10.068965517241379, "grad_norm": 16.375, "learning_rate": 1.455e-05, "loss": 0.194, "step": 292 }, { "epoch": 10.10344827586207, "grad_norm": 24.0, "learning_rate": 1.4599999999999999e-05, "loss": 0.131, "step": 293 }, { "epoch": 10.137931034482758, "grad_norm": 26.875, "learning_rate": 1.465e-05, "loss": 0.1969, "step": 294 }, { "epoch": 10.172413793103448, "grad_norm": 14.625, "learning_rate": 1.47e-05, "loss": 0.1139, "step": 295 }, { "epoch": 10.206896551724139, "grad_norm": 18.75, "learning_rate": 1.475e-05, "loss": 0.1825, "step": 296 }, { "epoch": 10.241379310344827, "grad_norm": 12.8125, "learning_rate": 1.48e-05, "loss": 0.1342, "step": 297 }, { "epoch": 10.275862068965518, "grad_norm": 12.25, "learning_rate": 1.485e-05, "loss": 0.1818, "step": 298 }, { "epoch": 10.310344827586206, "grad_norm": 14.125, "learning_rate": 1.49e-05, "loss": 0.1742, "step": 299 }, { "epoch": 10.344827586206897, "grad_norm": 13.3125, "learning_rate": 1.4950000000000001e-05, "loss": 0.1767, "step": 300 }, { "epoch": 10.379310344827585, "grad_norm": 17.125, "learning_rate": 1.5e-05, "loss": 0.1708, "step": 301 }, { "epoch": 10.413793103448276, "grad_norm": 26.0, "learning_rate": 1.505e-05, "loss": 0.2241, "step": 302 }, { "epoch": 10.448275862068966, "grad_norm": 9.875, "learning_rate": 1.51e-05, "loss": 0.1169, "step": 303 }, { "epoch": 10.482758620689655, "grad_norm": 6.03125, "learning_rate": 1.515e-05, "loss": 0.0911, "step": 304 }, { "epoch": 10.517241379310345, "grad_norm": 28.5, "learning_rate": 1.52e-05, "loss": 0.149, "step": 305 }, { "epoch": 10.551724137931034, "grad_norm": 25.0, "learning_rate": 1.525e-05, "loss": 0.1749, "step": 306 }, { "epoch": 10.586206896551724, "grad_norm": 28.5, "learning_rate": 1.53e-05, "loss": 0.2491, "step": 307 }, { "epoch": 10.620689655172415, "grad_norm": 19.125, "learning_rate": 1.535e-05, "loss": 0.2207, "step": 308 }, { "epoch": 10.655172413793103, "grad_norm": 10.4375, "learning_rate": 1.54e-05, "loss": 0.0713, "step": 309 }, { "epoch": 10.689655172413794, "grad_norm": 22.125, "learning_rate": 1.545e-05, "loss": 0.1765, "step": 310 }, { "epoch": 10.724137931034482, "grad_norm": 38.75, "learning_rate": 1.55e-05, "loss": 0.2481, "step": 311 }, { "epoch": 10.758620689655173, "grad_norm": 11.1875, "learning_rate": 1.5550000000000002e-05, "loss": 0.1622, "step": 312 }, { "epoch": 10.793103448275861, "grad_norm": 9.875, "learning_rate": 1.56e-05, "loss": 0.2097, "step": 313 }, { "epoch": 10.827586206896552, "grad_norm": 10.875, "learning_rate": 1.565e-05, "loss": 0.1912, "step": 314 }, { "epoch": 10.862068965517242, "grad_norm": 13.25, "learning_rate": 1.5700000000000002e-05, "loss": 0.1979, "step": 315 }, { "epoch": 10.89655172413793, "grad_norm": 6.84375, "learning_rate": 1.575e-05, "loss": 0.1604, "step": 316 }, { "epoch": 10.931034482758621, "grad_norm": 14.5, "learning_rate": 1.58e-05, "loss": 0.0979, "step": 317 }, { "epoch": 10.96551724137931, "grad_norm": 10.375, "learning_rate": 1.5850000000000002e-05, "loss": 0.1265, "step": 318 }, { "epoch": 11.0, "grad_norm": 11.8125, "learning_rate": 1.59e-05, "loss": 0.1491, "step": 319 }, { "epoch": 11.03448275862069, "grad_norm": 9.625, "learning_rate": 1.595e-05, "loss": 0.0802, "step": 320 }, { "epoch": 11.068965517241379, "grad_norm": 19.5, "learning_rate": 1.6000000000000003e-05, "loss": 0.1198, "step": 321 }, { "epoch": 11.10344827586207, "grad_norm": 12.4375, "learning_rate": 1.605e-05, "loss": 0.1194, "step": 322 }, { "epoch": 11.137931034482758, "grad_norm": 14.6875, "learning_rate": 1.6100000000000002e-05, "loss": 0.1542, "step": 323 }, { "epoch": 11.172413793103448, "grad_norm": 15.4375, "learning_rate": 1.6150000000000003e-05, "loss": 0.2049, "step": 324 }, { "epoch": 11.206896551724139, "grad_norm": 5.71875, "learning_rate": 1.62e-05, "loss": 0.1938, "step": 325 }, { "epoch": 11.241379310344827, "grad_norm": 5.125, "learning_rate": 1.6250000000000002e-05, "loss": 0.1508, "step": 326 }, { "epoch": 11.275862068965518, "grad_norm": 13.625, "learning_rate": 1.63e-05, "loss": 0.1507, "step": 327 }, { "epoch": 11.310344827586206, "grad_norm": 11.1875, "learning_rate": 1.635e-05, "loss": 0.0673, "step": 328 }, { "epoch": 11.344827586206897, "grad_norm": 6.1875, "learning_rate": 1.6400000000000002e-05, "loss": 0.0981, "step": 329 }, { "epoch": 11.379310344827585, "grad_norm": 21.25, "learning_rate": 1.645e-05, "loss": 0.2369, "step": 330 }, { "epoch": 11.413793103448276, "grad_norm": 10.8125, "learning_rate": 1.65e-05, "loss": 0.1376, "step": 331 }, { "epoch": 11.448275862068966, "grad_norm": 13.125, "learning_rate": 1.6550000000000002e-05, "loss": 0.1123, "step": 332 }, { "epoch": 11.482758620689655, "grad_norm": 8.6875, "learning_rate": 1.66e-05, "loss": 0.0977, "step": 333 }, { "epoch": 11.517241379310345, "grad_norm": 13.6875, "learning_rate": 1.665e-05, "loss": 0.1155, "step": 334 }, { "epoch": 11.551724137931034, "grad_norm": 10.125, "learning_rate": 1.6700000000000003e-05, "loss": 0.0793, "step": 335 }, { "epoch": 11.586206896551724, "grad_norm": 19.5, "learning_rate": 1.675e-05, "loss": 0.1009, "step": 336 }, { "epoch": 11.620689655172415, "grad_norm": 20.125, "learning_rate": 1.6800000000000002e-05, "loss": 0.0918, "step": 337 }, { "epoch": 11.655172413793103, "grad_norm": 8.1875, "learning_rate": 1.6850000000000003e-05, "loss": 0.156, "step": 338 }, { "epoch": 11.689655172413794, "grad_norm": 11.4375, "learning_rate": 1.69e-05, "loss": 0.1418, "step": 339 }, { "epoch": 11.724137931034482, "grad_norm": 8.1875, "learning_rate": 1.6950000000000002e-05, "loss": 0.124, "step": 340 }, { "epoch": 11.758620689655173, "grad_norm": 5.59375, "learning_rate": 1.7000000000000003e-05, "loss": 0.0767, "step": 341 }, { "epoch": 11.793103448275861, "grad_norm": 11.4375, "learning_rate": 1.705e-05, "loss": 0.0907, "step": 342 }, { "epoch": 11.827586206896552, "grad_norm": 8.8125, "learning_rate": 1.7100000000000002e-05, "loss": 0.0686, "step": 343 }, { "epoch": 11.862068965517242, "grad_norm": 6.15625, "learning_rate": 1.7150000000000004e-05, "loss": 0.0507, "step": 344 }, { "epoch": 11.89655172413793, "grad_norm": 15.75, "learning_rate": 1.7199999999999998e-05, "loss": 0.1582, "step": 345 }, { "epoch": 11.931034482758621, "grad_norm": 8.375, "learning_rate": 1.725e-05, "loss": 0.1571, "step": 346 }, { "epoch": 11.96551724137931, "grad_norm": 11.0625, "learning_rate": 1.73e-05, "loss": 0.056, "step": 347 }, { "epoch": 12.0, "grad_norm": 20.25, "learning_rate": 1.7349999999999998e-05, "loss": 0.244, "step": 348 }, { "epoch": 12.03448275862069, "grad_norm": 5.75, "learning_rate": 1.74e-05, "loss": 0.1215, "step": 349 }, { "epoch": 12.068965517241379, "grad_norm": 11.3125, "learning_rate": 1.745e-05, "loss": 0.1096, "step": 350 }, { "epoch": 12.10344827586207, "grad_norm": 15.4375, "learning_rate": 1.75e-05, "loss": 0.1726, "step": 351 }, { "epoch": 12.137931034482758, "grad_norm": 12.0, "learning_rate": 1.755e-05, "loss": 0.1236, "step": 352 }, { "epoch": 12.172413793103448, "grad_norm": 8.5, "learning_rate": 1.76e-05, "loss": 0.0745, "step": 353 }, { "epoch": 12.206896551724139, "grad_norm": 12.1875, "learning_rate": 1.765e-05, "loss": 0.251, "step": 354 }, { "epoch": 12.241379310344827, "grad_norm": 8.5625, "learning_rate": 1.77e-05, "loss": 0.0563, "step": 355 }, { "epoch": 12.275862068965518, "grad_norm": 7.75, "learning_rate": 1.775e-05, "loss": 0.0598, "step": 356 }, { "epoch": 12.310344827586206, "grad_norm": 15.4375, "learning_rate": 1.78e-05, "loss": 0.0929, "step": 357 }, { "epoch": 12.344827586206897, "grad_norm": 16.5, "learning_rate": 1.785e-05, "loss": 0.123, "step": 358 }, { "epoch": 12.379310344827585, "grad_norm": 18.0, "learning_rate": 1.79e-05, "loss": 0.1154, "step": 359 }, { "epoch": 12.413793103448276, "grad_norm": 13.0625, "learning_rate": 1.795e-05, "loss": 0.0995, "step": 360 }, { "epoch": 12.448275862068966, "grad_norm": 10.0, "learning_rate": 1.8e-05, "loss": 0.1705, "step": 361 }, { "epoch": 12.482758620689655, "grad_norm": 9.3125, "learning_rate": 1.805e-05, "loss": 0.0986, "step": 362 }, { "epoch": 12.517241379310345, "grad_norm": 7.5625, "learning_rate": 1.81e-05, "loss": 0.1274, "step": 363 }, { "epoch": 12.551724137931034, "grad_norm": 10.25, "learning_rate": 1.815e-05, "loss": 0.1215, "step": 364 }, { "epoch": 12.586206896551724, "grad_norm": 7.78125, "learning_rate": 1.8200000000000002e-05, "loss": 0.0702, "step": 365 }, { "epoch": 12.620689655172415, "grad_norm": 8.375, "learning_rate": 1.825e-05, "loss": 0.0633, "step": 366 }, { "epoch": 12.655172413793103, "grad_norm": 6.90625, "learning_rate": 1.83e-05, "loss": 0.11, "step": 367 }, { "epoch": 12.689655172413794, "grad_norm": 13.125, "learning_rate": 1.8350000000000002e-05, "loss": 0.0866, "step": 368 }, { "epoch": 12.724137931034482, "grad_norm": 13.125, "learning_rate": 1.84e-05, "loss": 0.1241, "step": 369 }, { "epoch": 12.758620689655173, "grad_norm": 6.40625, "learning_rate": 1.845e-05, "loss": 0.0847, "step": 370 }, { "epoch": 12.793103448275861, "grad_norm": 20.375, "learning_rate": 1.85e-05, "loss": 0.1581, "step": 371 }, { "epoch": 12.827586206896552, "grad_norm": 13.8125, "learning_rate": 1.855e-05, "loss": 0.11, "step": 372 }, { "epoch": 12.862068965517242, "grad_norm": 15.375, "learning_rate": 1.86e-05, "loss": 0.1661, "step": 373 }, { "epoch": 12.89655172413793, "grad_norm": 12.25, "learning_rate": 1.865e-05, "loss": 0.1962, "step": 374 }, { "epoch": 12.931034482758621, "grad_norm": 15.4375, "learning_rate": 1.87e-05, "loss": 0.1436, "step": 375 }, { "epoch": 12.96551724137931, "grad_norm": 9.25, "learning_rate": 1.8750000000000002e-05, "loss": 0.1182, "step": 376 }, { "epoch": 13.0, "grad_norm": 12.3125, "learning_rate": 1.88e-05, "loss": 0.1361, "step": 377 }, { "epoch": 13.03448275862069, "grad_norm": 9.3125, "learning_rate": 1.885e-05, "loss": 0.1018, "step": 378 }, { "epoch": 13.068965517241379, "grad_norm": 11.4375, "learning_rate": 1.8900000000000002e-05, "loss": 0.156, "step": 379 }, { "epoch": 13.10344827586207, "grad_norm": 22.125, "learning_rate": 1.895e-05, "loss": 0.1071, "step": 380 }, { "epoch": 13.137931034482758, "grad_norm": 18.25, "learning_rate": 1.9e-05, "loss": 0.0994, "step": 381 }, { "epoch": 13.172413793103448, "grad_norm": 7.09375, "learning_rate": 1.9050000000000002e-05, "loss": 0.0516, "step": 382 }, { "epoch": 13.206896551724139, "grad_norm": 14.5625, "learning_rate": 1.91e-05, "loss": 0.2332, "step": 383 }, { "epoch": 13.241379310344827, "grad_norm": 10.75, "learning_rate": 1.915e-05, "loss": 0.0853, "step": 384 }, { "epoch": 13.275862068965518, "grad_norm": 13.75, "learning_rate": 1.9200000000000003e-05, "loss": 0.0777, "step": 385 }, { "epoch": 13.310344827586206, "grad_norm": 20.625, "learning_rate": 1.925e-05, "loss": 0.0906, "step": 386 }, { "epoch": 13.344827586206897, "grad_norm": 13.3125, "learning_rate": 1.93e-05, "loss": 0.1627, "step": 387 }, { "epoch": 13.379310344827585, "grad_norm": 4.5, "learning_rate": 1.9350000000000003e-05, "loss": 0.0879, "step": 388 }, { "epoch": 13.413793103448276, "grad_norm": 4.6875, "learning_rate": 1.94e-05, "loss": 0.0781, "step": 389 }, { "epoch": 13.448275862068966, "grad_norm": 6.5, "learning_rate": 1.9450000000000002e-05, "loss": 0.1222, "step": 390 }, { "epoch": 13.482758620689655, "grad_norm": 11.25, "learning_rate": 1.9500000000000003e-05, "loss": 0.1702, "step": 391 }, { "epoch": 13.517241379310345, "grad_norm": 6.375, "learning_rate": 1.955e-05, "loss": 0.1278, "step": 392 }, { "epoch": 13.551724137931034, "grad_norm": 20.5, "learning_rate": 1.9600000000000002e-05, "loss": 0.1299, "step": 393 }, { "epoch": 13.586206896551724, "grad_norm": 7.71875, "learning_rate": 1.9650000000000003e-05, "loss": 0.0911, "step": 394 }, { "epoch": 13.620689655172415, "grad_norm": 8.9375, "learning_rate": 1.97e-05, "loss": 0.0854, "step": 395 }, { "epoch": 13.655172413793103, "grad_norm": 9.8125, "learning_rate": 1.9750000000000002e-05, "loss": 0.0823, "step": 396 }, { "epoch": 13.689655172413794, "grad_norm": 9.8125, "learning_rate": 1.9800000000000004e-05, "loss": 0.1113, "step": 397 }, { "epoch": 13.724137931034482, "grad_norm": 9.75, "learning_rate": 1.985e-05, "loss": 0.0763, "step": 398 }, { "epoch": 13.758620689655173, "grad_norm": 34.75, "learning_rate": 1.9900000000000003e-05, "loss": 0.2059, "step": 399 }, { "epoch": 13.793103448275861, "grad_norm": 10.625, "learning_rate": 1.995e-05, "loss": 0.0988, "step": 400 }, { "epoch": 13.827586206896552, "grad_norm": 13.5, "learning_rate": 2e-05, "loss": 0.0788, "step": 401 }, { "epoch": 13.862068965517242, "grad_norm": 8.0, "learning_rate": 2.0050000000000003e-05, "loss": 0.0628, "step": 402 }, { "epoch": 13.89655172413793, "grad_norm": 6.71875, "learning_rate": 2.01e-05, "loss": 0.0791, "step": 403 }, { "epoch": 13.931034482758621, "grad_norm": 9.25, "learning_rate": 2.0150000000000002e-05, "loss": 0.0993, "step": 404 }, { "epoch": 13.96551724137931, "grad_norm": 16.5, "learning_rate": 2.0200000000000003e-05, "loss": 0.0586, "step": 405 }, { "epoch": 14.0, "grad_norm": 12.875, "learning_rate": 2.025e-05, "loss": 0.079, "step": 406 }, { "epoch": 14.03448275862069, "grad_norm": 4.71875, "learning_rate": 2.0300000000000002e-05, "loss": 0.0735, "step": 407 }, { "epoch": 14.068965517241379, "grad_norm": 5.21875, "learning_rate": 2.035e-05, "loss": 0.1185, "step": 408 }, { "epoch": 14.10344827586207, "grad_norm": 14.375, "learning_rate": 2.04e-05, "loss": 0.159, "step": 409 }, { "epoch": 14.137931034482758, "grad_norm": 9.0625, "learning_rate": 2.045e-05, "loss": 0.1575, "step": 410 }, { "epoch": 14.172413793103448, "grad_norm": 6.75, "learning_rate": 2.05e-05, "loss": 0.1285, "step": 411 }, { "epoch": 14.206896551724139, "grad_norm": 10.375, "learning_rate": 2.055e-05, "loss": 0.0755, "step": 412 }, { "epoch": 14.241379310344827, "grad_norm": 8.3125, "learning_rate": 2.06e-05, "loss": 0.0915, "step": 413 }, { "epoch": 14.275862068965518, "grad_norm": 22.625, "learning_rate": 2.065e-05, "loss": 0.1124, "step": 414 }, { "epoch": 14.310344827586206, "grad_norm": 17.0, "learning_rate": 2.07e-05, "loss": 0.1152, "step": 415 }, { "epoch": 14.344827586206897, "grad_norm": 18.25, "learning_rate": 2.075e-05, "loss": 0.1321, "step": 416 }, { "epoch": 14.379310344827585, "grad_norm": 1.9140625, "learning_rate": 2.08e-05, "loss": 0.0793, "step": 417 }, { "epoch": 14.413793103448276, "grad_norm": 12.6875, "learning_rate": 2.085e-05, "loss": 0.0668, "step": 418 }, { "epoch": 14.448275862068966, "grad_norm": 12.3125, "learning_rate": 2.09e-05, "loss": 0.1602, "step": 419 }, { "epoch": 14.482758620689655, "grad_norm": 10.625, "learning_rate": 2.095e-05, "loss": 0.1085, "step": 420 }, { "epoch": 14.517241379310345, "grad_norm": 7.90625, "learning_rate": 2.1e-05, "loss": 0.0936, "step": 421 }, { "epoch": 14.551724137931034, "grad_norm": 11.625, "learning_rate": 2.105e-05, "loss": 0.1171, "step": 422 }, { "epoch": 14.586206896551724, "grad_norm": 6.15625, "learning_rate": 2.11e-05, "loss": 0.1356, "step": 423 }, { "epoch": 14.620689655172415, "grad_norm": 12.1875, "learning_rate": 2.115e-05, "loss": 0.11, "step": 424 }, { "epoch": 14.655172413793103, "grad_norm": 7.09375, "learning_rate": 2.12e-05, "loss": 0.1149, "step": 425 }, { "epoch": 14.689655172413794, "grad_norm": 8.375, "learning_rate": 2.125e-05, "loss": 0.1336, "step": 426 }, { "epoch": 14.724137931034482, "grad_norm": 6.1875, "learning_rate": 2.13e-05, "loss": 0.0762, "step": 427 }, { "epoch": 14.758620689655173, "grad_norm": 11.5, "learning_rate": 2.135e-05, "loss": 0.0914, "step": 428 }, { "epoch": 14.793103448275861, "grad_norm": 9.6875, "learning_rate": 2.1400000000000002e-05, "loss": 0.0788, "step": 429 }, { "epoch": 14.827586206896552, "grad_norm": 10.125, "learning_rate": 2.145e-05, "loss": 0.0666, "step": 430 }, { "epoch": 14.862068965517242, "grad_norm": 9.1875, "learning_rate": 2.15e-05, "loss": 0.0896, "step": 431 }, { "epoch": 14.89655172413793, "grad_norm": 18.375, "learning_rate": 2.1550000000000002e-05, "loss": 0.1313, "step": 432 }, { "epoch": 14.931034482758621, "grad_norm": 20.0, "learning_rate": 2.16e-05, "loss": 0.1791, "step": 433 }, { "epoch": 14.96551724137931, "grad_norm": 30.625, "learning_rate": 2.165e-05, "loss": 0.1444, "step": 434 }, { "epoch": 15.0, "grad_norm": 9.4375, "learning_rate": 2.1700000000000002e-05, "loss": 0.0575, "step": 435 }, { "epoch": 15.03448275862069, "grad_norm": 10.1875, "learning_rate": 2.175e-05, "loss": 0.108, "step": 436 }, { "epoch": 15.068965517241379, "grad_norm": 5.46875, "learning_rate": 2.18e-05, "loss": 0.1461, "step": 437 }, { "epoch": 15.10344827586207, "grad_norm": 8.9375, "learning_rate": 2.1850000000000003e-05, "loss": 0.1617, "step": 438 }, { "epoch": 15.137931034482758, "grad_norm": 7.09375, "learning_rate": 2.19e-05, "loss": 0.1034, "step": 439 }, { "epoch": 15.172413793103448, "grad_norm": 11.3125, "learning_rate": 2.195e-05, "loss": 0.1422, "step": 440 }, { "epoch": 15.206896551724139, "grad_norm": 5.65625, "learning_rate": 2.2000000000000003e-05, "loss": 0.0854, "step": 441 }, { "epoch": 15.241379310344827, "grad_norm": 9.4375, "learning_rate": 2.205e-05, "loss": 0.1378, "step": 442 }, { "epoch": 15.275862068965518, "grad_norm": 2.921875, "learning_rate": 2.2100000000000002e-05, "loss": 0.0643, "step": 443 }, { "epoch": 15.310344827586206, "grad_norm": 8.75, "learning_rate": 2.215e-05, "loss": 0.156, "step": 444 }, { "epoch": 15.344827586206897, "grad_norm": 6.28125, "learning_rate": 2.22e-05, "loss": 0.0975, "step": 445 }, { "epoch": 15.379310344827585, "grad_norm": 12.0, "learning_rate": 2.2250000000000002e-05, "loss": 0.1184, "step": 446 }, { "epoch": 15.413793103448276, "grad_norm": 12.4375, "learning_rate": 2.23e-05, "loss": 0.1377, "step": 447 }, { "epoch": 15.448275862068966, "grad_norm": 5.46875, "learning_rate": 2.235e-05, "loss": 0.0964, "step": 448 }, { "epoch": 15.482758620689655, "grad_norm": 9.75, "learning_rate": 2.2400000000000002e-05, "loss": 0.1424, "step": 449 }, { "epoch": 15.517241379310345, "grad_norm": 3.671875, "learning_rate": 2.245e-05, "loss": 0.06, "step": 450 }, { "epoch": 15.551724137931034, "grad_norm": 5.8125, "learning_rate": 2.25e-05, "loss": 0.0416, "step": 451 }, { "epoch": 15.586206896551724, "grad_norm": 9.1875, "learning_rate": 2.2550000000000003e-05, "loss": 0.0642, "step": 452 }, { "epoch": 15.620689655172415, "grad_norm": 5.375, "learning_rate": 2.26e-05, "loss": 0.033, "step": 453 }, { "epoch": 15.655172413793103, "grad_norm": 8.4375, "learning_rate": 2.265e-05, "loss": 0.0287, "step": 454 }, { "epoch": 15.689655172413794, "grad_norm": 10.625, "learning_rate": 2.2700000000000003e-05, "loss": 0.1769, "step": 455 }, { "epoch": 15.724137931034482, "grad_norm": 16.75, "learning_rate": 2.275e-05, "loss": 0.2303, "step": 456 }, { "epoch": 15.758620689655173, "grad_norm": 13.25, "learning_rate": 2.2800000000000002e-05, "loss": 0.0602, "step": 457 }, { "epoch": 15.793103448275861, "grad_norm": 19.625, "learning_rate": 2.2850000000000003e-05, "loss": 0.1136, "step": 458 }, { "epoch": 15.827586206896552, "grad_norm": 4.09375, "learning_rate": 2.29e-05, "loss": 0.1608, "step": 459 }, { "epoch": 15.862068965517242, "grad_norm": 5.5625, "learning_rate": 2.2950000000000002e-05, "loss": 0.03, "step": 460 }, { "epoch": 15.89655172413793, "grad_norm": 8.3125, "learning_rate": 2.3000000000000003e-05, "loss": 0.069, "step": 461 }, { "epoch": 15.931034482758621, "grad_norm": 15.0, "learning_rate": 2.305e-05, "loss": 0.1054, "step": 462 }, { "epoch": 15.96551724137931, "grad_norm": 6.375, "learning_rate": 2.3100000000000002e-05, "loss": 0.126, "step": 463 }, { "epoch": 16.0, "grad_norm": 8.9375, "learning_rate": 2.3150000000000004e-05, "loss": 0.1195, "step": 464 }, { "epoch": 16.03448275862069, "grad_norm": 5.5625, "learning_rate": 2.32e-05, "loss": 0.0993, "step": 465 }, { "epoch": 16.06896551724138, "grad_norm": 6.46875, "learning_rate": 2.3250000000000003e-05, "loss": 0.1264, "step": 466 }, { "epoch": 16.103448275862068, "grad_norm": 6.6875, "learning_rate": 2.3300000000000004e-05, "loss": 0.0932, "step": 467 }, { "epoch": 16.137931034482758, "grad_norm": 7.46875, "learning_rate": 2.3350000000000002e-05, "loss": 0.0495, "step": 468 }, { "epoch": 16.17241379310345, "grad_norm": 9.75, "learning_rate": 2.3400000000000003e-05, "loss": 0.1882, "step": 469 }, { "epoch": 16.20689655172414, "grad_norm": 8.8125, "learning_rate": 2.345e-05, "loss": 0.1539, "step": 470 }, { "epoch": 16.24137931034483, "grad_norm": 6.65625, "learning_rate": 2.35e-05, "loss": 0.0761, "step": 471 }, { "epoch": 16.275862068965516, "grad_norm": 4.78125, "learning_rate": 2.355e-05, "loss": 0.0614, "step": 472 }, { "epoch": 16.310344827586206, "grad_norm": 1.5859375, "learning_rate": 2.36e-05, "loss": 0.059, "step": 473 }, { "epoch": 16.344827586206897, "grad_norm": 4.25, "learning_rate": 2.365e-05, "loss": 0.0797, "step": 474 }, { "epoch": 16.379310344827587, "grad_norm": 6.96875, "learning_rate": 2.37e-05, "loss": 0.1647, "step": 475 }, { "epoch": 16.413793103448278, "grad_norm": 0.83203125, "learning_rate": 2.375e-05, "loss": 0.0428, "step": 476 }, { "epoch": 16.448275862068964, "grad_norm": 15.8125, "learning_rate": 2.38e-05, "loss": 0.1241, "step": 477 }, { "epoch": 16.482758620689655, "grad_norm": 13.125, "learning_rate": 2.385e-05, "loss": 0.0916, "step": 478 }, { "epoch": 16.517241379310345, "grad_norm": 22.5, "learning_rate": 2.39e-05, "loss": 0.0883, "step": 479 }, { "epoch": 16.551724137931036, "grad_norm": 10.5, "learning_rate": 2.395e-05, "loss": 0.1723, "step": 480 }, { "epoch": 16.586206896551722, "grad_norm": 15.4375, "learning_rate": 2.4e-05, "loss": 0.0899, "step": 481 }, { "epoch": 16.620689655172413, "grad_norm": 13.1875, "learning_rate": 2.4050000000000002e-05, "loss": 0.0934, "step": 482 }, { "epoch": 16.655172413793103, "grad_norm": 19.25, "learning_rate": 2.41e-05, "loss": 0.0765, "step": 483 }, { "epoch": 16.689655172413794, "grad_norm": 18.375, "learning_rate": 2.415e-05, "loss": 0.0626, "step": 484 }, { "epoch": 16.724137931034484, "grad_norm": 14.25, "learning_rate": 2.4200000000000002e-05, "loss": 0.0767, "step": 485 }, { "epoch": 16.75862068965517, "grad_norm": 18.625, "learning_rate": 2.425e-05, "loss": 0.1022, "step": 486 }, { "epoch": 16.79310344827586, "grad_norm": 6.28125, "learning_rate": 2.43e-05, "loss": 0.0547, "step": 487 }, { "epoch": 16.82758620689655, "grad_norm": 10.6875, "learning_rate": 2.435e-05, "loss": 0.0969, "step": 488 }, { "epoch": 16.862068965517242, "grad_norm": 10.8125, "learning_rate": 2.44e-05, "loss": 0.1287, "step": 489 }, { "epoch": 16.896551724137932, "grad_norm": 14.625, "learning_rate": 2.445e-05, "loss": 0.1214, "step": 490 }, { "epoch": 16.93103448275862, "grad_norm": 8.375, "learning_rate": 2.45e-05, "loss": 0.1637, "step": 491 }, { "epoch": 16.96551724137931, "grad_norm": 9.4375, "learning_rate": 2.455e-05, "loss": 0.1899, "step": 492 }, { "epoch": 17.0, "grad_norm": 5.9375, "learning_rate": 2.46e-05, "loss": 0.0579, "step": 493 }, { "epoch": 17.03448275862069, "grad_norm": 8.375, "learning_rate": 2.465e-05, "loss": 0.0568, "step": 494 }, { "epoch": 17.06896551724138, "grad_norm": 11.5625, "learning_rate": 2.47e-05, "loss": 0.0833, "step": 495 }, { "epoch": 17.103448275862068, "grad_norm": 5.375, "learning_rate": 2.4750000000000002e-05, "loss": 0.0462, "step": 496 }, { "epoch": 17.137931034482758, "grad_norm": 8.9375, "learning_rate": 2.48e-05, "loss": 0.1255, "step": 497 }, { "epoch": 17.17241379310345, "grad_norm": 7.03125, "learning_rate": 2.485e-05, "loss": 0.0568, "step": 498 }, { "epoch": 17.20689655172414, "grad_norm": 5.8125, "learning_rate": 2.4900000000000002e-05, "loss": 0.0692, "step": 499 }, { "epoch": 17.24137931034483, "grad_norm": 6.71875, "learning_rate": 2.495e-05, "loss": 0.0415, "step": 500 }, { "epoch": 17.275862068965516, "grad_norm": 3.375, "learning_rate": 2.5e-05, "loss": 0.0568, "step": 501 }, { "epoch": 17.310344827586206, "grad_norm": 11.1875, "learning_rate": 2.4999996953825926e-05, "loss": 0.1428, "step": 502 }, { "epoch": 17.344827586206897, "grad_norm": 10.25, "learning_rate": 2.499998781530519e-05, "loss": 0.1301, "step": 503 }, { "epoch": 17.379310344827587, "grad_norm": 6.9375, "learning_rate": 2.4999972584442245e-05, "loss": 0.058, "step": 504 }, { "epoch": 17.413793103448278, "grad_norm": 9.375, "learning_rate": 2.4999951261244508e-05, "loss": 0.0766, "step": 505 }, { "epoch": 17.448275862068964, "grad_norm": 3.59375, "learning_rate": 2.499992384572238e-05, "loss": 0.0328, "step": 506 }, { "epoch": 17.482758620689655, "grad_norm": 5.90625, "learning_rate": 2.499989033788922e-05, "loss": 0.0771, "step": 507 }, { "epoch": 17.517241379310345, "grad_norm": 13.8125, "learning_rate": 2.499985073776136e-05, "loss": 0.1904, "step": 508 }, { "epoch": 17.551724137931036, "grad_norm": 10.4375, "learning_rate": 2.4999805045358104e-05, "loss": 0.1024, "step": 509 }, { "epoch": 17.586206896551722, "grad_norm": 3.0625, "learning_rate": 2.4999753260701714e-05, "loss": 0.0579, "step": 510 }, { "epoch": 17.620689655172413, "grad_norm": 4.5625, "learning_rate": 2.4999695383817435e-05, "loss": 0.055, "step": 511 }, { "epoch": 17.655172413793103, "grad_norm": 6.6875, "learning_rate": 2.4999631414733476e-05, "loss": 0.0623, "step": 512 }, { "epoch": 17.689655172413794, "grad_norm": 6.28125, "learning_rate": 2.499956135348101e-05, "loss": 0.0509, "step": 513 }, { "epoch": 17.724137931034484, "grad_norm": 9.8125, "learning_rate": 2.4999485200094186e-05, "loss": 0.1501, "step": 514 }, { "epoch": 17.75862068965517, "grad_norm": 8.375, "learning_rate": 2.4999402954610125e-05, "loss": 0.0821, "step": 515 }, { "epoch": 17.79310344827586, "grad_norm": 7.25, "learning_rate": 2.4999314617068904e-05, "loss": 0.0756, "step": 516 }, { "epoch": 17.82758620689655, "grad_norm": 15.4375, "learning_rate": 2.4999220187513583e-05, "loss": 0.1124, "step": 517 }, { "epoch": 17.862068965517242, "grad_norm": 10.5, "learning_rate": 2.4999119665990183e-05, "loss": 0.2277, "step": 518 }, { "epoch": 17.896551724137932, "grad_norm": 6.96875, "learning_rate": 2.4999013052547703e-05, "loss": 0.0759, "step": 519 }, { "epoch": 17.93103448275862, "grad_norm": 6.125, "learning_rate": 2.4998900347238096e-05, "loss": 0.1428, "step": 520 }, { "epoch": 17.96551724137931, "grad_norm": 9.8125, "learning_rate": 2.4998781550116305e-05, "loss": 0.1024, "step": 521 }, { "epoch": 18.0, "grad_norm": 10.8125, "learning_rate": 2.499865666124022e-05, "loss": 0.0917, "step": 522 }, { "epoch": 18.03448275862069, "grad_norm": 5.84375, "learning_rate": 2.4998525680670712e-05, "loss": 0.0847, "step": 523 }, { "epoch": 18.06896551724138, "grad_norm": 7.46875, "learning_rate": 2.4998388608471625e-05, "loss": 0.173, "step": 524 }, { "epoch": 18.103448275862068, "grad_norm": 9.375, "learning_rate": 2.4998245444709757e-05, "loss": 0.1067, "step": 525 }, { "epoch": 18.137931034482758, "grad_norm": 7.59375, "learning_rate": 2.4998096189454893e-05, "loss": 0.1155, "step": 526 }, { "epoch": 18.17241379310345, "grad_norm": 3.90625, "learning_rate": 2.499794084277977e-05, "loss": 0.0493, "step": 527 }, { "epoch": 18.20689655172414, "grad_norm": 6.03125, "learning_rate": 2.4997779404760114e-05, "loss": 0.0602, "step": 528 }, { "epoch": 18.24137931034483, "grad_norm": 5.65625, "learning_rate": 2.4997611875474595e-05, "loss": 0.0602, "step": 529 }, { "epoch": 18.275862068965516, "grad_norm": 5.59375, "learning_rate": 2.4997438255004876e-05, "loss": 0.154, "step": 530 }, { "epoch": 18.310344827586206, "grad_norm": 12.75, "learning_rate": 2.499725854343557e-05, "loss": 0.0599, "step": 531 }, { "epoch": 18.344827586206897, "grad_norm": 7.28125, "learning_rate": 2.4997072740854265e-05, "loss": 0.077, "step": 532 }, { "epoch": 18.379310344827587, "grad_norm": 0.75, "learning_rate": 2.499688084735153e-05, "loss": 0.0523, "step": 533 }, { "epoch": 18.413793103448278, "grad_norm": 3.359375, "learning_rate": 2.4996682863020875e-05, "loss": 0.1114, "step": 534 }, { "epoch": 18.448275862068964, "grad_norm": 11.0, "learning_rate": 2.499647878795881e-05, "loss": 0.0806, "step": 535 }, { "epoch": 18.482758620689655, "grad_norm": 18.875, "learning_rate": 2.499626862226479e-05, "loss": 0.0865, "step": 536 }, { "epoch": 18.517241379310345, "grad_norm": 8.875, "learning_rate": 2.499605236604125e-05, "loss": 0.0644, "step": 537 }, { "epoch": 18.551724137931036, "grad_norm": 24.375, "learning_rate": 2.499583001939359e-05, "loss": 0.1566, "step": 538 }, { "epoch": 18.586206896551722, "grad_norm": 2.90625, "learning_rate": 2.4995601582430183e-05, "loss": 0.0517, "step": 539 }, { "epoch": 18.620689655172413, "grad_norm": 5.53125, "learning_rate": 2.499536705526236e-05, "loss": 0.0744, "step": 540 }, { "epoch": 18.655172413793103, "grad_norm": 18.25, "learning_rate": 2.499512643800443e-05, "loss": 0.2026, "step": 541 }, { "epoch": 18.689655172413794, "grad_norm": 7.9375, "learning_rate": 2.499487973077367e-05, "loss": 0.0891, "step": 542 }, { "epoch": 18.724137931034484, "grad_norm": 18.75, "learning_rate": 2.4994626933690314e-05, "loss": 0.1759, "step": 543 }, { "epoch": 18.75862068965517, "grad_norm": 5.40625, "learning_rate": 2.499436804687758e-05, "loss": 0.1206, "step": 544 }, { "epoch": 18.79310344827586, "grad_norm": 8.25, "learning_rate": 2.499410307046164e-05, "loss": 0.0894, "step": 545 }, { "epoch": 18.82758620689655, "grad_norm": 5.5, "learning_rate": 2.4993832004571646e-05, "loss": 0.078, "step": 546 }, { "epoch": 18.862068965517242, "grad_norm": 7.1875, "learning_rate": 2.499355484933971e-05, "loss": 0.0614, "step": 547 }, { "epoch": 18.896551724137932, "grad_norm": 5.125, "learning_rate": 2.4993271604900913e-05, "loss": 0.1131, "step": 548 }, { "epoch": 18.93103448275862, "grad_norm": 6.0, "learning_rate": 2.4992982271393307e-05, "loss": 0.0876, "step": 549 }, { "epoch": 18.96551724137931, "grad_norm": 3.46875, "learning_rate": 2.4992686848957905e-05, "loss": 0.0845, "step": 550 }, { "epoch": 19.0, "grad_norm": 3.578125, "learning_rate": 2.49923853377387e-05, "loss": 0.1, "step": 551 }, { "epoch": 19.03448275862069, "grad_norm": 3.625, "learning_rate": 2.4992077737882636e-05, "loss": 0.0898, "step": 552 }, { "epoch": 19.06896551724138, "grad_norm": 9.875, "learning_rate": 2.499176404953964e-05, "loss": 0.1156, "step": 553 }, { "epoch": 19.103448275862068, "grad_norm": 12.0, "learning_rate": 2.49914442728626e-05, "loss": 0.0626, "step": 554 }, { "epoch": 19.137931034482758, "grad_norm": 4.46875, "learning_rate": 2.4991118408007366e-05, "loss": 0.0882, "step": 555 }, { "epoch": 19.17241379310345, "grad_norm": 3.5625, "learning_rate": 2.4990786455132764e-05, "loss": 0.0683, "step": 556 }, { "epoch": 19.20689655172414, "grad_norm": 4.375, "learning_rate": 2.4990448414400585e-05, "loss": 0.0768, "step": 557 }, { "epoch": 19.24137931034483, "grad_norm": 5.59375, "learning_rate": 2.4990104285975586e-05, "loss": 0.083, "step": 558 }, { "epoch": 19.275862068965516, "grad_norm": 4.5625, "learning_rate": 2.498975407002549e-05, "loss": 0.0443, "step": 559 }, { "epoch": 19.310344827586206, "grad_norm": 4.71875, "learning_rate": 2.4989397766720983e-05, "loss": 0.0751, "step": 560 }, { "epoch": 19.344827586206897, "grad_norm": 3.09375, "learning_rate": 2.498903537623573e-05, "loss": 0.1097, "step": 561 }, { "epoch": 19.379310344827587, "grad_norm": 14.125, "learning_rate": 2.4988666898746356e-05, "loss": 0.1313, "step": 562 }, { "epoch": 19.413793103448278, "grad_norm": 8.0625, "learning_rate": 2.4988292334432444e-05, "loss": 0.1287, "step": 563 }, { "epoch": 19.448275862068964, "grad_norm": 7.15625, "learning_rate": 2.4987911683476562e-05, "loss": 0.1142, "step": 564 }, { "epoch": 19.482758620689655, "grad_norm": 3.375, "learning_rate": 2.4987524946064228e-05, "loss": 0.0711, "step": 565 }, { "epoch": 19.517241379310345, "grad_norm": 3.484375, "learning_rate": 2.4987132122383936e-05, "loss": 0.0365, "step": 566 }, { "epoch": 19.551724137931036, "grad_norm": 4.625, "learning_rate": 2.4986733212627143e-05, "loss": 0.0375, "step": 567 }, { "epoch": 19.586206896551722, "grad_norm": 4.09375, "learning_rate": 2.498632821698827e-05, "loss": 0.0426, "step": 568 }, { "epoch": 19.620689655172413, "grad_norm": 7.6875, "learning_rate": 2.4985917135664714e-05, "loss": 0.067, "step": 569 }, { "epoch": 19.655172413793103, "grad_norm": 4.3125, "learning_rate": 2.4985499968856824e-05, "loss": 0.0363, "step": 570 }, { "epoch": 19.689655172413794, "grad_norm": 8.6875, "learning_rate": 2.4985076716767927e-05, "loss": 0.105, "step": 571 }, { "epoch": 19.724137931034484, "grad_norm": 3.90625, "learning_rate": 2.498464737960431e-05, "loss": 0.0853, "step": 572 }, { "epoch": 19.75862068965517, "grad_norm": 4.15625, "learning_rate": 2.498421195757522e-05, "loss": 0.0419, "step": 573 }, { "epoch": 19.79310344827586, "grad_norm": 3.515625, "learning_rate": 2.4983770450892884e-05, "loss": 0.0292, "step": 574 }, { "epoch": 19.82758620689655, "grad_norm": 11.3125, "learning_rate": 2.4983322859772484e-05, "loss": 0.0539, "step": 575 }, { "epoch": 19.862068965517242, "grad_norm": 12.1875, "learning_rate": 2.4982869184432174e-05, "loss": 0.1276, "step": 576 }, { "epoch": 19.896551724137932, "grad_norm": 4.125, "learning_rate": 2.4982409425093067e-05, "loss": 0.0322, "step": 577 }, { "epoch": 19.93103448275862, "grad_norm": 7.90625, "learning_rate": 2.498194358197924e-05, "loss": 0.0992, "step": 578 }, { "epoch": 19.96551724137931, "grad_norm": 8.8125, "learning_rate": 2.498147165531775e-05, "loss": 0.1137, "step": 579 }, { "epoch": 20.0, "grad_norm": 5.9375, "learning_rate": 2.4980993645338595e-05, "loss": 0.0807, "step": 580 }, { "epoch": 20.03448275862069, "grad_norm": 3.328125, "learning_rate": 2.4980509552274765e-05, "loss": 0.064, "step": 581 }, { "epoch": 20.06896551724138, "grad_norm": 3.4375, "learning_rate": 2.498001937636219e-05, "loss": 0.0697, "step": 582 }, { "epoch": 20.103448275862068, "grad_norm": 4.21875, "learning_rate": 2.4979523117839783e-05, "loss": 0.0388, "step": 583 }, { "epoch": 20.137931034482758, "grad_norm": 4.96875, "learning_rate": 2.4979020776949405e-05, "loss": 0.0538, "step": 584 }, { "epoch": 20.17241379310345, "grad_norm": 4.46875, "learning_rate": 2.4978512353935903e-05, "loss": 0.0571, "step": 585 }, { "epoch": 20.20689655172414, "grad_norm": 0.890625, "learning_rate": 2.497799784904707e-05, "loss": 0.027, "step": 586 }, { "epoch": 20.24137931034483, "grad_norm": 3.578125, "learning_rate": 2.4977477262533667e-05, "loss": 0.0923, "step": 587 }, { "epoch": 20.275862068965516, "grad_norm": 8.25, "learning_rate": 2.4976950594649423e-05, "loss": 0.0662, "step": 588 }, { "epoch": 20.310344827586206, "grad_norm": 6.09375, "learning_rate": 2.4976417845651033e-05, "loss": 0.0733, "step": 589 }, { "epoch": 20.344827586206897, "grad_norm": 1.03125, "learning_rate": 2.4975879015798152e-05, "loss": 0.0816, "step": 590 }, { "epoch": 20.379310344827587, "grad_norm": 4.46875, "learning_rate": 2.4975334105353396e-05, "loss": 0.0322, "step": 591 }, { "epoch": 20.413793103448278, "grad_norm": 2.078125, "learning_rate": 2.4974783114582352e-05, "loss": 0.088, "step": 592 }, { "epoch": 20.448275862068964, "grad_norm": 1.46875, "learning_rate": 2.4974226043753558e-05, "loss": 0.1036, "step": 593 }, { "epoch": 20.482758620689655, "grad_norm": 2.140625, "learning_rate": 2.497366289313853e-05, "loss": 0.0668, "step": 594 }, { "epoch": 20.517241379310345, "grad_norm": 5.625, "learning_rate": 2.4973093663011743e-05, "loss": 0.0748, "step": 595 }, { "epoch": 20.551724137931036, "grad_norm": 9.8125, "learning_rate": 2.4972518353650626e-05, "loss": 0.1493, "step": 596 }, { "epoch": 20.586206896551722, "grad_norm": 8.25, "learning_rate": 2.4971936965335585e-05, "loss": 0.0578, "step": 597 }, { "epoch": 20.620689655172413, "grad_norm": 5.34375, "learning_rate": 2.497134949834997e-05, "loss": 0.0717, "step": 598 }, { "epoch": 20.655172413793103, "grad_norm": 5.4375, "learning_rate": 2.4970755952980125e-05, "loss": 0.0265, "step": 599 }, { "epoch": 20.689655172413794, "grad_norm": 5.9375, "learning_rate": 2.4970156329515315e-05, "loss": 0.0545, "step": 600 }, { "epoch": 20.724137931034484, "grad_norm": 1.734375, "learning_rate": 2.4969550628247805e-05, "loss": 0.041, "step": 601 }, { "epoch": 20.75862068965517, "grad_norm": 8.9375, "learning_rate": 2.4968938849472797e-05, "loss": 0.0981, "step": 602 }, { "epoch": 20.79310344827586, "grad_norm": 0.68359375, "learning_rate": 2.496832099348847e-05, "loss": 0.0305, "step": 603 }, { "epoch": 20.82758620689655, "grad_norm": 5.84375, "learning_rate": 2.4967697060595958e-05, "loss": 0.059, "step": 604 }, { "epoch": 20.862068965517242, "grad_norm": 2.609375, "learning_rate": 2.4967067051099356e-05, "loss": 0.0921, "step": 605 }, { "epoch": 20.896551724137932, "grad_norm": 19.375, "learning_rate": 2.4966430965305727e-05, "loss": 0.1143, "step": 606 }, { "epoch": 20.93103448275862, "grad_norm": 6.1875, "learning_rate": 2.4965788803525088e-05, "loss": 0.0914, "step": 607 }, { "epoch": 20.96551724137931, "grad_norm": 3.515625, "learning_rate": 2.4965140566070423e-05, "loss": 0.0968, "step": 608 }, { "epoch": 21.0, "grad_norm": 8.125, "learning_rate": 2.4964486253257674e-05, "loss": 0.0617, "step": 609 }, { "epoch": 21.03448275862069, "grad_norm": 3.765625, "learning_rate": 2.4963825865405747e-05, "loss": 0.0623, "step": 610 }, { "epoch": 21.06896551724138, "grad_norm": 5.59375, "learning_rate": 2.4963159402836506e-05, "loss": 0.055, "step": 611 }, { "epoch": 21.103448275862068, "grad_norm": 7.15625, "learning_rate": 2.4962486865874774e-05, "loss": 0.0802, "step": 612 }, { "epoch": 21.137931034482758, "grad_norm": 10.375, "learning_rate": 2.4961808254848342e-05, "loss": 0.0734, "step": 613 }, { "epoch": 21.17241379310345, "grad_norm": 2.828125, "learning_rate": 2.496112357008795e-05, "loss": 0.0474, "step": 614 }, { "epoch": 21.20689655172414, "grad_norm": 2.96875, "learning_rate": 2.4960432811927308e-05, "loss": 0.0594, "step": 615 }, { "epoch": 21.24137931034483, "grad_norm": 8.4375, "learning_rate": 2.495973598070309e-05, "loss": 0.121, "step": 616 }, { "epoch": 21.275862068965516, "grad_norm": 5.53125, "learning_rate": 2.4959033076754917e-05, "loss": 0.0759, "step": 617 }, { "epoch": 21.310344827586206, "grad_norm": 5.34375, "learning_rate": 2.495832410042537e-05, "loss": 0.0447, "step": 618 }, { "epoch": 21.344827586206897, "grad_norm": 1.9921875, "learning_rate": 2.495760905206001e-05, "loss": 0.0344, "step": 619 }, { "epoch": 21.379310344827587, "grad_norm": 8.4375, "learning_rate": 2.4956887932007332e-05, "loss": 0.0623, "step": 620 }, { "epoch": 21.413793103448278, "grad_norm": 4.875, "learning_rate": 2.4956160740618806e-05, "loss": 0.049, "step": 621 }, { "epoch": 21.448275862068964, "grad_norm": 2.890625, "learning_rate": 2.4955427478248848e-05, "loss": 0.093, "step": 622 }, { "epoch": 21.482758620689655, "grad_norm": 2.171875, "learning_rate": 2.4954688145254855e-05, "loss": 0.0238, "step": 623 }, { "epoch": 21.517241379310345, "grad_norm": 1.9609375, "learning_rate": 2.495394274199716e-05, "loss": 0.0665, "step": 624 }, { "epoch": 21.551724137931036, "grad_norm": 2.0, "learning_rate": 2.495319126883906e-05, "loss": 0.0393, "step": 625 }, { "epoch": 21.586206896551722, "grad_norm": 3.984375, "learning_rate": 2.495243372614682e-05, "loss": 0.0605, "step": 626 }, { "epoch": 21.620689655172413, "grad_norm": 5.40625, "learning_rate": 2.495167011428966e-05, "loss": 0.0788, "step": 627 }, { "epoch": 21.655172413793103, "grad_norm": 4.84375, "learning_rate": 2.4950900433639746e-05, "loss": 0.0887, "step": 628 }, { "epoch": 21.689655172413794, "grad_norm": 2.75, "learning_rate": 2.4950124684572217e-05, "loss": 0.0835, "step": 629 }, { "epoch": 21.724137931034484, "grad_norm": 3.328125, "learning_rate": 2.494934286746516e-05, "loss": 0.0763, "step": 630 }, { "epoch": 21.75862068965517, "grad_norm": 5.53125, "learning_rate": 2.494855498269963e-05, "loss": 0.1105, "step": 631 }, { "epoch": 21.79310344827586, "grad_norm": 4.3125, "learning_rate": 2.4947761030659627e-05, "loss": 0.099, "step": 632 }, { "epoch": 21.82758620689655, "grad_norm": 7.28125, "learning_rate": 2.4946961011732118e-05, "loss": 0.0621, "step": 633 }, { "epoch": 21.862068965517242, "grad_norm": 3.5625, "learning_rate": 2.4946154926307014e-05, "loss": 0.0652, "step": 634 }, { "epoch": 21.896551724137932, "grad_norm": 2.5, "learning_rate": 2.4945342774777196e-05, "loss": 0.07, "step": 635 }, { "epoch": 21.93103448275862, "grad_norm": 2.515625, "learning_rate": 2.4944524557538503e-05, "loss": 0.0327, "step": 636 }, { "epoch": 21.96551724137931, "grad_norm": 16.125, "learning_rate": 2.4943700274989714e-05, "loss": 0.0641, "step": 637 }, { "epoch": 22.0, "grad_norm": 3.6875, "learning_rate": 2.494286992753258e-05, "loss": 0.0602, "step": 638 }, { "epoch": 22.03448275862069, "grad_norm": 8.875, "learning_rate": 2.49420335155718e-05, "loss": 0.0638, "step": 639 }, { "epoch": 22.06896551724138, "grad_norm": 3.375, "learning_rate": 2.4941191039515032e-05, "loss": 0.0363, "step": 640 }, { "epoch": 22.103448275862068, "grad_norm": 4.3125, "learning_rate": 2.494034249977289e-05, "loss": 0.1158, "step": 641 }, { "epoch": 22.137931034482758, "grad_norm": 1.125, "learning_rate": 2.493948789675894e-05, "loss": 0.0566, "step": 642 }, { "epoch": 22.17241379310345, "grad_norm": 3.40625, "learning_rate": 2.4938627230889705e-05, "loss": 0.0468, "step": 643 }, { "epoch": 22.20689655172414, "grad_norm": 6.125, "learning_rate": 2.4937760502584666e-05, "loss": 0.0641, "step": 644 }, { "epoch": 22.24137931034483, "grad_norm": 5.0, "learning_rate": 2.4936887712266254e-05, "loss": 0.0758, "step": 645 }, { "epoch": 22.275862068965516, "grad_norm": 3.03125, "learning_rate": 2.4936008860359854e-05, "loss": 0.0778, "step": 646 }, { "epoch": 22.310344827586206, "grad_norm": 3.609375, "learning_rate": 2.493512394729381e-05, "loss": 0.0343, "step": 647 }, { "epoch": 22.344827586206897, "grad_norm": 3.296875, "learning_rate": 2.4934232973499415e-05, "loss": 0.0432, "step": 648 }, { "epoch": 22.379310344827587, "grad_norm": 4.0625, "learning_rate": 2.493333593941093e-05, "loss": 0.0493, "step": 649 }, { "epoch": 22.413793103448278, "grad_norm": 2.796875, "learning_rate": 2.4932432845465543e-05, "loss": 0.0382, "step": 650 }, { "epoch": 22.448275862068964, "grad_norm": 4.53125, "learning_rate": 2.4931523692103418e-05, "loss": 0.0493, "step": 651 }, { "epoch": 22.482758620689655, "grad_norm": 3.78125, "learning_rate": 2.4930608479767667e-05, "loss": 0.0186, "step": 652 }, { "epoch": 22.517241379310345, "grad_norm": 7.71875, "learning_rate": 2.492968720890435e-05, "loss": 0.0621, "step": 653 }, { "epoch": 22.551724137931036, "grad_norm": 7.65625, "learning_rate": 2.4928759879962487e-05, "loss": 0.0757, "step": 654 }, { "epoch": 22.586206896551722, "grad_norm": 5.59375, "learning_rate": 2.4927826493394046e-05, "loss": 0.0463, "step": 655 }, { "epoch": 22.620689655172413, "grad_norm": 2.9375, "learning_rate": 2.4926887049653943e-05, "loss": 0.0359, "step": 656 }, { "epoch": 22.655172413793103, "grad_norm": 1.0, "learning_rate": 2.492594154920006e-05, "loss": 0.0221, "step": 657 }, { "epoch": 22.689655172413794, "grad_norm": 1.4296875, "learning_rate": 2.4924989992493213e-05, "loss": 0.0659, "step": 658 }, { "epoch": 22.724137931034484, "grad_norm": 4.3125, "learning_rate": 2.4924032379997186e-05, "loss": 0.0955, "step": 659 }, { "epoch": 22.75862068965517, "grad_norm": 0.87109375, "learning_rate": 2.492306871217871e-05, "loss": 0.0412, "step": 660 }, { "epoch": 22.79310344827586, "grad_norm": 12.6875, "learning_rate": 2.4922098989507454e-05, "loss": 0.106, "step": 661 }, { "epoch": 22.82758620689655, "grad_norm": 0.87109375, "learning_rate": 2.492112321245606e-05, "loss": 0.0188, "step": 662 }, { "epoch": 22.862068965517242, "grad_norm": 1.0546875, "learning_rate": 2.4920141381500107e-05, "loss": 0.0458, "step": 663 }, { "epoch": 22.896551724137932, "grad_norm": 4.84375, "learning_rate": 2.4919153497118126e-05, "loss": 0.096, "step": 664 }, { "epoch": 22.93103448275862, "grad_norm": 6.78125, "learning_rate": 2.4918159559791596e-05, "loss": 0.118, "step": 665 }, { "epoch": 22.96551724137931, "grad_norm": 2.8125, "learning_rate": 2.4917159570004954e-05, "loss": 0.0464, "step": 666 }, { "epoch": 23.0, "grad_norm": 5.875, "learning_rate": 2.491615352824559e-05, "loss": 0.0772, "step": 667 }, { "epoch": 23.03448275862069, "grad_norm": 5.96875, "learning_rate": 2.4915141435003823e-05, "loss": 0.1011, "step": 668 }, { "epoch": 23.06896551724138, "grad_norm": 2.296875, "learning_rate": 2.4914123290772945e-05, "loss": 0.0547, "step": 669 }, { "epoch": 23.103448275862068, "grad_norm": 4.125, "learning_rate": 2.491309909604918e-05, "loss": 0.0575, "step": 670 }, { "epoch": 23.137931034482758, "grad_norm": 8.6875, "learning_rate": 2.491206885133171e-05, "loss": 0.0935, "step": 671 }, { "epoch": 23.17241379310345, "grad_norm": 3.78125, "learning_rate": 2.491103255712267e-05, "loss": 0.0451, "step": 672 }, { "epoch": 23.20689655172414, "grad_norm": 2.5625, "learning_rate": 2.4909990213927135e-05, "loss": 0.0392, "step": 673 }, { "epoch": 23.24137931034483, "grad_norm": 3.765625, "learning_rate": 2.4908941822253124e-05, "loss": 0.1074, "step": 674 }, { "epoch": 23.275862068965516, "grad_norm": 4.625, "learning_rate": 2.4907887382611613e-05, "loss": 0.0642, "step": 675 }, { "epoch": 23.310344827586206, "grad_norm": 2.296875, "learning_rate": 2.4906826895516528e-05, "loss": 0.0422, "step": 676 }, { "epoch": 23.344827586206897, "grad_norm": 3.3125, "learning_rate": 2.4905760361484733e-05, "loss": 0.06, "step": 677 }, { "epoch": 23.379310344827587, "grad_norm": 7.21875, "learning_rate": 2.490468778103604e-05, "loss": 0.0598, "step": 678 }, { "epoch": 23.413793103448278, "grad_norm": 2.5625, "learning_rate": 2.4903609154693222e-05, "loss": 0.033, "step": 679 }, { "epoch": 23.448275862068964, "grad_norm": 3.3125, "learning_rate": 2.490252448298198e-05, "loss": 0.0668, "step": 680 }, { "epoch": 23.482758620689655, "grad_norm": 18.875, "learning_rate": 2.4901433766430975e-05, "loss": 0.0487, "step": 681 }, { "epoch": 23.517241379310345, "grad_norm": 2.484375, "learning_rate": 2.4900337005571807e-05, "loss": 0.0262, "step": 682 }, { "epoch": 23.551724137931036, "grad_norm": 1.3359375, "learning_rate": 2.489923420093902e-05, "loss": 0.052, "step": 683 }, { "epoch": 23.586206896551722, "grad_norm": 7.3125, "learning_rate": 2.4898125353070112e-05, "loss": 0.0625, "step": 684 }, { "epoch": 23.620689655172413, "grad_norm": 5.40625, "learning_rate": 2.4897010462505522e-05, "loss": 0.0195, "step": 685 }, { "epoch": 23.655172413793103, "grad_norm": 4.28125, "learning_rate": 2.489588952978863e-05, "loss": 0.0419, "step": 686 }, { "epoch": 23.689655172413794, "grad_norm": 3.578125, "learning_rate": 2.4894762555465772e-05, "loss": 0.0263, "step": 687 }, { "epoch": 23.724137931034484, "grad_norm": 2.953125, "learning_rate": 2.4893629540086216e-05, "loss": 0.0673, "step": 688 }, { "epoch": 23.75862068965517, "grad_norm": 3.578125, "learning_rate": 2.489249048420218e-05, "loss": 0.025, "step": 689 }, { "epoch": 23.79310344827586, "grad_norm": 4.75, "learning_rate": 2.4891345388368834e-05, "loss": 0.1349, "step": 690 }, { "epoch": 23.82758620689655, "grad_norm": 2.859375, "learning_rate": 2.489019425314427e-05, "loss": 0.0605, "step": 691 }, { "epoch": 23.862068965517242, "grad_norm": 9.375, "learning_rate": 2.488903707908955e-05, "loss": 0.1087, "step": 692 }, { "epoch": 23.896551724137932, "grad_norm": 3.765625, "learning_rate": 2.488787386676866e-05, "loss": 0.0397, "step": 693 }, { "epoch": 23.93103448275862, "grad_norm": 10.4375, "learning_rate": 2.4886704616748536e-05, "loss": 0.09, "step": 694 }, { "epoch": 23.96551724137931, "grad_norm": 2.484375, "learning_rate": 2.488552932959906e-05, "loss": 0.0233, "step": 695 }, { "epoch": 24.0, "grad_norm": 7.84375, "learning_rate": 2.4884348005893045e-05, "loss": 0.0817, "step": 696 }, { "epoch": 24.03448275862069, "grad_norm": 4.375, "learning_rate": 2.488316064620626e-05, "loss": 0.1033, "step": 697 }, { "epoch": 24.06896551724138, "grad_norm": 7.65625, "learning_rate": 2.488196725111741e-05, "loss": 0.0722, "step": 698 }, { "epoch": 24.103448275862068, "grad_norm": 4.9375, "learning_rate": 2.488076782120814e-05, "loss": 0.046, "step": 699 }, { "epoch": 24.137931034482758, "grad_norm": 2.484375, "learning_rate": 2.4879562357063037e-05, "loss": 0.0237, "step": 700 }, { "epoch": 24.17241379310345, "grad_norm": 2.109375, "learning_rate": 2.487835085926963e-05, "loss": 0.0458, "step": 701 }, { "epoch": 24.20689655172414, "grad_norm": 2.265625, "learning_rate": 2.487713332841839e-05, "loss": 0.0185, "step": 702 }, { "epoch": 24.24137931034483, "grad_norm": 9.3125, "learning_rate": 2.487590976510272e-05, "loss": 0.1149, "step": 703 }, { "epoch": 24.275862068965516, "grad_norm": 2.453125, "learning_rate": 2.4874680169918976e-05, "loss": 0.0642, "step": 704 }, { "epoch": 24.310344827586206, "grad_norm": 5.90625, "learning_rate": 2.4873444543466448e-05, "loss": 0.0566, "step": 705 }, { "epoch": 24.344827586206897, "grad_norm": 0.9453125, "learning_rate": 2.4872202886347362e-05, "loss": 0.0395, "step": 706 }, { "epoch": 24.379310344827587, "grad_norm": 4.5625, "learning_rate": 2.487095519916689e-05, "loss": 0.0615, "step": 707 }, { "epoch": 24.413793103448278, "grad_norm": 4.90625, "learning_rate": 2.4869701482533138e-05, "loss": 0.0306, "step": 708 }, { "epoch": 24.448275862068964, "grad_norm": 6.4375, "learning_rate": 2.4868441737057153e-05, "loss": 0.0672, "step": 709 }, { "epoch": 24.482758620689655, "grad_norm": 5.03125, "learning_rate": 2.486717596335292e-05, "loss": 0.0355, "step": 710 }, { "epoch": 24.517241379310345, "grad_norm": 2.921875, "learning_rate": 2.4865904162037358e-05, "loss": 0.0563, "step": 711 }, { "epoch": 24.551724137931036, "grad_norm": 3.484375, "learning_rate": 2.4864626333730328e-05, "loss": 0.0656, "step": 712 }, { "epoch": 24.586206896551722, "grad_norm": 2.71875, "learning_rate": 2.4863342479054633e-05, "loss": 0.0577, "step": 713 }, { "epoch": 24.620689655172413, "grad_norm": 6.75, "learning_rate": 2.4862052598636005e-05, "loss": 0.0408, "step": 714 }, { "epoch": 24.655172413793103, "grad_norm": 11.0, "learning_rate": 2.4860756693103114e-05, "loss": 0.071, "step": 715 }, { "epoch": 24.689655172413794, "grad_norm": 4.65625, "learning_rate": 2.4859454763087577e-05, "loss": 0.0375, "step": 716 }, { "epoch": 24.724137931034484, "grad_norm": 1.0234375, "learning_rate": 2.4858146809223925e-05, "loss": 0.0076, "step": 717 }, { "epoch": 24.75862068965517, "grad_norm": 0.8671875, "learning_rate": 2.485683283214965e-05, "loss": 0.0375, "step": 718 }, { "epoch": 24.79310344827586, "grad_norm": 3.5, "learning_rate": 2.485551283250517e-05, "loss": 0.0377, "step": 719 }, { "epoch": 24.82758620689655, "grad_norm": 2.796875, "learning_rate": 2.4854186810933828e-05, "loss": 0.0456, "step": 720 }, { "epoch": 24.862068965517242, "grad_norm": 9.8125, "learning_rate": 2.4852854768081912e-05, "loss": 0.0904, "step": 721 }, { "epoch": 24.896551724137932, "grad_norm": 4.65625, "learning_rate": 2.485151670459865e-05, "loss": 0.0769, "step": 722 }, { "epoch": 24.93103448275862, "grad_norm": 4.15625, "learning_rate": 2.485017262113619e-05, "loss": 0.0526, "step": 723 }, { "epoch": 24.96551724137931, "grad_norm": 3.296875, "learning_rate": 2.484882251834963e-05, "loss": 0.0689, "step": 724 }, { "epoch": 25.0, "grad_norm": 1.59375, "learning_rate": 2.484746639689699e-05, "loss": 0.0189, "step": 725 }, { "epoch": 25.03448275862069, "grad_norm": 1.7109375, "learning_rate": 2.4846104257439222e-05, "loss": 0.0154, "step": 726 }, { "epoch": 25.06896551724138, "grad_norm": 2.78125, "learning_rate": 2.4844736100640224e-05, "loss": 0.0419, "step": 727 }, { "epoch": 25.103448275862068, "grad_norm": 1.96875, "learning_rate": 2.484336192716682e-05, "loss": 0.0297, "step": 728 }, { "epoch": 25.137931034482758, "grad_norm": 2.875, "learning_rate": 2.4841981737688754e-05, "loss": 0.0507, "step": 729 }, { "epoch": 25.17241379310345, "grad_norm": 1.1015625, "learning_rate": 2.4840595532878727e-05, "loss": 0.0334, "step": 730 }, { "epoch": 25.20689655172414, "grad_norm": 5.5, "learning_rate": 2.483920331341235e-05, "loss": 0.0506, "step": 731 }, { "epoch": 25.24137931034483, "grad_norm": 2.125, "learning_rate": 2.4837805079968176e-05, "loss": 0.0504, "step": 732 }, { "epoch": 25.275862068965516, "grad_norm": 10.3125, "learning_rate": 2.483640083322769e-05, "loss": 0.0515, "step": 733 }, { "epoch": 25.310344827586206, "grad_norm": 1.8515625, "learning_rate": 2.48349905738753e-05, "loss": 0.0107, "step": 734 }, { "epoch": 25.344827586206897, "grad_norm": 4.53125, "learning_rate": 2.483357430259835e-05, "loss": 0.0463, "step": 735 }, { "epoch": 25.379310344827587, "grad_norm": 2.71875, "learning_rate": 2.483215202008712e-05, "loss": 0.0617, "step": 736 }, { "epoch": 25.413793103448278, "grad_norm": 2.359375, "learning_rate": 2.4830723727034805e-05, "loss": 0.031, "step": 737 }, { "epoch": 25.448275862068964, "grad_norm": 1.046875, "learning_rate": 2.4829289424137538e-05, "loss": 0.0183, "step": 738 }, { "epoch": 25.482758620689655, "grad_norm": 1.8671875, "learning_rate": 2.4827849112094386e-05, "loss": 0.0379, "step": 739 }, { "epoch": 25.517241379310345, "grad_norm": 8.375, "learning_rate": 2.482640279160734e-05, "loss": 0.0301, "step": 740 }, { "epoch": 25.551724137931036, "grad_norm": 3.125, "learning_rate": 2.4824950463381314e-05, "loss": 0.0391, "step": 741 }, { "epoch": 25.586206896551722, "grad_norm": 6.59375, "learning_rate": 2.4823492128124154e-05, "loss": 0.0436, "step": 742 }, { "epoch": 25.620689655172413, "grad_norm": 1.40625, "learning_rate": 2.482202778654664e-05, "loss": 0.0167, "step": 743 }, { "epoch": 25.655172413793103, "grad_norm": 4.40625, "learning_rate": 2.4820557439362478e-05, "loss": 0.0349, "step": 744 }, { "epoch": 25.689655172413794, "grad_norm": 3.53125, "learning_rate": 2.4819081087288286e-05, "loss": 0.0874, "step": 745 }, { "epoch": 25.724137931034484, "grad_norm": 4.90625, "learning_rate": 2.481759873104363e-05, "loss": 0.0542, "step": 746 }, { "epoch": 25.75862068965517, "grad_norm": 4.78125, "learning_rate": 2.4816110371350982e-05, "loss": 0.031, "step": 747 }, { "epoch": 25.79310344827586, "grad_norm": 5.28125, "learning_rate": 2.4814616008935762e-05, "loss": 0.0243, "step": 748 }, { "epoch": 25.82758620689655, "grad_norm": 6.8125, "learning_rate": 2.4813115644526298e-05, "loss": 0.0685, "step": 749 }, { "epoch": 25.862068965517242, "grad_norm": 1.234375, "learning_rate": 2.4811609278853847e-05, "loss": 0.0138, "step": 750 }, { "epoch": 25.896551724137932, "grad_norm": 5.125, "learning_rate": 2.4810096912652604e-05, "loss": 0.0215, "step": 751 }, { "epoch": 25.93103448275862, "grad_norm": 7.5625, "learning_rate": 2.4808578546659663e-05, "loss": 0.0366, "step": 752 }, { "epoch": 25.96551724137931, "grad_norm": 9.375, "learning_rate": 2.4807054181615068e-05, "loss": 0.0499, "step": 753 }, { "epoch": 26.0, "grad_norm": 9.625, "learning_rate": 2.4805523818261774e-05, "loss": 0.0808, "step": 754 }, { "epoch": 26.03448275862069, "grad_norm": 2.9375, "learning_rate": 2.4803987457345658e-05, "loss": 0.0716, "step": 755 }, { "epoch": 26.06896551724138, "grad_norm": 9.8125, "learning_rate": 2.4802445099615525e-05, "loss": 0.0714, "step": 756 }, { "epoch": 26.103448275862068, "grad_norm": 3.703125, "learning_rate": 2.4800896745823105e-05, "loss": 0.0247, "step": 757 }, { "epoch": 26.137931034482758, "grad_norm": 4.5, "learning_rate": 2.4799342396723043e-05, "loss": 0.0511, "step": 758 }, { "epoch": 26.17241379310345, "grad_norm": 1.234375, "learning_rate": 2.479778205307291e-05, "loss": 0.0091, "step": 759 }, { "epoch": 26.20689655172414, "grad_norm": 4.21875, "learning_rate": 2.4796215715633196e-05, "loss": 0.0243, "step": 760 }, { "epoch": 26.24137931034483, "grad_norm": 3.71875, "learning_rate": 2.4794643385167327e-05, "loss": 0.0752, "step": 761 }, { "epoch": 26.275862068965516, "grad_norm": 3.921875, "learning_rate": 2.479306506244162e-05, "loss": 0.0386, "step": 762 }, { "epoch": 26.310344827586206, "grad_norm": 5.875, "learning_rate": 2.4791480748225344e-05, "loss": 0.1354, "step": 763 }, { "epoch": 26.344827586206897, "grad_norm": 3.015625, "learning_rate": 2.478989044329067e-05, "loss": 0.0309, "step": 764 }, { "epoch": 26.379310344827587, "grad_norm": 3.28125, "learning_rate": 2.4788294148412693e-05, "loss": 0.0278, "step": 765 }, { "epoch": 26.413793103448278, "grad_norm": 7.1875, "learning_rate": 2.478669186436943e-05, "loss": 0.0412, "step": 766 }, { "epoch": 26.448275862068964, "grad_norm": 3.015625, "learning_rate": 2.478508359194181e-05, "loss": 0.021, "step": 767 }, { "epoch": 26.482758620689655, "grad_norm": 10.0625, "learning_rate": 2.4783469331913688e-05, "loss": 0.1393, "step": 768 }, { "epoch": 26.517241379310345, "grad_norm": 3.296875, "learning_rate": 2.4781849085071833e-05, "loss": 0.0766, "step": 769 }, { "epoch": 26.551724137931036, "grad_norm": 4.5625, "learning_rate": 2.478022285220594e-05, "loss": 0.0229, "step": 770 }, { "epoch": 26.586206896551722, "grad_norm": 3.84375, "learning_rate": 2.4778590634108613e-05, "loss": 0.0686, "step": 771 }, { "epoch": 26.620689655172413, "grad_norm": 9.1875, "learning_rate": 2.4776952431575364e-05, "loss": 0.0645, "step": 772 }, { "epoch": 26.655172413793103, "grad_norm": 2.484375, "learning_rate": 2.477530824540465e-05, "loss": 0.0502, "step": 773 }, { "epoch": 26.689655172413794, "grad_norm": 3.6875, "learning_rate": 2.4773658076397816e-05, "loss": 0.0788, "step": 774 }, { "epoch": 26.724137931034484, "grad_norm": 2.921875, "learning_rate": 2.4772001925359133e-05, "loss": 0.0521, "step": 775 }, { "epoch": 26.75862068965517, "grad_norm": 1.7578125, "learning_rate": 2.4770339793095802e-05, "loss": 0.0191, "step": 776 }, { "epoch": 26.79310344827586, "grad_norm": 2.734375, "learning_rate": 2.4768671680417914e-05, "loss": 0.037, "step": 777 }, { "epoch": 26.82758620689655, "grad_norm": 3.765625, "learning_rate": 2.476699758813849e-05, "loss": 0.04, "step": 778 }, { "epoch": 26.862068965517242, "grad_norm": 4.1875, "learning_rate": 2.476531751707346e-05, "loss": 0.0563, "step": 779 }, { "epoch": 26.896551724137932, "grad_norm": 3.09375, "learning_rate": 2.4763631468041675e-05, "loss": 0.0769, "step": 780 }, { "epoch": 26.93103448275862, "grad_norm": 3.40625, "learning_rate": 2.4761939441864895e-05, "loss": 0.0424, "step": 781 }, { "epoch": 26.96551724137931, "grad_norm": 4.53125, "learning_rate": 2.476024143936779e-05, "loss": 0.0483, "step": 782 }, { "epoch": 27.0, "grad_norm": 1.53125, "learning_rate": 2.4758537461377947e-05, "loss": 0.0515, "step": 783 }, { "epoch": 27.03448275862069, "grad_norm": 4.8125, "learning_rate": 2.475682750872586e-05, "loss": 0.0417, "step": 784 }, { "epoch": 27.06896551724138, "grad_norm": 0.6328125, "learning_rate": 2.475511158224495e-05, "loss": 0.0136, "step": 785 }, { "epoch": 27.103448275862068, "grad_norm": 1.90625, "learning_rate": 2.4753389682771523e-05, "loss": 0.018, "step": 786 }, { "epoch": 27.137931034482758, "grad_norm": 2.890625, "learning_rate": 2.4751661811144828e-05, "loss": 0.017, "step": 787 }, { "epoch": 27.17241379310345, "grad_norm": 3.875, "learning_rate": 2.4749927968206996e-05, "loss": 0.0183, "step": 788 }, { "epoch": 27.20689655172414, "grad_norm": 7.25, "learning_rate": 2.4748188154803086e-05, "loss": 0.0604, "step": 789 }, { "epoch": 27.24137931034483, "grad_norm": 17.0, "learning_rate": 2.4746442371781066e-05, "loss": 0.0577, "step": 790 }, { "epoch": 27.275862068965516, "grad_norm": 3.484375, "learning_rate": 2.47446906199918e-05, "loss": 0.0244, "step": 791 }, { "epoch": 27.310344827586206, "grad_norm": 3.21875, "learning_rate": 2.474293290028908e-05, "loss": 0.0315, "step": 792 }, { "epoch": 27.344827586206897, "grad_norm": 11.9375, "learning_rate": 2.4741169213529594e-05, "loss": 0.1211, "step": 793 }, { "epoch": 27.379310344827587, "grad_norm": 0.734375, "learning_rate": 2.473939956057294e-05, "loss": 0.0057, "step": 794 }, { "epoch": 27.413793103448278, "grad_norm": 2.671875, "learning_rate": 2.4737623942281624e-05, "loss": 0.0419, "step": 795 }, { "epoch": 27.448275862068964, "grad_norm": 0.69921875, "learning_rate": 2.4735842359521064e-05, "loss": 0.0212, "step": 796 }, { "epoch": 27.482758620689655, "grad_norm": 10.4375, "learning_rate": 2.473405481315958e-05, "loss": 0.1407, "step": 797 }, { "epoch": 27.517241379310345, "grad_norm": 6.25, "learning_rate": 2.4732261304068403e-05, "loss": 0.0392, "step": 798 }, { "epoch": 27.551724137931036, "grad_norm": 4.21875, "learning_rate": 2.4730461833121664e-05, "loss": 0.0207, "step": 799 }, { "epoch": 27.586206896551722, "grad_norm": 4.1875, "learning_rate": 2.4728656401196406e-05, "loss": 0.0795, "step": 800 }, { "epoch": 27.620689655172413, "grad_norm": 5.3125, "learning_rate": 2.4726845009172572e-05, "loss": 0.0391, "step": 801 }, { "epoch": 27.655172413793103, "grad_norm": 0.7890625, "learning_rate": 2.4725027657933013e-05, "loss": 0.1046, "step": 802 }, { "epoch": 27.689655172413794, "grad_norm": 4.46875, "learning_rate": 2.4723204348363487e-05, "loss": 0.0937, "step": 803 }, { "epoch": 27.724137931034484, "grad_norm": 2.5, "learning_rate": 2.4721375081352648e-05, "loss": 0.0262, "step": 804 }, { "epoch": 27.75862068965517, "grad_norm": 5.46875, "learning_rate": 2.4719539857792068e-05, "loss": 0.0293, "step": 805 }, { "epoch": 27.79310344827586, "grad_norm": 4.28125, "learning_rate": 2.47176986785762e-05, "loss": 0.0283, "step": 806 }, { "epoch": 27.82758620689655, "grad_norm": 4.03125, "learning_rate": 2.471585154460242e-05, "loss": 0.0558, "step": 807 }, { "epoch": 27.862068965517242, "grad_norm": 4.65625, "learning_rate": 2.4713998456771e-05, "loss": 0.0288, "step": 808 }, { "epoch": 27.896551724137932, "grad_norm": 5.125, "learning_rate": 2.4712139415985104e-05, "loss": 0.0193, "step": 809 }, { "epoch": 27.93103448275862, "grad_norm": 0.50390625, "learning_rate": 2.471027442315081e-05, "loss": 0.0123, "step": 810 }, { "epoch": 27.96551724137931, "grad_norm": 6.84375, "learning_rate": 2.47084034791771e-05, "loss": 0.0779, "step": 811 }, { "epoch": 28.0, "grad_norm": 3.8125, "learning_rate": 2.4706526584975843e-05, "loss": 0.0844, "step": 812 }, { "epoch": 28.03448275862069, "grad_norm": 5.34375, "learning_rate": 2.4704643741461815e-05, "loss": 0.0571, "step": 813 }, { "epoch": 28.06896551724138, "grad_norm": 3.046875, "learning_rate": 2.4702754949552686e-05, "loss": 0.0509, "step": 814 }, { "epoch": 28.103448275862068, "grad_norm": 2.09375, "learning_rate": 2.470086021016904e-05, "loss": 0.0415, "step": 815 }, { "epoch": 28.137931034482758, "grad_norm": 6.1875, "learning_rate": 2.4698959524234346e-05, "loss": 0.0591, "step": 816 }, { "epoch": 28.17241379310345, "grad_norm": 7.9375, "learning_rate": 2.4697052892674972e-05, "loss": 0.0525, "step": 817 }, { "epoch": 28.20689655172414, "grad_norm": 5.71875, "learning_rate": 2.469514031642019e-05, "loss": 0.0381, "step": 818 }, { "epoch": 28.24137931034483, "grad_norm": 3.25, "learning_rate": 2.4693221796402166e-05, "loss": 0.0411, "step": 819 }, { "epoch": 28.275862068965516, "grad_norm": 6.1875, "learning_rate": 2.4691297333555964e-05, "loss": 0.0437, "step": 820 }, { "epoch": 28.310344827586206, "grad_norm": 1.2421875, "learning_rate": 2.468936692881954e-05, "loss": 0.0507, "step": 821 }, { "epoch": 28.344827586206897, "grad_norm": 19.375, "learning_rate": 2.468743058313376e-05, "loss": 0.0531, "step": 822 }, { "epoch": 28.379310344827587, "grad_norm": 5.90625, "learning_rate": 2.4685488297442362e-05, "loss": 0.0668, "step": 823 }, { "epoch": 28.413793103448278, "grad_norm": 1.984375, "learning_rate": 2.4683540072692003e-05, "loss": 0.0305, "step": 824 }, { "epoch": 28.448275862068964, "grad_norm": 3.75, "learning_rate": 2.4681585909832217e-05, "loss": 0.0159, "step": 825 }, { "epoch": 28.482758620689655, "grad_norm": 2.296875, "learning_rate": 2.4679625809815443e-05, "loss": 0.0338, "step": 826 }, { "epoch": 28.517241379310345, "grad_norm": 3.703125, "learning_rate": 2.4677659773597007e-05, "loss": 0.0255, "step": 827 }, { "epoch": 28.551724137931036, "grad_norm": 4.40625, "learning_rate": 2.4675687802135135e-05, "loss": 0.0945, "step": 828 }, { "epoch": 28.586206896551722, "grad_norm": 7.3125, "learning_rate": 2.4673709896390938e-05, "loss": 0.0497, "step": 829 }, { "epoch": 28.620689655172413, "grad_norm": 8.0625, "learning_rate": 2.467172605732843e-05, "loss": 0.0529, "step": 830 }, { "epoch": 28.655172413793103, "grad_norm": 2.734375, "learning_rate": 2.4669736285914505e-05, "loss": 0.0167, "step": 831 }, { "epoch": 28.689655172413794, "grad_norm": 4.96875, "learning_rate": 2.4667740583118952e-05, "loss": 0.0451, "step": 832 }, { "epoch": 28.724137931034484, "grad_norm": 10.5625, "learning_rate": 2.466573894991445e-05, "loss": 0.0549, "step": 833 }, { "epoch": 28.75862068965517, "grad_norm": 2.703125, "learning_rate": 2.4663731387276583e-05, "loss": 0.0223, "step": 834 }, { "epoch": 28.79310344827586, "grad_norm": 4.625, "learning_rate": 2.46617178961838e-05, "loss": 0.0813, "step": 835 }, { "epoch": 28.82758620689655, "grad_norm": 2.71875, "learning_rate": 2.465969847761746e-05, "loss": 0.0203, "step": 836 }, { "epoch": 28.862068965517242, "grad_norm": 1.484375, "learning_rate": 2.4657673132561797e-05, "loss": 0.0194, "step": 837 }, { "epoch": 28.896551724137932, "grad_norm": 1.65625, "learning_rate": 2.4655641862003943e-05, "loss": 0.0359, "step": 838 }, { "epoch": 28.93103448275862, "grad_norm": 2.390625, "learning_rate": 2.465360466693391e-05, "loss": 0.0249, "step": 839 }, { "epoch": 28.96551724137931, "grad_norm": 5.0625, "learning_rate": 2.465156154834461e-05, "loss": 0.0441, "step": 840 }, { "epoch": 29.0, "grad_norm": 2.03125, "learning_rate": 2.464951250723183e-05, "loss": 0.0377, "step": 841 }, { "epoch": 29.03448275862069, "grad_norm": 1.5390625, "learning_rate": 2.4647457544594246e-05, "loss": 0.0397, "step": 842 }, { "epoch": 29.06896551724138, "grad_norm": 7.78125, "learning_rate": 2.4645396661433424e-05, "loss": 0.0408, "step": 843 }, { "epoch": 29.103448275862068, "grad_norm": 7.1875, "learning_rate": 2.4643329858753818e-05, "loss": 0.0709, "step": 844 }, { "epoch": 29.137931034482758, "grad_norm": 5.3125, "learning_rate": 2.464125713756275e-05, "loss": 0.0732, "step": 845 }, { "epoch": 29.17241379310345, "grad_norm": 1.5234375, "learning_rate": 2.4639178498870452e-05, "loss": 0.0165, "step": 846 }, { "epoch": 29.20689655172414, "grad_norm": 4.78125, "learning_rate": 2.463709394369002e-05, "loss": 0.0522, "step": 847 }, { "epoch": 29.24137931034483, "grad_norm": 1.96875, "learning_rate": 2.4635003473037444e-05, "loss": 0.0609, "step": 848 }, { "epoch": 29.275862068965516, "grad_norm": 3.515625, "learning_rate": 2.4632907087931593e-05, "loss": 0.0543, "step": 849 }, { "epoch": 29.310344827586206, "grad_norm": 7.625, "learning_rate": 2.4630804789394218e-05, "loss": 0.066, "step": 850 }, { "epoch": 29.344827586206897, "grad_norm": 4.0625, "learning_rate": 2.4628696578449956e-05, "loss": 0.1002, "step": 851 }, { "epoch": 29.379310344827587, "grad_norm": 3.40625, "learning_rate": 2.4626582456126324e-05, "loss": 0.0305, "step": 852 }, { "epoch": 29.413793103448278, "grad_norm": 2.734375, "learning_rate": 2.4624462423453716e-05, "loss": 0.0214, "step": 853 }, { "epoch": 29.448275862068964, "grad_norm": 0.6640625, "learning_rate": 2.4622336481465414e-05, "loss": 0.0088, "step": 854 }, { "epoch": 29.482758620689655, "grad_norm": 1.3671875, "learning_rate": 2.4620204631197574e-05, "loss": 0.0223, "step": 855 }, { "epoch": 29.517241379310345, "grad_norm": 3.140625, "learning_rate": 2.4618066873689238e-05, "loss": 0.0345, "step": 856 }, { "epoch": 29.551724137931036, "grad_norm": 1.515625, "learning_rate": 2.4615923209982316e-05, "loss": 0.03, "step": 857 }, { "epoch": 29.586206896551722, "grad_norm": 5.34375, "learning_rate": 2.4613773641121607e-05, "loss": 0.0329, "step": 858 }, { "epoch": 29.620689655172413, "grad_norm": 7.65625, "learning_rate": 2.461161816815479e-05, "loss": 0.0534, "step": 859 }, { "epoch": 29.655172413793103, "grad_norm": 7.15625, "learning_rate": 2.4609456792132407e-05, "loss": 0.0471, "step": 860 }, { "epoch": 29.689655172413794, "grad_norm": 10.75, "learning_rate": 2.4607289514107888e-05, "loss": 0.0682, "step": 861 }, { "epoch": 29.724137931034484, "grad_norm": 3.609375, "learning_rate": 2.4605116335137547e-05, "loss": 0.0543, "step": 862 }, { "epoch": 29.75862068965517, "grad_norm": 0.8828125, "learning_rate": 2.4602937256280558e-05, "loss": 0.0311, "step": 863 }, { "epoch": 29.79310344827586, "grad_norm": 6.40625, "learning_rate": 2.4600752278598978e-05, "loss": 0.038, "step": 864 }, { "epoch": 29.82758620689655, "grad_norm": 3.34375, "learning_rate": 2.4598561403157734e-05, "loss": 0.0402, "step": 865 }, { "epoch": 29.862068965517242, "grad_norm": 1.4609375, "learning_rate": 2.4596364631024643e-05, "loss": 0.0305, "step": 866 }, { "epoch": 29.896551724137932, "grad_norm": 2.046875, "learning_rate": 2.4594161963270373e-05, "loss": 0.0324, "step": 867 }, { "epoch": 29.93103448275862, "grad_norm": 3.421875, "learning_rate": 2.459195340096849e-05, "loss": 0.0286, "step": 868 }, { "epoch": 29.96551724137931, "grad_norm": 7.15625, "learning_rate": 2.4589738945195413e-05, "loss": 0.0494, "step": 869 }, { "epoch": 30.0, "grad_norm": 6.71875, "learning_rate": 2.458751859703044e-05, "loss": 0.0373, "step": 870 }, { "epoch": 30.03448275862069, "grad_norm": 4.15625, "learning_rate": 2.4585292357555746e-05, "loss": 0.0119, "step": 871 }, { "epoch": 30.06896551724138, "grad_norm": 5.0625, "learning_rate": 2.4583060227856368e-05, "loss": 0.0248, "step": 872 }, { "epoch": 30.103448275862068, "grad_norm": 4.6875, "learning_rate": 2.458082220902022e-05, "loss": 0.019, "step": 873 }, { "epoch": 30.137931034482758, "grad_norm": 1.2734375, "learning_rate": 2.4578578302138087e-05, "loss": 0.0083, "step": 874 }, { "epoch": 30.17241379310345, "grad_norm": 0.7734375, "learning_rate": 2.457632850830363e-05, "loss": 0.0371, "step": 875 }, { "epoch": 30.20689655172414, "grad_norm": 8.125, "learning_rate": 2.4574072828613354e-05, "loss": 0.0944, "step": 876 }, { "epoch": 30.24137931034483, "grad_norm": 1.890625, "learning_rate": 2.457181126416666e-05, "loss": 0.0335, "step": 877 }, { "epoch": 30.275862068965516, "grad_norm": 2.765625, "learning_rate": 2.456954381606581e-05, "loss": 0.0137, "step": 878 }, { "epoch": 30.310344827586206, "grad_norm": 2.125, "learning_rate": 2.4567270485415926e-05, "loss": 0.093, "step": 879 }, { "epoch": 30.344827586206897, "grad_norm": 4.03125, "learning_rate": 2.4564991273325e-05, "loss": 0.0774, "step": 880 }, { "epoch": 30.379310344827587, "grad_norm": 0.6328125, "learning_rate": 2.4562706180903894e-05, "loss": 0.0123, "step": 881 }, { "epoch": 30.413793103448278, "grad_norm": 3.1875, "learning_rate": 2.4560415209266334e-05, "loss": 0.0246, "step": 882 }, { "epoch": 30.448275862068964, "grad_norm": 1.96875, "learning_rate": 2.455811835952892e-05, "loss": 0.0279, "step": 883 }, { "epoch": 30.482758620689655, "grad_norm": 10.75, "learning_rate": 2.4555815632811092e-05, "loss": 0.1412, "step": 884 }, { "epoch": 30.517241379310345, "grad_norm": 1.7109375, "learning_rate": 2.4553507030235184e-05, "loss": 0.0213, "step": 885 }, { "epoch": 30.551724137931036, "grad_norm": 3.1875, "learning_rate": 2.455119255292638e-05, "loss": 0.0227, "step": 886 }, { "epoch": 30.586206896551722, "grad_norm": 3.890625, "learning_rate": 2.4548872202012718e-05, "loss": 0.0512, "step": 887 }, { "epoch": 30.620689655172413, "grad_norm": 2.875, "learning_rate": 2.4546545978625126e-05, "loss": 0.0331, "step": 888 }, { "epoch": 30.655172413793103, "grad_norm": 7.03125, "learning_rate": 2.454421388389736e-05, "loss": 0.0586, "step": 889 }, { "epoch": 30.689655172413794, "grad_norm": 7.59375, "learning_rate": 2.4541875918966063e-05, "loss": 0.0847, "step": 890 }, { "epoch": 30.724137931034484, "grad_norm": 4.625, "learning_rate": 2.453953208497073e-05, "loss": 0.0295, "step": 891 }, { "epoch": 30.75862068965517, "grad_norm": 2.484375, "learning_rate": 2.4537182383053713e-05, "loss": 0.0409, "step": 892 }, { "epoch": 30.79310344827586, "grad_norm": 5.0625, "learning_rate": 2.4534826814360236e-05, "loss": 0.0569, "step": 893 }, { "epoch": 30.82758620689655, "grad_norm": 4.9375, "learning_rate": 2.453246538003837e-05, "loss": 0.0538, "step": 894 }, { "epoch": 30.862068965517242, "grad_norm": 2.34375, "learning_rate": 2.453009808123905e-05, "loss": 0.0185, "step": 895 }, { "epoch": 30.896551724137932, "grad_norm": 3.0, "learning_rate": 2.452772491911607e-05, "loss": 0.0202, "step": 896 }, { "epoch": 30.93103448275862, "grad_norm": 2.765625, "learning_rate": 2.4525345894826073e-05, "loss": 0.0213, "step": 897 }, { "epoch": 30.96551724137931, "grad_norm": 1.765625, "learning_rate": 2.4522961009528577e-05, "loss": 0.0161, "step": 898 }, { "epoch": 31.0, "grad_norm": 1.671875, "learning_rate": 2.452057026438594e-05, "loss": 0.0451, "step": 899 }, { "epoch": 31.03448275862069, "grad_norm": 1.71875, "learning_rate": 2.451817366056339e-05, "loss": 0.0509, "step": 900 }, { "epoch": 31.06896551724138, "grad_norm": 1.359375, "learning_rate": 2.4515771199228987e-05, "loss": 0.0253, "step": 901 }, { "epoch": 31.103448275862068, "grad_norm": 8.75, "learning_rate": 2.4513362881553674e-05, "loss": 0.0566, "step": 902 }, { "epoch": 31.137931034482758, "grad_norm": 4.75, "learning_rate": 2.4510948708711234e-05, "loss": 0.0387, "step": 903 }, { "epoch": 31.17241379310345, "grad_norm": 1.1875, "learning_rate": 2.45085286818783e-05, "loss": 0.0294, "step": 904 }, { "epoch": 31.20689655172414, "grad_norm": 7.53125, "learning_rate": 2.4506102802234372e-05, "loss": 0.0478, "step": 905 }, { "epoch": 31.24137931034483, "grad_norm": 5.4375, "learning_rate": 2.450367107096179e-05, "loss": 0.0924, "step": 906 }, { "epoch": 31.275862068965516, "grad_norm": 5.90625, "learning_rate": 2.450123348924575e-05, "loss": 0.072, "step": 907 }, { "epoch": 31.310344827586206, "grad_norm": 3.90625, "learning_rate": 2.4498790058274297e-05, "loss": 0.0452, "step": 908 }, { "epoch": 31.344827586206897, "grad_norm": 2.96875, "learning_rate": 2.4496340779238335e-05, "loss": 0.0238, "step": 909 }, { "epoch": 31.379310344827587, "grad_norm": 0.890625, "learning_rate": 2.449388565333161e-05, "loss": 0.0264, "step": 910 }, { "epoch": 31.413793103448278, "grad_norm": 0.78125, "learning_rate": 2.449142468175072e-05, "loss": 0.0353, "step": 911 }, { "epoch": 31.448275862068964, "grad_norm": 4.09375, "learning_rate": 2.4488957865695114e-05, "loss": 0.028, "step": 912 }, { "epoch": 31.482758620689655, "grad_norm": 18.75, "learning_rate": 2.4486485206367088e-05, "loss": 0.0983, "step": 913 }, { "epoch": 31.517241379310345, "grad_norm": 4.875, "learning_rate": 2.4484006704971784e-05, "loss": 0.0377, "step": 914 }, { "epoch": 31.551724137931036, "grad_norm": 8.0, "learning_rate": 2.44815223627172e-05, "loss": 0.0568, "step": 915 }, { "epoch": 31.586206896551722, "grad_norm": 0.84375, "learning_rate": 2.4479032180814166e-05, "loss": 0.0123, "step": 916 }, { "epoch": 31.620689655172413, "grad_norm": 6.53125, "learning_rate": 2.4476536160476372e-05, "loss": 0.064, "step": 917 }, { "epoch": 31.655172413793103, "grad_norm": 6.0, "learning_rate": 2.4474034302920347e-05, "loss": 0.03, "step": 918 }, { "epoch": 31.689655172413794, "grad_norm": 5.0, "learning_rate": 2.447152660936546e-05, "loss": 0.0262, "step": 919 }, { "epoch": 31.724137931034484, "grad_norm": 7.46875, "learning_rate": 2.4469013081033936e-05, "loss": 0.0619, "step": 920 }, { "epoch": 31.75862068965517, "grad_norm": 2.171875, "learning_rate": 2.446649371915084e-05, "loss": 0.0353, "step": 921 }, { "epoch": 31.79310344827586, "grad_norm": 1.921875, "learning_rate": 2.446396852494408e-05, "loss": 0.0779, "step": 922 }, { "epoch": 31.82758620689655, "grad_norm": 3.359375, "learning_rate": 2.4461437499644394e-05, "loss": 0.0226, "step": 923 }, { "epoch": 31.862068965517242, "grad_norm": 6.03125, "learning_rate": 2.4458900644485385e-05, "loss": 0.0668, "step": 924 }, { "epoch": 31.896551724137932, "grad_norm": 0.8515625, "learning_rate": 2.4456357960703475e-05, "loss": 0.0107, "step": 925 }, { "epoch": 31.93103448275862, "grad_norm": 2.546875, "learning_rate": 2.4453809449537947e-05, "loss": 0.0442, "step": 926 }, { "epoch": 31.96551724137931, "grad_norm": 2.953125, "learning_rate": 2.4451255112230904e-05, "loss": 0.0701, "step": 927 }, { "epoch": 32.0, "grad_norm": 2.078125, "learning_rate": 2.444869495002731e-05, "loss": 0.0438, "step": 928 }, { "epoch": 32.03448275862069, "grad_norm": 0.83984375, "learning_rate": 2.4446128964174954e-05, "loss": 0.0106, "step": 929 }, { "epoch": 32.06896551724138, "grad_norm": 3.125, "learning_rate": 2.4443557155924457e-05, "loss": 0.0474, "step": 930 }, { "epoch": 32.10344827586207, "grad_norm": 5.03125, "learning_rate": 2.4440979526529295e-05, "loss": 0.0321, "step": 931 }, { "epoch": 32.13793103448276, "grad_norm": 0.330078125, "learning_rate": 2.4438396077245778e-05, "loss": 0.0131, "step": 932 }, { "epoch": 32.172413793103445, "grad_norm": 5.3125, "learning_rate": 2.443580680933304e-05, "loss": 0.038, "step": 933 }, { "epoch": 32.206896551724135, "grad_norm": 6.09375, "learning_rate": 2.4433211724053055e-05, "loss": 0.0465, "step": 934 }, { "epoch": 32.241379310344826, "grad_norm": 5.84375, "learning_rate": 2.4430610822670654e-05, "loss": 0.0457, "step": 935 }, { "epoch": 32.275862068965516, "grad_norm": 0.6484375, "learning_rate": 2.4428004106453462e-05, "loss": 0.0267, "step": 936 }, { "epoch": 32.310344827586206, "grad_norm": 7.3125, "learning_rate": 2.4425391576671974e-05, "loss": 0.1117, "step": 937 }, { "epoch": 32.3448275862069, "grad_norm": 5.125, "learning_rate": 2.4422773234599506e-05, "loss": 0.0357, "step": 938 }, { "epoch": 32.37931034482759, "grad_norm": 1.75, "learning_rate": 2.44201490815122e-05, "loss": 0.0584, "step": 939 }, { "epoch": 32.41379310344828, "grad_norm": 4.4375, "learning_rate": 2.441751911868904e-05, "loss": 0.0463, "step": 940 }, { "epoch": 32.44827586206897, "grad_norm": 3.46875, "learning_rate": 2.4414883347411836e-05, "loss": 0.07, "step": 941 }, { "epoch": 32.48275862068966, "grad_norm": 5.34375, "learning_rate": 2.441224176896524e-05, "loss": 0.0621, "step": 942 }, { "epoch": 32.51724137931034, "grad_norm": 2.09375, "learning_rate": 2.4409594384636707e-05, "loss": 0.0182, "step": 943 }, { "epoch": 32.55172413793103, "grad_norm": 3.390625, "learning_rate": 2.4406941195716555e-05, "loss": 0.0634, "step": 944 }, { "epoch": 32.58620689655172, "grad_norm": 3.140625, "learning_rate": 2.440428220349791e-05, "loss": 0.045, "step": 945 }, { "epoch": 32.62068965517241, "grad_norm": 2.875, "learning_rate": 2.4401617409276735e-05, "loss": 0.0541, "step": 946 }, { "epoch": 32.6551724137931, "grad_norm": 2.265625, "learning_rate": 2.4398946814351818e-05, "loss": 0.0375, "step": 947 }, { "epoch": 32.689655172413794, "grad_norm": 3.015625, "learning_rate": 2.4396270420024774e-05, "loss": 0.0306, "step": 948 }, { "epoch": 32.724137931034484, "grad_norm": 2.609375, "learning_rate": 2.439358822760004e-05, "loss": 0.0326, "step": 949 }, { "epoch": 32.758620689655174, "grad_norm": 5.59375, "learning_rate": 2.439090023838489e-05, "loss": 0.0644, "step": 950 }, { "epoch": 32.793103448275865, "grad_norm": 2.59375, "learning_rate": 2.438820645368942e-05, "loss": 0.0228, "step": 951 }, { "epoch": 32.827586206896555, "grad_norm": 3.890625, "learning_rate": 2.438550687482654e-05, "loss": 0.0379, "step": 952 }, { "epoch": 32.86206896551724, "grad_norm": 3.734375, "learning_rate": 2.4382801503112003e-05, "loss": 0.0474, "step": 953 }, { "epoch": 32.89655172413793, "grad_norm": 2.53125, "learning_rate": 2.438009033986436e-05, "loss": 0.0143, "step": 954 }, { "epoch": 32.93103448275862, "grad_norm": 2.828125, "learning_rate": 2.437737338640501e-05, "loss": 0.0468, "step": 955 }, { "epoch": 32.96551724137931, "grad_norm": 2.109375, "learning_rate": 2.4374650644058156e-05, "loss": 0.0186, "step": 956 }, { "epoch": 33.0, "grad_norm": 3.328125, "learning_rate": 2.4371922114150835e-05, "loss": 0.0338, "step": 957 }, { "epoch": 33.03448275862069, "grad_norm": 2.234375, "learning_rate": 2.43691877980129e-05, "loss": 0.0308, "step": 958 }, { "epoch": 33.06896551724138, "grad_norm": 0.212890625, "learning_rate": 2.4366447696977018e-05, "loss": 0.0207, "step": 959 }, { "epoch": 33.10344827586207, "grad_norm": 0.87109375, "learning_rate": 2.4363701812378683e-05, "loss": 0.0054, "step": 960 }, { "epoch": 33.13793103448276, "grad_norm": 0.91796875, "learning_rate": 2.4360950145556208e-05, "loss": 0.0077, "step": 961 }, { "epoch": 33.172413793103445, "grad_norm": 6.9375, "learning_rate": 2.4358192697850718e-05, "loss": 0.0377, "step": 962 }, { "epoch": 33.206896551724135, "grad_norm": 0.65625, "learning_rate": 2.4355429470606166e-05, "loss": 0.0315, "step": 963 }, { "epoch": 33.241379310344826, "grad_norm": 1.265625, "learning_rate": 2.4352660465169307e-05, "loss": 0.0282, "step": 964 }, { "epoch": 33.275862068965516, "grad_norm": 3.265625, "learning_rate": 2.4349885682889727e-05, "loss": 0.0354, "step": 965 }, { "epoch": 33.310344827586206, "grad_norm": 0.90234375, "learning_rate": 2.4347105125119824e-05, "loss": 0.0438, "step": 966 }, { "epoch": 33.3448275862069, "grad_norm": 2.859375, "learning_rate": 2.4344318793214797e-05, "loss": 0.0148, "step": 967 }, { "epoch": 33.37931034482759, "grad_norm": 1.0234375, "learning_rate": 2.4341526688532678e-05, "loss": 0.0175, "step": 968 }, { "epoch": 33.41379310344828, "grad_norm": 10.1875, "learning_rate": 2.4338728812434305e-05, "loss": 0.1066, "step": 969 }, { "epoch": 33.44827586206897, "grad_norm": 7.0625, "learning_rate": 2.4335925166283325e-05, "loss": 0.0454, "step": 970 }, { "epoch": 33.48275862068966, "grad_norm": 0.9375, "learning_rate": 2.4333115751446208e-05, "loss": 0.0359, "step": 971 }, { "epoch": 33.51724137931034, "grad_norm": 3.203125, "learning_rate": 2.433030056929222e-05, "loss": 0.0149, "step": 972 }, { "epoch": 33.55172413793103, "grad_norm": 7.375, "learning_rate": 2.4327479621193448e-05, "loss": 0.1154, "step": 973 }, { "epoch": 33.58620689655172, "grad_norm": 1.9296875, "learning_rate": 2.4324652908524798e-05, "loss": 0.0151, "step": 974 }, { "epoch": 33.62068965517241, "grad_norm": 0.82421875, "learning_rate": 2.432182043266396e-05, "loss": 0.0227, "step": 975 }, { "epoch": 33.6551724137931, "grad_norm": 6.5625, "learning_rate": 2.4318982194991463e-05, "loss": 0.0438, "step": 976 }, { "epoch": 33.689655172413794, "grad_norm": 4.78125, "learning_rate": 2.431613819689062e-05, "loss": 0.0307, "step": 977 }, { "epoch": 33.724137931034484, "grad_norm": 5.46875, "learning_rate": 2.431328843974756e-05, "loss": 0.0518, "step": 978 }, { "epoch": 33.758620689655174, "grad_norm": 3.40625, "learning_rate": 2.4310432924951228e-05, "loss": 0.0423, "step": 979 }, { "epoch": 33.793103448275865, "grad_norm": 6.25, "learning_rate": 2.4307571653893364e-05, "loss": 0.0507, "step": 980 }, { "epoch": 33.827586206896555, "grad_norm": 3.46875, "learning_rate": 2.4304704627968515e-05, "loss": 0.0229, "step": 981 }, { "epoch": 33.86206896551724, "grad_norm": 2.75, "learning_rate": 2.4301831848574035e-05, "loss": 0.0295, "step": 982 }, { "epoch": 33.89655172413793, "grad_norm": 2.328125, "learning_rate": 2.429895331711008e-05, "loss": 0.0225, "step": 983 }, { "epoch": 33.93103448275862, "grad_norm": 1.3984375, "learning_rate": 2.429606903497962e-05, "loss": 0.0732, "step": 984 }, { "epoch": 33.96551724137931, "grad_norm": 0.9453125, "learning_rate": 2.4293179003588407e-05, "loss": 0.0366, "step": 985 }, { "epoch": 34.0, "grad_norm": 4.78125, "learning_rate": 2.429028322434501e-05, "loss": 0.0452, "step": 986 }, { "epoch": 34.03448275862069, "grad_norm": 6.09375, "learning_rate": 2.4287381698660807e-05, "loss": 0.0436, "step": 987 }, { "epoch": 34.06896551724138, "grad_norm": 2.84375, "learning_rate": 2.4284474427949956e-05, "loss": 0.0536, "step": 988 }, { "epoch": 34.10344827586207, "grad_norm": 2.796875, "learning_rate": 2.428156141362943e-05, "loss": 0.0378, "step": 989 }, { "epoch": 34.13793103448276, "grad_norm": 4.34375, "learning_rate": 2.427864265711899e-05, "loss": 0.0648, "step": 990 }, { "epoch": 34.172413793103445, "grad_norm": 6.1875, "learning_rate": 2.427571815984121e-05, "loss": 0.0624, "step": 991 }, { "epoch": 34.206896551724135, "grad_norm": 5.625, "learning_rate": 2.427278792322145e-05, "loss": 0.0526, "step": 992 }, { "epoch": 34.241379310344826, "grad_norm": 11.5625, "learning_rate": 2.4269851948687877e-05, "loss": 0.0716, "step": 993 }, { "epoch": 34.275862068965516, "grad_norm": 3.90625, "learning_rate": 2.426691023767145e-05, "loss": 0.0424, "step": 994 }, { "epoch": 34.310344827586206, "grad_norm": 10.125, "learning_rate": 2.426396279160591e-05, "loss": 0.0583, "step": 995 }, { "epoch": 34.3448275862069, "grad_norm": 4.90625, "learning_rate": 2.426100961192782e-05, "loss": 0.0318, "step": 996 }, { "epoch": 34.37931034482759, "grad_norm": 3.296875, "learning_rate": 2.425805070007652e-05, "loss": 0.0456, "step": 997 }, { "epoch": 34.41379310344828, "grad_norm": 2.171875, "learning_rate": 2.4255086057494146e-05, "loss": 0.0568, "step": 998 }, { "epoch": 34.44827586206897, "grad_norm": 4.125, "learning_rate": 2.4252115685625627e-05, "loss": 0.0582, "step": 999 }, { "epoch": 34.48275862068966, "grad_norm": 2.75, "learning_rate": 2.4249139585918695e-05, "loss": 0.0512, "step": 1000 }, { "epoch": 34.51724137931034, "grad_norm": 2.46875, "learning_rate": 2.4246157759823855e-05, "loss": 0.0435, "step": 1001 }, { "epoch": 34.55172413793103, "grad_norm": 3.796875, "learning_rate": 2.4243170208794418e-05, "loss": 0.0692, "step": 1002 }, { "epoch": 34.58620689655172, "grad_norm": 2.078125, "learning_rate": 2.424017693428648e-05, "loss": 0.0464, "step": 1003 }, { "epoch": 34.62068965517241, "grad_norm": 2.03125, "learning_rate": 2.4237177937758927e-05, "loss": 0.0259, "step": 1004 }, { "epoch": 34.6551724137931, "grad_norm": 4.9375, "learning_rate": 2.423417322067343e-05, "loss": 0.0322, "step": 1005 }, { "epoch": 34.689655172413794, "grad_norm": 1.3359375, "learning_rate": 2.423116278449445e-05, "loss": 0.021, "step": 1006 }, { "epoch": 34.724137931034484, "grad_norm": 9.9375, "learning_rate": 2.4228146630689245e-05, "loss": 0.0608, "step": 1007 }, { "epoch": 34.758620689655174, "grad_norm": 4.53125, "learning_rate": 2.4225124760727847e-05, "loss": 0.0241, "step": 1008 }, { "epoch": 34.793103448275865, "grad_norm": 2.765625, "learning_rate": 2.4222097176083074e-05, "loss": 0.0397, "step": 1009 }, { "epoch": 34.827586206896555, "grad_norm": 2.234375, "learning_rate": 2.421906387823055e-05, "loss": 0.026, "step": 1010 }, { "epoch": 34.86206896551724, "grad_norm": 2.359375, "learning_rate": 2.4216024868648644e-05, "loss": 0.0217, "step": 1011 }, { "epoch": 34.89655172413793, "grad_norm": 5.8125, "learning_rate": 2.421298014881855e-05, "loss": 0.051, "step": 1012 }, { "epoch": 34.93103448275862, "grad_norm": 13.75, "learning_rate": 2.4209929720224223e-05, "loss": 0.09, "step": 1013 }, { "epoch": 34.96551724137931, "grad_norm": 6.75, "learning_rate": 2.4206873584352404e-05, "loss": 0.0631, "step": 1014 }, { "epoch": 35.0, "grad_norm": 1.78125, "learning_rate": 2.4203811742692617e-05, "loss": 0.0356, "step": 1015 }, { "epoch": 35.03448275862069, "grad_norm": 4.28125, "learning_rate": 2.420074419673717e-05, "loss": 0.0418, "step": 1016 }, { "epoch": 35.06896551724138, "grad_norm": 2.6875, "learning_rate": 2.419767094798114e-05, "loss": 0.0454, "step": 1017 }, { "epoch": 35.10344827586207, "grad_norm": 3.328125, "learning_rate": 2.4194591997922393e-05, "loss": 0.0185, "step": 1018 }, { "epoch": 35.13793103448276, "grad_norm": 17.25, "learning_rate": 2.4191507348061572e-05, "loss": 0.0739, "step": 1019 }, { "epoch": 35.172413793103445, "grad_norm": 3.90625, "learning_rate": 2.4188416999902104e-05, "loss": 0.056, "step": 1020 }, { "epoch": 35.206896551724135, "grad_norm": 0.4921875, "learning_rate": 2.418532095495018e-05, "loss": 0.0133, "step": 1021 }, { "epoch": 35.241379310344826, "grad_norm": 2.734375, "learning_rate": 2.418221921471478e-05, "loss": 0.0498, "step": 1022 }, { "epoch": 35.275862068965516, "grad_norm": 3.171875, "learning_rate": 2.4179111780707652e-05, "loss": 0.0353, "step": 1023 }, { "epoch": 35.310344827586206, "grad_norm": 4.75, "learning_rate": 2.417599865444332e-05, "loss": 0.0153, "step": 1024 }, { "epoch": 35.3448275862069, "grad_norm": 3.453125, "learning_rate": 2.4172879837439086e-05, "loss": 0.0308, "step": 1025 }, { "epoch": 35.37931034482759, "grad_norm": 6.40625, "learning_rate": 2.4169755331215023e-05, "loss": 0.0417, "step": 1026 }, { "epoch": 35.41379310344828, "grad_norm": 6.8125, "learning_rate": 2.4166625137293976e-05, "loss": 0.0699, "step": 1027 }, { "epoch": 35.44827586206897, "grad_norm": 0.80078125, "learning_rate": 2.4163489257201564e-05, "loss": 0.0144, "step": 1028 }, { "epoch": 35.48275862068966, "grad_norm": 3.203125, "learning_rate": 2.4160347692466184e-05, "loss": 0.0288, "step": 1029 }, { "epoch": 35.51724137931034, "grad_norm": 2.890625, "learning_rate": 2.4157200444618985e-05, "loss": 0.029, "step": 1030 }, { "epoch": 35.55172413793103, "grad_norm": 5.09375, "learning_rate": 2.4154047515193904e-05, "loss": 0.035, "step": 1031 }, { "epoch": 35.58620689655172, "grad_norm": 11.9375, "learning_rate": 2.4150888905727638e-05, "loss": 0.0692, "step": 1032 }, { "epoch": 35.62068965517241, "grad_norm": 2.765625, "learning_rate": 2.4147724617759656e-05, "loss": 0.0272, "step": 1033 }, { "epoch": 35.6551724137931, "grad_norm": 3.9375, "learning_rate": 2.4144554652832196e-05, "loss": 0.0228, "step": 1034 }, { "epoch": 35.689655172413794, "grad_norm": 2.171875, "learning_rate": 2.4141379012490255e-05, "loss": 0.0213, "step": 1035 }, { "epoch": 35.724137931034484, "grad_norm": 2.28125, "learning_rate": 2.4138197698281606e-05, "loss": 0.0166, "step": 1036 }, { "epoch": 35.758620689655174, "grad_norm": 1.6015625, "learning_rate": 2.4135010711756777e-05, "loss": 0.012, "step": 1037 }, { "epoch": 35.793103448275865, "grad_norm": 3.96875, "learning_rate": 2.4131818054469078e-05, "loss": 0.0396, "step": 1038 }, { "epoch": 35.827586206896555, "grad_norm": 4.875, "learning_rate": 2.4128619727974562e-05, "loss": 0.0439, "step": 1039 }, { "epoch": 35.86206896551724, "grad_norm": 0.5703125, "learning_rate": 2.412541573383205e-05, "loss": 0.0134, "step": 1040 }, { "epoch": 35.89655172413793, "grad_norm": 2.46875, "learning_rate": 2.4122206073603142e-05, "loss": 0.0264, "step": 1041 }, { "epoch": 35.93103448275862, "grad_norm": 3.15625, "learning_rate": 2.4118990748852183e-05, "loss": 0.0174, "step": 1042 }, { "epoch": 35.96551724137931, "grad_norm": 2.703125, "learning_rate": 2.4115769761146282e-05, "loss": 0.0516, "step": 1043 }, { "epoch": 36.0, "grad_norm": 3.046875, "learning_rate": 2.4112543112055308e-05, "loss": 0.0556, "step": 1044 }, { "epoch": 36.03448275862069, "grad_norm": 3.84375, "learning_rate": 2.410931080315189e-05, "loss": 0.0275, "step": 1045 }, { "epoch": 36.06896551724138, "grad_norm": 0.93359375, "learning_rate": 2.4106072836011422e-05, "loss": 0.0489, "step": 1046 }, { "epoch": 36.10344827586207, "grad_norm": 4.5, "learning_rate": 2.410282921221204e-05, "loss": 0.0532, "step": 1047 }, { "epoch": 36.13793103448276, "grad_norm": 4.25, "learning_rate": 2.409957993333466e-05, "loss": 0.1283, "step": 1048 }, { "epoch": 36.172413793103445, "grad_norm": 2.078125, "learning_rate": 2.4096325000962927e-05, "loss": 0.0151, "step": 1049 }, { "epoch": 36.206896551724135, "grad_norm": 5.71875, "learning_rate": 2.409306441668327e-05, "loss": 0.0637, "step": 1050 }, { "epoch": 36.241379310344826, "grad_norm": 4.4375, "learning_rate": 2.4089798182084845e-05, "loss": 0.0193, "step": 1051 }, { "epoch": 36.275862068965516, "grad_norm": 5.375, "learning_rate": 2.408652629875958e-05, "loss": 0.0225, "step": 1052 }, { "epoch": 36.310344827586206, "grad_norm": 0.91796875, "learning_rate": 2.408324876830215e-05, "loss": 0.0164, "step": 1053 }, { "epoch": 36.3448275862069, "grad_norm": 11.0625, "learning_rate": 2.407996559230999e-05, "loss": 0.0204, "step": 1054 }, { "epoch": 36.37931034482759, "grad_norm": 3.203125, "learning_rate": 2.4076676772383277e-05, "loss": 0.0298, "step": 1055 }, { "epoch": 36.41379310344828, "grad_norm": 9.125, "learning_rate": 2.407338231012494e-05, "loss": 0.0472, "step": 1056 }, { "epoch": 36.44827586206897, "grad_norm": 4.4375, "learning_rate": 2.4070082207140653e-05, "loss": 0.0401, "step": 1057 }, { "epoch": 36.48275862068966, "grad_norm": 5.3125, "learning_rate": 2.4066776465038858e-05, "loss": 0.063, "step": 1058 }, { "epoch": 36.51724137931034, "grad_norm": 8.4375, "learning_rate": 2.4063465085430725e-05, "loss": 0.0335, "step": 1059 }, { "epoch": 36.55172413793103, "grad_norm": 1.0, "learning_rate": 2.4060148069930188e-05, "loss": 0.0068, "step": 1060 }, { "epoch": 36.58620689655172, "grad_norm": 1.90625, "learning_rate": 2.4056825420153917e-05, "loss": 0.0272, "step": 1061 }, { "epoch": 36.62068965517241, "grad_norm": 3.9375, "learning_rate": 2.4053497137721324e-05, "loss": 0.0277, "step": 1062 }, { "epoch": 36.6551724137931, "grad_norm": 4.6875, "learning_rate": 2.4050163224254586e-05, "loss": 0.0814, "step": 1063 }, { "epoch": 36.689655172413794, "grad_norm": 4.5625, "learning_rate": 2.4046823681378603e-05, "loss": 0.0625, "step": 1064 }, { "epoch": 36.724137931034484, "grad_norm": 4.90625, "learning_rate": 2.404347851072103e-05, "loss": 0.0841, "step": 1065 }, { "epoch": 36.758620689655174, "grad_norm": 2.96875, "learning_rate": 2.4040127713912264e-05, "loss": 0.0238, "step": 1066 }, { "epoch": 36.793103448275865, "grad_norm": 2.4375, "learning_rate": 2.4036771292585438e-05, "loss": 0.0153, "step": 1067 }, { "epoch": 36.827586206896555, "grad_norm": 7.5625, "learning_rate": 2.4033409248376437e-05, "loss": 0.0891, "step": 1068 }, { "epoch": 36.86206896551724, "grad_norm": 4.65625, "learning_rate": 2.4030041582923877e-05, "loss": 0.0308, "step": 1069 }, { "epoch": 36.89655172413793, "grad_norm": 2.828125, "learning_rate": 2.4026668297869118e-05, "loss": 0.0467, "step": 1070 }, { "epoch": 36.93103448275862, "grad_norm": 0.9296875, "learning_rate": 2.4023289394856257e-05, "loss": 0.0195, "step": 1071 }, { "epoch": 36.96551724137931, "grad_norm": 2.546875, "learning_rate": 2.4019904875532134e-05, "loss": 0.0264, "step": 1072 }, { "epoch": 37.0, "grad_norm": 7.21875, "learning_rate": 2.401651474154632e-05, "loss": 0.112, "step": 1073 }, { "epoch": 37.03448275862069, "grad_norm": 6.3125, "learning_rate": 2.401311899455112e-05, "loss": 0.0315, "step": 1074 }, { "epoch": 37.06896551724138, "grad_norm": 0.6875, "learning_rate": 2.400971763620159e-05, "loss": 0.0181, "step": 1075 }, { "epoch": 37.10344827586207, "grad_norm": 2.390625, "learning_rate": 2.4006310668155508e-05, "loss": 0.0697, "step": 1076 }, { "epoch": 37.13793103448276, "grad_norm": 1.1953125, "learning_rate": 2.400289809207338e-05, "loss": 0.0178, "step": 1077 }, { "epoch": 37.172413793103445, "grad_norm": 5.15625, "learning_rate": 2.3999479909618466e-05, "loss": 0.0471, "step": 1078 }, { "epoch": 37.206896551724135, "grad_norm": 2.78125, "learning_rate": 2.399605612245674e-05, "loss": 0.0264, "step": 1079 }, { "epoch": 37.241379310344826, "grad_norm": 2.015625, "learning_rate": 2.3992626732256912e-05, "loss": 0.0296, "step": 1080 }, { "epoch": 37.275862068965516, "grad_norm": 5.59375, "learning_rate": 2.398919174069043e-05, "loss": 0.0684, "step": 1081 }, { "epoch": 37.310344827586206, "grad_norm": 2.578125, "learning_rate": 2.3985751149431468e-05, "loss": 0.0371, "step": 1082 }, { "epoch": 37.3448275862069, "grad_norm": 3.265625, "learning_rate": 2.3982304960156925e-05, "loss": 0.0361, "step": 1083 }, { "epoch": 37.37931034482759, "grad_norm": 5.875, "learning_rate": 2.3978853174546433e-05, "loss": 0.0676, "step": 1084 }, { "epoch": 37.41379310344828, "grad_norm": 3.734375, "learning_rate": 2.397539579428235e-05, "loss": 0.0478, "step": 1085 }, { "epoch": 37.44827586206897, "grad_norm": 7.09375, "learning_rate": 2.3971932821049765e-05, "loss": 0.0619, "step": 1086 }, { "epoch": 37.48275862068966, "grad_norm": 2.796875, "learning_rate": 2.3968464256536482e-05, "loss": 0.0152, "step": 1087 }, { "epoch": 37.51724137931034, "grad_norm": 1.3828125, "learning_rate": 2.3964990102433044e-05, "loss": 0.0181, "step": 1088 }, { "epoch": 37.55172413793103, "grad_norm": 3.09375, "learning_rate": 2.3961510360432707e-05, "loss": 0.0363, "step": 1089 }, { "epoch": 37.58620689655172, "grad_norm": 7.4375, "learning_rate": 2.3958025032231455e-05, "loss": 0.045, "step": 1090 }, { "epoch": 37.62068965517241, "grad_norm": 4.40625, "learning_rate": 2.3954534119527996e-05, "loss": 0.0443, "step": 1091 }, { "epoch": 37.6551724137931, "grad_norm": 1.6796875, "learning_rate": 2.3951037624023758e-05, "loss": 0.0086, "step": 1092 }, { "epoch": 37.689655172413794, "grad_norm": 0.53125, "learning_rate": 2.3947535547422896e-05, "loss": 0.0149, "step": 1093 }, { "epoch": 37.724137931034484, "grad_norm": 3.21875, "learning_rate": 2.394402789143227e-05, "loss": 0.0135, "step": 1094 }, { "epoch": 37.758620689655174, "grad_norm": 9.375, "learning_rate": 2.394051465776147e-05, "loss": 0.0522, "step": 1095 }, { "epoch": 37.793103448275865, "grad_norm": 7.25, "learning_rate": 2.3936995848122812e-05, "loss": 0.0352, "step": 1096 }, { "epoch": 37.827586206896555, "grad_norm": 7.65625, "learning_rate": 2.3933471464231312e-05, "loss": 0.0583, "step": 1097 }, { "epoch": 37.86206896551724, "grad_norm": 4.375, "learning_rate": 2.392994150780472e-05, "loss": 0.0152, "step": 1098 }, { "epoch": 37.89655172413793, "grad_norm": 3.734375, "learning_rate": 2.392640598056349e-05, "loss": 0.0139, "step": 1099 }, { "epoch": 37.93103448275862, "grad_norm": 0.369140625, "learning_rate": 2.3922864884230788e-05, "loss": 0.0051, "step": 1100 }, { "epoch": 37.96551724137931, "grad_norm": 0.41015625, "learning_rate": 2.391931822053251e-05, "loss": 0.0268, "step": 1101 }, { "epoch": 38.0, "grad_norm": 6.28125, "learning_rate": 2.391576599119726e-05, "loss": 0.017, "step": 1102 }, { "epoch": 38.03448275862069, "grad_norm": 3.234375, "learning_rate": 2.3912208197956343e-05, "loss": 0.0205, "step": 1103 }, { "epoch": 38.06896551724138, "grad_norm": 4.625, "learning_rate": 2.3908644842543788e-05, "loss": 0.0295, "step": 1104 }, { "epoch": 38.10344827586207, "grad_norm": 13.1875, "learning_rate": 2.390507592669633e-05, "loss": 0.0794, "step": 1105 }, { "epoch": 38.13793103448276, "grad_norm": 15.25, "learning_rate": 2.390150145215341e-05, "loss": 0.0563, "step": 1106 }, { "epoch": 38.172413793103445, "grad_norm": 10.3125, "learning_rate": 2.38979214206572e-05, "loss": 0.0409, "step": 1107 }, { "epoch": 38.206896551724135, "grad_norm": 1.9765625, "learning_rate": 2.3894335833952548e-05, "loss": 0.0196, "step": 1108 }, { "epoch": 38.241379310344826, "grad_norm": 4.25, "learning_rate": 2.389074469378703e-05, "loss": 0.0887, "step": 1109 }, { "epoch": 38.275862068965516, "grad_norm": 7.84375, "learning_rate": 2.3887148001910925e-05, "loss": 0.0402, "step": 1110 }, { "epoch": 38.310344827586206, "grad_norm": 3.890625, "learning_rate": 2.3883545760077215e-05, "loss": 0.0362, "step": 1111 }, { "epoch": 38.3448275862069, "grad_norm": 3.34375, "learning_rate": 2.3879937970041595e-05, "loss": 0.0139, "step": 1112 }, { "epoch": 38.37931034482759, "grad_norm": 5.5, "learning_rate": 2.3876324633562446e-05, "loss": 0.0288, "step": 1113 }, { "epoch": 38.41379310344828, "grad_norm": 1.7890625, "learning_rate": 2.3872705752400874e-05, "loss": 0.0109, "step": 1114 }, { "epoch": 38.44827586206897, "grad_norm": 2.421875, "learning_rate": 2.386908132832068e-05, "loss": 0.078, "step": 1115 }, { "epoch": 38.48275862068966, "grad_norm": 2.5625, "learning_rate": 2.386545136308836e-05, "loss": 0.0401, "step": 1116 }, { "epoch": 38.51724137931034, "grad_norm": 0.79296875, "learning_rate": 2.386181585847311e-05, "loss": 0.0127, "step": 1117 }, { "epoch": 38.55172413793103, "grad_norm": 2.09375, "learning_rate": 2.3858174816246837e-05, "loss": 0.0339, "step": 1118 }, { "epoch": 38.58620689655172, "grad_norm": 2.765625, "learning_rate": 2.385452823818414e-05, "loss": 0.0223, "step": 1119 }, { "epoch": 38.62068965517241, "grad_norm": 6.15625, "learning_rate": 2.3850876126062314e-05, "loss": 0.0617, "step": 1120 }, { "epoch": 38.6551724137931, "grad_norm": 1.734375, "learning_rate": 2.384721848166136e-05, "loss": 0.0279, "step": 1121 }, { "epoch": 38.689655172413794, "grad_norm": 1.28125, "learning_rate": 2.3843555306763963e-05, "loss": 0.0694, "step": 1122 }, { "epoch": 38.724137931034484, "grad_norm": 1.6171875, "learning_rate": 2.3839886603155516e-05, "loss": 0.0743, "step": 1123 }, { "epoch": 38.758620689655174, "grad_norm": 0.5, "learning_rate": 2.3836212372624088e-05, "loss": 0.0146, "step": 1124 }, { "epoch": 38.793103448275865, "grad_norm": 3.09375, "learning_rate": 2.3832532616960473e-05, "loss": 0.0255, "step": 1125 }, { "epoch": 38.827586206896555, "grad_norm": 0.8359375, "learning_rate": 2.3828847337958126e-05, "loss": 0.0195, "step": 1126 }, { "epoch": 38.86206896551724, "grad_norm": 2.5625, "learning_rate": 2.3825156537413213e-05, "loss": 0.0191, "step": 1127 }, { "epoch": 38.89655172413793, "grad_norm": 3.1875, "learning_rate": 2.382146021712458e-05, "loss": 0.0417, "step": 1128 }, { "epoch": 38.93103448275862, "grad_norm": 2.265625, "learning_rate": 2.3817758378893774e-05, "loss": 0.0232, "step": 1129 }, { "epoch": 38.96551724137931, "grad_norm": 1.2109375, "learning_rate": 2.3814051024525018e-05, "loss": 0.0182, "step": 1130 }, { "epoch": 39.0, "grad_norm": 3.09375, "learning_rate": 2.3810338155825245e-05, "loss": 0.0122, "step": 1131 }, { "epoch": 39.03448275862069, "grad_norm": 2.234375, "learning_rate": 2.3806619774604053e-05, "loss": 0.0099, "step": 1132 }, { "epoch": 39.06896551724138, "grad_norm": 8.25, "learning_rate": 2.3802895882673734e-05, "loss": 0.0375, "step": 1133 }, { "epoch": 39.10344827586207, "grad_norm": 3.578125, "learning_rate": 2.3799166481849272e-05, "loss": 0.0419, "step": 1134 }, { "epoch": 39.13793103448276, "grad_norm": 6.59375, "learning_rate": 2.3795431573948333e-05, "loss": 0.0371, "step": 1135 }, { "epoch": 39.172413793103445, "grad_norm": 1.2109375, "learning_rate": 2.379169116079126e-05, "loss": 0.0172, "step": 1136 }, { "epoch": 39.206896551724135, "grad_norm": 8.6875, "learning_rate": 2.3787945244201092e-05, "loss": 0.041, "step": 1137 }, { "epoch": 39.241379310344826, "grad_norm": 4.3125, "learning_rate": 2.3784193826003538e-05, "loss": 0.0432, "step": 1138 }, { "epoch": 39.275862068965516, "grad_norm": 3.359375, "learning_rate": 2.3780436908026992e-05, "loss": 0.0179, "step": 1139 }, { "epoch": 39.310344827586206, "grad_norm": 1.0546875, "learning_rate": 2.3776674492102535e-05, "loss": 0.0113, "step": 1140 }, { "epoch": 39.3448275862069, "grad_norm": 7.40625, "learning_rate": 2.3772906580063924e-05, "loss": 0.0369, "step": 1141 }, { "epoch": 39.37931034482759, "grad_norm": 1.796875, "learning_rate": 2.376913317374759e-05, "loss": 0.018, "step": 1142 }, { "epoch": 39.41379310344828, "grad_norm": 0.2490234375, "learning_rate": 2.3765354274992643e-05, "loss": 0.0061, "step": 1143 }, { "epoch": 39.44827586206897, "grad_norm": 10.75, "learning_rate": 2.3761569885640877e-05, "loss": 0.0604, "step": 1144 }, { "epoch": 39.48275862068966, "grad_norm": 8.0625, "learning_rate": 2.3757780007536757e-05, "loss": 0.0576, "step": 1145 }, { "epoch": 39.51724137931034, "grad_norm": 0.796875, "learning_rate": 2.3753984642527423e-05, "loss": 0.0222, "step": 1146 }, { "epoch": 39.55172413793103, "grad_norm": 1.15625, "learning_rate": 2.3750183792462686e-05, "loss": 0.033, "step": 1147 }, { "epoch": 39.58620689655172, "grad_norm": 7.15625, "learning_rate": 2.3746377459195037e-05, "loss": 0.066, "step": 1148 }, { "epoch": 39.62068965517241, "grad_norm": 0.5546875, "learning_rate": 2.3742565644579638e-05, "loss": 0.011, "step": 1149 }, { "epoch": 39.6551724137931, "grad_norm": 2.40625, "learning_rate": 2.373874835047432e-05, "loss": 0.0128, "step": 1150 }, { "epoch": 39.689655172413794, "grad_norm": 3.0625, "learning_rate": 2.3734925578739588e-05, "loss": 0.0231, "step": 1151 }, { "epoch": 39.724137931034484, "grad_norm": 1.734375, "learning_rate": 2.373109733123861e-05, "loss": 0.0385, "step": 1152 }, { "epoch": 39.758620689655174, "grad_norm": 7.4375, "learning_rate": 2.372726360983723e-05, "loss": 0.0394, "step": 1153 }, { "epoch": 39.793103448275865, "grad_norm": 3.484375, "learning_rate": 2.372342441640396e-05, "loss": 0.0551, "step": 1154 }, { "epoch": 39.827586206896555, "grad_norm": 2.0625, "learning_rate": 2.371957975280997e-05, "loss": 0.0125, "step": 1155 }, { "epoch": 39.86206896551724, "grad_norm": 4.65625, "learning_rate": 2.3715729620929106e-05, "loss": 0.0343, "step": 1156 }, { "epoch": 39.89655172413793, "grad_norm": 4.34375, "learning_rate": 2.3711874022637875e-05, "loss": 0.075, "step": 1157 }, { "epoch": 39.93103448275862, "grad_norm": 2.96875, "learning_rate": 2.3708012959815452e-05, "loss": 0.0176, "step": 1158 }, { "epoch": 39.96551724137931, "grad_norm": 0.52734375, "learning_rate": 2.3704146434343664e-05, "loss": 0.0577, "step": 1159 }, { "epoch": 40.0, "grad_norm": 1.28125, "learning_rate": 2.3700274448107012e-05, "loss": 0.0175, "step": 1160 }, { "epoch": 40.03448275862069, "grad_norm": 3.703125, "learning_rate": 2.3696397002992663e-05, "loss": 0.0578, "step": 1161 }, { "epoch": 40.06896551724138, "grad_norm": 4.65625, "learning_rate": 2.3692514100890425e-05, "loss": 0.0166, "step": 1162 }, { "epoch": 40.10344827586207, "grad_norm": 8.9375, "learning_rate": 2.3688625743692784e-05, "loss": 0.0587, "step": 1163 }, { "epoch": 40.13793103448276, "grad_norm": 3.96875, "learning_rate": 2.368473193329488e-05, "loss": 0.0363, "step": 1164 }, { "epoch": 40.172413793103445, "grad_norm": 1.1484375, "learning_rate": 2.36808326715945e-05, "loss": 0.0205, "step": 1165 }, { "epoch": 40.206896551724135, "grad_norm": 5.40625, "learning_rate": 2.36769279604921e-05, "loss": 0.0663, "step": 1166 }, { "epoch": 40.241379310344826, "grad_norm": 4.0625, "learning_rate": 2.36730178018908e-05, "loss": 0.0376, "step": 1167 }, { "epoch": 40.275862068965516, "grad_norm": 2.90625, "learning_rate": 2.3669102197696347e-05, "loss": 0.0608, "step": 1168 }, { "epoch": 40.310344827586206, "grad_norm": 1.15625, "learning_rate": 2.3665181149817173e-05, "loss": 0.0686, "step": 1169 }, { "epoch": 40.3448275862069, "grad_norm": 1.4375, "learning_rate": 2.3661254660164337e-05, "loss": 0.0272, "step": 1170 }, { "epoch": 40.37931034482759, "grad_norm": 8.75, "learning_rate": 2.365732273065157e-05, "loss": 0.0693, "step": 1171 }, { "epoch": 40.41379310344828, "grad_norm": 1.6015625, "learning_rate": 2.3653385363195248e-05, "loss": 0.0567, "step": 1172 }, { "epoch": 40.44827586206897, "grad_norm": 4.375, "learning_rate": 2.3649442559714387e-05, "loss": 0.0409, "step": 1173 }, { "epoch": 40.48275862068966, "grad_norm": 1.140625, "learning_rate": 2.3645494322130668e-05, "loss": 0.023, "step": 1174 }, { "epoch": 40.51724137931034, "grad_norm": 4.25, "learning_rate": 2.3641540652368415e-05, "loss": 0.0484, "step": 1175 }, { "epoch": 40.55172413793103, "grad_norm": 1.4921875, "learning_rate": 2.36375815523546e-05, "loss": 0.0297, "step": 1176 }, { "epoch": 40.58620689655172, "grad_norm": 1.109375, "learning_rate": 2.3633617024018832e-05, "loss": 0.0321, "step": 1177 }, { "epoch": 40.62068965517241, "grad_norm": 3.453125, "learning_rate": 2.3629647069293384e-05, "loss": 0.0334, "step": 1178 }, { "epoch": 40.6551724137931, "grad_norm": 3.09375, "learning_rate": 2.3625671690113155e-05, "loss": 0.0398, "step": 1179 }, { "epoch": 40.689655172413794, "grad_norm": 6.65625, "learning_rate": 2.3621690888415705e-05, "loss": 0.0693, "step": 1180 }, { "epoch": 40.724137931034484, "grad_norm": 8.4375, "learning_rate": 2.361770466614122e-05, "loss": 0.0662, "step": 1181 }, { "epoch": 40.758620689655174, "grad_norm": 1.5390625, "learning_rate": 2.3613713025232542e-05, "loss": 0.0382, "step": 1182 }, { "epoch": 40.793103448275865, "grad_norm": 3.578125, "learning_rate": 2.3609715967635145e-05, "loss": 0.0371, "step": 1183 }, { "epoch": 40.827586206896555, "grad_norm": 2.828125, "learning_rate": 2.3605713495297148e-05, "loss": 0.0355, "step": 1184 }, { "epoch": 40.86206896551724, "grad_norm": 2.765625, "learning_rate": 2.360170561016931e-05, "loss": 0.0268, "step": 1185 }, { "epoch": 40.89655172413793, "grad_norm": 3.046875, "learning_rate": 2.3597692314205016e-05, "loss": 0.0372, "step": 1186 }, { "epoch": 40.93103448275862, "grad_norm": 2.828125, "learning_rate": 2.359367360936031e-05, "loss": 0.0304, "step": 1187 }, { "epoch": 40.96551724137931, "grad_norm": 2.40625, "learning_rate": 2.358964949759385e-05, "loss": 0.0453, "step": 1188 }, { "epoch": 41.0, "grad_norm": 0.81640625, "learning_rate": 2.358561998086695e-05, "loss": 0.0098, "step": 1189 }, { "epoch": 41.03448275862069, "grad_norm": 1.484375, "learning_rate": 2.3581585061143534e-05, "loss": 0.0226, "step": 1190 }, { "epoch": 41.06896551724138, "grad_norm": 6.625, "learning_rate": 2.3577544740390184e-05, "loss": 0.0202, "step": 1191 }, { "epoch": 41.10344827586207, "grad_norm": 3.3125, "learning_rate": 2.3573499020576102e-05, "loss": 0.0309, "step": 1192 }, { "epoch": 41.13793103448276, "grad_norm": 0.431640625, "learning_rate": 2.3569447903673116e-05, "loss": 0.0224, "step": 1193 }, { "epoch": 41.172413793103445, "grad_norm": 1.703125, "learning_rate": 2.356539139165569e-05, "loss": 0.0222, "step": 1194 }, { "epoch": 41.206896551724135, "grad_norm": 4.3125, "learning_rate": 2.356132948650093e-05, "loss": 0.0505, "step": 1195 }, { "epoch": 41.241379310344826, "grad_norm": 1.8984375, "learning_rate": 2.355726219018855e-05, "loss": 0.0455, "step": 1196 }, { "epoch": 41.275862068965516, "grad_norm": 4.125, "learning_rate": 2.3553189504700905e-05, "loss": 0.0306, "step": 1197 }, { "epoch": 41.310344827586206, "grad_norm": 0.91796875, "learning_rate": 2.3549111432022973e-05, "loss": 0.0142, "step": 1198 }, { "epoch": 41.3448275862069, "grad_norm": 1.171875, "learning_rate": 2.3545027974142353e-05, "loss": 0.034, "step": 1199 }, { "epoch": 41.37931034482759, "grad_norm": 0.515625, "learning_rate": 2.3540939133049273e-05, "loss": 0.0114, "step": 1200 }, { "epoch": 41.41379310344828, "grad_norm": 0.609375, "learning_rate": 2.353684491073659e-05, "loss": 0.0047, "step": 1201 }, { "epoch": 41.44827586206897, "grad_norm": 0.486328125, "learning_rate": 2.353274530919977e-05, "loss": 0.0032, "step": 1202 }, { "epoch": 41.48275862068966, "grad_norm": 2.546875, "learning_rate": 2.3528640330436917e-05, "loss": 0.0208, "step": 1203 }, { "epoch": 41.51724137931034, "grad_norm": 0.494140625, "learning_rate": 2.3524529976448744e-05, "loss": 0.0094, "step": 1204 }, { "epoch": 41.55172413793103, "grad_norm": 2.828125, "learning_rate": 2.352041424923859e-05, "loss": 0.0289, "step": 1205 }, { "epoch": 41.58620689655172, "grad_norm": 2.375, "learning_rate": 2.35162931508124e-05, "loss": 0.0175, "step": 1206 }, { "epoch": 41.62068965517241, "grad_norm": 1.25, "learning_rate": 2.3512166683178766e-05, "loss": 0.0103, "step": 1207 }, { "epoch": 41.6551724137931, "grad_norm": 2.671875, "learning_rate": 2.3508034848348866e-05, "loss": 0.0354, "step": 1208 }, { "epoch": 41.689655172413794, "grad_norm": 3.609375, "learning_rate": 2.3503897648336503e-05, "loss": 0.0344, "step": 1209 }, { "epoch": 41.724137931034484, "grad_norm": 1.171875, "learning_rate": 2.3499755085158106e-05, "loss": 0.0492, "step": 1210 }, { "epoch": 41.758620689655174, "grad_norm": 2.28125, "learning_rate": 2.3495607160832707e-05, "loss": 0.0236, "step": 1211 }, { "epoch": 41.793103448275865, "grad_norm": 5.34375, "learning_rate": 2.3491453877381953e-05, "loss": 0.0213, "step": 1212 }, { "epoch": 41.827586206896555, "grad_norm": 1.25, "learning_rate": 2.3487295236830108e-05, "loss": 0.0195, "step": 1213 }, { "epoch": 41.86206896551724, "grad_norm": 3.5, "learning_rate": 2.3483131241204037e-05, "loss": 0.0811, "step": 1214 }, { "epoch": 41.89655172413793, "grad_norm": 1.9453125, "learning_rate": 2.3478961892533223e-05, "loss": 0.0136, "step": 1215 }, { "epoch": 41.93103448275862, "grad_norm": 0.388671875, "learning_rate": 2.3474787192849756e-05, "loss": 0.0384, "step": 1216 }, { "epoch": 41.96551724137931, "grad_norm": 0.65234375, "learning_rate": 2.347060714418834e-05, "loss": 0.0127, "step": 1217 }, { "epoch": 42.0, "grad_norm": 2.765625, "learning_rate": 2.346642174858627e-05, "loss": 0.0352, "step": 1218 }, { "epoch": 42.03448275862069, "grad_norm": 3.015625, "learning_rate": 2.346223100808346e-05, "loss": 0.0242, "step": 1219 }, { "epoch": 42.06896551724138, "grad_norm": 6.96875, "learning_rate": 2.345803492472243e-05, "loss": 0.0368, "step": 1220 }, { "epoch": 42.10344827586207, "grad_norm": 4.6875, "learning_rate": 2.3453833500548295e-05, "loss": 0.0255, "step": 1221 }, { "epoch": 42.13793103448276, "grad_norm": 2.453125, "learning_rate": 2.3449626737608784e-05, "loss": 0.0338, "step": 1222 }, { "epoch": 42.172413793103445, "grad_norm": 0.2353515625, "learning_rate": 2.3445414637954217e-05, "loss": 0.0255, "step": 1223 }, { "epoch": 42.206896551724135, "grad_norm": 2.84375, "learning_rate": 2.3441197203637522e-05, "loss": 0.0386, "step": 1224 }, { "epoch": 42.241379310344826, "grad_norm": 0.5, "learning_rate": 2.3436974436714224e-05, "loss": 0.014, "step": 1225 }, { "epoch": 42.275862068965516, "grad_norm": 1.4453125, "learning_rate": 2.3432746339242448e-05, "loss": 0.0303, "step": 1226 }, { "epoch": 42.310344827586206, "grad_norm": 1.796875, "learning_rate": 2.342851291328292e-05, "loss": 0.0363, "step": 1227 }, { "epoch": 42.3448275862069, "grad_norm": 5.84375, "learning_rate": 2.342427416089896e-05, "loss": 0.0337, "step": 1228 }, { "epoch": 42.37931034482759, "grad_norm": 0.6953125, "learning_rate": 2.3420030084156486e-05, "loss": 0.0192, "step": 1229 }, { "epoch": 42.41379310344828, "grad_norm": 4.9375, "learning_rate": 2.3415780685124004e-05, "loss": 0.0545, "step": 1230 }, { "epoch": 42.44827586206897, "grad_norm": 5.46875, "learning_rate": 2.341152596587262e-05, "loss": 0.0623, "step": 1231 }, { "epoch": 42.48275862068966, "grad_norm": 3.09375, "learning_rate": 2.3407265928476042e-05, "loss": 0.0237, "step": 1232 }, { "epoch": 42.51724137931034, "grad_norm": 5.5, "learning_rate": 2.340300057501055e-05, "loss": 0.0879, "step": 1233 }, { "epoch": 42.55172413793103, "grad_norm": 3.078125, "learning_rate": 2.3398729907555026e-05, "loss": 0.023, "step": 1234 }, { "epoch": 42.58620689655172, "grad_norm": 2.21875, "learning_rate": 2.3394453928190947e-05, "loss": 0.0148, "step": 1235 }, { "epoch": 42.62068965517241, "grad_norm": 11.8125, "learning_rate": 2.339017263900237e-05, "loss": 0.0757, "step": 1236 }, { "epoch": 42.6551724137931, "grad_norm": 2.0625, "learning_rate": 2.3385886042075942e-05, "loss": 0.0122, "step": 1237 }, { "epoch": 42.689655172413794, "grad_norm": 2.1875, "learning_rate": 2.3381594139500902e-05, "loss": 0.0318, "step": 1238 }, { "epoch": 42.724137931034484, "grad_norm": 0.96875, "learning_rate": 2.3377296933369074e-05, "loss": 0.0118, "step": 1239 }, { "epoch": 42.758620689655174, "grad_norm": 1.6796875, "learning_rate": 2.337299442577485e-05, "loss": 0.0114, "step": 1240 }, { "epoch": 42.793103448275865, "grad_norm": 1.5, "learning_rate": 2.3368686618815238e-05, "loss": 0.0211, "step": 1241 }, { "epoch": 42.827586206896555, "grad_norm": 1.4921875, "learning_rate": 2.33643735145898e-05, "loss": 0.0186, "step": 1242 }, { "epoch": 42.86206896551724, "grad_norm": 4.09375, "learning_rate": 2.336005511520069e-05, "loss": 0.0401, "step": 1243 }, { "epoch": 42.89655172413793, "grad_norm": 5.25, "learning_rate": 2.3355731422752652e-05, "loss": 0.0523, "step": 1244 }, { "epoch": 42.93103448275862, "grad_norm": 4.4375, "learning_rate": 2.335140243935299e-05, "loss": 0.0667, "step": 1245 }, { "epoch": 42.96551724137931, "grad_norm": 0.4140625, "learning_rate": 2.334706816711161e-05, "loss": 0.0065, "step": 1246 }, { "epoch": 43.0, "grad_norm": 2.171875, "learning_rate": 2.3342728608140968e-05, "loss": 0.0184, "step": 1247 }, { "epoch": 43.03448275862069, "grad_norm": 2.796875, "learning_rate": 2.3338383764556126e-05, "loss": 0.0442, "step": 1248 }, { "epoch": 43.06896551724138, "grad_norm": 1.546875, "learning_rate": 2.3334033638474707e-05, "loss": 0.0369, "step": 1249 }, { "epoch": 43.10344827586207, "grad_norm": 7.0, "learning_rate": 2.3329678232016898e-05, "loss": 0.035, "step": 1250 }, { "epoch": 43.13793103448276, "grad_norm": 1.8359375, "learning_rate": 2.3325317547305485e-05, "loss": 0.0175, "step": 1251 }, { "epoch": 43.172413793103445, "grad_norm": 1.9765625, "learning_rate": 2.3320951586465804e-05, "loss": 0.0136, "step": 1252 }, { "epoch": 43.206896551724135, "grad_norm": 4.21875, "learning_rate": 2.3316580351625777e-05, "loss": 0.0177, "step": 1253 }, { "epoch": 43.241379310344826, "grad_norm": 5.09375, "learning_rate": 2.3312203844915885e-05, "loss": 0.0295, "step": 1254 }, { "epoch": 43.275862068965516, "grad_norm": 6.53125, "learning_rate": 2.3307822068469187e-05, "loss": 0.0624, "step": 1255 }, { "epoch": 43.310344827586206, "grad_norm": 0.5, "learning_rate": 2.3303435024421312e-05, "loss": 0.0066, "step": 1256 }, { "epoch": 43.3448275862069, "grad_norm": 3.53125, "learning_rate": 2.329904271491044e-05, "loss": 0.0309, "step": 1257 }, { "epoch": 43.37931034482759, "grad_norm": 1.4296875, "learning_rate": 2.329464514207734e-05, "loss": 0.0265, "step": 1258 }, { "epoch": 43.41379310344828, "grad_norm": 0.8203125, "learning_rate": 2.329024230806534e-05, "loss": 0.0097, "step": 1259 }, { "epoch": 43.44827586206897, "grad_norm": 4.09375, "learning_rate": 2.3285834215020313e-05, "loss": 0.0514, "step": 1260 }, { "epoch": 43.48275862068966, "grad_norm": 1.59375, "learning_rate": 2.328142086509072e-05, "loss": 0.013, "step": 1261 }, { "epoch": 43.51724137931034, "grad_norm": 2.34375, "learning_rate": 2.327700226042757e-05, "loss": 0.0194, "step": 1262 }, { "epoch": 43.55172413793103, "grad_norm": 3.78125, "learning_rate": 2.327257840318444e-05, "loss": 0.0811, "step": 1263 }, { "epoch": 43.58620689655172, "grad_norm": 5.65625, "learning_rate": 2.326814929551746e-05, "loss": 0.0328, "step": 1264 }, { "epoch": 43.62068965517241, "grad_norm": 2.125, "learning_rate": 2.326371493958533e-05, "loss": 0.0073, "step": 1265 }, { "epoch": 43.6551724137931, "grad_norm": 6.28125, "learning_rate": 2.32592753375493e-05, "loss": 0.0264, "step": 1266 }, { "epoch": 43.689655172413794, "grad_norm": 2.453125, "learning_rate": 2.3254830491573167e-05, "loss": 0.0268, "step": 1267 }, { "epoch": 43.724137931034484, "grad_norm": 1.2890625, "learning_rate": 2.325038040382331e-05, "loss": 0.015, "step": 1268 }, { "epoch": 43.758620689655174, "grad_norm": 2.640625, "learning_rate": 2.324592507646864e-05, "loss": 0.0091, "step": 1269 }, { "epoch": 43.793103448275865, "grad_norm": 4.03125, "learning_rate": 2.324146451168063e-05, "loss": 0.0449, "step": 1270 }, { "epoch": 43.827586206896555, "grad_norm": 1.1484375, "learning_rate": 2.3236998711633307e-05, "loss": 0.0092, "step": 1271 }, { "epoch": 43.86206896551724, "grad_norm": 4.84375, "learning_rate": 2.3232527678503246e-05, "loss": 0.0374, "step": 1272 }, { "epoch": 43.89655172413793, "grad_norm": 1.3125, "learning_rate": 2.3228051414469574e-05, "loss": 0.0089, "step": 1273 }, { "epoch": 43.93103448275862, "grad_norm": 2.78125, "learning_rate": 2.3223569921713966e-05, "loss": 0.0148, "step": 1274 }, { "epoch": 43.96551724137931, "grad_norm": 2.421875, "learning_rate": 2.3219083202420654e-05, "loss": 0.0667, "step": 1275 }, { "epoch": 44.0, "grad_norm": 2.1875, "learning_rate": 2.3214591258776404e-05, "loss": 0.0097, "step": 1276 }, { "epoch": 44.03448275862069, "grad_norm": 2.953125, "learning_rate": 2.3210094092970536e-05, "loss": 0.0381, "step": 1277 }, { "epoch": 44.06896551724138, "grad_norm": 4.4375, "learning_rate": 2.3205591707194918e-05, "loss": 0.016, "step": 1278 }, { "epoch": 44.10344827586207, "grad_norm": 0.8828125, "learning_rate": 2.320108410364395e-05, "loss": 0.027, "step": 1279 }, { "epoch": 44.13793103448276, "grad_norm": 0.318359375, "learning_rate": 2.3196571284514594e-05, "loss": 0.0287, "step": 1280 }, { "epoch": 44.172413793103445, "grad_norm": 3.921875, "learning_rate": 2.3192053252006335e-05, "loss": 0.0477, "step": 1281 }, { "epoch": 44.206896551724135, "grad_norm": 1.65625, "learning_rate": 2.3187530008321208e-05, "loss": 0.0138, "step": 1282 }, { "epoch": 44.241379310344826, "grad_norm": 7.0625, "learning_rate": 2.318300155566379e-05, "loss": 0.0458, "step": 1283 }, { "epoch": 44.275862068965516, "grad_norm": 1.765625, "learning_rate": 2.317846789624119e-05, "loss": 0.032, "step": 1284 }, { "epoch": 44.310344827586206, "grad_norm": 0.62890625, "learning_rate": 2.3173929032263063e-05, "loss": 0.0121, "step": 1285 }, { "epoch": 44.3448275862069, "grad_norm": 1.171875, "learning_rate": 2.3169384965941592e-05, "loss": 0.0336, "step": 1286 }, { "epoch": 44.37931034482759, "grad_norm": 6.03125, "learning_rate": 2.3164835699491505e-05, "loss": 0.0554, "step": 1287 }, { "epoch": 44.41379310344828, "grad_norm": 0.3671875, "learning_rate": 2.3160281235130055e-05, "loss": 0.0467, "step": 1288 }, { "epoch": 44.44827586206897, "grad_norm": 3.3125, "learning_rate": 2.3155721575077034e-05, "loss": 0.0219, "step": 1289 }, { "epoch": 44.48275862068966, "grad_norm": 3.0625, "learning_rate": 2.3151156721554763e-05, "loss": 0.0168, "step": 1290 }, { "epoch": 44.51724137931034, "grad_norm": 5.21875, "learning_rate": 2.3146586676788095e-05, "loss": 0.0316, "step": 1291 }, { "epoch": 44.55172413793103, "grad_norm": 1.1796875, "learning_rate": 2.3142011443004418e-05, "loss": 0.0365, "step": 1292 }, { "epoch": 44.58620689655172, "grad_norm": 4.375, "learning_rate": 2.3137431022433652e-05, "loss": 0.0347, "step": 1293 }, { "epoch": 44.62068965517241, "grad_norm": 4.59375, "learning_rate": 2.3132845417308223e-05, "loss": 0.0361, "step": 1294 }, { "epoch": 44.6551724137931, "grad_norm": 1.0390625, "learning_rate": 2.312825462986311e-05, "loss": 0.0204, "step": 1295 }, { "epoch": 44.689655172413794, "grad_norm": 0.59765625, "learning_rate": 2.3123658662335802e-05, "loss": 0.0064, "step": 1296 }, { "epoch": 44.724137931034484, "grad_norm": 3.65625, "learning_rate": 2.3119057516966318e-05, "loss": 0.028, "step": 1297 }, { "epoch": 44.758620689655174, "grad_norm": 3.6875, "learning_rate": 2.3114451195997206e-05, "loss": 0.0388, "step": 1298 }, { "epoch": 44.793103448275865, "grad_norm": 1.8828125, "learning_rate": 2.3109839701673525e-05, "loss": 0.0207, "step": 1299 }, { "epoch": 44.827586206896555, "grad_norm": 0.6484375, "learning_rate": 2.3105223036242863e-05, "loss": 0.0542, "step": 1300 }, { "epoch": 44.86206896551724, "grad_norm": 0.298828125, "learning_rate": 2.3100601201955324e-05, "loss": 0.0638, "step": 1301 }, { "epoch": 44.89655172413793, "grad_norm": 4.25, "learning_rate": 2.309597420106354e-05, "loss": 0.0356, "step": 1302 }, { "epoch": 44.93103448275862, "grad_norm": 12.875, "learning_rate": 2.309134203582265e-05, "loss": 0.0593, "step": 1303 }, { "epoch": 44.96551724137931, "grad_norm": 2.796875, "learning_rate": 2.308670470849032e-05, "loss": 0.0408, "step": 1304 }, { "epoch": 45.0, "grad_norm": 6.09375, "learning_rate": 2.3082062221326724e-05, "loss": 0.0156, "step": 1305 }, { "epoch": 45.03448275862069, "grad_norm": 0.44140625, "learning_rate": 2.3077414576594553e-05, "loss": 0.023, "step": 1306 }, { "epoch": 45.06896551724138, "grad_norm": 1.03125, "learning_rate": 2.307276177655901e-05, "loss": 0.0164, "step": 1307 }, { "epoch": 45.10344827586207, "grad_norm": 6.03125, "learning_rate": 2.306810382348782e-05, "loss": 0.032, "step": 1308 }, { "epoch": 45.13793103448276, "grad_norm": 2.765625, "learning_rate": 2.306344071965121e-05, "loss": 0.0367, "step": 1309 }, { "epoch": 45.172413793103445, "grad_norm": 9.9375, "learning_rate": 2.3058772467321914e-05, "loss": 0.056, "step": 1310 }, { "epoch": 45.206896551724135, "grad_norm": 1.2421875, "learning_rate": 2.305409906877519e-05, "loss": 0.0127, "step": 1311 }, { "epoch": 45.241379310344826, "grad_norm": 2.890625, "learning_rate": 2.3049420526288792e-05, "loss": 0.0488, "step": 1312 }, { "epoch": 45.275862068965516, "grad_norm": 2.296875, "learning_rate": 2.3044736842142987e-05, "loss": 0.0307, "step": 1313 }, { "epoch": 45.310344827586206, "grad_norm": 1.5625, "learning_rate": 2.3040048018620543e-05, "loss": 0.0207, "step": 1314 }, { "epoch": 45.3448275862069, "grad_norm": 14.625, "learning_rate": 2.3035354058006738e-05, "loss": 0.0774, "step": 1315 }, { "epoch": 45.37931034482759, "grad_norm": 10.875, "learning_rate": 2.3030654962589346e-05, "loss": 0.0359, "step": 1316 }, { "epoch": 45.41379310344828, "grad_norm": 5.34375, "learning_rate": 2.3025950734658654e-05, "loss": 0.0514, "step": 1317 }, { "epoch": 45.44827586206897, "grad_norm": 2.15625, "learning_rate": 2.3021241376507447e-05, "loss": 0.0393, "step": 1318 }, { "epoch": 45.48275862068966, "grad_norm": 4.40625, "learning_rate": 2.3016526890431006e-05, "loss": 0.0276, "step": 1319 }, { "epoch": 45.51724137931034, "grad_norm": 1.6171875, "learning_rate": 2.3011807278727116e-05, "loss": 0.0331, "step": 1320 }, { "epoch": 45.55172413793103, "grad_norm": 1.171875, "learning_rate": 2.3007082543696055e-05, "loss": 0.0521, "step": 1321 }, { "epoch": 45.58620689655172, "grad_norm": 7.5, "learning_rate": 2.3002352687640605e-05, "loss": 0.0693, "step": 1322 }, { "epoch": 45.62068965517241, "grad_norm": 1.703125, "learning_rate": 2.2997617712866037e-05, "loss": 0.0172, "step": 1323 }, { "epoch": 45.6551724137931, "grad_norm": 5.84375, "learning_rate": 2.2992877621680125e-05, "loss": 0.0388, "step": 1324 }, { "epoch": 45.689655172413794, "grad_norm": 2.046875, "learning_rate": 2.2988132416393125e-05, "loss": 0.0285, "step": 1325 }, { "epoch": 45.724137931034484, "grad_norm": 13.4375, "learning_rate": 2.2983382099317803e-05, "loss": 0.0653, "step": 1326 }, { "epoch": 45.758620689655174, "grad_norm": 1.234375, "learning_rate": 2.2978626672769395e-05, "loss": 0.0115, "step": 1327 }, { "epoch": 45.793103448275865, "grad_norm": 2.0, "learning_rate": 2.2973866139065643e-05, "loss": 0.0278, "step": 1328 }, { "epoch": 45.827586206896555, "grad_norm": 2.015625, "learning_rate": 2.2969100500526775e-05, "loss": 0.0176, "step": 1329 }, { "epoch": 45.86206896551724, "grad_norm": 8.8125, "learning_rate": 2.29643297594755e-05, "loss": 0.0428, "step": 1330 }, { "epoch": 45.89655172413793, "grad_norm": 3.765625, "learning_rate": 2.295955391823702e-05, "loss": 0.0418, "step": 1331 }, { "epoch": 45.93103448275862, "grad_norm": 3.4375, "learning_rate": 2.295477297913903e-05, "loss": 0.0309, "step": 1332 }, { "epoch": 45.96551724137931, "grad_norm": 2.46875, "learning_rate": 2.294998694451169e-05, "loss": 0.0348, "step": 1333 }, { "epoch": 46.0, "grad_norm": 3.453125, "learning_rate": 2.294519581668767e-05, "loss": 0.0571, "step": 1334 }, { "epoch": 46.03448275862069, "grad_norm": 2.515625, "learning_rate": 2.2940399598002095e-05, "loss": 0.0188, "step": 1335 }, { "epoch": 46.06896551724138, "grad_norm": 2.0625, "learning_rate": 2.2935598290792583e-05, "loss": 0.0388, "step": 1336 }, { "epoch": 46.10344827586207, "grad_norm": 2.109375, "learning_rate": 2.293079189739924e-05, "loss": 0.0195, "step": 1337 }, { "epoch": 46.13793103448276, "grad_norm": 1.171875, "learning_rate": 2.292598042016464e-05, "loss": 0.0308, "step": 1338 }, { "epoch": 46.172413793103445, "grad_norm": 1.6171875, "learning_rate": 2.292116386143384e-05, "loss": 0.0358, "step": 1339 }, { "epoch": 46.206896551724135, "grad_norm": 1.8828125, "learning_rate": 2.2916342223554374e-05, "loss": 0.0235, "step": 1340 }, { "epoch": 46.241379310344826, "grad_norm": 2.421875, "learning_rate": 2.2911515508876243e-05, "loss": 0.0291, "step": 1341 }, { "epoch": 46.275862068965516, "grad_norm": 0.78515625, "learning_rate": 2.2906683719751938e-05, "loss": 0.0106, "step": 1342 }, { "epoch": 46.310344827586206, "grad_norm": 5.84375, "learning_rate": 2.2901846858536408e-05, "loss": 0.0511, "step": 1343 }, { "epoch": 46.3448275862069, "grad_norm": 3.75, "learning_rate": 2.289700492758708e-05, "loss": 0.0225, "step": 1344 }, { "epoch": 46.37931034482759, "grad_norm": 2.453125, "learning_rate": 2.2892157929263858e-05, "loss": 0.0351, "step": 1345 }, { "epoch": 46.41379310344828, "grad_norm": 4.0, "learning_rate": 2.2887305865929104e-05, "loss": 0.0611, "step": 1346 }, { "epoch": 46.44827586206897, "grad_norm": 2.09375, "learning_rate": 2.2882448739947658e-05, "loss": 0.0207, "step": 1347 }, { "epoch": 46.48275862068966, "grad_norm": 3.875, "learning_rate": 2.2877586553686824e-05, "loss": 0.0512, "step": 1348 }, { "epoch": 46.51724137931034, "grad_norm": 4.6875, "learning_rate": 2.287271930951637e-05, "loss": 0.0326, "step": 1349 }, { "epoch": 46.55172413793103, "grad_norm": 0.99609375, "learning_rate": 2.286784700980854e-05, "loss": 0.0095, "step": 1350 }, { "epoch": 46.58620689655172, "grad_norm": 0.49609375, "learning_rate": 2.286296965693802e-05, "loss": 0.0121, "step": 1351 }, { "epoch": 46.62068965517241, "grad_norm": 4.5, "learning_rate": 2.2858087253281986e-05, "loss": 0.0369, "step": 1352 }, { "epoch": 46.6551724137931, "grad_norm": 0.318359375, "learning_rate": 2.2853199801220053e-05, "loss": 0.0051, "step": 1353 }, { "epoch": 46.689655172413794, "grad_norm": 3.34375, "learning_rate": 2.284830730313431e-05, "loss": 0.018, "step": 1354 }, { "epoch": 46.724137931034484, "grad_norm": 4.03125, "learning_rate": 2.2843409761409297e-05, "loss": 0.0421, "step": 1355 }, { "epoch": 46.758620689655174, "grad_norm": 4.3125, "learning_rate": 2.283850717843202e-05, "loss": 0.0186, "step": 1356 }, { "epoch": 46.793103448275865, "grad_norm": 5.0, "learning_rate": 2.2833599556591945e-05, "loss": 0.0353, "step": 1357 }, { "epoch": 46.827586206896555, "grad_norm": 3.4375, "learning_rate": 2.2828686898280974e-05, "loss": 0.0251, "step": 1358 }, { "epoch": 46.86206896551724, "grad_norm": 1.0234375, "learning_rate": 2.2823769205893484e-05, "loss": 0.0176, "step": 1359 }, { "epoch": 46.89655172413793, "grad_norm": 2.140625, "learning_rate": 2.2818846481826298e-05, "loss": 0.0114, "step": 1360 }, { "epoch": 46.93103448275862, "grad_norm": 8.8125, "learning_rate": 2.281391872847869e-05, "loss": 0.0567, "step": 1361 }, { "epoch": 46.96551724137931, "grad_norm": 1.140625, "learning_rate": 2.280898594825239e-05, "loss": 0.0663, "step": 1362 }, { "epoch": 47.0, "grad_norm": 2.125, "learning_rate": 2.2804048143551575e-05, "loss": 0.0182, "step": 1363 }, { "epoch": 47.03448275862069, "grad_norm": 2.625, "learning_rate": 2.279910531678287e-05, "loss": 0.0203, "step": 1364 }, { "epoch": 47.06896551724138, "grad_norm": 2.25, "learning_rate": 2.2794157470355344e-05, "loss": 0.0156, "step": 1365 }, { "epoch": 47.10344827586207, "grad_norm": 2.109375, "learning_rate": 2.2789204606680524e-05, "loss": 0.0148, "step": 1366 }, { "epoch": 47.13793103448276, "grad_norm": 0.2373046875, "learning_rate": 2.2784246728172375e-05, "loss": 0.0095, "step": 1367 }, { "epoch": 47.172413793103445, "grad_norm": 3.140625, "learning_rate": 2.2779283837247305e-05, "loss": 0.0358, "step": 1368 }, { "epoch": 47.206896551724135, "grad_norm": 4.59375, "learning_rate": 2.2774315936324163e-05, "loss": 0.0447, "step": 1369 }, { "epoch": 47.241379310344826, "grad_norm": 0.65234375, "learning_rate": 2.2769343027824253e-05, "loss": 0.0055, "step": 1370 }, { "epoch": 47.275862068965516, "grad_norm": 1.96875, "learning_rate": 2.2764365114171303e-05, "loss": 0.017, "step": 1371 }, { "epoch": 47.310344827586206, "grad_norm": 6.5, "learning_rate": 2.275938219779149e-05, "loss": 0.0164, "step": 1372 }, { "epoch": 47.3448275862069, "grad_norm": 7.6875, "learning_rate": 2.2754394281113424e-05, "loss": 0.0773, "step": 1373 }, { "epoch": 47.37931034482759, "grad_norm": 1.6171875, "learning_rate": 2.2749401366568154e-05, "loss": 0.0083, "step": 1374 }, { "epoch": 47.41379310344828, "grad_norm": 7.96875, "learning_rate": 2.2744403456589173e-05, "loss": 0.0257, "step": 1375 }, { "epoch": 47.44827586206897, "grad_norm": 3.5625, "learning_rate": 2.27394005536124e-05, "loss": 0.0301, "step": 1376 }, { "epoch": 47.48275862068966, "grad_norm": 0.470703125, "learning_rate": 2.2734392660076183e-05, "loss": 0.0539, "step": 1377 }, { "epoch": 47.51724137931034, "grad_norm": 0.64453125, "learning_rate": 2.2729379778421312e-05, "loss": 0.0218, "step": 1378 }, { "epoch": 47.55172413793103, "grad_norm": 4.5625, "learning_rate": 2.272436191109101e-05, "loss": 0.0449, "step": 1379 }, { "epoch": 47.58620689655172, "grad_norm": 2.59375, "learning_rate": 2.2719339060530914e-05, "loss": 0.0313, "step": 1380 }, { "epoch": 47.62068965517241, "grad_norm": 1.2265625, "learning_rate": 2.2714311229189105e-05, "loss": 0.0317, "step": 1381 }, { "epoch": 47.6551724137931, "grad_norm": 3.421875, "learning_rate": 2.2709278419516096e-05, "loss": 0.0264, "step": 1382 }, { "epoch": 47.689655172413794, "grad_norm": 1.671875, "learning_rate": 2.2704240633964802e-05, "loss": 0.0174, "step": 1383 }, { "epoch": 47.724137931034484, "grad_norm": 4.28125, "learning_rate": 2.2699197874990584e-05, "loss": 0.0256, "step": 1384 }, { "epoch": 47.758620689655174, "grad_norm": 1.5390625, "learning_rate": 2.2694150145051228e-05, "loss": 0.0378, "step": 1385 }, { "epoch": 47.793103448275865, "grad_norm": 1.171875, "learning_rate": 2.268909744660693e-05, "loss": 0.0358, "step": 1386 }, { "epoch": 47.827586206896555, "grad_norm": 1.6328125, "learning_rate": 2.2684039782120315e-05, "loss": 0.026, "step": 1387 }, { "epoch": 47.86206896551724, "grad_norm": 3.296875, "learning_rate": 2.2678977154056426e-05, "loss": 0.0532, "step": 1388 }, { "epoch": 47.89655172413793, "grad_norm": 2.984375, "learning_rate": 2.267390956488273e-05, "loss": 0.0313, "step": 1389 }, { "epoch": 47.93103448275862, "grad_norm": 4.03125, "learning_rate": 2.26688370170691e-05, "loss": 0.0335, "step": 1390 }, { "epoch": 47.96551724137931, "grad_norm": 2.578125, "learning_rate": 2.2663759513087846e-05, "loss": 0.0191, "step": 1391 }, { "epoch": 48.0, "grad_norm": 3.546875, "learning_rate": 2.2658677055413676e-05, "loss": 0.0292, "step": 1392 }, { "epoch": 48.03448275862069, "grad_norm": 0.8046875, "learning_rate": 2.2653589646523714e-05, "loss": 0.0205, "step": 1393 }, { "epoch": 48.06896551724138, "grad_norm": 2.953125, "learning_rate": 2.2648497288897514e-05, "loss": 0.0247, "step": 1394 }, { "epoch": 48.10344827586207, "grad_norm": 1.875, "learning_rate": 2.2643399985017015e-05, "loss": 0.0372, "step": 1395 }, { "epoch": 48.13793103448276, "grad_norm": 3.890625, "learning_rate": 2.2638297737366583e-05, "loss": 0.0361, "step": 1396 }, { "epoch": 48.172413793103445, "grad_norm": 1.1328125, "learning_rate": 2.2633190548433e-05, "loss": 0.009, "step": 1397 }, { "epoch": 48.206896551724135, "grad_norm": 1.3359375, "learning_rate": 2.2628078420705443e-05, "loss": 0.0114, "step": 1398 }, { "epoch": 48.241379310344826, "grad_norm": 4.125, "learning_rate": 2.2622961356675498e-05, "loss": 0.073, "step": 1399 }, { "epoch": 48.275862068965516, "grad_norm": 3.359375, "learning_rate": 2.261783935883717e-05, "loss": 0.0295, "step": 1400 }, { "epoch": 48.310344827586206, "grad_norm": 4.03125, "learning_rate": 2.2612712429686845e-05, "loss": 0.0381, "step": 1401 }, { "epoch": 48.3448275862069, "grad_norm": 2.015625, "learning_rate": 2.260758057172333e-05, "loss": 0.0198, "step": 1402 }, { "epoch": 48.37931034482759, "grad_norm": 1.3125, "learning_rate": 2.2602443787447842e-05, "loss": 0.0169, "step": 1403 }, { "epoch": 48.41379310344828, "grad_norm": 2.90625, "learning_rate": 2.259730207936397e-05, "loss": 0.0203, "step": 1404 }, { "epoch": 48.44827586206897, "grad_norm": 1.5234375, "learning_rate": 2.2592155449977733e-05, "loss": 0.0255, "step": 1405 }, { "epoch": 48.48275862068966, "grad_norm": 2.28125, "learning_rate": 2.2587003901797528e-05, "loss": 0.0173, "step": 1406 }, { "epoch": 48.51724137931034, "grad_norm": 1.0859375, "learning_rate": 2.258184743733416e-05, "loss": 0.0163, "step": 1407 }, { "epoch": 48.55172413793103, "grad_norm": 6.59375, "learning_rate": 2.2576686059100825e-05, "loss": 0.0223, "step": 1408 }, { "epoch": 48.58620689655172, "grad_norm": 3.484375, "learning_rate": 2.2571519769613117e-05, "loss": 0.0306, "step": 1409 }, { "epoch": 48.62068965517241, "grad_norm": 0.921875, "learning_rate": 2.2566348571389025e-05, "loss": 0.0096, "step": 1410 }, { "epoch": 48.6551724137931, "grad_norm": 1.6796875, "learning_rate": 2.2561172466948925e-05, "loss": 0.0074, "step": 1411 }, { "epoch": 48.689655172413794, "grad_norm": 1.96875, "learning_rate": 2.2555991458815597e-05, "loss": 0.0086, "step": 1412 }, { "epoch": 48.724137931034484, "grad_norm": 2.796875, "learning_rate": 2.2550805549514184e-05, "loss": 0.0208, "step": 1413 }, { "epoch": 48.758620689655174, "grad_norm": 1.6484375, "learning_rate": 2.2545614741572248e-05, "loss": 0.0187, "step": 1414 }, { "epoch": 48.793103448275865, "grad_norm": 0.9140625, "learning_rate": 2.2540419037519725e-05, "loss": 0.0046, "step": 1415 }, { "epoch": 48.827586206896555, "grad_norm": 1.1796875, "learning_rate": 2.2535218439888933e-05, "loss": 0.0449, "step": 1416 }, { "epoch": 48.86206896551724, "grad_norm": 2.6875, "learning_rate": 2.2530012951214587e-05, "loss": 0.0153, "step": 1417 }, { "epoch": 48.89655172413793, "grad_norm": 0.59765625, "learning_rate": 2.2524802574033765e-05, "loss": 0.0102, "step": 1418 }, { "epoch": 48.93103448275862, "grad_norm": 8.1875, "learning_rate": 2.251958731088596e-05, "loss": 0.0445, "step": 1419 }, { "epoch": 48.96551724137931, "grad_norm": 0.85546875, "learning_rate": 2.2514367164313017e-05, "loss": 0.0058, "step": 1420 }, { "epoch": 49.0, "grad_norm": 1.5859375, "learning_rate": 2.2509142136859168e-05, "loss": 0.006, "step": 1421 }, { "epoch": 49.03448275862069, "grad_norm": 0.75390625, "learning_rate": 2.2503912231071038e-05, "loss": 0.0177, "step": 1422 }, { "epoch": 49.06896551724138, "grad_norm": 10.6875, "learning_rate": 2.249867744949762e-05, "loss": 0.042, "step": 1423 }, { "epoch": 49.10344827586207, "grad_norm": 4.15625, "learning_rate": 2.2493437794690274e-05, "loss": 0.0477, "step": 1424 }, { "epoch": 49.13793103448276, "grad_norm": 2.9375, "learning_rate": 2.2488193269202748e-05, "loss": 0.0223, "step": 1425 }, { "epoch": 49.172413793103445, "grad_norm": 0.24609375, "learning_rate": 2.248294387559116e-05, "loss": 0.0469, "step": 1426 }, { "epoch": 49.206896551724135, "grad_norm": 0.37890625, "learning_rate": 2.2477689616414004e-05, "loss": 0.0106, "step": 1427 }, { "epoch": 49.241379310344826, "grad_norm": 5.625, "learning_rate": 2.2472430494232138e-05, "loss": 0.0463, "step": 1428 }, { "epoch": 49.275862068965516, "grad_norm": 2.828125, "learning_rate": 2.2467166511608796e-05, "loss": 0.0181, "step": 1429 }, { "epoch": 49.310344827586206, "grad_norm": 2.109375, "learning_rate": 2.246189767110958e-05, "loss": 0.0053, "step": 1430 }, { "epoch": 49.3448275862069, "grad_norm": 5.09375, "learning_rate": 2.2456623975302454e-05, "loss": 0.0327, "step": 1431 }, { "epoch": 49.37931034482759, "grad_norm": 7.84375, "learning_rate": 2.245134542675776e-05, "loss": 0.0498, "step": 1432 }, { "epoch": 49.41379310344828, "grad_norm": 14.4375, "learning_rate": 2.2446062028048193e-05, "loss": 0.0382, "step": 1433 }, { "epoch": 49.44827586206897, "grad_norm": 3.46875, "learning_rate": 2.244077378174882e-05, "loss": 0.0206, "step": 1434 }, { "epoch": 49.48275862068966, "grad_norm": 1.4296875, "learning_rate": 2.243548069043707e-05, "loss": 0.0233, "step": 1435 }, { "epoch": 49.51724137931034, "grad_norm": 1.453125, "learning_rate": 2.2430182756692728e-05, "loss": 0.0178, "step": 1436 }, { "epoch": 49.55172413793103, "grad_norm": 2.375, "learning_rate": 2.2424879983097943e-05, "loss": 0.0185, "step": 1437 }, { "epoch": 49.58620689655172, "grad_norm": 4.9375, "learning_rate": 2.2419572372237226e-05, "loss": 0.0222, "step": 1438 }, { "epoch": 49.62068965517241, "grad_norm": 1.8125, "learning_rate": 2.2414259926697434e-05, "loss": 0.0249, "step": 1439 }, { "epoch": 49.6551724137931, "grad_norm": 2.265625, "learning_rate": 2.2408942649067795e-05, "loss": 0.0845, "step": 1440 }, { "epoch": 49.689655172413794, "grad_norm": 1.7109375, "learning_rate": 2.2403620541939884e-05, "loss": 0.036, "step": 1441 }, { "epoch": 49.724137931034484, "grad_norm": 2.0625, "learning_rate": 2.2398293607907634e-05, "loss": 0.0664, "step": 1442 }, { "epoch": 49.758620689655174, "grad_norm": 2.828125, "learning_rate": 2.2392961849567325e-05, "loss": 0.016, "step": 1443 }, { "epoch": 49.793103448275865, "grad_norm": 1.859375, "learning_rate": 2.2387625269517587e-05, "loss": 0.0627, "step": 1444 }, { "epoch": 49.827586206896555, "grad_norm": 4.78125, "learning_rate": 2.2382283870359412e-05, "loss": 0.0488, "step": 1445 }, { "epoch": 49.86206896551724, "grad_norm": 2.5625, "learning_rate": 2.2376937654696133e-05, "loss": 0.0284, "step": 1446 }, { "epoch": 49.89655172413793, "grad_norm": 1.8828125, "learning_rate": 2.237158662513342e-05, "loss": 0.0228, "step": 1447 }, { "epoch": 49.93103448275862, "grad_norm": 2.28125, "learning_rate": 2.2366230784279313e-05, "loss": 0.0196, "step": 1448 }, { "epoch": 49.96551724137931, "grad_norm": 4.78125, "learning_rate": 2.2360870134744178e-05, "loss": 0.0427, "step": 1449 }, { "epoch": 50.0, "grad_norm": 1.3515625, "learning_rate": 2.2355504679140728e-05, "loss": 0.0182, "step": 1450 }, { "epoch": 50.03448275862069, "grad_norm": 1.9453125, "learning_rate": 2.2350134420084023e-05, "loss": 0.0283, "step": 1451 }, { "epoch": 50.06896551724138, "grad_norm": 0.2578125, "learning_rate": 2.2344759360191468e-05, "loss": 0.0296, "step": 1452 }, { "epoch": 50.10344827586207, "grad_norm": 0.244140625, "learning_rate": 2.2339379502082792e-05, "loss": 0.0071, "step": 1453 }, { "epoch": 50.13793103448276, "grad_norm": 1.15625, "learning_rate": 2.2333994848380075e-05, "loss": 0.0213, "step": 1454 }, { "epoch": 50.172413793103445, "grad_norm": 8.875, "learning_rate": 2.2328605401707735e-05, "loss": 0.0378, "step": 1455 }, { "epoch": 50.206896551724135, "grad_norm": 1.859375, "learning_rate": 2.2323211164692526e-05, "loss": 0.0444, "step": 1456 }, { "epoch": 50.241379310344826, "grad_norm": 1.40625, "learning_rate": 2.2317812139963522e-05, "loss": 0.0113, "step": 1457 }, { "epoch": 50.275862068965516, "grad_norm": 2.828125, "learning_rate": 2.2312408330152157e-05, "loss": 0.026, "step": 1458 }, { "epoch": 50.310344827586206, "grad_norm": 3.984375, "learning_rate": 2.2306999737892168e-05, "loss": 0.0252, "step": 1459 }, { "epoch": 50.3448275862069, "grad_norm": 1.8125, "learning_rate": 2.2301586365819644e-05, "loss": 0.0315, "step": 1460 }, { "epoch": 50.37931034482759, "grad_norm": 4.21875, "learning_rate": 2.2296168216573e-05, "loss": 0.0258, "step": 1461 }, { "epoch": 50.41379310344828, "grad_norm": 0.291015625, "learning_rate": 2.229074529279297e-05, "loss": 0.0235, "step": 1462 }, { "epoch": 50.44827586206897, "grad_norm": 0.83984375, "learning_rate": 2.2285317597122623e-05, "loss": 0.0104, "step": 1463 }, { "epoch": 50.48275862068966, "grad_norm": 2.765625, "learning_rate": 2.227988513220735e-05, "loss": 0.03, "step": 1464 }, { "epoch": 50.51724137931034, "grad_norm": 1.0625, "learning_rate": 2.2274447900694876e-05, "loss": 0.0243, "step": 1465 }, { "epoch": 50.55172413793103, "grad_norm": 2.84375, "learning_rate": 2.2269005905235234e-05, "loss": 0.0195, "step": 1466 }, { "epoch": 50.58620689655172, "grad_norm": 2.609375, "learning_rate": 2.226355914848079e-05, "loss": 0.0364, "step": 1467 }, { "epoch": 50.62068965517241, "grad_norm": 0.259765625, "learning_rate": 2.2258107633086227e-05, "loss": 0.0086, "step": 1468 }, { "epoch": 50.6551724137931, "grad_norm": 0.66015625, "learning_rate": 2.2252651361708544e-05, "loss": 0.0169, "step": 1469 }, { "epoch": 50.689655172413794, "grad_norm": 1.0546875, "learning_rate": 2.2247190337007066e-05, "loss": 0.0258, "step": 1470 }, { "epoch": 50.724137931034484, "grad_norm": 4.90625, "learning_rate": 2.224172456164343e-05, "loss": 0.0774, "step": 1471 }, { "epoch": 50.758620689655174, "grad_norm": 2.78125, "learning_rate": 2.2236254038281583e-05, "loss": 0.0294, "step": 1472 }, { "epoch": 50.793103448275865, "grad_norm": 0.5234375, "learning_rate": 2.2230778769587797e-05, "loss": 0.0281, "step": 1473 }, { "epoch": 50.827586206896555, "grad_norm": 2.03125, "learning_rate": 2.2225298758230643e-05, "loss": 0.0443, "step": 1474 }, { "epoch": 50.86206896551724, "grad_norm": 1.8671875, "learning_rate": 2.2219814006881027e-05, "loss": 0.0394, "step": 1475 }, { "epoch": 50.89655172413793, "grad_norm": 0.4375, "learning_rate": 2.221432451821214e-05, "loss": 0.0169, "step": 1476 }, { "epoch": 50.93103448275862, "grad_norm": 4.03125, "learning_rate": 2.220883029489949e-05, "loss": 0.0179, "step": 1477 }, { "epoch": 50.96551724137931, "grad_norm": 5.71875, "learning_rate": 2.22033313396209e-05, "loss": 0.0156, "step": 1478 }, { "epoch": 51.0, "grad_norm": 0.6015625, "learning_rate": 2.2197827655056494e-05, "loss": 0.0203, "step": 1479 }, { "epoch": 51.03448275862069, "grad_norm": 4.90625, "learning_rate": 2.21923192438887e-05, "loss": 0.0481, "step": 1480 }, { "epoch": 51.06896551724138, "grad_norm": 0.380859375, "learning_rate": 2.2186806108802248e-05, "loss": 0.0351, "step": 1481 }, { "epoch": 51.10344827586207, "grad_norm": 1.5078125, "learning_rate": 2.218128825248417e-05, "loss": 0.0055, "step": 1482 }, { "epoch": 51.13793103448276, "grad_norm": 4.5625, "learning_rate": 2.2175765677623816e-05, "loss": 0.0498, "step": 1483 }, { "epoch": 51.172413793103445, "grad_norm": 10.8125, "learning_rate": 2.2170238386912813e-05, "loss": 0.0505, "step": 1484 }, { "epoch": 51.206896551724135, "grad_norm": 1.09375, "learning_rate": 2.216470638304509e-05, "loss": 0.0155, "step": 1485 }, { "epoch": 51.241379310344826, "grad_norm": 2.890625, "learning_rate": 2.215916966871689e-05, "loss": 0.0367, "step": 1486 }, { "epoch": 51.275862068965516, "grad_norm": 3.578125, "learning_rate": 2.2153628246626728e-05, "loss": 0.0353, "step": 1487 }, { "epoch": 51.310344827586206, "grad_norm": 3.21875, "learning_rate": 2.2148082119475437e-05, "loss": 0.0185, "step": 1488 }, { "epoch": 51.3448275862069, "grad_norm": 6.09375, "learning_rate": 2.214253128996613e-05, "loss": 0.0138, "step": 1489 }, { "epoch": 51.37931034482759, "grad_norm": 1.5859375, "learning_rate": 2.2136975760804204e-05, "loss": 0.0315, "step": 1490 }, { "epoch": 51.41379310344828, "grad_norm": 13.75, "learning_rate": 2.213141553469737e-05, "loss": 0.0413, "step": 1491 }, { "epoch": 51.44827586206897, "grad_norm": 6.46875, "learning_rate": 2.21258506143556e-05, "loss": 0.0528, "step": 1492 }, { "epoch": 51.48275862068966, "grad_norm": 0.50390625, "learning_rate": 2.2120281002491182e-05, "loss": 0.0084, "step": 1493 }, { "epoch": 51.51724137931034, "grad_norm": 2.546875, "learning_rate": 2.2114706701818672e-05, "loss": 0.0132, "step": 1494 }, { "epoch": 51.55172413793103, "grad_norm": 3.34375, "learning_rate": 2.2109127715054916e-05, "loss": 0.0193, "step": 1495 }, { "epoch": 51.58620689655172, "grad_norm": 6.78125, "learning_rate": 2.2103544044919045e-05, "loss": 0.0564, "step": 1496 }, { "epoch": 51.62068965517241, "grad_norm": 4.28125, "learning_rate": 2.2097955694132473e-05, "loss": 0.0411, "step": 1497 }, { "epoch": 51.6551724137931, "grad_norm": 6.28125, "learning_rate": 2.2092362665418887e-05, "loss": 0.0357, "step": 1498 }, { "epoch": 51.689655172413794, "grad_norm": 4.8125, "learning_rate": 2.208676496150427e-05, "loss": 0.0174, "step": 1499 }, { "epoch": 51.724137931034484, "grad_norm": 0.96875, "learning_rate": 2.2081162585116872e-05, "loss": 0.0106, "step": 1500 }, { "epoch": 51.758620689655174, "grad_norm": 1.3828125, "learning_rate": 2.2075555538987227e-05, "loss": 0.0191, "step": 1501 }, { "epoch": 51.793103448275865, "grad_norm": 1.84375, "learning_rate": 2.2069943825848133e-05, "loss": 0.0146, "step": 1502 }, { "epoch": 51.827586206896555, "grad_norm": 1.78125, "learning_rate": 2.206432744843468e-05, "loss": 0.0114, "step": 1503 }, { "epoch": 51.86206896551724, "grad_norm": 0.609375, "learning_rate": 2.2058706409484213e-05, "loss": 0.0107, "step": 1504 }, { "epoch": 51.89655172413793, "grad_norm": 2.0, "learning_rate": 2.2053080711736364e-05, "loss": 0.0183, "step": 1505 }, { "epoch": 51.93103448275862, "grad_norm": 2.171875, "learning_rate": 2.2047450357933032e-05, "loss": 0.0329, "step": 1506 }, { "epoch": 51.96551724137931, "grad_norm": 3.59375, "learning_rate": 2.2041815350818374e-05, "loss": 0.0201, "step": 1507 }, { "epoch": 52.0, "grad_norm": 4.34375, "learning_rate": 2.2036175693138827e-05, "loss": 0.0313, "step": 1508 }, { "epoch": 52.03448275862069, "grad_norm": 0.59375, "learning_rate": 2.20305313876431e-05, "loss": 0.0108, "step": 1509 }, { "epoch": 52.06896551724138, "grad_norm": 4.03125, "learning_rate": 2.202488243708215e-05, "loss": 0.0275, "step": 1510 }, { "epoch": 52.10344827586207, "grad_norm": 1.90625, "learning_rate": 2.201922884420921e-05, "loss": 0.0213, "step": 1511 }, { "epoch": 52.13793103448276, "grad_norm": 1.46875, "learning_rate": 2.2013570611779775e-05, "loss": 0.0386, "step": 1512 }, { "epoch": 52.172413793103445, "grad_norm": 1.4296875, "learning_rate": 2.200790774255159e-05, "loss": 0.0702, "step": 1513 }, { "epoch": 52.206896551724135, "grad_norm": 3.125, "learning_rate": 2.2002240239284677e-05, "loss": 0.0584, "step": 1514 }, { "epoch": 52.241379310344826, "grad_norm": 5.40625, "learning_rate": 2.199656810474131e-05, "loss": 0.0093, "step": 1515 }, { "epoch": 52.275862068965516, "grad_norm": 0.7890625, "learning_rate": 2.1990891341686008e-05, "loss": 0.004, "step": 1516 }, { "epoch": 52.310344827586206, "grad_norm": 2.4375, "learning_rate": 2.1985209952885573e-05, "loss": 0.0117, "step": 1517 }, { "epoch": 52.3448275862069, "grad_norm": 1.1796875, "learning_rate": 2.1979523941109026e-05, "loss": 0.0261, "step": 1518 }, { "epoch": 52.37931034482759, "grad_norm": 3.09375, "learning_rate": 2.1973833309127666e-05, "loss": 0.0378, "step": 1519 }, { "epoch": 52.41379310344828, "grad_norm": 4.0625, "learning_rate": 2.1968138059715048e-05, "loss": 0.032, "step": 1520 }, { "epoch": 52.44827586206897, "grad_norm": 1.625, "learning_rate": 2.1962438195646958e-05, "loss": 0.024, "step": 1521 }, { "epoch": 52.48275862068966, "grad_norm": 1.2890625, "learning_rate": 2.195673371970144e-05, "loss": 0.0487, "step": 1522 }, { "epoch": 52.51724137931034, "grad_norm": 2.828125, "learning_rate": 2.195102463465879e-05, "loss": 0.0187, "step": 1523 }, { "epoch": 52.55172413793103, "grad_norm": 0.453125, "learning_rate": 2.1945310943301544e-05, "loss": 0.0153, "step": 1524 }, { "epoch": 52.58620689655172, "grad_norm": 1.4296875, "learning_rate": 2.193959264841449e-05, "loss": 0.0104, "step": 1525 }, { "epoch": 52.62068965517241, "grad_norm": 0.78515625, "learning_rate": 2.1933869752784654e-05, "loss": 0.0111, "step": 1526 }, { "epoch": 52.6551724137931, "grad_norm": 3.171875, "learning_rate": 2.1928142259201297e-05, "loss": 0.0249, "step": 1527 }, { "epoch": 52.689655172413794, "grad_norm": 5.6875, "learning_rate": 2.1922410170455944e-05, "loss": 0.0591, "step": 1528 }, { "epoch": 52.724137931034484, "grad_norm": 13.8125, "learning_rate": 2.1916673489342337e-05, "loss": 0.0636, "step": 1529 }, { "epoch": 52.758620689655174, "grad_norm": 0.265625, "learning_rate": 2.1910932218656464e-05, "loss": 0.0125, "step": 1530 }, { "epoch": 52.793103448275865, "grad_norm": 2.375, "learning_rate": 2.1905186361196556e-05, "loss": 0.0252, "step": 1531 }, { "epoch": 52.827586206896555, "grad_norm": 0.9921875, "learning_rate": 2.189943591976307e-05, "loss": 0.016, "step": 1532 }, { "epoch": 52.86206896551724, "grad_norm": 4.875, "learning_rate": 2.1893680897158702e-05, "loss": 0.0475, "step": 1533 }, { "epoch": 52.89655172413793, "grad_norm": 5.71875, "learning_rate": 2.188792129618838e-05, "loss": 0.0275, "step": 1534 }, { "epoch": 52.93103448275862, "grad_norm": 1.359375, "learning_rate": 2.1882157119659265e-05, "loss": 0.0136, "step": 1535 }, { "epoch": 52.96551724137931, "grad_norm": 0.20703125, "learning_rate": 2.1876388370380745e-05, "loss": 0.0076, "step": 1536 }, { "epoch": 53.0, "grad_norm": 0.6015625, "learning_rate": 2.187061505116444e-05, "loss": 0.0206, "step": 1537 }, { "epoch": 53.03448275862069, "grad_norm": 0.23828125, "learning_rate": 2.186483716482419e-05, "loss": 0.0326, "step": 1538 }, { "epoch": 53.06896551724138, "grad_norm": 0.72265625, "learning_rate": 2.1859054714176076e-05, "loss": 0.0107, "step": 1539 }, { "epoch": 53.10344827586207, "grad_norm": 1.203125, "learning_rate": 2.1853267702038392e-05, "loss": 0.0106, "step": 1540 }, { "epoch": 53.13793103448276, "grad_norm": 9.25, "learning_rate": 2.184747613123165e-05, "loss": 0.0757, "step": 1541 }, { "epoch": 53.172413793103445, "grad_norm": 3.84375, "learning_rate": 2.1841680004578596e-05, "loss": 0.0079, "step": 1542 }, { "epoch": 53.206896551724135, "grad_norm": 1.65625, "learning_rate": 2.1835879324904195e-05, "loss": 0.0131, "step": 1543 }, { "epoch": 53.241379310344826, "grad_norm": 1.0390625, "learning_rate": 2.183007409503562e-05, "loss": 0.0069, "step": 1544 }, { "epoch": 53.275862068965516, "grad_norm": 2.90625, "learning_rate": 2.1824264317802278e-05, "loss": 0.019, "step": 1545 }, { "epoch": 53.310344827586206, "grad_norm": 3.4375, "learning_rate": 2.181844999603578e-05, "loss": 0.0247, "step": 1546 }, { "epoch": 53.3448275862069, "grad_norm": 7.8125, "learning_rate": 2.1812631132569953e-05, "loss": 0.03, "step": 1547 }, { "epoch": 53.37931034482759, "grad_norm": 2.21875, "learning_rate": 2.1806807730240845e-05, "loss": 0.0522, "step": 1548 }, { "epoch": 53.41379310344828, "grad_norm": 1.515625, "learning_rate": 2.1800979791886707e-05, "loss": 0.0231, "step": 1549 }, { "epoch": 53.44827586206897, "grad_norm": 2.84375, "learning_rate": 2.1795147320348012e-05, "loss": 0.0218, "step": 1550 }, { "epoch": 53.48275862068966, "grad_norm": 1.84375, "learning_rate": 2.178931031846743e-05, "loss": 0.047, "step": 1551 }, { "epoch": 53.51724137931034, "grad_norm": 4.8125, "learning_rate": 2.1783468789089844e-05, "loss": 0.0348, "step": 1552 }, { "epoch": 53.55172413793103, "grad_norm": 0.345703125, "learning_rate": 2.1777622735062354e-05, "loss": 0.016, "step": 1553 }, { "epoch": 53.58620689655172, "grad_norm": 1.0546875, "learning_rate": 2.1771772159234243e-05, "loss": 0.0219, "step": 1554 }, { "epoch": 53.62068965517241, "grad_norm": 0.78125, "learning_rate": 2.1765917064457013e-05, "loss": 0.012, "step": 1555 }, { "epoch": 53.6551724137931, "grad_norm": 1.5078125, "learning_rate": 2.1760057453584376e-05, "loss": 0.0196, "step": 1556 }, { "epoch": 53.689655172413794, "grad_norm": 6.8125, "learning_rate": 2.175419332947223e-05, "loss": 0.0505, "step": 1557 }, { "epoch": 53.724137931034484, "grad_norm": 1.34375, "learning_rate": 2.174832469497867e-05, "loss": 0.0216, "step": 1558 }, { "epoch": 53.758620689655174, "grad_norm": 3.234375, "learning_rate": 2.1742451552964006e-05, "loss": 0.0337, "step": 1559 }, { "epoch": 53.793103448275865, "grad_norm": 1.7421875, "learning_rate": 2.1736573906290736e-05, "loss": 0.0093, "step": 1560 }, { "epoch": 53.827586206896555, "grad_norm": 1.15625, "learning_rate": 2.1730691757823553e-05, "loss": 0.0075, "step": 1561 }, { "epoch": 53.86206896551724, "grad_norm": 0.330078125, "learning_rate": 2.1724805110429335e-05, "loss": 0.0053, "step": 1562 }, { "epoch": 53.89655172413793, "grad_norm": 3.4375, "learning_rate": 2.1718913966977175e-05, "loss": 0.0166, "step": 1563 }, { "epoch": 53.93103448275862, "grad_norm": 5.1875, "learning_rate": 2.1713018330338343e-05, "loss": 0.0104, "step": 1564 }, { "epoch": 53.96551724137931, "grad_norm": 0.34765625, "learning_rate": 2.1707118203386292e-05, "loss": 0.028, "step": 1565 }, { "epoch": 54.0, "grad_norm": 4.9375, "learning_rate": 2.1701213588996683e-05, "loss": 0.0283, "step": 1566 }, { "epoch": 54.03448275862069, "grad_norm": 0.146484375, "learning_rate": 2.1695304490047344e-05, "loss": 0.0251, "step": 1567 }, { "epoch": 54.06896551724138, "grad_norm": 4.5625, "learning_rate": 2.1689390909418305e-05, "loss": 0.0448, "step": 1568 }, { "epoch": 54.10344827586207, "grad_norm": 0.48828125, "learning_rate": 2.168347284999177e-05, "loss": 0.0046, "step": 1569 }, { "epoch": 54.13793103448276, "grad_norm": 4.21875, "learning_rate": 2.1677550314652127e-05, "loss": 0.0578, "step": 1570 }, { "epoch": 54.172413793103445, "grad_norm": 0.9921875, "learning_rate": 2.1671623306285956e-05, "loss": 0.0103, "step": 1571 }, { "epoch": 54.206896551724135, "grad_norm": 1.8046875, "learning_rate": 2.1665691827782003e-05, "loss": 0.0149, "step": 1572 }, { "epoch": 54.241379310344826, "grad_norm": 0.47265625, "learning_rate": 2.1659755882031196e-05, "loss": 0.0029, "step": 1573 }, { "epoch": 54.275862068965516, "grad_norm": 0.376953125, "learning_rate": 2.1653815471926646e-05, "loss": 0.0327, "step": 1574 }, { "epoch": 54.310344827586206, "grad_norm": 0.71875, "learning_rate": 2.164787060036364e-05, "loss": 0.0097, "step": 1575 }, { "epoch": 54.3448275862069, "grad_norm": 2.6875, "learning_rate": 2.1641921270239632e-05, "loss": 0.0129, "step": 1576 }, { "epoch": 54.37931034482759, "grad_norm": 0.53125, "learning_rate": 2.1635967484454257e-05, "loss": 0.028, "step": 1577 }, { "epoch": 54.41379310344828, "grad_norm": 2.234375, "learning_rate": 2.163000924590931e-05, "loss": 0.021, "step": 1578 }, { "epoch": 54.44827586206897, "grad_norm": 0.7265625, "learning_rate": 2.162404655750877e-05, "loss": 0.0072, "step": 1579 }, { "epoch": 54.48275862068966, "grad_norm": 3.703125, "learning_rate": 2.161807942215878e-05, "loss": 0.0466, "step": 1580 }, { "epoch": 54.51724137931034, "grad_norm": 3.4375, "learning_rate": 2.1612107842767647e-05, "loss": 0.0453, "step": 1581 }, { "epoch": 54.55172413793103, "grad_norm": 2.65625, "learning_rate": 2.1606131822245844e-05, "loss": 0.0206, "step": 1582 }, { "epoch": 54.58620689655172, "grad_norm": 2.25, "learning_rate": 2.1600151363506015e-05, "loss": 0.0409, "step": 1583 }, { "epoch": 54.62068965517241, "grad_norm": 4.15625, "learning_rate": 2.159416646946296e-05, "loss": 0.0479, "step": 1584 }, { "epoch": 54.6551724137931, "grad_norm": 3.875, "learning_rate": 2.158817714303365e-05, "loss": 0.0644, "step": 1585 }, { "epoch": 54.689655172413794, "grad_norm": 1.2109375, "learning_rate": 2.15821833871372e-05, "loss": 0.0272, "step": 1586 }, { "epoch": 54.724137931034484, "grad_norm": 2.296875, "learning_rate": 2.1576185204694898e-05, "loss": 0.032, "step": 1587 }, { "epoch": 54.758620689655174, "grad_norm": 3.953125, "learning_rate": 2.157018259863019e-05, "loss": 0.0321, "step": 1588 }, { "epoch": 54.793103448275865, "grad_norm": 0.6875, "learning_rate": 2.1564175571868668e-05, "loss": 0.0197, "step": 1589 }, { "epoch": 54.827586206896555, "grad_norm": 0.2734375, "learning_rate": 2.1558164127338086e-05, "loss": 0.0178, "step": 1590 }, { "epoch": 54.86206896551724, "grad_norm": 1.0078125, "learning_rate": 2.1552148267968347e-05, "loss": 0.0164, "step": 1591 }, { "epoch": 54.89655172413793, "grad_norm": 8.9375, "learning_rate": 2.1546127996691506e-05, "loss": 0.0356, "step": 1592 }, { "epoch": 54.93103448275862, "grad_norm": 0.1181640625, "learning_rate": 2.1540103316441777e-05, "loss": 0.0167, "step": 1593 }, { "epoch": 54.96551724137931, "grad_norm": 0.166015625, "learning_rate": 2.153407423015551e-05, "loss": 0.0141, "step": 1594 }, { "epoch": 55.0, "grad_norm": 5.03125, "learning_rate": 2.152804074077121e-05, "loss": 0.0535, "step": 1595 }, { "epoch": 55.03448275862069, "grad_norm": 0.50390625, "learning_rate": 2.152200285122953e-05, "loss": 0.0155, "step": 1596 }, { "epoch": 55.06896551724138, "grad_norm": 3.890625, "learning_rate": 2.151596056447326e-05, "loss": 0.0168, "step": 1597 }, { "epoch": 55.10344827586207, "grad_norm": 7.4375, "learning_rate": 2.1509913883447335e-05, "loss": 0.0348, "step": 1598 }, { "epoch": 55.13793103448276, "grad_norm": 3.75, "learning_rate": 2.1503862811098837e-05, "loss": 0.0121, "step": 1599 }, { "epoch": 55.172413793103445, "grad_norm": 1.1015625, "learning_rate": 2.1497807350376986e-05, "loss": 0.016, "step": 1600 }, { "epoch": 55.206896551724135, "grad_norm": 3.359375, "learning_rate": 2.1491747504233138e-05, "loss": 0.0288, "step": 1601 }, { "epoch": 55.241379310344826, "grad_norm": 4.40625, "learning_rate": 2.1485683275620794e-05, "loss": 0.0113, "step": 1602 }, { "epoch": 55.275862068965516, "grad_norm": 0.546875, "learning_rate": 2.1479614667495578e-05, "loss": 0.0313, "step": 1603 }, { "epoch": 55.310344827586206, "grad_norm": 0.99609375, "learning_rate": 2.1473541682815257e-05, "loss": 0.0043, "step": 1604 }, { "epoch": 55.3448275862069, "grad_norm": 0.61328125, "learning_rate": 2.1467464324539734e-05, "loss": 0.0044, "step": 1605 }, { "epoch": 55.37931034482759, "grad_norm": 2.953125, "learning_rate": 2.1461382595631036e-05, "loss": 0.0158, "step": 1606 }, { "epoch": 55.41379310344828, "grad_norm": 5.75, "learning_rate": 2.1455296499053324e-05, "loss": 0.0142, "step": 1607 }, { "epoch": 55.44827586206897, "grad_norm": 2.6875, "learning_rate": 2.144920603777289e-05, "loss": 0.0155, "step": 1608 }, { "epoch": 55.48275862068966, "grad_norm": 1.3515625, "learning_rate": 2.1443111214758148e-05, "loss": 0.0069, "step": 1609 }, { "epoch": 55.51724137931034, "grad_norm": 1.359375, "learning_rate": 2.1437012032979646e-05, "loss": 0.019, "step": 1610 }, { "epoch": 55.55172413793103, "grad_norm": 10.8125, "learning_rate": 2.1430908495410042e-05, "loss": 0.041, "step": 1611 }, { "epoch": 55.58620689655172, "grad_norm": 2.578125, "learning_rate": 2.1424800605024137e-05, "loss": 0.0172, "step": 1612 }, { "epoch": 55.62068965517241, "grad_norm": 2.265625, "learning_rate": 2.141868836479883e-05, "loss": 0.0168, "step": 1613 }, { "epoch": 55.6551724137931, "grad_norm": 2.53125, "learning_rate": 2.1412571777713163e-05, "loss": 0.0374, "step": 1614 }, { "epoch": 55.689655172413794, "grad_norm": 11.375, "learning_rate": 2.1406450846748282e-05, "loss": 0.0826, "step": 1615 }, { "epoch": 55.724137931034484, "grad_norm": 1.2265625, "learning_rate": 2.140032557488746e-05, "loss": 0.0081, "step": 1616 }, { "epoch": 55.758620689655174, "grad_norm": 1.3515625, "learning_rate": 2.139419596511607e-05, "loss": 0.0083, "step": 1617 }, { "epoch": 55.793103448275865, "grad_norm": 8.75, "learning_rate": 2.138806202042161e-05, "loss": 0.0849, "step": 1618 }, { "epoch": 55.827586206896555, "grad_norm": 3.96875, "learning_rate": 2.1381923743793704e-05, "loss": 0.047, "step": 1619 }, { "epoch": 55.86206896551724, "grad_norm": 0.72265625, "learning_rate": 2.1375781138224062e-05, "loss": 0.0319, "step": 1620 }, { "epoch": 55.89655172413793, "grad_norm": 0.1748046875, "learning_rate": 2.136963420670651e-05, "loss": 0.004, "step": 1621 }, { "epoch": 55.93103448275862, "grad_norm": 4.625, "learning_rate": 2.1363482952237e-05, "loss": 0.0397, "step": 1622 }, { "epoch": 55.96551724137931, "grad_norm": 6.21875, "learning_rate": 2.135732737781358e-05, "loss": 0.0428, "step": 1623 }, { "epoch": 56.0, "grad_norm": 0.6015625, "learning_rate": 2.1351167486436388e-05, "loss": 0.0084, "step": 1624 }, { "epoch": 56.03448275862069, "grad_norm": 2.90625, "learning_rate": 2.1345003281107692e-05, "loss": 0.0428, "step": 1625 }, { "epoch": 56.06896551724138, "grad_norm": 5.25, "learning_rate": 2.1338834764831845e-05, "loss": 0.037, "step": 1626 }, { "epoch": 56.10344827586207, "grad_norm": 2.640625, "learning_rate": 2.1332661940615312e-05, "loss": 0.0182, "step": 1627 }, { "epoch": 56.13793103448276, "grad_norm": 1.875, "learning_rate": 2.1326484811466646e-05, "loss": 0.0224, "step": 1628 }, { "epoch": 56.172413793103445, "grad_norm": 0.310546875, "learning_rate": 2.132030338039651e-05, "loss": 0.013, "step": 1629 }, { "epoch": 56.206896551724135, "grad_norm": 4.9375, "learning_rate": 2.1314117650417657e-05, "loss": 0.0304, "step": 1630 }, { "epoch": 56.241379310344826, "grad_norm": 5.84375, "learning_rate": 2.1307927624544934e-05, "loss": 0.0655, "step": 1631 }, { "epoch": 56.275862068965516, "grad_norm": 0.1943359375, "learning_rate": 2.130173330579529e-05, "loss": 0.0307, "step": 1632 }, { "epoch": 56.310344827586206, "grad_norm": 3.71875, "learning_rate": 2.1295534697187756e-05, "loss": 0.0172, "step": 1633 }, { "epoch": 56.3448275862069, "grad_norm": 1.109375, "learning_rate": 2.128933180174346e-05, "loss": 0.0191, "step": 1634 }, { "epoch": 56.37931034482759, "grad_norm": 3.828125, "learning_rate": 2.1283124622485616e-05, "loss": 0.0428, "step": 1635 }, { "epoch": 56.41379310344828, "grad_norm": 1.5703125, "learning_rate": 2.1276913162439532e-05, "loss": 0.0561, "step": 1636 }, { "epoch": 56.44827586206897, "grad_norm": 1.2734375, "learning_rate": 2.1270697424632593e-05, "loss": 0.0191, "step": 1637 }, { "epoch": 56.48275862068966, "grad_norm": 0.41796875, "learning_rate": 2.126447741209428e-05, "loss": 0.0198, "step": 1638 }, { "epoch": 56.51724137931034, "grad_norm": 6.5, "learning_rate": 2.1258253127856148e-05, "loss": 0.0475, "step": 1639 }, { "epoch": 56.55172413793103, "grad_norm": 0.9765625, "learning_rate": 2.1252024574951834e-05, "loss": 0.0216, "step": 1640 }, { "epoch": 56.58620689655172, "grad_norm": 1.328125, "learning_rate": 2.124579175641707e-05, "loss": 0.0083, "step": 1641 }, { "epoch": 56.62068965517241, "grad_norm": 5.75, "learning_rate": 2.1239554675289645e-05, "loss": 0.0338, "step": 1642 }, { "epoch": 56.6551724137931, "grad_norm": 3.703125, "learning_rate": 2.123331333460944e-05, "loss": 0.0227, "step": 1643 }, { "epoch": 56.689655172413794, "grad_norm": 1.0625, "learning_rate": 2.122706773741841e-05, "loss": 0.0257, "step": 1644 }, { "epoch": 56.724137931034484, "grad_norm": 3.21875, "learning_rate": 2.122081788676058e-05, "loss": 0.0214, "step": 1645 }, { "epoch": 56.758620689655174, "grad_norm": 0.87109375, "learning_rate": 2.121456378568206e-05, "loss": 0.0103, "step": 1646 }, { "epoch": 56.793103448275865, "grad_norm": 5.21875, "learning_rate": 2.120830543723101e-05, "loss": 0.0517, "step": 1647 }, { "epoch": 56.827586206896555, "grad_norm": 0.87890625, "learning_rate": 2.1202042844457682e-05, "loss": 0.0208, "step": 1648 }, { "epoch": 56.86206896551724, "grad_norm": 1.5625, "learning_rate": 2.1195776010414388e-05, "loss": 0.0107, "step": 1649 }, { "epoch": 56.89655172413793, "grad_norm": 1.2109375, "learning_rate": 2.11895049381555e-05, "loss": 0.0195, "step": 1650 }, { "epoch": 56.93103448275862, "grad_norm": 1.6875, "learning_rate": 2.1183229630737467e-05, "loss": 0.0125, "step": 1651 }, { "epoch": 56.96551724137931, "grad_norm": 3.9375, "learning_rate": 2.11769500912188e-05, "loss": 0.028, "step": 1652 }, { "epoch": 57.0, "grad_norm": 6.28125, "learning_rate": 2.117066632266006e-05, "loss": 0.0342, "step": 1653 }, { "epoch": 57.03448275862069, "grad_norm": 1.0078125, "learning_rate": 2.1164378328123898e-05, "loss": 0.0083, "step": 1654 }, { "epoch": 57.06896551724138, "grad_norm": 2.25, "learning_rate": 2.1158086110674987e-05, "loss": 0.0183, "step": 1655 }, { "epoch": 57.10344827586207, "grad_norm": 2.34375, "learning_rate": 2.1151789673380086e-05, "loss": 0.0194, "step": 1656 }, { "epoch": 57.13793103448276, "grad_norm": 2.296875, "learning_rate": 2.1145489019308e-05, "loss": 0.0129, "step": 1657 }, { "epoch": 57.172413793103445, "grad_norm": 6.96875, "learning_rate": 2.11391841515296e-05, "loss": 0.0205, "step": 1658 }, { "epoch": 57.206896551724135, "grad_norm": 0.96875, "learning_rate": 2.1132875073117785e-05, "loss": 0.0242, "step": 1659 }, { "epoch": 57.241379310344826, "grad_norm": 1.09375, "learning_rate": 2.112656178714754e-05, "loss": 0.0381, "step": 1660 }, { "epoch": 57.275862068965516, "grad_norm": 2.296875, "learning_rate": 2.1120244296695874e-05, "loss": 0.0406, "step": 1661 }, { "epoch": 57.310344827586206, "grad_norm": 1.6484375, "learning_rate": 2.1113922604841858e-05, "loss": 0.009, "step": 1662 }, { "epoch": 57.3448275862069, "grad_norm": 1.5625, "learning_rate": 2.1107596714666604e-05, "loss": 0.0085, "step": 1663 }, { "epoch": 57.37931034482759, "grad_norm": 4.9375, "learning_rate": 2.1101266629253286e-05, "loss": 0.0281, "step": 1664 }, { "epoch": 57.41379310344828, "grad_norm": 0.4140625, "learning_rate": 2.1094932351687095e-05, "loss": 0.0038, "step": 1665 }, { "epoch": 57.44827586206897, "grad_norm": 11.0, "learning_rate": 2.1088593885055288e-05, "loss": 0.0357, "step": 1666 }, { "epoch": 57.48275862068966, "grad_norm": 2.203125, "learning_rate": 2.1082251232447158e-05, "loss": 0.0259, "step": 1667 }, { "epoch": 57.51724137931034, "grad_norm": 0.5546875, "learning_rate": 2.107590439695404e-05, "loss": 0.0094, "step": 1668 }, { "epoch": 57.55172413793103, "grad_norm": 2.875, "learning_rate": 2.1069553381669296e-05, "loss": 0.0642, "step": 1669 }, { "epoch": 57.58620689655172, "grad_norm": 0.66796875, "learning_rate": 2.106319818968834e-05, "loss": 0.0051, "step": 1670 }, { "epoch": 57.62068965517241, "grad_norm": 2.6875, "learning_rate": 2.105683882410861e-05, "loss": 0.041, "step": 1671 }, { "epoch": 57.6551724137931, "grad_norm": 2.09375, "learning_rate": 2.1050475288029586e-05, "loss": 0.0295, "step": 1672 }, { "epoch": 57.689655172413794, "grad_norm": 2.296875, "learning_rate": 2.1044107584552785e-05, "loss": 0.0241, "step": 1673 }, { "epoch": 57.724137931034484, "grad_norm": 4.21875, "learning_rate": 2.1037735716781736e-05, "loss": 0.0204, "step": 1674 }, { "epoch": 57.758620689655174, "grad_norm": 3.078125, "learning_rate": 2.1031359687822023e-05, "loss": 0.0149, "step": 1675 }, { "epoch": 57.793103448275865, "grad_norm": 0.1767578125, "learning_rate": 2.1024979500781232e-05, "loss": 0.0048, "step": 1676 }, { "epoch": 57.827586206896555, "grad_norm": 3.796875, "learning_rate": 2.1018595158769e-05, "loss": 0.0209, "step": 1677 }, { "epoch": 57.86206896551724, "grad_norm": 4.5, "learning_rate": 2.101220666489697e-05, "loss": 0.0354, "step": 1678 }, { "epoch": 57.89655172413793, "grad_norm": 0.32421875, "learning_rate": 2.1005814022278816e-05, "loss": 0.0111, "step": 1679 }, { "epoch": 57.93103448275862, "grad_norm": 1.0859375, "learning_rate": 2.099941723403024e-05, "loss": 0.0168, "step": 1680 }, { "epoch": 57.96551724137931, "grad_norm": 4.75, "learning_rate": 2.099301630326896e-05, "loss": 0.0666, "step": 1681 }, { "epoch": 58.0, "grad_norm": 1.0078125, "learning_rate": 2.0986611233114702e-05, "loss": 0.0157, "step": 1682 }, { "epoch": 58.03448275862069, "grad_norm": 1.2734375, "learning_rate": 2.0980202026689227e-05, "loss": 0.0311, "step": 1683 }, { "epoch": 58.06896551724138, "grad_norm": 3.328125, "learning_rate": 2.0973788687116304e-05, "loss": 0.0326, "step": 1684 }, { "epoch": 58.10344827586207, "grad_norm": 10.5625, "learning_rate": 2.0967371217521716e-05, "loss": 0.0375, "step": 1685 }, { "epoch": 58.13793103448276, "grad_norm": 0.7890625, "learning_rate": 2.096094962103326e-05, "loss": 0.0126, "step": 1686 }, { "epoch": 58.172413793103445, "grad_norm": 1.546875, "learning_rate": 2.0954523900780742e-05, "loss": 0.0147, "step": 1687 }, { "epoch": 58.206896551724135, "grad_norm": 6.78125, "learning_rate": 2.0948094059895984e-05, "loss": 0.0607, "step": 1688 }, { "epoch": 58.241379310344826, "grad_norm": 3.609375, "learning_rate": 2.0941660101512806e-05, "loss": 0.0202, "step": 1689 }, { "epoch": 58.275862068965516, "grad_norm": 7.625, "learning_rate": 2.0935222028767047e-05, "loss": 0.0935, "step": 1690 }, { "epoch": 58.310344827586206, "grad_norm": 1.6875, "learning_rate": 2.092877984479654e-05, "loss": 0.0145, "step": 1691 }, { "epoch": 58.3448275862069, "grad_norm": 1.65625, "learning_rate": 2.0922333552741136e-05, "loss": 0.0181, "step": 1692 }, { "epoch": 58.37931034482759, "grad_norm": 3.59375, "learning_rate": 2.091588315574267e-05, "loss": 0.0573, "step": 1693 }, { "epoch": 58.41379310344828, "grad_norm": 0.4609375, "learning_rate": 2.0909428656944995e-05, "loss": 0.0073, "step": 1694 }, { "epoch": 58.44827586206897, "grad_norm": 8.625, "learning_rate": 2.0902970059493955e-05, "loss": 0.0883, "step": 1695 }, { "epoch": 58.48275862068966, "grad_norm": 1.8984375, "learning_rate": 2.089650736653738e-05, "loss": 0.013, "step": 1696 }, { "epoch": 58.51724137931034, "grad_norm": 1.40625, "learning_rate": 2.089004058122512e-05, "loss": 0.0106, "step": 1697 }, { "epoch": 58.55172413793103, "grad_norm": 2.3125, "learning_rate": 2.0883569706709e-05, "loss": 0.0214, "step": 1698 }, { "epoch": 58.58620689655172, "grad_norm": 2.09375, "learning_rate": 2.0877094746142857e-05, "loss": 0.0169, "step": 1699 }, { "epoch": 58.62068965517241, "grad_norm": 2.84375, "learning_rate": 2.0870615702682494e-05, "loss": 0.0467, "step": 1700 }, { "epoch": 58.6551724137931, "grad_norm": 1.078125, "learning_rate": 2.086413257948573e-05, "loss": 0.0211, "step": 1701 }, { "epoch": 58.689655172413794, "grad_norm": 1.96875, "learning_rate": 2.0857645379712353e-05, "loss": 0.0217, "step": 1702 }, { "epoch": 58.724137931034484, "grad_norm": 0.7109375, "learning_rate": 2.0851154106524146e-05, "loss": 0.0468, "step": 1703 }, { "epoch": 58.758620689655174, "grad_norm": 2.953125, "learning_rate": 2.0844658763084883e-05, "loss": 0.0312, "step": 1704 }, { "epoch": 58.793103448275865, "grad_norm": 3.375, "learning_rate": 2.0838159352560308e-05, "loss": 0.0296, "step": 1705 }, { "epoch": 58.827586206896555, "grad_norm": 1.1953125, "learning_rate": 2.0831655878118155e-05, "loss": 0.0166, "step": 1706 }, { "epoch": 58.86206896551724, "grad_norm": 4.375, "learning_rate": 2.0825148342928146e-05, "loss": 0.0419, "step": 1707 }, { "epoch": 58.89655172413793, "grad_norm": 1.421875, "learning_rate": 2.081863675016197e-05, "loss": 0.015, "step": 1708 }, { "epoch": 58.93103448275862, "grad_norm": 1.359375, "learning_rate": 2.081212110299329e-05, "loss": 0.0166, "step": 1709 }, { "epoch": 58.96551724137931, "grad_norm": 1.4375, "learning_rate": 2.080560140459777e-05, "loss": 0.0453, "step": 1710 }, { "epoch": 59.0, "grad_norm": 3.265625, "learning_rate": 2.0799077658153022e-05, "loss": 0.0223, "step": 1711 }, { "epoch": 59.03448275862069, "grad_norm": 3.265625, "learning_rate": 2.079254986683864e-05, "loss": 0.0448, "step": 1712 }, { "epoch": 59.06896551724138, "grad_norm": 3.140625, "learning_rate": 2.078601803383619e-05, "loss": 0.0173, "step": 1713 }, { "epoch": 59.10344827586207, "grad_norm": 0.40234375, "learning_rate": 2.077948216232921e-05, "loss": 0.0118, "step": 1714 }, { "epoch": 59.13793103448276, "grad_norm": 3.40625, "learning_rate": 2.077294225550321e-05, "loss": 0.045, "step": 1715 }, { "epoch": 59.172413793103445, "grad_norm": 7.6875, "learning_rate": 2.0766398316545648e-05, "loss": 0.043, "step": 1716 }, { "epoch": 59.206896551724135, "grad_norm": 1.078125, "learning_rate": 2.0759850348645972e-05, "loss": 0.016, "step": 1717 }, { "epoch": 59.241379310344826, "grad_norm": 1.1953125, "learning_rate": 2.075329835499558e-05, "loss": 0.0109, "step": 1718 }, { "epoch": 59.275862068965516, "grad_norm": 0.64453125, "learning_rate": 2.074674233878783e-05, "loss": 0.0142, "step": 1719 }, { "epoch": 59.310344827586206, "grad_norm": 5.46875, "learning_rate": 2.074018230321804e-05, "loss": 0.0501, "step": 1720 }, { "epoch": 59.3448275862069, "grad_norm": 1.0859375, "learning_rate": 2.0733618251483506e-05, "loss": 0.0114, "step": 1721 }, { "epoch": 59.37931034482759, "grad_norm": 1.65625, "learning_rate": 2.072705018678346e-05, "loss": 0.0318, "step": 1722 }, { "epoch": 59.41379310344828, "grad_norm": 3.71875, "learning_rate": 2.0720478112319092e-05, "loss": 0.0206, "step": 1723 }, { "epoch": 59.44827586206897, "grad_norm": 1.546875, "learning_rate": 2.0713902031293558e-05, "loss": 0.013, "step": 1724 }, { "epoch": 59.48275862068966, "grad_norm": 1.7578125, "learning_rate": 2.0707321946911957e-05, "loss": 0.0089, "step": 1725 }, { "epoch": 59.51724137931034, "grad_norm": 2.1875, "learning_rate": 2.070073786238134e-05, "loss": 0.0181, "step": 1726 }, { "epoch": 59.55172413793103, "grad_norm": 3.46875, "learning_rate": 2.069414978091072e-05, "loss": 0.0591, "step": 1727 }, { "epoch": 59.58620689655172, "grad_norm": 2.875, "learning_rate": 2.0687557705711033e-05, "loss": 0.019, "step": 1728 }, { "epoch": 59.62068965517241, "grad_norm": 2.796875, "learning_rate": 2.0680961639995185e-05, "loss": 0.0366, "step": 1729 }, { "epoch": 59.6551724137931, "grad_norm": 2.6875, "learning_rate": 2.067436158697802e-05, "loss": 0.0353, "step": 1730 }, { "epoch": 59.689655172413794, "grad_norm": 4.21875, "learning_rate": 2.066775754987632e-05, "loss": 0.0714, "step": 1731 }, { "epoch": 59.724137931034484, "grad_norm": 7.84375, "learning_rate": 2.0661149531908812e-05, "loss": 0.0442, "step": 1732 }, { "epoch": 59.758620689655174, "grad_norm": 3.90625, "learning_rate": 2.0654537536296164e-05, "loss": 0.0391, "step": 1733 }, { "epoch": 59.793103448275865, "grad_norm": 1.4609375, "learning_rate": 2.0647921566260987e-05, "loss": 0.0123, "step": 1734 }, { "epoch": 59.827586206896555, "grad_norm": 1.609375, "learning_rate": 2.064130162502782e-05, "loss": 0.0316, "step": 1735 }, { "epoch": 59.86206896551724, "grad_norm": 2.0625, "learning_rate": 2.0634677715823137e-05, "loss": 0.0352, "step": 1736 }, { "epoch": 59.89655172413793, "grad_norm": 0.1884765625, "learning_rate": 2.062804984187536e-05, "loss": 0.0036, "step": 1737 }, { "epoch": 59.93103448275862, "grad_norm": 0.890625, "learning_rate": 2.0621418006414833e-05, "loss": 0.0067, "step": 1738 }, { "epoch": 59.96551724137931, "grad_norm": 0.421875, "learning_rate": 2.0614782212673827e-05, "loss": 0.0065, "step": 1739 }, { "epoch": 60.0, "grad_norm": 2.203125, "learning_rate": 2.0608142463886553e-05, "loss": 0.0213, "step": 1740 }, { "epoch": 60.03448275862069, "grad_norm": 0.283203125, "learning_rate": 2.0601498763289138e-05, "loss": 0.0048, "step": 1741 }, { "epoch": 60.06896551724138, "grad_norm": 0.359375, "learning_rate": 2.0594851114119645e-05, "loss": 0.0483, "step": 1742 }, { "epoch": 60.10344827586207, "grad_norm": 5.4375, "learning_rate": 2.0588199519618058e-05, "loss": 0.0173, "step": 1743 }, { "epoch": 60.13793103448276, "grad_norm": 4.6875, "learning_rate": 2.0581543983026276e-05, "loss": 0.0306, "step": 1744 }, { "epoch": 60.172413793103445, "grad_norm": 0.3828125, "learning_rate": 2.0574884507588137e-05, "loss": 0.0041, "step": 1745 }, { "epoch": 60.206896551724135, "grad_norm": 5.65625, "learning_rate": 2.0568221096549384e-05, "loss": 0.0195, "step": 1746 }, { "epoch": 60.241379310344826, "grad_norm": 1.34375, "learning_rate": 2.056155375315768e-05, "loss": 0.0146, "step": 1747 }, { "epoch": 60.275862068965516, "grad_norm": 0.72265625, "learning_rate": 2.055488248066261e-05, "loss": 0.0245, "step": 1748 }, { "epoch": 60.310344827586206, "grad_norm": 1.4921875, "learning_rate": 2.0548207282315675e-05, "loss": 0.0138, "step": 1749 }, { "epoch": 60.3448275862069, "grad_norm": 0.296875, "learning_rate": 2.0541528161370276e-05, "loss": 0.01, "step": 1750 }, { "epoch": 60.37931034482759, "grad_norm": 1.015625, "learning_rate": 2.0534845121081742e-05, "loss": 0.0068, "step": 1751 }, { "epoch": 60.41379310344828, "grad_norm": 0.60546875, "learning_rate": 2.0528158164707308e-05, "loss": 0.0297, "step": 1752 }, { "epoch": 60.44827586206897, "grad_norm": 3.3125, "learning_rate": 2.052146729550611e-05, "loss": 0.077, "step": 1753 }, { "epoch": 60.48275862068966, "grad_norm": 1.2421875, "learning_rate": 2.0514772516739194e-05, "loss": 0.0374, "step": 1754 }, { "epoch": 60.51724137931034, "grad_norm": 1.4765625, "learning_rate": 2.050807383166952e-05, "loss": 0.034, "step": 1755 }, { "epoch": 60.55172413793103, "grad_norm": 5.375, "learning_rate": 2.0501371243561946e-05, "loss": 0.0627, "step": 1756 }, { "epoch": 60.58620689655172, "grad_norm": 5.96875, "learning_rate": 2.0494664755683226e-05, "loss": 0.0308, "step": 1757 }, { "epoch": 60.62068965517241, "grad_norm": 2.703125, "learning_rate": 2.0487954371302027e-05, "loss": 0.0093, "step": 1758 }, { "epoch": 60.6551724137931, "grad_norm": 0.74609375, "learning_rate": 2.0481240093688904e-05, "loss": 0.0222, "step": 1759 }, { "epoch": 60.689655172413794, "grad_norm": 0.51953125, "learning_rate": 2.047452192611632e-05, "loss": 0.0152, "step": 1760 }, { "epoch": 60.724137931034484, "grad_norm": 0.49609375, "learning_rate": 2.0467799871858624e-05, "loss": 0.0309, "step": 1761 }, { "epoch": 60.758620689655174, "grad_norm": 2.21875, "learning_rate": 2.0461073934192064e-05, "loss": 0.0155, "step": 1762 }, { "epoch": 60.793103448275865, "grad_norm": 3.0, "learning_rate": 2.045434411639478e-05, "loss": 0.0144, "step": 1763 }, { "epoch": 60.827586206896555, "grad_norm": 0.890625, "learning_rate": 2.0447610421746803e-05, "loss": 0.006, "step": 1764 }, { "epoch": 60.86206896551724, "grad_norm": 2.046875, "learning_rate": 2.0440872853530055e-05, "loss": 0.0235, "step": 1765 }, { "epoch": 60.89655172413793, "grad_norm": 0.291015625, "learning_rate": 2.0434131415028346e-05, "loss": 0.0263, "step": 1766 }, { "epoch": 60.93103448275862, "grad_norm": 7.40625, "learning_rate": 2.042738610952737e-05, "loss": 0.024, "step": 1767 }, { "epoch": 60.96551724137931, "grad_norm": 0.58203125, "learning_rate": 2.0420636940314708e-05, "loss": 0.0212, "step": 1768 }, { "epoch": 61.0, "grad_norm": 0.609375, "learning_rate": 2.041388391067982e-05, "loss": 0.0397, "step": 1769 }, { "epoch": 61.03448275862069, "grad_norm": 5.96875, "learning_rate": 2.0407127023914058e-05, "loss": 0.0368, "step": 1770 }, { "epoch": 61.06896551724138, "grad_norm": 0.201171875, "learning_rate": 2.0400366283310636e-05, "loss": 0.0026, "step": 1771 }, { "epoch": 61.10344827586207, "grad_norm": 0.75390625, "learning_rate": 2.0393601692164665e-05, "loss": 0.0204, "step": 1772 }, { "epoch": 61.13793103448276, "grad_norm": 7.21875, "learning_rate": 2.038683325377312e-05, "loss": 0.0592, "step": 1773 }, { "epoch": 61.172413793103445, "grad_norm": 11.375, "learning_rate": 2.0380060971434862e-05, "loss": 0.0889, "step": 1774 }, { "epoch": 61.206896551724135, "grad_norm": 4.09375, "learning_rate": 2.037328484845061e-05, "loss": 0.0271, "step": 1775 }, { "epoch": 61.241379310344826, "grad_norm": 2.78125, "learning_rate": 2.036650488812297e-05, "loss": 0.0177, "step": 1776 }, { "epoch": 61.275862068965516, "grad_norm": 0.90234375, "learning_rate": 2.035972109375641e-05, "loss": 0.0111, "step": 1777 }, { "epoch": 61.310344827586206, "grad_norm": 2.65625, "learning_rate": 2.035293346865727e-05, "loss": 0.0189, "step": 1778 }, { "epoch": 61.3448275862069, "grad_norm": 1.6640625, "learning_rate": 2.0346142016133758e-05, "loss": 0.0119, "step": 1779 }, { "epoch": 61.37931034482759, "grad_norm": 5.5625, "learning_rate": 2.033934673949594e-05, "loss": 0.0473, "step": 1780 }, { "epoch": 61.41379310344828, "grad_norm": 6.0, "learning_rate": 2.033254764205576e-05, "loss": 0.026, "step": 1781 }, { "epoch": 61.44827586206897, "grad_norm": 1.34375, "learning_rate": 2.0325744727127005e-05, "loss": 0.0313, "step": 1782 }, { "epoch": 61.48275862068966, "grad_norm": 4.28125, "learning_rate": 2.031893799802534e-05, "loss": 0.0183, "step": 1783 }, { "epoch": 61.51724137931034, "grad_norm": 1.8125, "learning_rate": 2.031212745806828e-05, "loss": 0.0135, "step": 1784 }, { "epoch": 61.55172413793103, "grad_norm": 0.8203125, "learning_rate": 2.0305313110575197e-05, "loss": 0.0054, "step": 1785 }, { "epoch": 61.58620689655172, "grad_norm": 1.34375, "learning_rate": 2.029849495886733e-05, "loss": 0.0131, "step": 1786 }, { "epoch": 61.62068965517241, "grad_norm": 0.34375, "learning_rate": 2.0291673006267753e-05, "loss": 0.0186, "step": 1787 }, { "epoch": 61.6551724137931, "grad_norm": 2.109375, "learning_rate": 2.028484725610141e-05, "loss": 0.0212, "step": 1788 }, { "epoch": 61.689655172413794, "grad_norm": 1.5234375, "learning_rate": 2.027801771169508e-05, "loss": 0.0159, "step": 1789 }, { "epoch": 61.724137931034484, "grad_norm": 0.404296875, "learning_rate": 2.0271184376377413e-05, "loss": 0.0075, "step": 1790 }, { "epoch": 61.758620689655174, "grad_norm": 2.015625, "learning_rate": 2.026434725347888e-05, "loss": 0.0109, "step": 1791 }, { "epoch": 61.793103448275865, "grad_norm": 0.54296875, "learning_rate": 2.0257506346331823e-05, "loss": 0.003, "step": 1792 }, { "epoch": 61.827586206896555, "grad_norm": 6.0, "learning_rate": 2.025066165827041e-05, "loss": 0.0464, "step": 1793 }, { "epoch": 61.86206896551724, "grad_norm": 2.578125, "learning_rate": 2.024381319263066e-05, "loss": 0.0229, "step": 1794 }, { "epoch": 61.89655172413793, "grad_norm": 6.5625, "learning_rate": 2.0236960952750435e-05, "loss": 0.0541, "step": 1795 }, { "epoch": 61.93103448275862, "grad_norm": 1.8359375, "learning_rate": 2.0230104941969426e-05, "loss": 0.0179, "step": 1796 }, { "epoch": 61.96551724137931, "grad_norm": 1.921875, "learning_rate": 2.022324516362918e-05, "loss": 0.0168, "step": 1797 }, { "epoch": 62.0, "grad_norm": 2.234375, "learning_rate": 2.0216381621073064e-05, "loss": 0.0293, "step": 1798 }, { "epoch": 62.03448275862069, "grad_norm": 3.21875, "learning_rate": 2.020951431764628e-05, "loss": 0.036, "step": 1799 }, { "epoch": 62.06896551724138, "grad_norm": 4.84375, "learning_rate": 2.020264325669588e-05, "loss": 0.029, "step": 1800 }, { "epoch": 62.10344827586207, "grad_norm": 1.7421875, "learning_rate": 2.0195768441570727e-05, "loss": 0.0116, "step": 1801 }, { "epoch": 62.13793103448276, "grad_norm": 3.875, "learning_rate": 2.018888987562153e-05, "loss": 0.0657, "step": 1802 }, { "epoch": 62.172413793103445, "grad_norm": 2.453125, "learning_rate": 2.0182007562200805e-05, "loss": 0.0341, "step": 1803 }, { "epoch": 62.206896551724135, "grad_norm": 1.3046875, "learning_rate": 2.017512150466292e-05, "loss": 0.0066, "step": 1804 }, { "epoch": 62.241379310344826, "grad_norm": 2.25, "learning_rate": 2.0168231706364054e-05, "loss": 0.0229, "step": 1805 }, { "epoch": 62.275862068965516, "grad_norm": 2.1875, "learning_rate": 2.0161338170662208e-05, "loss": 0.0245, "step": 1806 }, { "epoch": 62.310344827586206, "grad_norm": 6.46875, "learning_rate": 2.0154440900917206e-05, "loss": 0.0305, "step": 1807 }, { "epoch": 62.3448275862069, "grad_norm": 5.9375, "learning_rate": 2.0147539900490697e-05, "loss": 0.0283, "step": 1808 }, { "epoch": 62.37931034482759, "grad_norm": 1.1015625, "learning_rate": 2.0140635172746146e-05, "loss": 0.0054, "step": 1809 }, { "epoch": 62.41379310344828, "grad_norm": 1.125, "learning_rate": 2.0133726721048832e-05, "loss": 0.0079, "step": 1810 }, { "epoch": 62.44827586206897, "grad_norm": 0.9375, "learning_rate": 2.012681454876585e-05, "loss": 0.0168, "step": 1811 }, { "epoch": 62.48275862068966, "grad_norm": 0.52734375, "learning_rate": 2.01198986592661e-05, "loss": 0.017, "step": 1812 }, { "epoch": 62.51724137931034, "grad_norm": 0.5390625, "learning_rate": 2.011297905592032e-05, "loss": 0.0045, "step": 1813 }, { "epoch": 62.55172413793103, "grad_norm": 3.734375, "learning_rate": 2.0106055742101026e-05, "loss": 0.0171, "step": 1814 }, { "epoch": 62.58620689655172, "grad_norm": 2.84375, "learning_rate": 2.009912872118257e-05, "loss": 0.0097, "step": 1815 }, { "epoch": 62.62068965517241, "grad_norm": 2.90625, "learning_rate": 2.0092197996541085e-05, "loss": 0.079, "step": 1816 }, { "epoch": 62.6551724137931, "grad_norm": 3.796875, "learning_rate": 2.008526357155453e-05, "loss": 0.0186, "step": 1817 }, { "epoch": 62.689655172413794, "grad_norm": 0.58984375, "learning_rate": 2.0078325449602655e-05, "loss": 0.0063, "step": 1818 }, { "epoch": 62.724137931034484, "grad_norm": 6.53125, "learning_rate": 2.007138363406702e-05, "loss": 0.0325, "step": 1819 }, { "epoch": 62.758620689655174, "grad_norm": 3.28125, "learning_rate": 2.0064438128330975e-05, "loss": 0.0289, "step": 1820 }, { "epoch": 62.793103448275865, "grad_norm": 2.28125, "learning_rate": 2.005748893577969e-05, "loss": 0.0175, "step": 1821 }, { "epoch": 62.827586206896555, "grad_norm": 1.2265625, "learning_rate": 2.0050536059800098e-05, "loss": 0.0406, "step": 1822 }, { "epoch": 62.86206896551724, "grad_norm": 4.59375, "learning_rate": 2.0043579503780958e-05, "loss": 0.0573, "step": 1823 }, { "epoch": 62.89655172413793, "grad_norm": 0.78515625, "learning_rate": 2.003661927111281e-05, "loss": 0.0233, "step": 1824 }, { "epoch": 62.93103448275862, "grad_norm": 1.6875, "learning_rate": 2.002965536518798e-05, "loss": 0.0124, "step": 1825 }, { "epoch": 62.96551724137931, "grad_norm": 0.83203125, "learning_rate": 2.0022687789400607e-05, "loss": 0.0049, "step": 1826 }, { "epoch": 63.0, "grad_norm": 1.4296875, "learning_rate": 2.0015716547146586e-05, "loss": 0.0232, "step": 1827 }, { "epoch": 63.03448275862069, "grad_norm": 5.28125, "learning_rate": 2.000874164182362e-05, "loss": 0.023, "step": 1828 }, { "epoch": 63.06896551724138, "grad_norm": 0.41796875, "learning_rate": 2.0001763076831195e-05, "loss": 0.0071, "step": 1829 }, { "epoch": 63.10344827586207, "grad_norm": 0.49609375, "learning_rate": 1.999478085557058e-05, "loss": 0.0041, "step": 1830 }, { "epoch": 63.13793103448276, "grad_norm": 3.28125, "learning_rate": 1.9987794981444823e-05, "loss": 0.0206, "step": 1831 }, { "epoch": 63.172413793103445, "grad_norm": 0.318359375, "learning_rate": 1.9980805457858758e-05, "loss": 0.0061, "step": 1832 }, { "epoch": 63.206896551724135, "grad_norm": 0.80078125, "learning_rate": 1.9973812288218987e-05, "loss": 0.0138, "step": 1833 }, { "epoch": 63.241379310344826, "grad_norm": 1.8046875, "learning_rate": 1.9966815475933902e-05, "loss": 0.0089, "step": 1834 }, { "epoch": 63.275862068965516, "grad_norm": 6.15625, "learning_rate": 1.9959815024413657e-05, "loss": 0.0063, "step": 1835 }, { "epoch": 63.310344827586206, "grad_norm": 1.359375, "learning_rate": 1.99528109370702e-05, "loss": 0.0064, "step": 1836 }, { "epoch": 63.3448275862069, "grad_norm": 3.578125, "learning_rate": 1.9945803217317226e-05, "loss": 0.0288, "step": 1837 }, { "epoch": 63.37931034482759, "grad_norm": 14.6875, "learning_rate": 1.993879186857022e-05, "loss": 0.0416, "step": 1838 }, { "epoch": 63.41379310344828, "grad_norm": 2.9375, "learning_rate": 1.9931776894246417e-05, "loss": 0.0212, "step": 1839 }, { "epoch": 63.44827586206897, "grad_norm": 2.75, "learning_rate": 1.992475829776484e-05, "loss": 0.0353, "step": 1840 }, { "epoch": 63.48275862068966, "grad_norm": 10.8125, "learning_rate": 1.991773608254627e-05, "loss": 0.0701, "step": 1841 }, { "epoch": 63.51724137931034, "grad_norm": 3.46875, "learning_rate": 1.9910710252013244e-05, "loss": 0.0098, "step": 1842 }, { "epoch": 63.55172413793103, "grad_norm": 12.625, "learning_rate": 1.9903680809590065e-05, "loss": 0.0452, "step": 1843 }, { "epoch": 63.58620689655172, "grad_norm": 0.388671875, "learning_rate": 1.98966477587028e-05, "loss": 0.0065, "step": 1844 }, { "epoch": 63.62068965517241, "grad_norm": 6.6875, "learning_rate": 1.9889611102779276e-05, "loss": 0.0448, "step": 1845 }, { "epoch": 63.6551724137931, "grad_norm": 2.109375, "learning_rate": 1.988257084524907e-05, "loss": 0.0133, "step": 1846 }, { "epoch": 63.689655172413794, "grad_norm": 5.875, "learning_rate": 1.9875526989543517e-05, "loss": 0.0444, "step": 1847 }, { "epoch": 63.724137931034484, "grad_norm": 1.5546875, "learning_rate": 1.986847953909571e-05, "loss": 0.0128, "step": 1848 }, { "epoch": 63.758620689655174, "grad_norm": 5.71875, "learning_rate": 1.9861428497340483e-05, "loss": 0.0381, "step": 1849 }, { "epoch": 63.793103448275865, "grad_norm": 5.4375, "learning_rate": 1.9854373867714443e-05, "loss": 0.0746, "step": 1850 }, { "epoch": 63.827586206896555, "grad_norm": 2.96875, "learning_rate": 1.9847315653655915e-05, "loss": 0.0197, "step": 1851 }, { "epoch": 63.86206896551724, "grad_norm": 2.03125, "learning_rate": 1.9840253858604995e-05, "loss": 0.0156, "step": 1852 }, { "epoch": 63.89655172413793, "grad_norm": 1.03125, "learning_rate": 1.9833188486003516e-05, "loss": 0.0088, "step": 1853 }, { "epoch": 63.93103448275862, "grad_norm": 2.921875, "learning_rate": 1.9826119539295054e-05, "loss": 0.0348, "step": 1854 }, { "epoch": 63.96551724137931, "grad_norm": 1.3203125, "learning_rate": 1.9819047021924926e-05, "loss": 0.0053, "step": 1855 }, { "epoch": 64.0, "grad_norm": 0.59765625, "learning_rate": 1.9811970937340196e-05, "loss": 0.0096, "step": 1856 }, { "epoch": 64.03448275862068, "grad_norm": 2.03125, "learning_rate": 1.9804891288989653e-05, "loss": 0.0445, "step": 1857 }, { "epoch": 64.06896551724138, "grad_norm": 7.375, "learning_rate": 1.9797808080323837e-05, "loss": 0.0326, "step": 1858 }, { "epoch": 64.10344827586206, "grad_norm": 5.6875, "learning_rate": 1.979072131479502e-05, "loss": 0.0323, "step": 1859 }, { "epoch": 64.13793103448276, "grad_norm": 1.859375, "learning_rate": 1.9783630995857202e-05, "loss": 0.0266, "step": 1860 }, { "epoch": 64.17241379310344, "grad_norm": 5.1875, "learning_rate": 1.977653712696612e-05, "loss": 0.0317, "step": 1861 }, { "epoch": 64.20689655172414, "grad_norm": 0.443359375, "learning_rate": 1.9769439711579234e-05, "loss": 0.03, "step": 1862 }, { "epoch": 64.24137931034483, "grad_norm": 2.25, "learning_rate": 1.976233875315575e-05, "loss": 0.024, "step": 1863 }, { "epoch": 64.27586206896552, "grad_norm": 1.453125, "learning_rate": 1.9755234255156576e-05, "loss": 0.0099, "step": 1864 }, { "epoch": 64.3103448275862, "grad_norm": 2.25, "learning_rate": 1.9748126221044367e-05, "loss": 0.0331, "step": 1865 }, { "epoch": 64.34482758620689, "grad_norm": 0.76953125, "learning_rate": 1.9741014654283486e-05, "loss": 0.0099, "step": 1866 }, { "epoch": 64.37931034482759, "grad_norm": 3.375, "learning_rate": 1.973389955834003e-05, "loss": 0.0161, "step": 1867 }, { "epoch": 64.41379310344827, "grad_norm": 1.671875, "learning_rate": 1.9726780936681807e-05, "loss": 0.0119, "step": 1868 }, { "epoch": 64.44827586206897, "grad_norm": 4.15625, "learning_rate": 1.9719658792778344e-05, "loss": 0.0295, "step": 1869 }, { "epoch": 64.48275862068965, "grad_norm": 0.73828125, "learning_rate": 1.9712533130100898e-05, "loss": 0.0176, "step": 1870 }, { "epoch": 64.51724137931035, "grad_norm": 1.7734375, "learning_rate": 1.970540395212242e-05, "loss": 0.0208, "step": 1871 }, { "epoch": 64.55172413793103, "grad_norm": 1.046875, "learning_rate": 1.9698271262317587e-05, "loss": 0.0101, "step": 1872 }, { "epoch": 64.58620689655173, "grad_norm": 0.796875, "learning_rate": 1.9691135064162784e-05, "loss": 0.013, "step": 1873 }, { "epoch": 64.62068965517241, "grad_norm": 2.0625, "learning_rate": 1.968399536113611e-05, "loss": 0.0086, "step": 1874 }, { "epoch": 64.65517241379311, "grad_norm": 1.8359375, "learning_rate": 1.9676852156717375e-05, "loss": 0.0328, "step": 1875 }, { "epoch": 64.6896551724138, "grad_norm": 8.25, "learning_rate": 1.9669705454388076e-05, "loss": 0.0986, "step": 1876 }, { "epoch": 64.72413793103448, "grad_norm": 3.28125, "learning_rate": 1.9662555257631445e-05, "loss": 0.035, "step": 1877 }, { "epoch": 64.75862068965517, "grad_norm": 1.0234375, "learning_rate": 1.9655401569932388e-05, "loss": 0.0083, "step": 1878 }, { "epoch": 64.79310344827586, "grad_norm": 1.5390625, "learning_rate": 1.9648244394777533e-05, "loss": 0.0088, "step": 1879 }, { "epoch": 64.82758620689656, "grad_norm": 0.40625, "learning_rate": 1.9641083735655193e-05, "loss": 0.0079, "step": 1880 }, { "epoch": 64.86206896551724, "grad_norm": 3.625, "learning_rate": 1.96339195960554e-05, "loss": 0.0345, "step": 1881 }, { "epoch": 64.89655172413794, "grad_norm": 0.6796875, "learning_rate": 1.9626751979469855e-05, "loss": 0.0171, "step": 1882 }, { "epoch": 64.93103448275862, "grad_norm": 2.859375, "learning_rate": 1.961958088939197e-05, "loss": 0.0285, "step": 1883 }, { "epoch": 64.96551724137932, "grad_norm": 2.0625, "learning_rate": 1.961240632931685e-05, "loss": 0.0109, "step": 1884 }, { "epoch": 65.0, "grad_norm": 0.76953125, "learning_rate": 1.9605228302741288e-05, "loss": 0.0061, "step": 1885 }, { "epoch": 65.03448275862068, "grad_norm": 1.203125, "learning_rate": 1.9598046813163766e-05, "loss": 0.0126, "step": 1886 }, { "epoch": 65.06896551724138, "grad_norm": 0.376953125, "learning_rate": 1.9590861864084458e-05, "loss": 0.0036, "step": 1887 }, { "epoch": 65.10344827586206, "grad_norm": 2.71875, "learning_rate": 1.9583673459005212e-05, "loss": 0.0276, "step": 1888 }, { "epoch": 65.13793103448276, "grad_norm": 1.1328125, "learning_rate": 1.9576481601429584e-05, "loss": 0.0101, "step": 1889 }, { "epoch": 65.17241379310344, "grad_norm": 1.6875, "learning_rate": 1.9569286294862788e-05, "loss": 0.0225, "step": 1890 }, { "epoch": 65.20689655172414, "grad_norm": 1.75, "learning_rate": 1.9562087542811725e-05, "loss": 0.0061, "step": 1891 }, { "epoch": 65.24137931034483, "grad_norm": 0.47265625, "learning_rate": 1.955488534878499e-05, "loss": 0.007, "step": 1892 }, { "epoch": 65.27586206896552, "grad_norm": 2.21875, "learning_rate": 1.9547679716292834e-05, "loss": 0.0077, "step": 1893 }, { "epoch": 65.3103448275862, "grad_norm": 2.203125, "learning_rate": 1.954047064884721e-05, "loss": 0.0327, "step": 1894 }, { "epoch": 65.34482758620689, "grad_norm": 1.8828125, "learning_rate": 1.953325814996171e-05, "loss": 0.0277, "step": 1895 }, { "epoch": 65.37931034482759, "grad_norm": 2.515625, "learning_rate": 1.9526042223151634e-05, "loss": 0.0244, "step": 1896 }, { "epoch": 65.41379310344827, "grad_norm": 0.466796875, "learning_rate": 1.951882287193393e-05, "loss": 0.0272, "step": 1897 }, { "epoch": 65.44827586206897, "grad_norm": 3.625, "learning_rate": 1.9511600099827225e-05, "loss": 0.0239, "step": 1898 }, { "epoch": 65.48275862068965, "grad_norm": 0.890625, "learning_rate": 1.950437391035181e-05, "loss": 0.0134, "step": 1899 }, { "epoch": 65.51724137931035, "grad_norm": 1.3828125, "learning_rate": 1.949714430702964e-05, "loss": 0.027, "step": 1900 }, { "epoch": 65.55172413793103, "grad_norm": 0.81640625, "learning_rate": 1.9489911293384337e-05, "loss": 0.012, "step": 1901 }, { "epoch": 65.58620689655173, "grad_norm": 6.15625, "learning_rate": 1.9482674872941183e-05, "loss": 0.0378, "step": 1902 }, { "epoch": 65.62068965517241, "grad_norm": 10.4375, "learning_rate": 1.9475435049227125e-05, "loss": 0.0581, "step": 1903 }, { "epoch": 65.65517241379311, "grad_norm": 9.0625, "learning_rate": 1.946819182577076e-05, "loss": 0.0259, "step": 1904 }, { "epoch": 65.6896551724138, "grad_norm": 0.515625, "learning_rate": 1.9460945206102352e-05, "loss": 0.0249, "step": 1905 }, { "epoch": 65.72413793103448, "grad_norm": 1.375, "learning_rate": 1.9453695193753812e-05, "loss": 0.0258, "step": 1906 }, { "epoch": 65.75862068965517, "grad_norm": 5.1875, "learning_rate": 1.9446441792258707e-05, "loss": 0.0434, "step": 1907 }, { "epoch": 65.79310344827586, "grad_norm": 1.421875, "learning_rate": 1.943918500515226e-05, "loss": 0.0082, "step": 1908 }, { "epoch": 65.82758620689656, "grad_norm": 5.65625, "learning_rate": 1.9431924835971342e-05, "loss": 0.0412, "step": 1909 }, { "epoch": 65.86206896551724, "grad_norm": 3.5, "learning_rate": 1.9424661288254467e-05, "loss": 0.0321, "step": 1910 }, { "epoch": 65.89655172413794, "grad_norm": 0.396484375, "learning_rate": 1.9417394365541803e-05, "loss": 0.0084, "step": 1911 }, { "epoch": 65.93103448275862, "grad_norm": 1.3203125, "learning_rate": 1.9410124071375158e-05, "loss": 0.0102, "step": 1912 }, { "epoch": 65.96551724137932, "grad_norm": 0.3046875, "learning_rate": 1.9402850409297986e-05, "loss": 0.0159, "step": 1913 }, { "epoch": 66.0, "grad_norm": 1.40625, "learning_rate": 1.939557338285538e-05, "loss": 0.0329, "step": 1914 }, { "epoch": 66.03448275862068, "grad_norm": 2.765625, "learning_rate": 1.9388292995594076e-05, "loss": 0.0185, "step": 1915 }, { "epoch": 66.06896551724138, "grad_norm": 0.515625, "learning_rate": 1.9381009251062447e-05, "loss": 0.0057, "step": 1916 }, { "epoch": 66.10344827586206, "grad_norm": 0.443359375, "learning_rate": 1.93737221528105e-05, "loss": 0.0239, "step": 1917 }, { "epoch": 66.13793103448276, "grad_norm": 4.3125, "learning_rate": 1.9366431704389874e-05, "loss": 0.0118, "step": 1918 }, { "epoch": 66.17241379310344, "grad_norm": 1.6875, "learning_rate": 1.9359137909353857e-05, "loss": 0.0136, "step": 1919 }, { "epoch": 66.20689655172414, "grad_norm": 5.21875, "learning_rate": 1.9351840771257345e-05, "loss": 0.0298, "step": 1920 }, { "epoch": 66.24137931034483, "grad_norm": 3.96875, "learning_rate": 1.934454029365688e-05, "loss": 0.0129, "step": 1921 }, { "epoch": 66.27586206896552, "grad_norm": 3.09375, "learning_rate": 1.933723648011062e-05, "loss": 0.0177, "step": 1922 }, { "epoch": 66.3103448275862, "grad_norm": 1.734375, "learning_rate": 1.9329929334178366e-05, "loss": 0.0429, "step": 1923 }, { "epoch": 66.34482758620689, "grad_norm": 2.5625, "learning_rate": 1.9322618859421522e-05, "loss": 0.0345, "step": 1924 }, { "epoch": 66.37931034482759, "grad_norm": 2.203125, "learning_rate": 1.9315305059403127e-05, "loss": 0.0442, "step": 1925 }, { "epoch": 66.41379310344827, "grad_norm": 2.328125, "learning_rate": 1.930798793768784e-05, "loss": 0.0103, "step": 1926 }, { "epoch": 66.44827586206897, "grad_norm": 4.46875, "learning_rate": 1.9300667497841934e-05, "loss": 0.0294, "step": 1927 }, { "epoch": 66.48275862068965, "grad_norm": 4.375, "learning_rate": 1.929334374343331e-05, "loss": 0.0398, "step": 1928 }, { "epoch": 66.51724137931035, "grad_norm": 1.9609375, "learning_rate": 1.9286016678031472e-05, "loss": 0.0289, "step": 1929 }, { "epoch": 66.55172413793103, "grad_norm": 0.69140625, "learning_rate": 1.9278686305207537e-05, "loss": 0.008, "step": 1930 }, { "epoch": 66.58620689655173, "grad_norm": 4.46875, "learning_rate": 1.927135262853425e-05, "loss": 0.0138, "step": 1931 }, { "epoch": 66.62068965517241, "grad_norm": 2.375, "learning_rate": 1.9264015651585948e-05, "loss": 0.0141, "step": 1932 }, { "epoch": 66.65517241379311, "grad_norm": 2.875, "learning_rate": 1.925667537793859e-05, "loss": 0.0207, "step": 1933 }, { "epoch": 66.6896551724138, "grad_norm": 2.84375, "learning_rate": 1.9249331811169735e-05, "loss": 0.0284, "step": 1934 }, { "epoch": 66.72413793103448, "grad_norm": 1.84375, "learning_rate": 1.9241984954858543e-05, "loss": 0.038, "step": 1935 }, { "epoch": 66.75862068965517, "grad_norm": 0.369140625, "learning_rate": 1.9234634812585788e-05, "loss": 0.0034, "step": 1936 }, { "epoch": 66.79310344827586, "grad_norm": 0.2890625, "learning_rate": 1.9227281387933842e-05, "loss": 0.0345, "step": 1937 }, { "epoch": 66.82758620689656, "grad_norm": 0.95703125, "learning_rate": 1.921992468448667e-05, "loss": 0.0056, "step": 1938 }, { "epoch": 66.86206896551724, "grad_norm": 1.1640625, "learning_rate": 1.921256470582984e-05, "loss": 0.0211, "step": 1939 }, { "epoch": 66.89655172413794, "grad_norm": 1.078125, "learning_rate": 1.9205201455550514e-05, "loss": 0.0133, "step": 1940 }, { "epoch": 66.93103448275862, "grad_norm": 5.8125, "learning_rate": 1.9197834937237457e-05, "loss": 0.0251, "step": 1941 }, { "epoch": 66.96551724137932, "grad_norm": 4.6875, "learning_rate": 1.9190465154481018e-05, "loss": 0.0156, "step": 1942 }, { "epoch": 67.0, "grad_norm": 0.78125, "learning_rate": 1.918309211087314e-05, "loss": 0.0265, "step": 1943 }, { "epoch": 67.03448275862068, "grad_norm": 3.90625, "learning_rate": 1.917571581000736e-05, "loss": 0.037, "step": 1944 }, { "epoch": 67.06896551724138, "grad_norm": 4.03125, "learning_rate": 1.9168336255478787e-05, "loss": 0.0374, "step": 1945 }, { "epoch": 67.10344827586206, "grad_norm": 7.96875, "learning_rate": 1.9160953450884125e-05, "loss": 0.0154, "step": 1946 }, { "epoch": 67.13793103448276, "grad_norm": 5.4375, "learning_rate": 1.9153567399821674e-05, "loss": 0.0295, "step": 1947 }, { "epoch": 67.17241379310344, "grad_norm": 7.15625, "learning_rate": 1.91461781058913e-05, "loss": 0.0519, "step": 1948 }, { "epoch": 67.20689655172414, "grad_norm": 8.1875, "learning_rate": 1.913878557269446e-05, "loss": 0.0203, "step": 1949 }, { "epoch": 67.24137931034483, "grad_norm": 0.96484375, "learning_rate": 1.9131389803834177e-05, "loss": 0.0081, "step": 1950 }, { "epoch": 67.27586206896552, "grad_norm": 5.25, "learning_rate": 1.912399080291506e-05, "loss": 0.0353, "step": 1951 }, { "epoch": 67.3103448275862, "grad_norm": 3.875, "learning_rate": 1.9116588573543305e-05, "loss": 0.02, "step": 1952 }, { "epoch": 67.34482758620689, "grad_norm": 3.53125, "learning_rate": 1.9109183119326644e-05, "loss": 0.0213, "step": 1953 }, { "epoch": 67.37931034482759, "grad_norm": 1.7734375, "learning_rate": 1.9101774443874432e-05, "loss": 0.0145, "step": 1954 }, { "epoch": 67.41379310344827, "grad_norm": 0.53125, "learning_rate": 1.9094362550797547e-05, "loss": 0.0042, "step": 1955 }, { "epoch": 67.44827586206897, "grad_norm": 0.57421875, "learning_rate": 1.9086947443708472e-05, "loss": 0.0287, "step": 1956 }, { "epoch": 67.48275862068965, "grad_norm": 5.5625, "learning_rate": 1.9079529126221223e-05, "loss": 0.0363, "step": 1957 }, { "epoch": 67.51724137931035, "grad_norm": 2.1875, "learning_rate": 1.9072107601951417e-05, "loss": 0.0176, "step": 1958 }, { "epoch": 67.55172413793103, "grad_norm": 2.40625, "learning_rate": 1.90646828745162e-05, "loss": 0.0214, "step": 1959 }, { "epoch": 67.58620689655173, "grad_norm": 1.390625, "learning_rate": 1.9057254947534293e-05, "loss": 0.0594, "step": 1960 }, { "epoch": 67.62068965517241, "grad_norm": 1.546875, "learning_rate": 1.9049823824625993e-05, "loss": 0.0285, "step": 1961 }, { "epoch": 67.65517241379311, "grad_norm": 1.734375, "learning_rate": 1.904238950941312e-05, "loss": 0.0229, "step": 1962 }, { "epoch": 67.6896551724138, "grad_norm": 1.40625, "learning_rate": 1.9034952005519082e-05, "loss": 0.0345, "step": 1963 }, { "epoch": 67.72413793103448, "grad_norm": 2.359375, "learning_rate": 1.9027511316568824e-05, "loss": 0.0259, "step": 1964 }, { "epoch": 67.75862068965517, "grad_norm": 1.375, "learning_rate": 1.902006744618885e-05, "loss": 0.0068, "step": 1965 }, { "epoch": 67.79310344827586, "grad_norm": 0.58203125, "learning_rate": 1.901262039800721e-05, "loss": 0.0121, "step": 1966 }, { "epoch": 67.82758620689656, "grad_norm": 3.75, "learning_rate": 1.9005170175653498e-05, "loss": 0.0504, "step": 1967 }, { "epoch": 67.86206896551724, "grad_norm": 2.609375, "learning_rate": 1.8997716782758867e-05, "loss": 0.0187, "step": 1968 }, { "epoch": 67.89655172413794, "grad_norm": 1.1328125, "learning_rate": 1.899026022295602e-05, "loss": 0.0353, "step": 1969 }, { "epoch": 67.93103448275862, "grad_norm": 0.369140625, "learning_rate": 1.8982800499879178e-05, "loss": 0.0051, "step": 1970 }, { "epoch": 67.96551724137932, "grad_norm": 0.85546875, "learning_rate": 1.897533761716413e-05, "loss": 0.0036, "step": 1971 }, { "epoch": 68.0, "grad_norm": 1.5390625, "learning_rate": 1.8967871578448187e-05, "loss": 0.0138, "step": 1972 }, { "epoch": 68.03448275862068, "grad_norm": 1.8046875, "learning_rate": 1.8960402387370214e-05, "loss": 0.0113, "step": 1973 }, { "epoch": 68.06896551724138, "grad_norm": 1.9609375, "learning_rate": 1.8952930047570595e-05, "loss": 0.0105, "step": 1974 }, { "epoch": 68.10344827586206, "grad_norm": 0.279296875, "learning_rate": 1.8945454562691266e-05, "loss": 0.012, "step": 1975 }, { "epoch": 68.13793103448276, "grad_norm": 5.34375, "learning_rate": 1.893797593637568e-05, "loss": 0.0217, "step": 1976 }, { "epoch": 68.17241379310344, "grad_norm": 1.953125, "learning_rate": 1.893049417226883e-05, "loss": 0.0128, "step": 1977 }, { "epoch": 68.20689655172414, "grad_norm": 2.59375, "learning_rate": 1.8923009274017245e-05, "loss": 0.0111, "step": 1978 }, { "epoch": 68.24137931034483, "grad_norm": 1.234375, "learning_rate": 1.8915521245268964e-05, "loss": 0.0247, "step": 1979 }, { "epoch": 68.27586206896552, "grad_norm": 5.5625, "learning_rate": 1.8908030089673563e-05, "loss": 0.0309, "step": 1980 }, { "epoch": 68.3103448275862, "grad_norm": 1.5, "learning_rate": 1.8900535810882146e-05, "loss": 0.0274, "step": 1981 }, { "epoch": 68.34482758620689, "grad_norm": 2.34375, "learning_rate": 1.8893038412547327e-05, "loss": 0.0268, "step": 1982 }, { "epoch": 68.37931034482759, "grad_norm": 5.84375, "learning_rate": 1.888553789832325e-05, "loss": 0.0575, "step": 1983 }, { "epoch": 68.41379310344827, "grad_norm": 1.78125, "learning_rate": 1.8878034271865576e-05, "loss": 0.0126, "step": 1984 }, { "epoch": 68.44827586206897, "grad_norm": 3.921875, "learning_rate": 1.8870527536831473e-05, "loss": 0.0167, "step": 1985 }, { "epoch": 68.48275862068965, "grad_norm": 1.109375, "learning_rate": 1.8863017696879642e-05, "loss": 0.0106, "step": 1986 }, { "epoch": 68.51724137931035, "grad_norm": 5.71875, "learning_rate": 1.8855504755670287e-05, "loss": 0.0214, "step": 1987 }, { "epoch": 68.55172413793103, "grad_norm": 2.875, "learning_rate": 1.8847988716865115e-05, "loss": 0.0175, "step": 1988 }, { "epoch": 68.58620689655173, "grad_norm": 0.59765625, "learning_rate": 1.8840469584127367e-05, "loss": 0.0168, "step": 1989 }, { "epoch": 68.62068965517241, "grad_norm": 5.78125, "learning_rate": 1.8832947361121763e-05, "loss": 0.0422, "step": 1990 }, { "epoch": 68.65517241379311, "grad_norm": 2.796875, "learning_rate": 1.8825422051514546e-05, "loss": 0.0226, "step": 1991 }, { "epoch": 68.6896551724138, "grad_norm": 0.796875, "learning_rate": 1.8817893658973475e-05, "loss": 0.0102, "step": 1992 }, { "epoch": 68.72413793103448, "grad_norm": 11.9375, "learning_rate": 1.8810362187167774e-05, "loss": 0.0945, "step": 1993 }, { "epoch": 68.75862068965517, "grad_norm": 3.015625, "learning_rate": 1.8802827639768206e-05, "loss": 0.0264, "step": 1994 }, { "epoch": 68.79310344827586, "grad_norm": 3.109375, "learning_rate": 1.879529002044701e-05, "loss": 0.0223, "step": 1995 }, { "epoch": 68.82758620689656, "grad_norm": 4.4375, "learning_rate": 1.878774933287794e-05, "loss": 0.0177, "step": 1996 }, { "epoch": 68.86206896551724, "grad_norm": 4.53125, "learning_rate": 1.8780205580736228e-05, "loss": 0.0344, "step": 1997 }, { "epoch": 68.89655172413794, "grad_norm": 3.171875, "learning_rate": 1.8772658767698607e-05, "loss": 0.0167, "step": 1998 }, { "epoch": 68.93103448275862, "grad_norm": 1.0390625, "learning_rate": 1.8765108897443303e-05, "loss": 0.0216, "step": 1999 }, { "epoch": 68.96551724137932, "grad_norm": 1.3125, "learning_rate": 1.8757555973650036e-05, "loss": 0.0064, "step": 2000 }, { "epoch": 69.0, "grad_norm": 1.5390625, "learning_rate": 1.8750000000000002e-05, "loss": 0.0087, "step": 2001 }, { "epoch": 69.03448275862068, "grad_norm": 0.435546875, "learning_rate": 1.8742440980175896e-05, "loss": 0.0238, "step": 2002 }, { "epoch": 69.06896551724138, "grad_norm": 0.96484375, "learning_rate": 1.8734878917861885e-05, "loss": 0.0082, "step": 2003 }, { "epoch": 69.10344827586206, "grad_norm": 0.421875, "learning_rate": 1.8727313816743637e-05, "loss": 0.004, "step": 2004 }, { "epoch": 69.13793103448276, "grad_norm": 2.546875, "learning_rate": 1.8719745680508282e-05, "loss": 0.0213, "step": 2005 }, { "epoch": 69.17241379310344, "grad_norm": 1.15625, "learning_rate": 1.8712174512844445e-05, "loss": 0.0108, "step": 2006 }, { "epoch": 69.20689655172414, "grad_norm": 0.419921875, "learning_rate": 1.8704600317442213e-05, "loss": 0.005, "step": 2007 }, { "epoch": 69.24137931034483, "grad_norm": 3.03125, "learning_rate": 1.869702309799316e-05, "loss": 0.0427, "step": 2008 }, { "epoch": 69.27586206896552, "grad_norm": 1.78125, "learning_rate": 1.8689442858190332e-05, "loss": 0.0902, "step": 2009 }, { "epoch": 69.3103448275862, "grad_norm": 0.69140625, "learning_rate": 1.8681859601728248e-05, "loss": 0.0069, "step": 2010 }, { "epoch": 69.34482758620689, "grad_norm": 0.23828125, "learning_rate": 1.8674273332302888e-05, "loss": 0.0135, "step": 2011 }, { "epoch": 69.37931034482759, "grad_norm": 0.84765625, "learning_rate": 1.8666684053611712e-05, "loss": 0.0187, "step": 2012 }, { "epoch": 69.41379310344827, "grad_norm": 1.703125, "learning_rate": 1.8659091769353644e-05, "loss": 0.0077, "step": 2013 }, { "epoch": 69.44827586206897, "grad_norm": 0.34375, "learning_rate": 1.865149648322907e-05, "loss": 0.0106, "step": 2014 }, { "epoch": 69.48275862068965, "grad_norm": 4.46875, "learning_rate": 1.8643898198939837e-05, "loss": 0.0553, "step": 2015 }, { "epoch": 69.51724137931035, "grad_norm": 0.6328125, "learning_rate": 1.863629692018926e-05, "loss": 0.0033, "step": 2016 }, { "epoch": 69.55172413793103, "grad_norm": 0.388671875, "learning_rate": 1.862869265068211e-05, "loss": 0.0046, "step": 2017 }, { "epoch": 69.58620689655173, "grad_norm": 1.3671875, "learning_rate": 1.8621085394124616e-05, "loss": 0.0135, "step": 2018 }, { "epoch": 69.62068965517241, "grad_norm": 0.314453125, "learning_rate": 1.8613475154224454e-05, "loss": 0.0168, "step": 2019 }, { "epoch": 69.65517241379311, "grad_norm": 2.1875, "learning_rate": 1.8605861934690775e-05, "loss": 0.0066, "step": 2020 }, { "epoch": 69.6896551724138, "grad_norm": 1.4375, "learning_rate": 1.859824573923416e-05, "loss": 0.0096, "step": 2021 }, { "epoch": 69.72413793103448, "grad_norm": 1.203125, "learning_rate": 1.8590626571566654e-05, "loss": 0.0219, "step": 2022 }, { "epoch": 69.75862068965517, "grad_norm": 1.234375, "learning_rate": 1.858300443540175e-05, "loss": 0.0071, "step": 2023 }, { "epoch": 69.79310344827586, "grad_norm": 1.1953125, "learning_rate": 1.857537933445438e-05, "loss": 0.0154, "step": 2024 }, { "epoch": 69.82758620689656, "grad_norm": 1.671875, "learning_rate": 1.8567751272440925e-05, "loss": 0.0589, "step": 2025 }, { "epoch": 69.86206896551724, "grad_norm": 0.2451171875, "learning_rate": 1.8560120253079215e-05, "loss": 0.0043, "step": 2026 }, { "epoch": 69.89655172413794, "grad_norm": 1.5234375, "learning_rate": 1.855248628008851e-05, "loss": 0.0084, "step": 2027 }, { "epoch": 69.93103448275862, "grad_norm": 1.328125, "learning_rate": 1.854484935718952e-05, "loss": 0.0535, "step": 2028 }, { "epoch": 69.96551724137932, "grad_norm": 1.8203125, "learning_rate": 1.8537209488104383e-05, "loss": 0.0201, "step": 2029 }, { "epoch": 70.0, "grad_norm": 2.078125, "learning_rate": 1.852956667655669e-05, "loss": 0.0068, "step": 2030 }, { "epoch": 70.03448275862068, "grad_norm": 2.71875, "learning_rate": 1.852192092627144e-05, "loss": 0.0118, "step": 2031 }, { "epoch": 70.06896551724138, "grad_norm": 0.91015625, "learning_rate": 1.8514272240975093e-05, "loss": 0.0276, "step": 2032 }, { "epoch": 70.10344827586206, "grad_norm": 5.0, "learning_rate": 1.850662062439551e-05, "loss": 0.0284, "step": 2033 }, { "epoch": 70.13793103448276, "grad_norm": 1.8671875, "learning_rate": 1.849896608026201e-05, "loss": 0.0346, "step": 2034 }, { "epoch": 70.17241379310344, "grad_norm": 3.40625, "learning_rate": 1.8491308612305314e-05, "loss": 0.035, "step": 2035 }, { "epoch": 70.20689655172414, "grad_norm": 0.294921875, "learning_rate": 1.848364822425759e-05, "loss": 0.0086, "step": 2036 }, { "epoch": 70.24137931034483, "grad_norm": 0.494140625, "learning_rate": 1.8475984919852404e-05, "loss": 0.0205, "step": 2037 }, { "epoch": 70.27586206896552, "grad_norm": 0.5234375, "learning_rate": 1.8468318702824765e-05, "loss": 0.0103, "step": 2038 }, { "epoch": 70.3103448275862, "grad_norm": 0.404296875, "learning_rate": 1.84606495769111e-05, "loss": 0.0304, "step": 2039 }, { "epoch": 70.34482758620689, "grad_norm": 5.75, "learning_rate": 1.845297754584924e-05, "loss": 0.0431, "step": 2040 }, { "epoch": 70.37931034482759, "grad_norm": 0.314453125, "learning_rate": 1.8445302613378444e-05, "loss": 0.0039, "step": 2041 }, { "epoch": 70.41379310344827, "grad_norm": 10.375, "learning_rate": 1.8437624783239372e-05, "loss": 0.079, "step": 2042 }, { "epoch": 70.44827586206897, "grad_norm": 2.625, "learning_rate": 1.8429944059174116e-05, "loss": 0.0134, "step": 2043 }, { "epoch": 70.48275862068965, "grad_norm": 4.75, "learning_rate": 1.8422260444926165e-05, "loss": 0.0283, "step": 2044 }, { "epoch": 70.51724137931035, "grad_norm": 3.515625, "learning_rate": 1.8414573944240414e-05, "loss": 0.0317, "step": 2045 }, { "epoch": 70.55172413793103, "grad_norm": 0.35546875, "learning_rate": 1.840688456086318e-05, "loss": 0.0063, "step": 2046 }, { "epoch": 70.58620689655173, "grad_norm": 0.77734375, "learning_rate": 1.8399192298542163e-05, "loss": 0.0043, "step": 2047 }, { "epoch": 70.62068965517241, "grad_norm": 5.0, "learning_rate": 1.839149716102648e-05, "loss": 0.0401, "step": 2048 }, { "epoch": 70.65517241379311, "grad_norm": 2.65625, "learning_rate": 1.8383799152066662e-05, "loss": 0.0517, "step": 2049 }, { "epoch": 70.6896551724138, "grad_norm": 3.765625, "learning_rate": 1.8376098275414606e-05, "loss": 0.0212, "step": 2050 }, { "epoch": 70.72413793103448, "grad_norm": 1.7109375, "learning_rate": 1.8368394534823635e-05, "loss": 0.0354, "step": 2051 }, { "epoch": 70.75862068965517, "grad_norm": 3.625, "learning_rate": 1.8360687934048462e-05, "loss": 0.017, "step": 2052 }, { "epoch": 70.79310344827586, "grad_norm": 1.75, "learning_rate": 1.8352978476845182e-05, "loss": 0.0119, "step": 2053 }, { "epoch": 70.82758620689656, "grad_norm": 3.84375, "learning_rate": 1.83452661669713e-05, "loss": 0.039, "step": 2054 }, { "epoch": 70.86206896551724, "grad_norm": 3.015625, "learning_rate": 1.8337551008185694e-05, "loss": 0.0133, "step": 2055 }, { "epoch": 70.89655172413794, "grad_norm": 0.7578125, "learning_rate": 1.832983300424864e-05, "loss": 0.0132, "step": 2056 }, { "epoch": 70.93103448275862, "grad_norm": 1.6015625, "learning_rate": 1.8322112158921804e-05, "loss": 0.0201, "step": 2057 }, { "epoch": 70.96551724137932, "grad_norm": 0.66796875, "learning_rate": 1.831438847596823e-05, "loss": 0.0086, "step": 2058 }, { "epoch": 71.0, "grad_norm": 2.8125, "learning_rate": 1.830666195915235e-05, "loss": 0.0251, "step": 2059 }, { "epoch": 71.03448275862068, "grad_norm": 0.5078125, "learning_rate": 1.8298932612239964e-05, "loss": 0.0132, "step": 2060 }, { "epoch": 71.06896551724138, "grad_norm": 0.83984375, "learning_rate": 1.8291200438998273e-05, "loss": 0.0086, "step": 2061 }, { "epoch": 71.10344827586206, "grad_norm": 3.125, "learning_rate": 1.8283465443195837e-05, "loss": 0.0149, "step": 2062 }, { "epoch": 71.13793103448276, "grad_norm": 0.84765625, "learning_rate": 1.827572762860261e-05, "loss": 0.025, "step": 2063 }, { "epoch": 71.17241379310344, "grad_norm": 1.625, "learning_rate": 1.8267986998989896e-05, "loss": 0.0196, "step": 2064 }, { "epoch": 71.20689655172414, "grad_norm": 1.5703125, "learning_rate": 1.8260243558130388e-05, "loss": 0.0113, "step": 2065 }, { "epoch": 71.24137931034483, "grad_norm": 2.03125, "learning_rate": 1.8252497309798147e-05, "loss": 0.0136, "step": 2066 }, { "epoch": 71.27586206896552, "grad_norm": 2.171875, "learning_rate": 1.8244748257768602e-05, "loss": 0.0084, "step": 2067 }, { "epoch": 71.3103448275862, "grad_norm": 0.28515625, "learning_rate": 1.823699640581854e-05, "loss": 0.014, "step": 2068 }, { "epoch": 71.34482758620689, "grad_norm": 0.6640625, "learning_rate": 1.8229241757726124e-05, "loss": 0.0352, "step": 2069 }, { "epoch": 71.37931034482759, "grad_norm": 2.46875, "learning_rate": 1.8221484317270876e-05, "loss": 0.0229, "step": 2070 }, { "epoch": 71.41379310344827, "grad_norm": 0.2001953125, "learning_rate": 1.821372408823368e-05, "loss": 0.0022, "step": 2071 }, { "epoch": 71.44827586206897, "grad_norm": 1.265625, "learning_rate": 1.8205961074396772e-05, "loss": 0.0047, "step": 2072 }, { "epoch": 71.48275862068965, "grad_norm": 1.578125, "learning_rate": 1.8198195279543754e-05, "loss": 0.0099, "step": 2073 }, { "epoch": 71.51724137931035, "grad_norm": 1.7265625, "learning_rate": 1.8190426707459572e-05, "loss": 0.0086, "step": 2074 }, { "epoch": 71.55172413793103, "grad_norm": 0.98828125, "learning_rate": 1.818265536193055e-05, "loss": 0.0043, "step": 2075 }, { "epoch": 71.58620689655173, "grad_norm": 2.46875, "learning_rate": 1.8174881246744335e-05, "loss": 0.0101, "step": 2076 }, { "epoch": 71.62068965517241, "grad_norm": 3.109375, "learning_rate": 1.816710436568994e-05, "loss": 0.0182, "step": 2077 }, { "epoch": 71.65517241379311, "grad_norm": 1.1640625, "learning_rate": 1.8159324722557725e-05, "loss": 0.0157, "step": 2078 }, { "epoch": 71.6896551724138, "grad_norm": 1.9921875, "learning_rate": 1.8151542321139385e-05, "loss": 0.039, "step": 2079 }, { "epoch": 71.72413793103448, "grad_norm": 4.71875, "learning_rate": 1.814375716522798e-05, "loss": 0.0234, "step": 2080 }, { "epoch": 71.75862068965517, "grad_norm": 0.2275390625, "learning_rate": 1.8135969258617884e-05, "loss": 0.0049, "step": 2081 }, { "epoch": 71.79310344827586, "grad_norm": 0.73828125, "learning_rate": 1.8128178605104844e-05, "loss": 0.0107, "step": 2082 }, { "epoch": 71.82758620689656, "grad_norm": 1.5078125, "learning_rate": 1.8120385208485924e-05, "loss": 0.0093, "step": 2083 }, { "epoch": 71.86206896551724, "grad_norm": 2.234375, "learning_rate": 1.811258907255953e-05, "loss": 0.0526, "step": 2084 }, { "epoch": 71.89655172413794, "grad_norm": 3.296875, "learning_rate": 1.8104790201125408e-05, "loss": 0.0166, "step": 2085 }, { "epoch": 71.93103448275862, "grad_norm": 4.9375, "learning_rate": 1.809698859798462e-05, "loss": 0.0256, "step": 2086 }, { "epoch": 71.96551724137932, "grad_norm": 6.75, "learning_rate": 1.808918426693959e-05, "loss": 0.0487, "step": 2087 }, { "epoch": 72.0, "grad_norm": 1.359375, "learning_rate": 1.8081377211794042e-05, "loss": 0.0126, "step": 2088 }, { "epoch": 72.03448275862068, "grad_norm": 0.44140625, "learning_rate": 1.8073567436353054e-05, "loss": 0.0783, "step": 2089 }, { "epoch": 72.06896551724138, "grad_norm": 0.35546875, "learning_rate": 1.8065754944422996e-05, "loss": 0.0067, "step": 2090 }, { "epoch": 72.10344827586206, "grad_norm": 1.546875, "learning_rate": 1.8057939739811595e-05, "loss": 0.0333, "step": 2091 }, { "epoch": 72.13793103448276, "grad_norm": 1.578125, "learning_rate": 1.8050121826327883e-05, "loss": 0.0632, "step": 2092 }, { "epoch": 72.17241379310344, "grad_norm": 0.408203125, "learning_rate": 1.8042301207782214e-05, "loss": 0.0078, "step": 2093 }, { "epoch": 72.20689655172414, "grad_norm": 0.326171875, "learning_rate": 1.8034477887986274e-05, "loss": 0.0024, "step": 2094 }, { "epoch": 72.24137931034483, "grad_norm": 1.0078125, "learning_rate": 1.8026651870753035e-05, "loss": 0.0235, "step": 2095 }, { "epoch": 72.27586206896552, "grad_norm": 0.212890625, "learning_rate": 1.8018823159896816e-05, "loss": 0.0078, "step": 2096 }, { "epoch": 72.3103448275862, "grad_norm": 3.703125, "learning_rate": 1.8010991759233235e-05, "loss": 0.0404, "step": 2097 }, { "epoch": 72.34482758620689, "grad_norm": 7.28125, "learning_rate": 1.8003157672579217e-05, "loss": 0.0549, "step": 2098 }, { "epoch": 72.37931034482759, "grad_norm": 4.78125, "learning_rate": 1.7995320903753e-05, "loss": 0.0253, "step": 2099 }, { "epoch": 72.41379310344827, "grad_norm": 2.125, "learning_rate": 1.7987481456574135e-05, "loss": 0.0194, "step": 2100 }, { "epoch": 72.44827586206897, "grad_norm": 1.78125, "learning_rate": 1.7979639334863467e-05, "loss": 0.011, "step": 2101 }, { "epoch": 72.48275862068965, "grad_norm": 1.96875, "learning_rate": 1.797179454244316e-05, "loss": 0.02, "step": 2102 }, { "epoch": 72.51724137931035, "grad_norm": 0.470703125, "learning_rate": 1.796394708313666e-05, "loss": 0.0248, "step": 2103 }, { "epoch": 72.55172413793103, "grad_norm": 0.734375, "learning_rate": 1.7956096960768738e-05, "loss": 0.0079, "step": 2104 }, { "epoch": 72.58620689655173, "grad_norm": 2.578125, "learning_rate": 1.7948244179165428e-05, "loss": 0.0258, "step": 2105 }, { "epoch": 72.62068965517241, "grad_norm": 1.984375, "learning_rate": 1.7940388742154096e-05, "loss": 0.0538, "step": 2106 }, { "epoch": 72.65517241379311, "grad_norm": 0.3671875, "learning_rate": 1.793253065356338e-05, "loss": 0.0201, "step": 2107 }, { "epoch": 72.6896551724138, "grad_norm": 4.03125, "learning_rate": 1.792466991722322e-05, "loss": 0.0107, "step": 2108 }, { "epoch": 72.72413793103448, "grad_norm": 0.90625, "learning_rate": 1.7916806536964842e-05, "loss": 0.0078, "step": 2109 }, { "epoch": 72.75862068965517, "grad_norm": 0.2236328125, "learning_rate": 1.790894051662076e-05, "loss": 0.0057, "step": 2110 }, { "epoch": 72.79310344827586, "grad_norm": 0.5859375, "learning_rate": 1.790107186002478e-05, "loss": 0.0115, "step": 2111 }, { "epoch": 72.82758620689656, "grad_norm": 6.0, "learning_rate": 1.7893200571011988e-05, "loss": 0.0471, "step": 2112 }, { "epoch": 72.86206896551724, "grad_norm": 4.0, "learning_rate": 1.7885326653418757e-05, "loss": 0.0258, "step": 2113 }, { "epoch": 72.89655172413794, "grad_norm": 7.25, "learning_rate": 1.7877450111082733e-05, "loss": 0.0232, "step": 2114 }, { "epoch": 72.93103448275862, "grad_norm": 1.1953125, "learning_rate": 1.7869570947842855e-05, "loss": 0.015, "step": 2115 }, { "epoch": 72.96551724137932, "grad_norm": 2.546875, "learning_rate": 1.7861689167539324e-05, "loss": 0.0151, "step": 2116 }, { "epoch": 73.0, "grad_norm": 0.83984375, "learning_rate": 1.785380477401363e-05, "loss": 0.0154, "step": 2117 }, { "epoch": 73.03448275862068, "grad_norm": 4.03125, "learning_rate": 1.7845917771108527e-05, "loss": 0.0198, "step": 2118 }, { "epoch": 73.06896551724138, "grad_norm": 0.27734375, "learning_rate": 1.783802816266804e-05, "loss": 0.0142, "step": 2119 }, { "epoch": 73.10344827586206, "grad_norm": 0.3515625, "learning_rate": 1.7830135952537487e-05, "loss": 0.0365, "step": 2120 }, { "epoch": 73.13793103448276, "grad_norm": 3.8125, "learning_rate": 1.782224114456341e-05, "loss": 0.0103, "step": 2121 }, { "epoch": 73.17241379310344, "grad_norm": 14.5, "learning_rate": 1.781434374259366e-05, "loss": 0.0434, "step": 2122 }, { "epoch": 73.20689655172414, "grad_norm": 0.91015625, "learning_rate": 1.7806443750477324e-05, "loss": 0.0099, "step": 2123 }, { "epoch": 73.24137931034483, "grad_norm": 3.5, "learning_rate": 1.779854117206477e-05, "loss": 0.0373, "step": 2124 }, { "epoch": 73.27586206896552, "grad_norm": 0.40625, "learning_rate": 1.7790636011207622e-05, "loss": 0.007, "step": 2125 }, { "epoch": 73.3103448275862, "grad_norm": 3.53125, "learning_rate": 1.7782728271758746e-05, "loss": 0.0223, "step": 2126 }, { "epoch": 73.34482758620689, "grad_norm": 1.484375, "learning_rate": 1.7774817957572286e-05, "loss": 0.0134, "step": 2127 }, { "epoch": 73.37931034482759, "grad_norm": 0.294921875, "learning_rate": 1.7766905072503636e-05, "loss": 0.0035, "step": 2128 }, { "epoch": 73.41379310344827, "grad_norm": 5.46875, "learning_rate": 1.7758989620409435e-05, "loss": 0.0188, "step": 2129 }, { "epoch": 73.44827586206897, "grad_norm": 5.59375, "learning_rate": 1.7751071605147582e-05, "loss": 0.0529, "step": 2130 }, { "epoch": 73.48275862068965, "grad_norm": 2.109375, "learning_rate": 1.7743151030577214e-05, "loss": 0.0348, "step": 2131 }, { "epoch": 73.51724137931035, "grad_norm": 1.4453125, "learning_rate": 1.773522790055873e-05, "loss": 0.0093, "step": 2132 }, { "epoch": 73.55172413793103, "grad_norm": 0.421875, "learning_rate": 1.7727302218953766e-05, "loss": 0.0156, "step": 2133 }, { "epoch": 73.58620689655173, "grad_norm": 2.921875, "learning_rate": 1.771937398962521e-05, "loss": 0.0201, "step": 2134 }, { "epoch": 73.62068965517241, "grad_norm": 2.359375, "learning_rate": 1.771144321643717e-05, "loss": 0.0206, "step": 2135 }, { "epoch": 73.65517241379311, "grad_norm": 0.19140625, "learning_rate": 1.7703509903255015e-05, "loss": 0.0148, "step": 2136 }, { "epoch": 73.6896551724138, "grad_norm": 2.71875, "learning_rate": 1.769557405394535e-05, "loss": 0.0332, "step": 2137 }, { "epoch": 73.72413793103448, "grad_norm": 7.71875, "learning_rate": 1.7687635672376006e-05, "loss": 0.0348, "step": 2138 }, { "epoch": 73.75862068965517, "grad_norm": 1.59375, "learning_rate": 1.7679694762416056e-05, "loss": 0.006, "step": 2139 }, { "epoch": 73.79310344827586, "grad_norm": 2.546875, "learning_rate": 1.7671751327935797e-05, "loss": 0.0293, "step": 2140 }, { "epoch": 73.82758620689656, "grad_norm": 5.875, "learning_rate": 1.7663805372806775e-05, "loss": 0.0429, "step": 2141 }, { "epoch": 73.86206896551724, "grad_norm": 3.296875, "learning_rate": 1.765585690090174e-05, "loss": 0.0355, "step": 2142 }, { "epoch": 73.89655172413794, "grad_norm": 1.0078125, "learning_rate": 1.7647905916094685e-05, "loss": 0.0054, "step": 2143 }, { "epoch": 73.93103448275862, "grad_norm": 3.953125, "learning_rate": 1.763995242226083e-05, "loss": 0.0433, "step": 2144 }, { "epoch": 73.96551724137932, "grad_norm": 0.32421875, "learning_rate": 1.76319964232766e-05, "loss": 0.0235, "step": 2145 }, { "epoch": 74.0, "grad_norm": 2.84375, "learning_rate": 1.762403792301966e-05, "loss": 0.032, "step": 2146 }, { "epoch": 74.03448275862068, "grad_norm": 2.8125, "learning_rate": 1.7616076925368885e-05, "loss": 0.0464, "step": 2147 }, { "epoch": 74.06896551724138, "grad_norm": 1.0234375, "learning_rate": 1.7608113434204366e-05, "loss": 0.0131, "step": 2148 }, { "epoch": 74.10344827586206, "grad_norm": 3.78125, "learning_rate": 1.7600147453407412e-05, "loss": 0.0499, "step": 2149 }, { "epoch": 74.13793103448276, "grad_norm": 0.6015625, "learning_rate": 1.7592178986860544e-05, "loss": 0.0152, "step": 2150 }, { "epoch": 74.17241379310344, "grad_norm": 2.671875, "learning_rate": 1.7584208038447505e-05, "loss": 0.0183, "step": 2151 }, { "epoch": 74.20689655172414, "grad_norm": 0.875, "learning_rate": 1.7576234612053227e-05, "loss": 0.0042, "step": 2152 }, { "epoch": 74.24137931034483, "grad_norm": 0.359375, "learning_rate": 1.756825871156387e-05, "loss": 0.0177, "step": 2153 }, { "epoch": 74.27586206896552, "grad_norm": 2.671875, "learning_rate": 1.756028034086678e-05, "loss": 0.0159, "step": 2154 }, { "epoch": 74.3103448275862, "grad_norm": 2.984375, "learning_rate": 1.7552299503850527e-05, "loss": 0.064, "step": 2155 }, { "epoch": 74.34482758620689, "grad_norm": 0.392578125, "learning_rate": 1.7544316204404877e-05, "loss": 0.0032, "step": 2156 }, { "epoch": 74.37931034482759, "grad_norm": 0.78125, "learning_rate": 1.7536330446420785e-05, "loss": 0.0116, "step": 2157 }, { "epoch": 74.41379310344827, "grad_norm": 2.078125, "learning_rate": 1.7528342233790415e-05, "loss": 0.0429, "step": 2158 }, { "epoch": 74.44827586206897, "grad_norm": 1.2890625, "learning_rate": 1.7520351570407127e-05, "loss": 0.0184, "step": 2159 }, { "epoch": 74.48275862068965, "grad_norm": 1.8125, "learning_rate": 1.751235846016547e-05, "loss": 0.0199, "step": 2160 }, { "epoch": 74.51724137931035, "grad_norm": 2.0625, "learning_rate": 1.7504362906961193e-05, "loss": 0.0174, "step": 2161 }, { "epoch": 74.55172413793103, "grad_norm": 3.171875, "learning_rate": 1.749636491469122e-05, "loss": 0.0651, "step": 2162 }, { "epoch": 74.58620689655173, "grad_norm": 0.51171875, "learning_rate": 1.7488364487253692e-05, "loss": 0.0065, "step": 2163 }, { "epoch": 74.62068965517241, "grad_norm": 4.21875, "learning_rate": 1.748036162854791e-05, "loss": 0.0371, "step": 2164 }, { "epoch": 74.65517241379311, "grad_norm": 1.8515625, "learning_rate": 1.7472356342474376e-05, "loss": 0.0149, "step": 2165 }, { "epoch": 74.6896551724138, "grad_norm": 4.75, "learning_rate": 1.7464348632934758e-05, "loss": 0.0167, "step": 2166 }, { "epoch": 74.72413793103448, "grad_norm": 2.078125, "learning_rate": 1.745633850383193e-05, "loss": 0.0175, "step": 2167 }, { "epoch": 74.75862068965517, "grad_norm": 2.28125, "learning_rate": 1.7448325959069923e-05, "loss": 0.0271, "step": 2168 }, { "epoch": 74.79310344827586, "grad_norm": 0.5, "learning_rate": 1.7440311002553957e-05, "loss": 0.0104, "step": 2169 }, { "epoch": 74.82758620689656, "grad_norm": 0.1591796875, "learning_rate": 1.7432293638190425e-05, "loss": 0.0207, "step": 2170 }, { "epoch": 74.86206896551724, "grad_norm": 1.0625, "learning_rate": 1.742427386988689e-05, "loss": 0.0062, "step": 2171 }, { "epoch": 74.89655172413794, "grad_norm": 0.875, "learning_rate": 1.741625170155209e-05, "loss": 0.0063, "step": 2172 }, { "epoch": 74.93103448275862, "grad_norm": 1.5078125, "learning_rate": 1.740822713709594e-05, "loss": 0.0097, "step": 2173 }, { "epoch": 74.96551724137932, "grad_norm": 1.734375, "learning_rate": 1.7400200180429503e-05, "loss": 0.0108, "step": 2174 }, { "epoch": 75.0, "grad_norm": 0.73828125, "learning_rate": 1.7392170835465032e-05, "loss": 0.0285, "step": 2175 }, { "epoch": 75.03448275862068, "grad_norm": 1.3125, "learning_rate": 1.738413910611592e-05, "loss": 0.0353, "step": 2176 }, { "epoch": 75.06896551724138, "grad_norm": 2.421875, "learning_rate": 1.7376104996296746e-05, "loss": 0.0217, "step": 2177 }, { "epoch": 75.10344827586206, "grad_norm": 0.458984375, "learning_rate": 1.7368068509923233e-05, "loss": 0.009, "step": 2178 }, { "epoch": 75.13793103448276, "grad_norm": 0.68359375, "learning_rate": 1.7360029650912263e-05, "loss": 0.0063, "step": 2179 }, { "epoch": 75.17241379310344, "grad_norm": 1.15625, "learning_rate": 1.7351988423181885e-05, "loss": 0.0163, "step": 2180 }, { "epoch": 75.20689655172414, "grad_norm": 0.8203125, "learning_rate": 1.7343944830651288e-05, "loss": 0.0087, "step": 2181 }, { "epoch": 75.24137931034483, "grad_norm": 0.75390625, "learning_rate": 1.733589887724083e-05, "loss": 0.0073, "step": 2182 }, { "epoch": 75.27586206896552, "grad_norm": 2.203125, "learning_rate": 1.7327850566872e-05, "loss": 0.0248, "step": 2183 }, { "epoch": 75.3103448275862, "grad_norm": 4.09375, "learning_rate": 1.731979990346746e-05, "loss": 0.0106, "step": 2184 }, { "epoch": 75.34482758620689, "grad_norm": 1.828125, "learning_rate": 1.7311746890950994e-05, "loss": 0.0295, "step": 2185 }, { "epoch": 75.37931034482759, "grad_norm": 0.54296875, "learning_rate": 1.7303691533247546e-05, "loss": 0.046, "step": 2186 }, { "epoch": 75.41379310344827, "grad_norm": 0.2001953125, "learning_rate": 1.7295633834283205e-05, "loss": 0.0038, "step": 2187 }, { "epoch": 75.44827586206897, "grad_norm": 3.6875, "learning_rate": 1.728757379798519e-05, "loss": 0.044, "step": 2188 }, { "epoch": 75.48275862068965, "grad_norm": 1.15625, "learning_rate": 1.7279511428281864e-05, "loss": 0.0187, "step": 2189 }, { "epoch": 75.51724137931035, "grad_norm": 0.52734375, "learning_rate": 1.727144672910273e-05, "loss": 0.0037, "step": 2190 }, { "epoch": 75.55172413793103, "grad_norm": 0.28515625, "learning_rate": 1.726337970437843e-05, "loss": 0.0026, "step": 2191 }, { "epoch": 75.58620689655173, "grad_norm": 3.640625, "learning_rate": 1.7255310358040723e-05, "loss": 0.0286, "step": 2192 }, { "epoch": 75.62068965517241, "grad_norm": 0.90625, "learning_rate": 1.7247238694022517e-05, "loss": 0.0036, "step": 2193 }, { "epoch": 75.65517241379311, "grad_norm": 0.40234375, "learning_rate": 1.7239164716257837e-05, "loss": 0.0128, "step": 2194 }, { "epoch": 75.6896551724138, "grad_norm": 3.1875, "learning_rate": 1.7231088428681852e-05, "loss": 0.0226, "step": 2195 }, { "epoch": 75.72413793103448, "grad_norm": 2.453125, "learning_rate": 1.722300983523084e-05, "loss": 0.0304, "step": 2196 }, { "epoch": 75.75862068965517, "grad_norm": 3.9375, "learning_rate": 1.721492893984221e-05, "loss": 0.0146, "step": 2197 }, { "epoch": 75.79310344827586, "grad_norm": 0.26171875, "learning_rate": 1.7206845746454492e-05, "loss": 0.0069, "step": 2198 }, { "epoch": 75.82758620689656, "grad_norm": 0.55859375, "learning_rate": 1.7198760259007336e-05, "loss": 0.0434, "step": 2199 }, { "epoch": 75.86206896551724, "grad_norm": 3.5, "learning_rate": 1.719067248144151e-05, "loss": 0.0229, "step": 2200 }, { "epoch": 75.89655172413794, "grad_norm": 0.419921875, "learning_rate": 1.7182582417698903e-05, "loss": 0.0165, "step": 2201 }, { "epoch": 75.93103448275862, "grad_norm": 1.2734375, "learning_rate": 1.7174490071722504e-05, "loss": 0.0502, "step": 2202 }, { "epoch": 75.96551724137932, "grad_norm": 0.22265625, "learning_rate": 1.716639544745644e-05, "loss": 0.0239, "step": 2203 }, { "epoch": 76.0, "grad_norm": 4.84375, "learning_rate": 1.7158298548845914e-05, "loss": 0.0276, "step": 2204 }, { "epoch": 76.03448275862068, "grad_norm": 2.046875, "learning_rate": 1.715019937983727e-05, "loss": 0.0109, "step": 2205 }, { "epoch": 76.06896551724138, "grad_norm": 1.53125, "learning_rate": 1.714209794437794e-05, "loss": 0.036, "step": 2206 }, { "epoch": 76.10344827586206, "grad_norm": 7.8125, "learning_rate": 1.7133994246416456e-05, "loss": 0.0301, "step": 2207 }, { "epoch": 76.13793103448276, "grad_norm": 0.275390625, "learning_rate": 1.712588828990248e-05, "loss": 0.0083, "step": 2208 }, { "epoch": 76.17241379310344, "grad_norm": 1.75, "learning_rate": 1.7117780078786745e-05, "loss": 0.0155, "step": 2209 }, { "epoch": 76.20689655172414, "grad_norm": 2.953125, "learning_rate": 1.7109669617021095e-05, "loss": 0.011, "step": 2210 }, { "epoch": 76.24137931034483, "grad_norm": 1.2421875, "learning_rate": 1.7101556908558478e-05, "loss": 0.0192, "step": 2211 }, { "epoch": 76.27586206896552, "grad_norm": 7.59375, "learning_rate": 1.7093441957352917e-05, "loss": 0.0452, "step": 2212 }, { "epoch": 76.3103448275862, "grad_norm": 1.796875, "learning_rate": 1.7085324767359554e-05, "loss": 0.0182, "step": 2213 }, { "epoch": 76.34482758620689, "grad_norm": 3.421875, "learning_rate": 1.7077205342534598e-05, "loss": 0.0128, "step": 2214 }, { "epoch": 76.37931034482759, "grad_norm": 4.46875, "learning_rate": 1.7069083686835363e-05, "loss": 0.0226, "step": 2215 }, { "epoch": 76.41379310344827, "grad_norm": 2.1875, "learning_rate": 1.7060959804220246e-05, "loss": 0.0109, "step": 2216 }, { "epoch": 76.44827586206897, "grad_norm": 3.046875, "learning_rate": 1.705283369864873e-05, "loss": 0.0184, "step": 2217 }, { "epoch": 76.48275862068965, "grad_norm": 7.0, "learning_rate": 1.7044705374081374e-05, "loss": 0.0308, "step": 2218 }, { "epoch": 76.51724137931035, "grad_norm": 0.58203125, "learning_rate": 1.703657483447983e-05, "loss": 0.0045, "step": 2219 }, { "epoch": 76.55172413793103, "grad_norm": 0.173828125, "learning_rate": 1.702844208380682e-05, "loss": 0.0255, "step": 2220 }, { "epoch": 76.58620689655173, "grad_norm": 2.9375, "learning_rate": 1.7020307126026152e-05, "loss": 0.0188, "step": 2221 }, { "epoch": 76.62068965517241, "grad_norm": 8.6875, "learning_rate": 1.7012169965102708e-05, "loss": 0.078, "step": 2222 }, { "epoch": 76.65517241379311, "grad_norm": 6.4375, "learning_rate": 1.7004030605002434e-05, "loss": 0.0592, "step": 2223 }, { "epoch": 76.6896551724138, "grad_norm": 1.9140625, "learning_rate": 1.699588904969236e-05, "loss": 0.0107, "step": 2224 }, { "epoch": 76.72413793103448, "grad_norm": 4.3125, "learning_rate": 1.698774530314058e-05, "loss": 0.05, "step": 2225 }, { "epoch": 76.75862068965517, "grad_norm": 0.4453125, "learning_rate": 1.697959936931625e-05, "loss": 0.0064, "step": 2226 }, { "epoch": 76.79310344827586, "grad_norm": 3.4375, "learning_rate": 1.6971451252189617e-05, "loss": 0.022, "step": 2227 }, { "epoch": 76.82758620689656, "grad_norm": 1.953125, "learning_rate": 1.6963300955731958e-05, "loss": 0.0191, "step": 2228 }, { "epoch": 76.86206896551724, "grad_norm": 0.68359375, "learning_rate": 1.6955148483915633e-05, "loss": 0.0136, "step": 2229 }, { "epoch": 76.89655172413794, "grad_norm": 8.9375, "learning_rate": 1.694699384071406e-05, "loss": 0.0386, "step": 2230 }, { "epoch": 76.93103448275862, "grad_norm": 1.78125, "learning_rate": 1.6938837030101714e-05, "loss": 0.0235, "step": 2231 }, { "epoch": 76.96551724137932, "grad_norm": 3.0, "learning_rate": 1.6930678056054123e-05, "loss": 0.0317, "step": 2232 }, { "epoch": 77.0, "grad_norm": 1.5234375, "learning_rate": 1.692251692254787e-05, "loss": 0.0343, "step": 2233 }, { "epoch": 77.03448275862068, "grad_norm": 1.7421875, "learning_rate": 1.6914353633560594e-05, "loss": 0.0166, "step": 2234 }, { "epoch": 77.06896551724138, "grad_norm": 1.921875, "learning_rate": 1.690618819307098e-05, "loss": 0.0171, "step": 2235 }, { "epoch": 77.10344827586206, "grad_norm": 2.09375, "learning_rate": 1.6898020605058773e-05, "loss": 0.0253, "step": 2236 }, { "epoch": 77.13793103448276, "grad_norm": 1.6953125, "learning_rate": 1.6889850873504752e-05, "loss": 0.022, "step": 2237 }, { "epoch": 77.17241379310344, "grad_norm": 1.359375, "learning_rate": 1.6881679002390738e-05, "loss": 0.0256, "step": 2238 }, { "epoch": 77.20689655172414, "grad_norm": 4.5, "learning_rate": 1.6873504995699614e-05, "loss": 0.0313, "step": 2239 }, { "epoch": 77.24137931034483, "grad_norm": 1.2890625, "learning_rate": 1.686532885741528e-05, "loss": 0.0169, "step": 2240 }, { "epoch": 77.27586206896552, "grad_norm": 0.6875, "learning_rate": 1.6857150591522692e-05, "loss": 0.0097, "step": 2241 }, { "epoch": 77.3103448275862, "grad_norm": 2.96875, "learning_rate": 1.6848970202007836e-05, "loss": 0.0195, "step": 2242 }, { "epoch": 77.34482758620689, "grad_norm": 2.359375, "learning_rate": 1.684078769285774e-05, "loss": 0.0149, "step": 2243 }, { "epoch": 77.37931034482759, "grad_norm": 2.75, "learning_rate": 1.6832603068060453e-05, "loss": 0.0488, "step": 2244 }, { "epoch": 77.41379310344827, "grad_norm": 1.7109375, "learning_rate": 1.6824416331605062e-05, "loss": 0.0099, "step": 2245 }, { "epoch": 77.44827586206897, "grad_norm": 0.28515625, "learning_rate": 1.6816227487481685e-05, "loss": 0.0027, "step": 2246 }, { "epoch": 77.48275862068965, "grad_norm": 1.6328125, "learning_rate": 1.6808036539681463e-05, "loss": 0.0323, "step": 2247 }, { "epoch": 77.51724137931035, "grad_norm": 1.7109375, "learning_rate": 1.6799843492196567e-05, "loss": 0.042, "step": 2248 }, { "epoch": 77.55172413793103, "grad_norm": 1.7421875, "learning_rate": 1.6791648349020192e-05, "loss": 0.0122, "step": 2249 }, { "epoch": 77.58620689655173, "grad_norm": 4.71875, "learning_rate": 1.6783451114146545e-05, "loss": 0.0409, "step": 2250 }, { "epoch": 77.62068965517241, "grad_norm": 1.953125, "learning_rate": 1.677525179157086e-05, "loss": 0.0146, "step": 2251 }, { "epoch": 77.65517241379311, "grad_norm": 0.6875, "learning_rate": 1.6767050385289386e-05, "loss": 0.0137, "step": 2252 }, { "epoch": 77.6896551724138, "grad_norm": 0.61328125, "learning_rate": 1.67588468992994e-05, "loss": 0.0096, "step": 2253 }, { "epoch": 77.72413793103448, "grad_norm": 1.90625, "learning_rate": 1.675064133759916e-05, "loss": 0.021, "step": 2254 }, { "epoch": 77.75862068965517, "grad_norm": 0.52734375, "learning_rate": 1.6742433704187977e-05, "loss": 0.0176, "step": 2255 }, { "epoch": 77.79310344827586, "grad_norm": 1.6484375, "learning_rate": 1.6734224003066146e-05, "loss": 0.0107, "step": 2256 }, { "epoch": 77.82758620689656, "grad_norm": 2.9375, "learning_rate": 1.6726012238234967e-05, "loss": 0.0234, "step": 2257 }, { "epoch": 77.86206896551724, "grad_norm": 0.53515625, "learning_rate": 1.6717798413696765e-05, "loss": 0.0135, "step": 2258 }, { "epoch": 77.89655172413794, "grad_norm": 1.765625, "learning_rate": 1.6709582533454853e-05, "loss": 0.0079, "step": 2259 }, { "epoch": 77.93103448275862, "grad_norm": 1.9296875, "learning_rate": 1.6701364601513557e-05, "loss": 0.0134, "step": 2260 }, { "epoch": 77.96551724137932, "grad_norm": 1.3125, "learning_rate": 1.669314462187819e-05, "loss": 0.007, "step": 2261 }, { "epoch": 78.0, "grad_norm": 7.0, "learning_rate": 1.6684922598555074e-05, "loss": 0.039, "step": 2262 }, { "epoch": 78.03448275862068, "grad_norm": 1.109375, "learning_rate": 1.6676698535551523e-05, "loss": 0.0075, "step": 2263 }, { "epoch": 78.06896551724138, "grad_norm": 1.4765625, "learning_rate": 1.666847243687584e-05, "loss": 0.0194, "step": 2264 }, { "epoch": 78.10344827586206, "grad_norm": 6.5625, "learning_rate": 1.6660244306537335e-05, "loss": 0.035, "step": 2265 }, { "epoch": 78.13793103448276, "grad_norm": 0.353515625, "learning_rate": 1.665201414854629e-05, "loss": 0.0046, "step": 2266 }, { "epoch": 78.17241379310344, "grad_norm": 1.859375, "learning_rate": 1.6643781966913997e-05, "loss": 0.0275, "step": 2267 }, { "epoch": 78.20689655172414, "grad_norm": 3.15625, "learning_rate": 1.6635547765652704e-05, "loss": 0.0184, "step": 2268 }, { "epoch": 78.24137931034483, "grad_norm": 0.765625, "learning_rate": 1.662731154877567e-05, "loss": 0.0134, "step": 2269 }, { "epoch": 78.27586206896552, "grad_norm": 3.234375, "learning_rate": 1.6619073320297128e-05, "loss": 0.0257, "step": 2270 }, { "epoch": 78.3103448275862, "grad_norm": 2.9375, "learning_rate": 1.661083308423229e-05, "loss": 0.0159, "step": 2271 }, { "epoch": 78.34482758620689, "grad_norm": 1.9296875, "learning_rate": 1.660259084459735e-05, "loss": 0.0272, "step": 2272 }, { "epoch": 78.37931034482759, "grad_norm": 0.828125, "learning_rate": 1.659434660540946e-05, "loss": 0.0334, "step": 2273 }, { "epoch": 78.41379310344827, "grad_norm": 1.6796875, "learning_rate": 1.6586100370686782e-05, "loss": 0.0196, "step": 2274 }, { "epoch": 78.44827586206897, "grad_norm": 3.109375, "learning_rate": 1.657785214444842e-05, "loss": 0.0204, "step": 2275 }, { "epoch": 78.48275862068965, "grad_norm": 4.21875, "learning_rate": 1.656960193071446e-05, "loss": 0.0236, "step": 2276 }, { "epoch": 78.51724137931035, "grad_norm": 0.1767578125, "learning_rate": 1.6561349733505956e-05, "loss": 0.0058, "step": 2277 }, { "epoch": 78.55172413793103, "grad_norm": 3.34375, "learning_rate": 1.655309555684493e-05, "loss": 0.0483, "step": 2278 }, { "epoch": 78.58620689655173, "grad_norm": 5.6875, "learning_rate": 1.654483940475437e-05, "loss": 0.0301, "step": 2279 }, { "epoch": 78.62068965517241, "grad_norm": 1.171875, "learning_rate": 1.653658128125822e-05, "loss": 0.019, "step": 2280 }, { "epoch": 78.65517241379311, "grad_norm": 0.7109375, "learning_rate": 1.652832119038139e-05, "loss": 0.008, "step": 2281 }, { "epoch": 78.6896551724138, "grad_norm": 0.5703125, "learning_rate": 1.6520059136149747e-05, "loss": 0.0044, "step": 2282 }, { "epoch": 78.72413793103448, "grad_norm": 5.34375, "learning_rate": 1.651179512259012e-05, "loss": 0.0336, "step": 2283 }, { "epoch": 78.75862068965517, "grad_norm": 9.3125, "learning_rate": 1.6503529153730285e-05, "loss": 0.0479, "step": 2284 }, { "epoch": 78.79310344827586, "grad_norm": 0.2392578125, "learning_rate": 1.6495261233598976e-05, "loss": 0.0216, "step": 2285 }, { "epoch": 78.82758620689656, "grad_norm": 0.61328125, "learning_rate": 1.6486991366225878e-05, "loss": 0.004, "step": 2286 }, { "epoch": 78.86206896551724, "grad_norm": 0.189453125, "learning_rate": 1.647871955564162e-05, "loss": 0.0062, "step": 2287 }, { "epoch": 78.89655172413794, "grad_norm": 1.34375, "learning_rate": 1.6470445805877782e-05, "loss": 0.0107, "step": 2288 }, { "epoch": 78.93103448275862, "grad_norm": 9.25, "learning_rate": 1.64621701209669e-05, "loss": 0.0764, "step": 2289 }, { "epoch": 78.96551724137932, "grad_norm": 4.21875, "learning_rate": 1.6453892504942425e-05, "loss": 0.0342, "step": 2290 }, { "epoch": 79.0, "grad_norm": 0.92578125, "learning_rate": 1.6445612961838783e-05, "loss": 0.0318, "step": 2291 }, { "epoch": 79.03448275862068, "grad_norm": 7.46875, "learning_rate": 1.643733149569131e-05, "loss": 0.0176, "step": 2292 }, { "epoch": 79.06896551724138, "grad_norm": 0.546875, "learning_rate": 1.64290481105363e-05, "loss": 0.0078, "step": 2293 }, { "epoch": 79.10344827586206, "grad_norm": 0.9609375, "learning_rate": 1.6420762810410973e-05, "loss": 0.0107, "step": 2294 }, { "epoch": 79.13793103448276, "grad_norm": 2.921875, "learning_rate": 1.6412475599353477e-05, "loss": 0.0262, "step": 2295 }, { "epoch": 79.17241379310344, "grad_norm": 2.65625, "learning_rate": 1.6404186481402908e-05, "loss": 0.0148, "step": 2296 }, { "epoch": 79.20689655172414, "grad_norm": 2.046875, "learning_rate": 1.6395895460599274e-05, "loss": 0.0317, "step": 2297 }, { "epoch": 79.24137931034483, "grad_norm": 0.96484375, "learning_rate": 1.6387602540983525e-05, "loss": 0.0106, "step": 2298 }, { "epoch": 79.27586206896552, "grad_norm": 2.90625, "learning_rate": 1.6379307726597517e-05, "loss": 0.0113, "step": 2299 }, { "epoch": 79.3103448275862, "grad_norm": 3.1875, "learning_rate": 1.6371011021484052e-05, "loss": 0.0177, "step": 2300 }, { "epoch": 79.34482758620689, "grad_norm": 0.447265625, "learning_rate": 1.6362712429686846e-05, "loss": 0.0098, "step": 2301 }, { "epoch": 79.37931034482759, "grad_norm": 1.578125, "learning_rate": 1.635441195525052e-05, "loss": 0.0095, "step": 2302 }, { "epoch": 79.41379310344827, "grad_norm": 4.09375, "learning_rate": 1.6346109602220636e-05, "loss": 0.0158, "step": 2303 }, { "epoch": 79.44827586206897, "grad_norm": 2.796875, "learning_rate": 1.633780537464365e-05, "loss": 0.0136, "step": 2304 }, { "epoch": 79.48275862068965, "grad_norm": 2.6875, "learning_rate": 1.6329499276566948e-05, "loss": 0.0296, "step": 2305 }, { "epoch": 79.51724137931035, "grad_norm": 2.765625, "learning_rate": 1.632119131203882e-05, "loss": 0.0091, "step": 2306 }, { "epoch": 79.55172413793103, "grad_norm": 0.8828125, "learning_rate": 1.6312881485108473e-05, "loss": 0.0043, "step": 2307 }, { "epoch": 79.58620689655173, "grad_norm": 0.287109375, "learning_rate": 1.6304569799826004e-05, "loss": 0.014, "step": 2308 }, { "epoch": 79.62068965517241, "grad_norm": 0.373046875, "learning_rate": 1.6296256260242428e-05, "loss": 0.0174, "step": 2309 }, { "epoch": 79.65517241379311, "grad_norm": 2.03125, "learning_rate": 1.6287940870409677e-05, "loss": 0.019, "step": 2310 }, { "epoch": 79.6896551724138, "grad_norm": 3.0625, "learning_rate": 1.6279623634380557e-05, "loss": 0.0187, "step": 2311 }, { "epoch": 79.72413793103448, "grad_norm": 1.546875, "learning_rate": 1.6271304556208795e-05, "loss": 0.0241, "step": 2312 }, { "epoch": 79.75862068965517, "grad_norm": 1.9765625, "learning_rate": 1.6262983639949008e-05, "loss": 0.0147, "step": 2313 }, { "epoch": 79.79310344827586, "grad_norm": 2.984375, "learning_rate": 1.62546608896567e-05, "loss": 0.0311, "step": 2314 }, { "epoch": 79.82758620689656, "grad_norm": 1.28125, "learning_rate": 1.62463363093883e-05, "loss": 0.0213, "step": 2315 }, { "epoch": 79.86206896551724, "grad_norm": 5.375, "learning_rate": 1.6238009903201085e-05, "loss": 0.0204, "step": 2316 }, { "epoch": 79.89655172413794, "grad_norm": 0.7109375, "learning_rate": 1.6229681675153255e-05, "loss": 0.0091, "step": 2317 }, { "epoch": 79.93103448275862, "grad_norm": 2.71875, "learning_rate": 1.6221351629303883e-05, "loss": 0.0103, "step": 2318 }, { "epoch": 79.96551724137932, "grad_norm": 1.6171875, "learning_rate": 1.6213019769712938e-05, "loss": 0.011, "step": 2319 }, { "epoch": 80.0, "grad_norm": 0.703125, "learning_rate": 1.6204686100441264e-05, "loss": 0.0071, "step": 2320 }, { "epoch": 80.03448275862068, "grad_norm": 2.703125, "learning_rate": 1.6196350625550585e-05, "loss": 0.0249, "step": 2321 }, { "epoch": 80.06896551724138, "grad_norm": 1.6171875, "learning_rate": 1.618801334910352e-05, "loss": 0.0122, "step": 2322 }, { "epoch": 80.10344827586206, "grad_norm": 1.5703125, "learning_rate": 1.6179674275163545e-05, "loss": 0.0296, "step": 2323 }, { "epoch": 80.13793103448276, "grad_norm": 4.15625, "learning_rate": 1.6171333407795036e-05, "loss": 0.0273, "step": 2324 }, { "epoch": 80.17241379310344, "grad_norm": 5.375, "learning_rate": 1.6162990751063217e-05, "loss": 0.0555, "step": 2325 }, { "epoch": 80.20689655172414, "grad_norm": 1.859375, "learning_rate": 1.615464630903421e-05, "loss": 0.0279, "step": 2326 }, { "epoch": 80.24137931034483, "grad_norm": 0.51953125, "learning_rate": 1.6146300085774986e-05, "loss": 0.0073, "step": 2327 }, { "epoch": 80.27586206896552, "grad_norm": 0.2060546875, "learning_rate": 1.6137952085353398e-05, "loss": 0.0053, "step": 2328 }, { "epoch": 80.3103448275862, "grad_norm": 3.203125, "learning_rate": 1.612960231183816e-05, "loss": 0.021, "step": 2329 }, { "epoch": 80.34482758620689, "grad_norm": 3.265625, "learning_rate": 1.6121250769298844e-05, "loss": 0.0081, "step": 2330 }, { "epoch": 80.37931034482759, "grad_norm": 1.953125, "learning_rate": 1.6112897461805895e-05, "loss": 0.012, "step": 2331 }, { "epoch": 80.41379310344827, "grad_norm": 0.75, "learning_rate": 1.6104542393430615e-05, "loss": 0.0067, "step": 2332 }, { "epoch": 80.44827586206897, "grad_norm": 1.109375, "learning_rate": 1.609618556824516e-05, "loss": 0.0263, "step": 2333 }, { "epoch": 80.48275862068965, "grad_norm": 1.5390625, "learning_rate": 1.6087826990322546e-05, "loss": 0.0063, "step": 2334 }, { "epoch": 80.51724137931035, "grad_norm": 2.046875, "learning_rate": 1.607946666373664e-05, "loss": 0.0137, "step": 2335 }, { "epoch": 80.55172413793103, "grad_norm": 0.859375, "learning_rate": 1.607110459256217e-05, "loss": 0.0159, "step": 2336 }, { "epoch": 80.58620689655173, "grad_norm": 0.330078125, "learning_rate": 1.60627407808747e-05, "loss": 0.0084, "step": 2337 }, { "epoch": 80.62068965517241, "grad_norm": 5.25, "learning_rate": 1.605437523275066e-05, "loss": 0.0399, "step": 2338 }, { "epoch": 80.65517241379311, "grad_norm": 1.7265625, "learning_rate": 1.6046007952267305e-05, "loss": 0.0111, "step": 2339 }, { "epoch": 80.6896551724138, "grad_norm": 0.3203125, "learning_rate": 1.6037638943502754e-05, "loss": 0.0044, "step": 2340 }, { "epoch": 80.72413793103448, "grad_norm": 1.71875, "learning_rate": 1.602926821053596e-05, "loss": 0.0067, "step": 2341 }, { "epoch": 80.75862068965517, "grad_norm": 3.78125, "learning_rate": 1.6020895757446708e-05, "loss": 0.0324, "step": 2342 }, { "epoch": 80.79310344827586, "grad_norm": 0.453125, "learning_rate": 1.601252158831564e-05, "loss": 0.0083, "step": 2343 }, { "epoch": 80.82758620689656, "grad_norm": 0.9921875, "learning_rate": 1.600414570722422e-05, "loss": 0.0049, "step": 2344 }, { "epoch": 80.86206896551724, "grad_norm": 0.4453125, "learning_rate": 1.599576811825475e-05, "loss": 0.0171, "step": 2345 }, { "epoch": 80.89655172413794, "grad_norm": 3.84375, "learning_rate": 1.5987388825490368e-05, "loss": 0.0108, "step": 2346 }, { "epoch": 80.93103448275862, "grad_norm": 0.265625, "learning_rate": 1.5979007833015036e-05, "loss": 0.003, "step": 2347 }, { "epoch": 80.96551724137932, "grad_norm": 0.138671875, "learning_rate": 1.5970625144913552e-05, "loss": 0.0057, "step": 2348 }, { "epoch": 81.0, "grad_norm": 0.765625, "learning_rate": 1.596224076527153e-05, "loss": 0.0131, "step": 2349 }, { "epoch": 81.03448275862068, "grad_norm": 3.890625, "learning_rate": 1.5953854698175422e-05, "loss": 0.0223, "step": 2350 }, { "epoch": 81.06896551724138, "grad_norm": 0.76953125, "learning_rate": 1.5945466947712488e-05, "loss": 0.0049, "step": 2351 }, { "epoch": 81.10344827586206, "grad_norm": 1.4609375, "learning_rate": 1.593707751797083e-05, "loss": 0.0178, "step": 2352 }, { "epoch": 81.13793103448276, "grad_norm": 3.6875, "learning_rate": 1.5928686413039334e-05, "loss": 0.0204, "step": 2353 }, { "epoch": 81.17241379310344, "grad_norm": 0.216796875, "learning_rate": 1.592029363700773e-05, "loss": 0.0097, "step": 2354 }, { "epoch": 81.20689655172414, "grad_norm": 2.359375, "learning_rate": 1.5911899193966567e-05, "loss": 0.0106, "step": 2355 }, { "epoch": 81.24137931034483, "grad_norm": 4.09375, "learning_rate": 1.590350308800718e-05, "loss": 0.0382, "step": 2356 }, { "epoch": 81.27586206896552, "grad_norm": 4.65625, "learning_rate": 1.5895105323221735e-05, "loss": 0.018, "step": 2357 }, { "epoch": 81.3103448275862, "grad_norm": 8.125, "learning_rate": 1.58867059037032e-05, "loss": 0.0435, "step": 2358 }, { "epoch": 81.34482758620689, "grad_norm": 0.91015625, "learning_rate": 1.5878304833545348e-05, "loss": 0.004, "step": 2359 }, { "epoch": 81.37931034482759, "grad_norm": 0.4765625, "learning_rate": 1.5869902116842765e-05, "loss": 0.008, "step": 2360 }, { "epoch": 81.41379310344827, "grad_norm": 6.46875, "learning_rate": 1.586149775769082e-05, "loss": 0.0511, "step": 2361 }, { "epoch": 81.44827586206897, "grad_norm": 1.921875, "learning_rate": 1.585309176018571e-05, "loss": 0.0355, "step": 2362 }, { "epoch": 81.48275862068965, "grad_norm": 1.046875, "learning_rate": 1.584468412842441e-05, "loss": 0.0139, "step": 2363 }, { "epoch": 81.51724137931035, "grad_norm": 1.3203125, "learning_rate": 1.5836274866504698e-05, "loss": 0.0056, "step": 2364 }, { "epoch": 81.55172413793103, "grad_norm": 1.4921875, "learning_rate": 1.5827863978525143e-05, "loss": 0.0119, "step": 2365 }, { "epoch": 81.58620689655173, "grad_norm": 1.109375, "learning_rate": 1.581945146858511e-05, "loss": 0.017, "step": 2366 }, { "epoch": 81.62068965517241, "grad_norm": 0.208984375, "learning_rate": 1.581103734078476e-05, "loss": 0.0046, "step": 2367 }, { "epoch": 81.65517241379311, "grad_norm": 0.3359375, "learning_rate": 1.5802621599225033e-05, "loss": 0.018, "step": 2368 }, { "epoch": 81.6896551724138, "grad_norm": 0.765625, "learning_rate": 1.579420424800766e-05, "loss": 0.006, "step": 2369 }, { "epoch": 81.72413793103448, "grad_norm": 0.84765625, "learning_rate": 1.578578529123515e-05, "loss": 0.0231, "step": 2370 }, { "epoch": 81.75862068965517, "grad_norm": 0.73046875, "learning_rate": 1.577736473301081e-05, "loss": 0.0159, "step": 2371 }, { "epoch": 81.79310344827586, "grad_norm": 1.515625, "learning_rate": 1.576894257743871e-05, "loss": 0.0112, "step": 2372 }, { "epoch": 81.82758620689656, "grad_norm": 0.259765625, "learning_rate": 1.5760518828623715e-05, "loss": 0.0028, "step": 2373 }, { "epoch": 81.86206896551724, "grad_norm": 0.79296875, "learning_rate": 1.5752093490671447e-05, "loss": 0.0047, "step": 2374 }, { "epoch": 81.89655172413794, "grad_norm": 5.03125, "learning_rate": 1.574366656768832e-05, "loss": 0.0234, "step": 2375 }, { "epoch": 81.93103448275862, "grad_norm": 2.90625, "learning_rate": 1.573523806378151e-05, "loss": 0.0336, "step": 2376 }, { "epoch": 81.96551724137932, "grad_norm": 1.0859375, "learning_rate": 1.5726807983058974e-05, "loss": 0.0078, "step": 2377 }, { "epoch": 82.0, "grad_norm": 0.34375, "learning_rate": 1.5718376329629424e-05, "loss": 0.0037, "step": 2378 }, { "epoch": 82.03448275862068, "grad_norm": 0.2314453125, "learning_rate": 1.5709943107602347e-05, "loss": 0.0555, "step": 2379 }, { "epoch": 82.06896551724138, "grad_norm": 1.5859375, "learning_rate": 1.5701508321087995e-05, "loss": 0.0279, "step": 2380 }, { "epoch": 82.10344827586206, "grad_norm": 4.75, "learning_rate": 1.569307197419738e-05, "loss": 0.0503, "step": 2381 }, { "epoch": 82.13793103448276, "grad_norm": 0.54296875, "learning_rate": 1.5684634071042278e-05, "loss": 0.0049, "step": 2382 }, { "epoch": 82.17241379310344, "grad_norm": 2.265625, "learning_rate": 1.5676194615735217e-05, "loss": 0.0154, "step": 2383 }, { "epoch": 82.20689655172414, "grad_norm": 4.25, "learning_rate": 1.566775361238948e-05, "loss": 0.032, "step": 2384 }, { "epoch": 82.24137931034483, "grad_norm": 0.50390625, "learning_rate": 1.5659311065119118e-05, "loss": 0.0068, "step": 2385 }, { "epoch": 82.27586206896552, "grad_norm": 0.49609375, "learning_rate": 1.5650866978038925e-05, "loss": 0.0095, "step": 2386 }, { "epoch": 82.3103448275862, "grad_norm": 0.75390625, "learning_rate": 1.5642421355264438e-05, "loss": 0.0069, "step": 2387 }, { "epoch": 82.34482758620689, "grad_norm": 0.330078125, "learning_rate": 1.563397420091196e-05, "loss": 0.0183, "step": 2388 }, { "epoch": 82.37931034482759, "grad_norm": 0.1591796875, "learning_rate": 1.5625525519098527e-05, "loss": 0.0054, "step": 2389 }, { "epoch": 82.41379310344827, "grad_norm": 0.205078125, "learning_rate": 1.5617075313941925e-05, "loss": 0.0032, "step": 2390 }, { "epoch": 82.44827586206897, "grad_norm": 4.625, "learning_rate": 1.560862358956069e-05, "loss": 0.0469, "step": 2391 }, { "epoch": 82.48275862068965, "grad_norm": 0.365234375, "learning_rate": 1.560017035007407e-05, "loss": 0.0067, "step": 2392 }, { "epoch": 82.51724137931035, "grad_norm": 1.1328125, "learning_rate": 1.559171559960209e-05, "loss": 0.0086, "step": 2393 }, { "epoch": 82.55172413793103, "grad_norm": 0.47265625, "learning_rate": 1.5583259342265483e-05, "loss": 0.0028, "step": 2394 }, { "epoch": 82.58620689655173, "grad_norm": 0.142578125, "learning_rate": 1.5574801582185726e-05, "loss": 0.0165, "step": 2395 }, { "epoch": 82.62068965517241, "grad_norm": 2.953125, "learning_rate": 1.5566342323485032e-05, "loss": 0.0295, "step": 2396 }, { "epoch": 82.65517241379311, "grad_norm": 1.8671875, "learning_rate": 1.555788157028634e-05, "loss": 0.0075, "step": 2397 }, { "epoch": 82.6896551724138, "grad_norm": 2.34375, "learning_rate": 1.5549419326713323e-05, "loss": 0.0422, "step": 2398 }, { "epoch": 82.72413793103448, "grad_norm": 3.765625, "learning_rate": 1.5540955596890365e-05, "loss": 0.0206, "step": 2399 }, { "epoch": 82.75862068965517, "grad_norm": 6.40625, "learning_rate": 1.5532490384942593e-05, "loss": 0.0676, "step": 2400 }, { "epoch": 82.79310344827586, "grad_norm": 0.9765625, "learning_rate": 1.5524023694995844e-05, "loss": 0.0607, "step": 2401 }, { "epoch": 82.82758620689656, "grad_norm": 0.427734375, "learning_rate": 1.5515555531176687e-05, "loss": 0.0037, "step": 2402 }, { "epoch": 82.86206896551724, "grad_norm": 2.625, "learning_rate": 1.5507085897612396e-05, "loss": 0.0205, "step": 2403 }, { "epoch": 82.89655172413794, "grad_norm": 5.65625, "learning_rate": 1.5498614798430972e-05, "loss": 0.0294, "step": 2404 }, { "epoch": 82.93103448275862, "grad_norm": 2.390625, "learning_rate": 1.549014223776112e-05, "loss": 0.0212, "step": 2405 }, { "epoch": 82.96551724137932, "grad_norm": 1.453125, "learning_rate": 1.5481668219732263e-05, "loss": 0.0232, "step": 2406 }, { "epoch": 83.0, "grad_norm": 3.546875, "learning_rate": 1.547319274847454e-05, "loss": 0.0207, "step": 2407 }, { "epoch": 83.03448275862068, "grad_norm": 0.9140625, "learning_rate": 1.546471582811879e-05, "loss": 0.0126, "step": 2408 }, { "epoch": 83.06896551724138, "grad_norm": 1.671875, "learning_rate": 1.545623746279656e-05, "loss": 0.0113, "step": 2409 }, { "epoch": 83.10344827586206, "grad_norm": 1.2578125, "learning_rate": 1.5447757656640104e-05, "loss": 0.0178, "step": 2410 }, { "epoch": 83.13793103448276, "grad_norm": 7.8125, "learning_rate": 1.5439276413782375e-05, "loss": 0.0296, "step": 2411 }, { "epoch": 83.17241379310344, "grad_norm": 0.64453125, "learning_rate": 1.543079373835703e-05, "loss": 0.0077, "step": 2412 }, { "epoch": 83.20689655172414, "grad_norm": 2.484375, "learning_rate": 1.5422309634498423e-05, "loss": 0.025, "step": 2413 }, { "epoch": 83.24137931034483, "grad_norm": 0.478515625, "learning_rate": 1.54138241063416e-05, "loss": 0.0351, "step": 2414 }, { "epoch": 83.27586206896552, "grad_norm": 1.2578125, "learning_rate": 1.54053371580223e-05, "loss": 0.0175, "step": 2415 }, { "epoch": 83.3103448275862, "grad_norm": 7.375, "learning_rate": 1.5396848793676966e-05, "loss": 0.0723, "step": 2416 }, { "epoch": 83.34482758620689, "grad_norm": 0.482421875, "learning_rate": 1.5388359017442727e-05, "loss": 0.0039, "step": 2417 }, { "epoch": 83.37931034482759, "grad_norm": 0.474609375, "learning_rate": 1.5379867833457386e-05, "loss": 0.0053, "step": 2418 }, { "epoch": 83.41379310344827, "grad_norm": 10.125, "learning_rate": 1.5371375245859446e-05, "loss": 0.041, "step": 2419 }, { "epoch": 83.44827586206897, "grad_norm": 0.8359375, "learning_rate": 1.5362881258788098e-05, "loss": 0.0089, "step": 2420 }, { "epoch": 83.48275862068965, "grad_norm": 0.9375, "learning_rate": 1.5354385876383197e-05, "loss": 0.0056, "step": 2421 }, { "epoch": 83.51724137931035, "grad_norm": 0.98828125, "learning_rate": 1.53458891027853e-05, "loss": 0.0325, "step": 2422 }, { "epoch": 83.55172413793103, "grad_norm": 1.5546875, "learning_rate": 1.5337390942135627e-05, "loss": 0.0102, "step": 2423 }, { "epoch": 83.58620689655173, "grad_norm": 1.7578125, "learning_rate": 1.5328891398576077e-05, "loss": 0.0168, "step": 2424 }, { "epoch": 83.62068965517241, "grad_norm": 2.53125, "learning_rate": 1.532039047624922e-05, "loss": 0.0177, "step": 2425 }, { "epoch": 83.65517241379311, "grad_norm": 0.53125, "learning_rate": 1.5311888179298313e-05, "loss": 0.0278, "step": 2426 }, { "epoch": 83.6896551724138, "grad_norm": 0.88671875, "learning_rate": 1.5303384511867268e-05, "loss": 0.0093, "step": 2427 }, { "epoch": 83.72413793103448, "grad_norm": 0.84375, "learning_rate": 1.5294879478100663e-05, "loss": 0.0281, "step": 2428 }, { "epoch": 83.75862068965517, "grad_norm": 3.0625, "learning_rate": 1.5286373082143757e-05, "loss": 0.0749, "step": 2429 }, { "epoch": 83.79310344827586, "grad_norm": 4.1875, "learning_rate": 1.5277865328142457e-05, "loss": 0.0615, "step": 2430 }, { "epoch": 83.82758620689656, "grad_norm": 1.75, "learning_rate": 1.5269356220243345e-05, "loss": 0.0054, "step": 2431 }, { "epoch": 83.86206896551724, "grad_norm": 1.03125, "learning_rate": 1.5260845762593647e-05, "loss": 0.0153, "step": 2432 }, { "epoch": 83.89655172413794, "grad_norm": 1.3046875, "learning_rate": 1.525233395934127e-05, "loss": 0.0375, "step": 2433 }, { "epoch": 83.93103448275862, "grad_norm": 1.8984375, "learning_rate": 1.5243820814634758e-05, "loss": 0.0265, "step": 2434 }, { "epoch": 83.96551724137932, "grad_norm": 1.8359375, "learning_rate": 1.5235306332623309e-05, "loss": 0.0176, "step": 2435 }, { "epoch": 84.0, "grad_norm": 5.6875, "learning_rate": 1.5226790517456785e-05, "loss": 0.0322, "step": 2436 }, { "epoch": 84.03448275862068, "grad_norm": 1.9921875, "learning_rate": 1.5218273373285686e-05, "loss": 0.0162, "step": 2437 }, { "epoch": 84.06896551724138, "grad_norm": 1.6796875, "learning_rate": 1.5209754904261167e-05, "loss": 0.009, "step": 2438 }, { "epoch": 84.10344827586206, "grad_norm": 0.84765625, "learning_rate": 1.5201235114535028e-05, "loss": 0.0106, "step": 2439 }, { "epoch": 84.13793103448276, "grad_norm": 0.296875, "learning_rate": 1.5192714008259706e-05, "loss": 0.0199, "step": 2440 }, { "epoch": 84.17241379310344, "grad_norm": 2.390625, "learning_rate": 1.5184191589588292e-05, "loss": 0.0158, "step": 2441 }, { "epoch": 84.20689655172414, "grad_norm": 1.28125, "learning_rate": 1.51756678626745e-05, "loss": 0.0187, "step": 2442 }, { "epoch": 84.24137931034483, "grad_norm": 2.140625, "learning_rate": 1.5167142831672702e-05, "loss": 0.0285, "step": 2443 }, { "epoch": 84.27586206896552, "grad_norm": 1.3984375, "learning_rate": 1.5158616500737885e-05, "loss": 0.0105, "step": 2444 }, { "epoch": 84.3103448275862, "grad_norm": 2.359375, "learning_rate": 1.5150088874025686e-05, "loss": 0.0108, "step": 2445 }, { "epoch": 84.34482758620689, "grad_norm": 0.25, "learning_rate": 1.514155995569236e-05, "loss": 0.0034, "step": 2446 }, { "epoch": 84.37931034482759, "grad_norm": 1.25, "learning_rate": 1.51330297498948e-05, "loss": 0.0067, "step": 2447 }, { "epoch": 84.41379310344827, "grad_norm": 1.90625, "learning_rate": 1.5124498260790535e-05, "loss": 0.0176, "step": 2448 }, { "epoch": 84.44827586206897, "grad_norm": 1.15625, "learning_rate": 1.511596549253769e-05, "loss": 0.0287, "step": 2449 }, { "epoch": 84.48275862068965, "grad_norm": 1.78125, "learning_rate": 1.5107431449295047e-05, "loss": 0.0514, "step": 2450 }, { "epoch": 84.51724137931035, "grad_norm": 0.490234375, "learning_rate": 1.5098896135221991e-05, "loss": 0.0049, "step": 2451 }, { "epoch": 84.55172413793103, "grad_norm": 1.9375, "learning_rate": 1.5090359554478531e-05, "loss": 0.0369, "step": 2452 }, { "epoch": 84.58620689655173, "grad_norm": 0.357421875, "learning_rate": 1.5081821711225295e-05, "loss": 0.0179, "step": 2453 }, { "epoch": 84.62068965517241, "grad_norm": 0.515625, "learning_rate": 1.5073282609623518e-05, "loss": 0.01, "step": 2454 }, { "epoch": 84.65517241379311, "grad_norm": 1.84375, "learning_rate": 1.5064742253835057e-05, "loss": 0.0241, "step": 2455 }, { "epoch": 84.6896551724138, "grad_norm": 0.2451171875, "learning_rate": 1.5056200648022378e-05, "loss": 0.0042, "step": 2456 }, { "epoch": 84.72413793103448, "grad_norm": 0.5078125, "learning_rate": 1.5047657796348563e-05, "loss": 0.0058, "step": 2457 }, { "epoch": 84.75862068965517, "grad_norm": 4.25, "learning_rate": 1.5039113702977278e-05, "loss": 0.0439, "step": 2458 }, { "epoch": 84.79310344827586, "grad_norm": 0.9296875, "learning_rate": 1.5030568372072826e-05, "loss": 0.0102, "step": 2459 }, { "epoch": 84.82758620689656, "grad_norm": 0.703125, "learning_rate": 1.5022021807800087e-05, "loss": 0.0712, "step": 2460 }, { "epoch": 84.86206896551724, "grad_norm": 2.75, "learning_rate": 1.501347401432456e-05, "loss": 0.0128, "step": 2461 }, { "epoch": 84.89655172413794, "grad_norm": 0.494140625, "learning_rate": 1.5004924995812331e-05, "loss": 0.0035, "step": 2462 }, { "epoch": 84.93103448275862, "grad_norm": 0.671875, "learning_rate": 1.4996374756430087e-05, "loss": 0.0204, "step": 2463 }, { "epoch": 84.96551724137932, "grad_norm": 4.03125, "learning_rate": 1.4987823300345117e-05, "loss": 0.0206, "step": 2464 }, { "epoch": 85.0, "grad_norm": 2.8125, "learning_rate": 1.4979270631725292e-05, "loss": 0.0216, "step": 2465 }, { "epoch": 85.03448275862068, "grad_norm": 0.7578125, "learning_rate": 1.4970716754739078e-05, "loss": 0.0098, "step": 2466 }, { "epoch": 85.06896551724138, "grad_norm": 2.75, "learning_rate": 1.4962161673555537e-05, "loss": 0.0206, "step": 2467 }, { "epoch": 85.10344827586206, "grad_norm": 1.28125, "learning_rate": 1.49536053923443e-05, "loss": 0.0282, "step": 2468 }, { "epoch": 85.13793103448276, "grad_norm": 4.96875, "learning_rate": 1.4945047915275604e-05, "loss": 0.0239, "step": 2469 }, { "epoch": 85.17241379310344, "grad_norm": 4.25, "learning_rate": 1.4936489246520257e-05, "loss": 0.0206, "step": 2470 }, { "epoch": 85.20689655172414, "grad_norm": 0.1328125, "learning_rate": 1.492792939024965e-05, "loss": 0.0174, "step": 2471 }, { "epoch": 85.24137931034483, "grad_norm": 3.4375, "learning_rate": 1.4919368350635757e-05, "loss": 0.0183, "step": 2472 }, { "epoch": 85.27586206896552, "grad_norm": 0.46875, "learning_rate": 1.4910806131851113e-05, "loss": 0.0133, "step": 2473 }, { "epoch": 85.3103448275862, "grad_norm": 0.5625, "learning_rate": 1.4902242738068853e-05, "loss": 0.0201, "step": 2474 }, { "epoch": 85.34482758620689, "grad_norm": 1.640625, "learning_rate": 1.4893678173462664e-05, "loss": 0.0091, "step": 2475 }, { "epoch": 85.37931034482759, "grad_norm": 3.765625, "learning_rate": 1.4885112442206812e-05, "loss": 0.0242, "step": 2476 }, { "epoch": 85.41379310344827, "grad_norm": 2.46875, "learning_rate": 1.4876545548476128e-05, "loss": 0.0209, "step": 2477 }, { "epoch": 85.44827586206897, "grad_norm": 1.578125, "learning_rate": 1.4867977496446017e-05, "loss": 0.0319, "step": 2478 }, { "epoch": 85.48275862068965, "grad_norm": 0.28125, "learning_rate": 1.4859408290292442e-05, "loss": 0.0062, "step": 2479 }, { "epoch": 85.51724137931035, "grad_norm": 3.59375, "learning_rate": 1.4850837934191924e-05, "loss": 0.0263, "step": 2480 }, { "epoch": 85.55172413793103, "grad_norm": 0.4921875, "learning_rate": 1.4842266432321561e-05, "loss": 0.0067, "step": 2481 }, { "epoch": 85.58620689655173, "grad_norm": 1.7890625, "learning_rate": 1.4833693788858988e-05, "loss": 0.0064, "step": 2482 }, { "epoch": 85.62068965517241, "grad_norm": 1.546875, "learning_rate": 1.4825120007982413e-05, "loss": 0.0357, "step": 2483 }, { "epoch": 85.65517241379311, "grad_norm": 0.71875, "learning_rate": 1.4816545093870593e-05, "loss": 0.0137, "step": 2484 }, { "epoch": 85.6896551724138, "grad_norm": 0.4140625, "learning_rate": 1.4807969050702838e-05, "loss": 0.0055, "step": 2485 }, { "epoch": 85.72413793103448, "grad_norm": 0.2177734375, "learning_rate": 1.4799391882659004e-05, "loss": 0.0225, "step": 2486 }, { "epoch": 85.75862068965517, "grad_norm": 0.1845703125, "learning_rate": 1.4790813593919497e-05, "loss": 0.0084, "step": 2487 }, { "epoch": 85.79310344827586, "grad_norm": 6.21875, "learning_rate": 1.4782234188665278e-05, "loss": 0.0339, "step": 2488 }, { "epoch": 85.82758620689656, "grad_norm": 5.0625, "learning_rate": 1.4773653671077836e-05, "loss": 0.0208, "step": 2489 }, { "epoch": 85.86206896551724, "grad_norm": 2.15625, "learning_rate": 1.4765072045339218e-05, "loss": 0.0227, "step": 2490 }, { "epoch": 85.89655172413794, "grad_norm": 0.341796875, "learning_rate": 1.4756489315632003e-05, "loss": 0.0021, "step": 2491 }, { "epoch": 85.93103448275862, "grad_norm": 8.6875, "learning_rate": 1.4747905486139305e-05, "loss": 0.0277, "step": 2492 }, { "epoch": 85.96551724137932, "grad_norm": 4.0, "learning_rate": 1.4739320561044781e-05, "loss": 0.0452, "step": 2493 }, { "epoch": 86.0, "grad_norm": 2.546875, "learning_rate": 1.4730734544532617e-05, "loss": 0.0469, "step": 2494 }, { "epoch": 86.03448275862068, "grad_norm": 3.171875, "learning_rate": 1.472214744078754e-05, "loss": 0.0271, "step": 2495 }, { "epoch": 86.06896551724138, "grad_norm": 1.640625, "learning_rate": 1.4713559253994793e-05, "loss": 0.0191, "step": 2496 }, { "epoch": 86.10344827586206, "grad_norm": 1.609375, "learning_rate": 1.4704969988340156e-05, "loss": 0.0154, "step": 2497 }, { "epoch": 86.13793103448276, "grad_norm": 1.453125, "learning_rate": 1.4696379648009933e-05, "loss": 0.0102, "step": 2498 }, { "epoch": 86.17241379310344, "grad_norm": 2.65625, "learning_rate": 1.4687788237190952e-05, "loss": 0.0141, "step": 2499 }, { "epoch": 86.20689655172414, "grad_norm": 1.546875, "learning_rate": 1.4679195760070561e-05, "loss": 0.0099, "step": 2500 }, { "epoch": 86.24137931034483, "grad_norm": 1.5859375, "learning_rate": 1.4670602220836633e-05, "loss": 0.0184, "step": 2501 }, { "epoch": 86.27586206896552, "grad_norm": 2.828125, "learning_rate": 1.4662007623677549e-05, "loss": 0.0195, "step": 2502 }, { "epoch": 86.3103448275862, "grad_norm": 0.177734375, "learning_rate": 1.4653411972782211e-05, "loss": 0.0042, "step": 2503 }, { "epoch": 86.34482758620689, "grad_norm": 4.53125, "learning_rate": 1.4644815272340037e-05, "loss": 0.0268, "step": 2504 }, { "epoch": 86.37931034482759, "grad_norm": 1.828125, "learning_rate": 1.4636217526540961e-05, "loss": 0.0402, "step": 2505 }, { "epoch": 86.41379310344827, "grad_norm": 0.462890625, "learning_rate": 1.4627618739575408e-05, "loss": 0.0122, "step": 2506 }, { "epoch": 86.44827586206897, "grad_norm": 0.25390625, "learning_rate": 1.461901891563433e-05, "loss": 0.0033, "step": 2507 }, { "epoch": 86.48275862068965, "grad_norm": 3.046875, "learning_rate": 1.4610418058909172e-05, "loss": 0.0377, "step": 2508 }, { "epoch": 86.51724137931035, "grad_norm": 2.203125, "learning_rate": 1.4601816173591895e-05, "loss": 0.0177, "step": 2509 }, { "epoch": 86.55172413793103, "grad_norm": 0.39453125, "learning_rate": 1.4593213263874944e-05, "loss": 0.066, "step": 2510 }, { "epoch": 86.58620689655173, "grad_norm": 0.87109375, "learning_rate": 1.458460933395128e-05, "loss": 0.0232, "step": 2511 }, { "epoch": 86.62068965517241, "grad_norm": 2.296875, "learning_rate": 1.457600438801435e-05, "loss": 0.0414, "step": 2512 }, { "epoch": 86.65517241379311, "grad_norm": 1.1875, "learning_rate": 1.4567398430258097e-05, "loss": 0.0056, "step": 2513 }, { "epoch": 86.6896551724138, "grad_norm": 0.205078125, "learning_rate": 1.4558791464876966e-05, "loss": 0.0186, "step": 2514 }, { "epoch": 86.72413793103448, "grad_norm": 0.56640625, "learning_rate": 1.455018349606589e-05, "loss": 0.0045, "step": 2515 }, { "epoch": 86.75862068965517, "grad_norm": 1.6171875, "learning_rate": 1.454157452802028e-05, "loss": 0.008, "step": 2516 }, { "epoch": 86.79310344827586, "grad_norm": 0.490234375, "learning_rate": 1.4532964564936047e-05, "loss": 0.0041, "step": 2517 }, { "epoch": 86.82758620689656, "grad_norm": 0.1806640625, "learning_rate": 1.4524353611009577e-05, "loss": 0.0216, "step": 2518 }, { "epoch": 86.86206896551724, "grad_norm": 1.65625, "learning_rate": 1.451574167043776e-05, "loss": 0.0062, "step": 2519 }, { "epoch": 86.89655172413794, "grad_norm": 0.8046875, "learning_rate": 1.4507128747417931e-05, "loss": 0.0052, "step": 2520 }, { "epoch": 86.93103448275862, "grad_norm": 0.16015625, "learning_rate": 1.4498514846147937e-05, "loss": 0.0249, "step": 2521 }, { "epoch": 86.96551724137932, "grad_norm": 1.015625, "learning_rate": 1.4489899970826085e-05, "loss": 0.0128, "step": 2522 }, { "epoch": 87.0, "grad_norm": 0.083984375, "learning_rate": 1.4481284125651162e-05, "loss": 0.0022, "step": 2523 }, { "epoch": 87.03448275862068, "grad_norm": 1.203125, "learning_rate": 1.4472667314822425e-05, "loss": 0.0274, "step": 2524 }, { "epoch": 87.06896551724138, "grad_norm": 1.6015625, "learning_rate": 1.4464049542539598e-05, "loss": 0.0086, "step": 2525 }, { "epoch": 87.10344827586206, "grad_norm": 8.8125, "learning_rate": 1.4455430813002888e-05, "loss": 0.04, "step": 2526 }, { "epoch": 87.13793103448276, "grad_norm": 0.1630859375, "learning_rate": 1.4446811130412952e-05, "loss": 0.004, "step": 2527 }, { "epoch": 87.17241379310344, "grad_norm": 1.84375, "learning_rate": 1.4438190498970924e-05, "loss": 0.0228, "step": 2528 }, { "epoch": 87.20689655172414, "grad_norm": 2.1875, "learning_rate": 1.4429568922878392e-05, "loss": 0.0093, "step": 2529 }, { "epoch": 87.24137931034483, "grad_norm": 0.298828125, "learning_rate": 1.4420946406337404e-05, "loss": 0.0059, "step": 2530 }, { "epoch": 87.27586206896552, "grad_norm": 3.640625, "learning_rate": 1.4412322953550477e-05, "loss": 0.0249, "step": 2531 }, { "epoch": 87.3103448275862, "grad_norm": 1.6484375, "learning_rate": 1.4403698568720572e-05, "loss": 0.0071, "step": 2532 }, { "epoch": 87.34482758620689, "grad_norm": 0.57421875, "learning_rate": 1.4395073256051114e-05, "loss": 0.0058, "step": 2533 }, { "epoch": 87.37931034482759, "grad_norm": 0.68359375, "learning_rate": 1.438644701974597e-05, "loss": 0.018, "step": 2534 }, { "epoch": 87.41379310344827, "grad_norm": 1.4921875, "learning_rate": 1.4377819864009465e-05, "loss": 0.013, "step": 2535 }, { "epoch": 87.44827586206897, "grad_norm": 3.640625, "learning_rate": 1.4369191793046372e-05, "loss": 0.0238, "step": 2536 }, { "epoch": 87.48275862068965, "grad_norm": 4.09375, "learning_rate": 1.4360562811061903e-05, "loss": 0.0263, "step": 2537 }, { "epoch": 87.51724137931035, "grad_norm": 2.03125, "learning_rate": 1.4351932922261726e-05, "loss": 0.0254, "step": 2538 }, { "epoch": 87.55172413793103, "grad_norm": 1.890625, "learning_rate": 1.4343302130851938e-05, "loss": 0.0181, "step": 2539 }, { "epoch": 87.58620689655173, "grad_norm": 4.1875, "learning_rate": 1.4334670441039083e-05, "loss": 0.0375, "step": 2540 }, { "epoch": 87.62068965517241, "grad_norm": 0.423828125, "learning_rate": 1.4326037857030148e-05, "loss": 0.012, "step": 2541 }, { "epoch": 87.65517241379311, "grad_norm": 3.859375, "learning_rate": 1.431740438303254e-05, "loss": 0.0484, "step": 2542 }, { "epoch": 87.6896551724138, "grad_norm": 1.453125, "learning_rate": 1.4308770023254114e-05, "loss": 0.0145, "step": 2543 }, { "epoch": 87.72413793103448, "grad_norm": 5.21875, "learning_rate": 1.4300134781903151e-05, "loss": 0.035, "step": 2544 }, { "epoch": 87.75862068965517, "grad_norm": 1.2109375, "learning_rate": 1.429149866318837e-05, "loss": 0.0085, "step": 2545 }, { "epoch": 87.79310344827586, "grad_norm": 5.09375, "learning_rate": 1.4282861671318898e-05, "loss": 0.0477, "step": 2546 }, { "epoch": 87.82758620689656, "grad_norm": 0.66015625, "learning_rate": 1.4274223810504305e-05, "loss": 0.0107, "step": 2547 }, { "epoch": 87.86206896551724, "grad_norm": 3.078125, "learning_rate": 1.4265585084954586e-05, "loss": 0.0242, "step": 2548 }, { "epoch": 87.89655172413794, "grad_norm": 0.70703125, "learning_rate": 1.4256945498880136e-05, "loss": 0.0176, "step": 2549 }, { "epoch": 87.93103448275862, "grad_norm": 1.6796875, "learning_rate": 1.4248305056491804e-05, "loss": 0.0282, "step": 2550 }, { "epoch": 87.96551724137932, "grad_norm": 1.8984375, "learning_rate": 1.4239663762000818e-05, "loss": 0.0306, "step": 2551 }, { "epoch": 88.0, "grad_norm": 0.6484375, "learning_rate": 1.4231021619618854e-05, "loss": 0.0127, "step": 2552 }, { "epoch": 88.03448275862068, "grad_norm": 7.25, "learning_rate": 1.4222378633557976e-05, "loss": 0.043, "step": 2553 }, { "epoch": 88.06896551724138, "grad_norm": 0.478515625, "learning_rate": 1.4213734808030682e-05, "loss": 0.007, "step": 2554 }, { "epoch": 88.10344827586206, "grad_norm": 3.0625, "learning_rate": 1.4205090147249856e-05, "loss": 0.0169, "step": 2555 }, { "epoch": 88.13793103448276, "grad_norm": 0.453125, "learning_rate": 1.4196444655428806e-05, "loss": 0.0065, "step": 2556 }, { "epoch": 88.17241379310344, "grad_norm": 1.671875, "learning_rate": 1.4187798336781241e-05, "loss": 0.0093, "step": 2557 }, { "epoch": 88.20689655172414, "grad_norm": 0.2294921875, "learning_rate": 1.4179151195521268e-05, "loss": 0.0081, "step": 2558 }, { "epoch": 88.24137931034483, "grad_norm": 0.74609375, "learning_rate": 1.4170503235863408e-05, "loss": 0.0144, "step": 2559 }, { "epoch": 88.27586206896552, "grad_norm": 0.251953125, "learning_rate": 1.4161854462022556e-05, "loss": 0.0051, "step": 2560 }, { "epoch": 88.3103448275862, "grad_norm": 0.380859375, "learning_rate": 1.4153204878214032e-05, "loss": 0.0062, "step": 2561 }, { "epoch": 88.34482758620689, "grad_norm": 0.494140625, "learning_rate": 1.4144554488653533e-05, "loss": 0.0025, "step": 2562 }, { "epoch": 88.37931034482759, "grad_norm": 1.1875, "learning_rate": 1.4135903297557155e-05, "loss": 0.0093, "step": 2563 }, { "epoch": 88.41379310344827, "grad_norm": 1.4765625, "learning_rate": 1.4127251309141381e-05, "loss": 0.0181, "step": 2564 }, { "epoch": 88.44827586206897, "grad_norm": 6.4375, "learning_rate": 1.4118598527623086e-05, "loss": 0.0204, "step": 2565 }, { "epoch": 88.48275862068965, "grad_norm": 2.0, "learning_rate": 1.4109944957219537e-05, "loss": 0.0282, "step": 2566 }, { "epoch": 88.51724137931035, "grad_norm": 0.76171875, "learning_rate": 1.4101290602148371e-05, "loss": 0.0248, "step": 2567 }, { "epoch": 88.55172413793103, "grad_norm": 0.9609375, "learning_rate": 1.4092635466627618e-05, "loss": 0.0068, "step": 2568 }, { "epoch": 88.58620689655173, "grad_norm": 1.546875, "learning_rate": 1.4083979554875685e-05, "loss": 0.0061, "step": 2569 }, { "epoch": 88.62068965517241, "grad_norm": 2.265625, "learning_rate": 1.4075322871111357e-05, "loss": 0.0135, "step": 2570 }, { "epoch": 88.65517241379311, "grad_norm": 1.40625, "learning_rate": 1.4066665419553804e-05, "loss": 0.0344, "step": 2571 }, { "epoch": 88.6896551724138, "grad_norm": 6.28125, "learning_rate": 1.4058007204422558e-05, "loss": 0.0231, "step": 2572 }, { "epoch": 88.72413793103448, "grad_norm": 0.16015625, "learning_rate": 1.4049348229937523e-05, "loss": 0.003, "step": 2573 }, { "epoch": 88.75862068965517, "grad_norm": 2.28125, "learning_rate": 1.4040688500318985e-05, "loss": 0.0207, "step": 2574 }, { "epoch": 88.79310344827586, "grad_norm": 1.703125, "learning_rate": 1.4032028019787585e-05, "loss": 0.0135, "step": 2575 }, { "epoch": 88.82758620689656, "grad_norm": 0.337890625, "learning_rate": 1.4023366792564343e-05, "loss": 0.0045, "step": 2576 }, { "epoch": 88.86206896551724, "grad_norm": 1.3671875, "learning_rate": 1.4014704822870634e-05, "loss": 0.0075, "step": 2577 }, { "epoch": 88.89655172413794, "grad_norm": 0.765625, "learning_rate": 1.4006042114928191e-05, "loss": 0.0193, "step": 2578 }, { "epoch": 88.93103448275862, "grad_norm": 3.84375, "learning_rate": 1.3997378672959122e-05, "loss": 0.0129, "step": 2579 }, { "epoch": 88.96551724137932, "grad_norm": 2.625, "learning_rate": 1.3988714501185873e-05, "loss": 0.0113, "step": 2580 }, { "epoch": 89.0, "grad_norm": 0.77734375, "learning_rate": 1.3980049603831266e-05, "loss": 0.0252, "step": 2581 }, { "epoch": 89.03448275862068, "grad_norm": 0.67578125, "learning_rate": 1.3971383985118455e-05, "loss": 0.0039, "step": 2582 }, { "epoch": 89.06896551724138, "grad_norm": 0.36328125, "learning_rate": 1.396271764927097e-05, "loss": 0.0166, "step": 2583 }, { "epoch": 89.10344827586206, "grad_norm": 2.65625, "learning_rate": 1.3954050600512669e-05, "loss": 0.0216, "step": 2584 }, { "epoch": 89.13793103448276, "grad_norm": 3.453125, "learning_rate": 1.3945382843067772e-05, "loss": 0.0298, "step": 2585 }, { "epoch": 89.17241379310344, "grad_norm": 4.3125, "learning_rate": 1.3936714381160835e-05, "loss": 0.0317, "step": 2586 }, { "epoch": 89.20689655172414, "grad_norm": 5.15625, "learning_rate": 1.3928045219016761e-05, "loss": 0.0432, "step": 2587 }, { "epoch": 89.24137931034483, "grad_norm": 8.4375, "learning_rate": 1.3919375360860795e-05, "loss": 0.0367, "step": 2588 }, { "epoch": 89.27586206896552, "grad_norm": 0.96875, "learning_rate": 1.391070481091852e-05, "loss": 0.0122, "step": 2589 }, { "epoch": 89.3103448275862, "grad_norm": 0.408203125, "learning_rate": 1.3902033573415862e-05, "loss": 0.0055, "step": 2590 }, { "epoch": 89.34482758620689, "grad_norm": 1.171875, "learning_rate": 1.3893361652579068e-05, "loss": 0.0088, "step": 2591 }, { "epoch": 89.37931034482759, "grad_norm": 1.2265625, "learning_rate": 1.3884689052634733e-05, "loss": 0.0098, "step": 2592 }, { "epoch": 89.41379310344827, "grad_norm": 1.078125, "learning_rate": 1.3876015777809778e-05, "loss": 0.0218, "step": 2593 }, { "epoch": 89.44827586206897, "grad_norm": 0.75390625, "learning_rate": 1.3867341832331446e-05, "loss": 0.0045, "step": 2594 }, { "epoch": 89.48275862068965, "grad_norm": 3.078125, "learning_rate": 1.3858667220427315e-05, "loss": 0.0092, "step": 2595 }, { "epoch": 89.51724137931035, "grad_norm": 1.59375, "learning_rate": 1.3849991946325287e-05, "loss": 0.023, "step": 2596 }, { "epoch": 89.55172413793103, "grad_norm": 0.431640625, "learning_rate": 1.3841316014253584e-05, "loss": 0.0069, "step": 2597 }, { "epoch": 89.58620689655173, "grad_norm": 2.671875, "learning_rate": 1.383263942844075e-05, "loss": 0.0234, "step": 2598 }, { "epoch": 89.62068965517241, "grad_norm": 0.296875, "learning_rate": 1.382396219311565e-05, "loss": 0.0369, "step": 2599 }, { "epoch": 89.65517241379311, "grad_norm": 2.375, "learning_rate": 1.3815284312507459e-05, "loss": 0.0207, "step": 2600 }, { "epoch": 89.6896551724138, "grad_norm": 1.1953125, "learning_rate": 1.380660579084567e-05, "loss": 0.0115, "step": 2601 }, { "epoch": 89.72413793103448, "grad_norm": 3.46875, "learning_rate": 1.3797926632360092e-05, "loss": 0.0192, "step": 2602 }, { "epoch": 89.75862068965517, "grad_norm": 4.09375, "learning_rate": 1.3789246841280842e-05, "loss": 0.0048, "step": 2603 }, { "epoch": 89.79310344827586, "grad_norm": 3.296875, "learning_rate": 1.3780566421838346e-05, "loss": 0.0269, "step": 2604 }, { "epoch": 89.82758620689656, "grad_norm": 0.625, "learning_rate": 1.377188537826333e-05, "loss": 0.0119, "step": 2605 }, { "epoch": 89.86206896551724, "grad_norm": 0.94921875, "learning_rate": 1.3763203714786829e-05, "loss": 0.0054, "step": 2606 }, { "epoch": 89.89655172413794, "grad_norm": 1.265625, "learning_rate": 1.3754521435640186e-05, "loss": 0.0217, "step": 2607 }, { "epoch": 89.93103448275862, "grad_norm": 3.25, "learning_rate": 1.3745838545055038e-05, "loss": 0.0183, "step": 2608 }, { "epoch": 89.96551724137932, "grad_norm": 0.62109375, "learning_rate": 1.3737155047263315e-05, "loss": 0.0247, "step": 2609 }, { "epoch": 90.0, "grad_norm": 0.69140625, "learning_rate": 1.3728470946497248e-05, "loss": 0.0146, "step": 2610 }, { "epoch": 90.03448275862068, "grad_norm": 11.1875, "learning_rate": 1.3719786246989368e-05, "loss": 0.0261, "step": 2611 }, { "epoch": 90.06896551724138, "grad_norm": 0.8515625, "learning_rate": 1.3711100952972491e-05, "loss": 0.0313, "step": 2612 }, { "epoch": 90.10344827586206, "grad_norm": 0.19140625, "learning_rate": 1.3702415068679722e-05, "loss": 0.004, "step": 2613 }, { "epoch": 90.13793103448276, "grad_norm": 3.90625, "learning_rate": 1.3693728598344454e-05, "loss": 0.0367, "step": 2614 }, { "epoch": 90.17241379310344, "grad_norm": 2.171875, "learning_rate": 1.3685041546200367e-05, "loss": 0.0107, "step": 2615 }, { "epoch": 90.20689655172414, "grad_norm": 0.8046875, "learning_rate": 1.3676353916481433e-05, "loss": 0.0142, "step": 2616 }, { "epoch": 90.24137931034483, "grad_norm": 0.68359375, "learning_rate": 1.3667665713421884e-05, "loss": 0.0038, "step": 2617 }, { "epoch": 90.27586206896552, "grad_norm": 0.89453125, "learning_rate": 1.3658976941256252e-05, "loss": 0.0099, "step": 2618 }, { "epoch": 90.3103448275862, "grad_norm": 1.7578125, "learning_rate": 1.3650287604219342e-05, "loss": 0.0433, "step": 2619 }, { "epoch": 90.34482758620689, "grad_norm": 1.5390625, "learning_rate": 1.3641597706546223e-05, "loss": 0.0064, "step": 2620 }, { "epoch": 90.37931034482759, "grad_norm": 1.3125, "learning_rate": 1.3632907252472255e-05, "loss": 0.008, "step": 2621 }, { "epoch": 90.41379310344827, "grad_norm": 3.421875, "learning_rate": 1.3624216246233048e-05, "loss": 0.032, "step": 2622 }, { "epoch": 90.44827586206897, "grad_norm": 0.5390625, "learning_rate": 1.3615524692064508e-05, "loss": 0.0085, "step": 2623 }, { "epoch": 90.48275862068965, "grad_norm": 0.466796875, "learning_rate": 1.3606832594202778e-05, "loss": 0.0085, "step": 2624 }, { "epoch": 90.51724137931035, "grad_norm": 0.69921875, "learning_rate": 1.3598139956884292e-05, "loss": 0.0042, "step": 2625 }, { "epoch": 90.55172413793103, "grad_norm": 0.5546875, "learning_rate": 1.358944678434573e-05, "loss": 0.0059, "step": 2626 }, { "epoch": 90.58620689655173, "grad_norm": 0.1572265625, "learning_rate": 1.3580753080824038e-05, "loss": 0.0405, "step": 2627 }, { "epoch": 90.62068965517241, "grad_norm": 1.25, "learning_rate": 1.3572058850556423e-05, "loss": 0.0361, "step": 2628 }, { "epoch": 90.65517241379311, "grad_norm": 1.578125, "learning_rate": 1.3563364097780351e-05, "loss": 0.0165, "step": 2629 }, { "epoch": 90.6896551724138, "grad_norm": 5.0, "learning_rate": 1.3554668826733535e-05, "loss": 0.0238, "step": 2630 }, { "epoch": 90.72413793103448, "grad_norm": 3.03125, "learning_rate": 1.3545973041653948e-05, "loss": 0.046, "step": 2631 }, { "epoch": 90.75862068965517, "grad_norm": 0.27734375, "learning_rate": 1.35372767467798e-05, "loss": 0.0033, "step": 2632 }, { "epoch": 90.79310344827586, "grad_norm": 1.28125, "learning_rate": 1.352857994634957e-05, "loss": 0.014, "step": 2633 }, { "epoch": 90.82758620689656, "grad_norm": 0.86328125, "learning_rate": 1.3519882644601973e-05, "loss": 0.0041, "step": 2634 }, { "epoch": 90.86206896551724, "grad_norm": 0.84765625, "learning_rate": 1.3511184845775959e-05, "loss": 0.007, "step": 2635 }, { "epoch": 90.89655172413794, "grad_norm": 3.125, "learning_rate": 1.3502486554110736e-05, "loss": 0.0231, "step": 2636 }, { "epoch": 90.93103448275862, "grad_norm": 1.046875, "learning_rate": 1.3493787773845749e-05, "loss": 0.0078, "step": 2637 }, { "epoch": 90.96551724137932, "grad_norm": 0.20703125, "learning_rate": 1.348508850922067e-05, "loss": 0.0031, "step": 2638 }, { "epoch": 91.0, "grad_norm": 0.212890625, "learning_rate": 1.347638876447542e-05, "loss": 0.0193, "step": 2639 }, { "epoch": 91.03448275862068, "grad_norm": 3.546875, "learning_rate": 1.3467688543850149e-05, "loss": 0.0086, "step": 2640 }, { "epoch": 91.06896551724138, "grad_norm": 4.90625, "learning_rate": 1.3458987851585233e-05, "loss": 0.0408, "step": 2641 }, { "epoch": 91.10344827586206, "grad_norm": 0.248046875, "learning_rate": 1.3450286691921291e-05, "loss": 0.0085, "step": 2642 }, { "epoch": 91.13793103448276, "grad_norm": 0.271484375, "learning_rate": 1.3441585069099161e-05, "loss": 0.006, "step": 2643 }, { "epoch": 91.17241379310344, "grad_norm": 0.51953125, "learning_rate": 1.3432882987359905e-05, "loss": 0.0027, "step": 2644 }, { "epoch": 91.20689655172414, "grad_norm": 8.375, "learning_rate": 1.3424180450944815e-05, "loss": 0.0318, "step": 2645 }, { "epoch": 91.24137931034483, "grad_norm": 3.421875, "learning_rate": 1.3415477464095397e-05, "loss": 0.0352, "step": 2646 }, { "epoch": 91.27586206896552, "grad_norm": 5.375, "learning_rate": 1.3406774031053387e-05, "loss": 0.0269, "step": 2647 }, { "epoch": 91.3103448275862, "grad_norm": 0.412109375, "learning_rate": 1.3398070156060729e-05, "loss": 0.0058, "step": 2648 }, { "epoch": 91.34482758620689, "grad_norm": 1.1953125, "learning_rate": 1.3389365843359589e-05, "loss": 0.0135, "step": 2649 }, { "epoch": 91.37931034482759, "grad_norm": 2.21875, "learning_rate": 1.3380661097192337e-05, "loss": 0.0081, "step": 2650 }, { "epoch": 91.41379310344827, "grad_norm": 0.419921875, "learning_rate": 1.3371955921801566e-05, "loss": 0.0101, "step": 2651 }, { "epoch": 91.44827586206897, "grad_norm": 1.4609375, "learning_rate": 1.3363250321430077e-05, "loss": 0.0192, "step": 2652 }, { "epoch": 91.48275862068965, "grad_norm": 0.1953125, "learning_rate": 1.3354544300320859e-05, "loss": 0.0183, "step": 2653 }, { "epoch": 91.51724137931035, "grad_norm": 2.671875, "learning_rate": 1.3345837862717136e-05, "loss": 0.0199, "step": 2654 }, { "epoch": 91.55172413793103, "grad_norm": 1.9921875, "learning_rate": 1.3337131012862313e-05, "loss": 0.0293, "step": 2655 }, { "epoch": 91.58620689655173, "grad_norm": 0.89453125, "learning_rate": 1.3328423755000006e-05, "loss": 0.0059, "step": 2656 }, { "epoch": 91.62068965517241, "grad_norm": 8.0, "learning_rate": 1.3319716093374021e-05, "loss": 0.0542, "step": 2657 }, { "epoch": 91.65517241379311, "grad_norm": 13.1875, "learning_rate": 1.3311008032228369e-05, "loss": 0.0474, "step": 2658 }, { "epoch": 91.6896551724138, "grad_norm": 3.265625, "learning_rate": 1.3302299575807257e-05, "loss": 0.0249, "step": 2659 }, { "epoch": 91.72413793103448, "grad_norm": 0.259765625, "learning_rate": 1.3293590728355077e-05, "loss": 0.0042, "step": 2660 }, { "epoch": 91.75862068965517, "grad_norm": 0.4453125, "learning_rate": 1.328488149411642e-05, "loss": 0.0061, "step": 2661 }, { "epoch": 91.79310344827586, "grad_norm": 0.66796875, "learning_rate": 1.3276171877336051e-05, "loss": 0.0149, "step": 2662 }, { "epoch": 91.82758620689656, "grad_norm": 1.2265625, "learning_rate": 1.3267461882258945e-05, "loss": 0.0075, "step": 2663 }, { "epoch": 91.86206896551724, "grad_norm": 1.3515625, "learning_rate": 1.3258751513130239e-05, "loss": 0.011, "step": 2664 }, { "epoch": 91.89655172413794, "grad_norm": 4.375, "learning_rate": 1.325004077419526e-05, "loss": 0.0426, "step": 2665 }, { "epoch": 91.93103448275862, "grad_norm": 0.25390625, "learning_rate": 1.3241329669699523e-05, "loss": 0.0038, "step": 2666 }, { "epoch": 91.96551724137932, "grad_norm": 0.44140625, "learning_rate": 1.3232618203888705e-05, "loss": 0.0082, "step": 2667 }, { "epoch": 92.0, "grad_norm": 4.0, "learning_rate": 1.322390638100868e-05, "loss": 0.0209, "step": 2668 }, { "epoch": 92.03448275862068, "grad_norm": 7.03125, "learning_rate": 1.3215194205305478e-05, "loss": 0.0812, "step": 2669 }, { "epoch": 92.06896551724138, "grad_norm": 0.48828125, "learning_rate": 1.3206481681025309e-05, "loss": 0.0053, "step": 2670 }, { "epoch": 92.10344827586206, "grad_norm": 1.3046875, "learning_rate": 1.3197768812414548e-05, "loss": 0.043, "step": 2671 }, { "epoch": 92.13793103448276, "grad_norm": 6.03125, "learning_rate": 1.3189055603719747e-05, "loss": 0.0303, "step": 2672 }, { "epoch": 92.17241379310344, "grad_norm": 10.125, "learning_rate": 1.3180342059187614e-05, "loss": 0.0651, "step": 2673 }, { "epoch": 92.20689655172414, "grad_norm": 1.203125, "learning_rate": 1.3171628183065027e-05, "loss": 0.0078, "step": 2674 }, { "epoch": 92.24137931034483, "grad_norm": 4.09375, "learning_rate": 1.3162913979599022e-05, "loss": 0.0155, "step": 2675 }, { "epoch": 92.27586206896552, "grad_norm": 9.0625, "learning_rate": 1.31541994530368e-05, "loss": 0.0218, "step": 2676 }, { "epoch": 92.3103448275862, "grad_norm": 3.453125, "learning_rate": 1.3145484607625708e-05, "loss": 0.0695, "step": 2677 }, { "epoch": 92.34482758620689, "grad_norm": 0.1123046875, "learning_rate": 1.3136769447613261e-05, "loss": 0.0263, "step": 2678 }, { "epoch": 92.37931034482759, "grad_norm": 0.26171875, "learning_rate": 1.3128053977247121e-05, "loss": 0.0026, "step": 2679 }, { "epoch": 92.41379310344827, "grad_norm": 0.3046875, "learning_rate": 1.3119338200775103e-05, "loss": 0.0026, "step": 2680 }, { "epoch": 92.44827586206897, "grad_norm": 7.84375, "learning_rate": 1.3110622122445168e-05, "loss": 0.0296, "step": 2681 }, { "epoch": 92.48275862068965, "grad_norm": 1.125, "learning_rate": 1.3101905746505428e-05, "loss": 0.0086, "step": 2682 }, { "epoch": 92.51724137931035, "grad_norm": 0.94140625, "learning_rate": 1.3093189077204143e-05, "loss": 0.0039, "step": 2683 }, { "epoch": 92.55172413793103, "grad_norm": 2.765625, "learning_rate": 1.3084472118789703e-05, "loss": 0.0262, "step": 2684 }, { "epoch": 92.58620689655173, "grad_norm": 1.015625, "learning_rate": 1.3075754875510654e-05, "loss": 0.0051, "step": 2685 }, { "epoch": 92.62068965517241, "grad_norm": 0.341796875, "learning_rate": 1.3067037351615672e-05, "loss": 0.0045, "step": 2686 }, { "epoch": 92.65517241379311, "grad_norm": 0.1962890625, "learning_rate": 1.3058319551353578e-05, "loss": 0.0049, "step": 2687 }, { "epoch": 92.6896551724138, "grad_norm": 3.4375, "learning_rate": 1.3049601478973312e-05, "loss": 0.0206, "step": 2688 }, { "epoch": 92.72413793103448, "grad_norm": 0.7734375, "learning_rate": 1.3040883138723964e-05, "loss": 0.0323, "step": 2689 }, { "epoch": 92.75862068965517, "grad_norm": 9.0625, "learning_rate": 1.3032164534854747e-05, "loss": 0.0776, "step": 2690 }, { "epoch": 92.79310344827586, "grad_norm": 3.171875, "learning_rate": 1.3023445671614995e-05, "loss": 0.0301, "step": 2691 }, { "epoch": 92.82758620689656, "grad_norm": 0.578125, "learning_rate": 1.301472655325419e-05, "loss": 0.0066, "step": 2692 }, { "epoch": 92.86206896551724, "grad_norm": 7.0625, "learning_rate": 1.3006007184021907e-05, "loss": 0.0191, "step": 2693 }, { "epoch": 92.89655172413794, "grad_norm": 2.59375, "learning_rate": 1.2997287568167874e-05, "loss": 0.026, "step": 2694 }, { "epoch": 92.93103448275862, "grad_norm": 4.375, "learning_rate": 1.2988567709941921e-05, "loss": 0.0196, "step": 2695 }, { "epoch": 92.96551724137932, "grad_norm": 5.5625, "learning_rate": 1.2979847613594002e-05, "loss": 0.0332, "step": 2696 }, { "epoch": 93.0, "grad_norm": 1.1015625, "learning_rate": 1.2971127283374185e-05, "loss": 0.0104, "step": 2697 }, { "epoch": 93.03448275862068, "grad_norm": 3.96875, "learning_rate": 1.296240672353265e-05, "loss": 0.0537, "step": 2698 }, { "epoch": 93.06896551724138, "grad_norm": 4.46875, "learning_rate": 1.2953685938319699e-05, "loss": 0.0419, "step": 2699 }, { "epoch": 93.10344827586206, "grad_norm": 3.0625, "learning_rate": 1.2944964931985731e-05, "loss": 0.0236, "step": 2700 }, { "epoch": 93.13793103448276, "grad_norm": 6.71875, "learning_rate": 1.2936243708781266e-05, "loss": 0.0445, "step": 2701 }, { "epoch": 93.17241379310344, "grad_norm": 1.484375, "learning_rate": 1.2927522272956912e-05, "loss": 0.0349, "step": 2702 }, { "epoch": 93.20689655172414, "grad_norm": 1.171875, "learning_rate": 1.2918800628763394e-05, "loss": 0.0083, "step": 2703 }, { "epoch": 93.24137931034483, "grad_norm": 2.03125, "learning_rate": 1.291007878045154e-05, "loss": 0.0232, "step": 2704 }, { "epoch": 93.27586206896552, "grad_norm": 0.2412109375, "learning_rate": 1.2901356732272271e-05, "loss": 0.0145, "step": 2705 }, { "epoch": 93.3103448275862, "grad_norm": 3.296875, "learning_rate": 1.2892634488476605e-05, "loss": 0.0167, "step": 2706 }, { "epoch": 93.34482758620689, "grad_norm": 1.109375, "learning_rate": 1.288391205331566e-05, "loss": 0.0333, "step": 2707 }, { "epoch": 93.37931034482759, "grad_norm": 1.6015625, "learning_rate": 1.287518943104064e-05, "loss": 0.0083, "step": 2708 }, { "epoch": 93.41379310344827, "grad_norm": 12.25, "learning_rate": 1.2866466625902857e-05, "loss": 0.0194, "step": 2709 }, { "epoch": 93.44827586206897, "grad_norm": 1.5234375, "learning_rate": 1.2857743642153685e-05, "loss": 0.0171, "step": 2710 }, { "epoch": 93.48275862068965, "grad_norm": 0.28125, "learning_rate": 1.2849020484044613e-05, "loss": 0.0196, "step": 2711 }, { "epoch": 93.51724137931035, "grad_norm": 0.88671875, "learning_rate": 1.284029715582719e-05, "loss": 0.0102, "step": 2712 }, { "epoch": 93.55172413793103, "grad_norm": 1.0703125, "learning_rate": 1.2831573661753074e-05, "loss": 0.0089, "step": 2713 }, { "epoch": 93.58620689655173, "grad_norm": 0.7890625, "learning_rate": 1.2822850006073986e-05, "loss": 0.0068, "step": 2714 }, { "epoch": 93.62068965517241, "grad_norm": 1.3515625, "learning_rate": 1.281412619304172e-05, "loss": 0.0093, "step": 2715 }, { "epoch": 93.65517241379311, "grad_norm": 4.78125, "learning_rate": 1.2805402226908166e-05, "loss": 0.0317, "step": 2716 }, { "epoch": 93.6896551724138, "grad_norm": 1.7890625, "learning_rate": 1.2796678111925276e-05, "loss": 0.0146, "step": 2717 }, { "epoch": 93.72413793103448, "grad_norm": 4.65625, "learning_rate": 1.278795385234508e-05, "loss": 0.041, "step": 2718 }, { "epoch": 93.75862068965517, "grad_norm": 0.94921875, "learning_rate": 1.2779229452419673e-05, "loss": 0.0073, "step": 2719 }, { "epoch": 93.79310344827586, "grad_norm": 1.9375, "learning_rate": 1.2770504916401224e-05, "loss": 0.0195, "step": 2720 }, { "epoch": 93.82758620689656, "grad_norm": 6.75, "learning_rate": 1.2761780248541963e-05, "loss": 0.0384, "step": 2721 }, { "epoch": 93.86206896551724, "grad_norm": 2.25, "learning_rate": 1.275305545309419e-05, "loss": 0.0119, "step": 2722 }, { "epoch": 93.89655172413794, "grad_norm": 1.09375, "learning_rate": 1.2744330534310267e-05, "loss": 0.012, "step": 2723 }, { "epoch": 93.93103448275862, "grad_norm": 1.8828125, "learning_rate": 1.2735605496442604e-05, "loss": 0.013, "step": 2724 }, { "epoch": 93.96551724137932, "grad_norm": 13.375, "learning_rate": 1.2726880343743685e-05, "loss": 0.0353, "step": 2725 }, { "epoch": 94.0, "grad_norm": 0.36328125, "learning_rate": 1.2718155080466046e-05, "loss": 0.0037, "step": 2726 }, { "epoch": 94.03448275862068, "grad_norm": 0.6171875, "learning_rate": 1.2709429710862269e-05, "loss": 0.0253, "step": 2727 }, { "epoch": 94.06896551724138, "grad_norm": 6.59375, "learning_rate": 1.2700704239184995e-05, "loss": 0.0335, "step": 2728 }, { "epoch": 94.10344827586206, "grad_norm": 3.765625, "learning_rate": 1.2691978669686913e-05, "loss": 0.0495, "step": 2729 }, { "epoch": 94.13793103448276, "grad_norm": 0.8203125, "learning_rate": 1.268325300662076e-05, "loss": 0.0046, "step": 2730 }, { "epoch": 94.17241379310344, "grad_norm": 3.96875, "learning_rate": 1.2674527254239318e-05, "loss": 0.0355, "step": 2731 }, { "epoch": 94.20689655172414, "grad_norm": 1.921875, "learning_rate": 1.2665801416795412e-05, "loss": 0.016, "step": 2732 }, { "epoch": 94.24137931034483, "grad_norm": 1.171875, "learning_rate": 1.265707549854191e-05, "loss": 0.0112, "step": 2733 }, { "epoch": 94.27586206896552, "grad_norm": 1.3359375, "learning_rate": 1.2648349503731716e-05, "loss": 0.0082, "step": 2734 }, { "epoch": 94.3103448275862, "grad_norm": 2.84375, "learning_rate": 1.263962343661778e-05, "loss": 0.0139, "step": 2735 }, { "epoch": 94.34482758620689, "grad_norm": 3.40625, "learning_rate": 1.2630897301453073e-05, "loss": 0.0416, "step": 2736 }, { "epoch": 94.37931034482759, "grad_norm": 1.5390625, "learning_rate": 1.2622171102490615e-05, "loss": 0.0172, "step": 2737 }, { "epoch": 94.41379310344827, "grad_norm": 2.890625, "learning_rate": 1.2613444843983441e-05, "loss": 0.0215, "step": 2738 }, { "epoch": 94.44827586206897, "grad_norm": 0.53125, "learning_rate": 1.2604718530184634e-05, "loss": 0.043, "step": 2739 }, { "epoch": 94.48275862068965, "grad_norm": 0.49609375, "learning_rate": 1.2595992165347287e-05, "loss": 0.0109, "step": 2740 }, { "epoch": 94.51724137931035, "grad_norm": 0.76171875, "learning_rate": 1.258726575372452e-05, "loss": 0.0087, "step": 2741 }, { "epoch": 94.55172413793103, "grad_norm": 0.3671875, "learning_rate": 1.2578539299569488e-05, "loss": 0.0307, "step": 2742 }, { "epoch": 94.58620689655173, "grad_norm": 3.609375, "learning_rate": 1.2569812807135353e-05, "loss": 0.0401, "step": 2743 }, { "epoch": 94.62068965517241, "grad_norm": 0.38671875, "learning_rate": 1.2561086280675308e-05, "loss": 0.0051, "step": 2744 }, { "epoch": 94.65517241379311, "grad_norm": 1.03125, "learning_rate": 1.255235972444255e-05, "loss": 0.0072, "step": 2745 }, { "epoch": 94.6896551724138, "grad_norm": 0.8515625, "learning_rate": 1.2543633142690298e-05, "loss": 0.0056, "step": 2746 }, { "epoch": 94.72413793103448, "grad_norm": 2.140625, "learning_rate": 1.2534906539671785e-05, "loss": 0.0392, "step": 2747 }, { "epoch": 94.75862068965517, "grad_norm": 2.5625, "learning_rate": 1.252617991964024e-05, "loss": 0.0078, "step": 2748 }, { "epoch": 94.79310344827586, "grad_norm": 1.5546875, "learning_rate": 1.2517453286848932e-05, "loss": 0.0127, "step": 2749 }, { "epoch": 94.82758620689656, "grad_norm": 0.81640625, "learning_rate": 1.2508726645551094e-05, "loss": 0.0063, "step": 2750 }, { "epoch": 94.86206896551724, "grad_norm": 0.6953125, "learning_rate": 1.25e-05, "loss": 0.0046, "step": 2751 }, { "epoch": 94.89655172413794, "grad_norm": 1.140625, "learning_rate": 1.2491273354448907e-05, "loss": 0.0072, "step": 2752 }, { "epoch": 94.93103448275862, "grad_norm": 3.34375, "learning_rate": 1.248254671315107e-05, "loss": 0.0213, "step": 2753 }, { "epoch": 94.96551724137932, "grad_norm": 0.470703125, "learning_rate": 1.2473820080359758e-05, "loss": 0.003, "step": 2754 }, { "epoch": 95.0, "grad_norm": 0.73828125, "learning_rate": 1.246509346032822e-05, "loss": 0.0161, "step": 2755 }, { "epoch": 95.03448275862068, "grad_norm": 2.3125, "learning_rate": 1.2456366857309709e-05, "loss": 0.0177, "step": 2756 }, { "epoch": 95.06896551724138, "grad_norm": 2.640625, "learning_rate": 1.2447640275557453e-05, "loss": 0.0424, "step": 2757 }, { "epoch": 95.10344827586206, "grad_norm": 1.125, "learning_rate": 1.2438913719324696e-05, "loss": 0.0214, "step": 2758 }, { "epoch": 95.13793103448276, "grad_norm": 4.0625, "learning_rate": 1.243018719286465e-05, "loss": 0.0267, "step": 2759 }, { "epoch": 95.17241379310344, "grad_norm": 4.875, "learning_rate": 1.2421460700430514e-05, "loss": 0.0186, "step": 2760 }, { "epoch": 95.20689655172414, "grad_norm": 0.8984375, "learning_rate": 1.2412734246275482e-05, "loss": 0.0101, "step": 2761 }, { "epoch": 95.24137931034483, "grad_norm": 0.84375, "learning_rate": 1.2404007834652716e-05, "loss": 0.012, "step": 2762 }, { "epoch": 95.27586206896552, "grad_norm": 0.7578125, "learning_rate": 1.2395281469815367e-05, "loss": 0.0042, "step": 2763 }, { "epoch": 95.3103448275862, "grad_norm": 1.6171875, "learning_rate": 1.2386555156016558e-05, "loss": 0.0303, "step": 2764 }, { "epoch": 95.34482758620689, "grad_norm": 1.1953125, "learning_rate": 1.237782889750939e-05, "loss": 0.0291, "step": 2765 }, { "epoch": 95.37931034482759, "grad_norm": 2.671875, "learning_rate": 1.2369102698546932e-05, "loss": 0.0123, "step": 2766 }, { "epoch": 95.41379310344827, "grad_norm": 1.2890625, "learning_rate": 1.2360376563382224e-05, "loss": 0.0325, "step": 2767 }, { "epoch": 95.44827586206897, "grad_norm": 2.953125, "learning_rate": 1.2351650496268287e-05, "loss": 0.0215, "step": 2768 }, { "epoch": 95.48275862068965, "grad_norm": 1.875, "learning_rate": 1.2342924501458091e-05, "loss": 0.0103, "step": 2769 }, { "epoch": 95.51724137931035, "grad_norm": 0.96875, "learning_rate": 1.233419858320459e-05, "loss": 0.0432, "step": 2770 }, { "epoch": 95.55172413793103, "grad_norm": 0.56640625, "learning_rate": 1.2325472745760686e-05, "loss": 0.005, "step": 2771 }, { "epoch": 95.58620689655173, "grad_norm": 0.51171875, "learning_rate": 1.2316746993379242e-05, "loss": 0.0031, "step": 2772 }, { "epoch": 95.62068965517241, "grad_norm": 6.75, "learning_rate": 1.2308021330313089e-05, "loss": 0.0652, "step": 2773 }, { "epoch": 95.65517241379311, "grad_norm": 0.31640625, "learning_rate": 1.229929576081501e-05, "loss": 0.0028, "step": 2774 }, { "epoch": 95.6896551724138, "grad_norm": 0.349609375, "learning_rate": 1.2290570289137736e-05, "loss": 0.003, "step": 2775 }, { "epoch": 95.72413793103448, "grad_norm": 5.34375, "learning_rate": 1.2281844919533957e-05, "loss": 0.0081, "step": 2776 }, { "epoch": 95.75862068965517, "grad_norm": 0.4921875, "learning_rate": 1.2273119656256318e-05, "loss": 0.0175, "step": 2777 }, { "epoch": 95.79310344827586, "grad_norm": 2.25, "learning_rate": 1.22643945035574e-05, "loss": 0.0264, "step": 2778 }, { "epoch": 95.82758620689656, "grad_norm": 0.39453125, "learning_rate": 1.2255669465689736e-05, "loss": 0.0121, "step": 2779 }, { "epoch": 95.86206896551724, "grad_norm": 0.96484375, "learning_rate": 1.2246944546905811e-05, "loss": 0.0119, "step": 2780 }, { "epoch": 95.89655172413794, "grad_norm": 0.232421875, "learning_rate": 1.223821975145804e-05, "loss": 0.002, "step": 2781 }, { "epoch": 95.93103448275862, "grad_norm": 0.51171875, "learning_rate": 1.2229495083598777e-05, "loss": 0.029, "step": 2782 }, { "epoch": 95.96551724137932, "grad_norm": 0.98828125, "learning_rate": 1.222077054758033e-05, "loss": 0.0111, "step": 2783 }, { "epoch": 96.0, "grad_norm": 0.240234375, "learning_rate": 1.2212046147654926e-05, "loss": 0.0028, "step": 2784 }, { "epoch": 96.03448275862068, "grad_norm": 0.2216796875, "learning_rate": 1.2203321888074729e-05, "loss": 0.0268, "step": 2785 }, { "epoch": 96.06896551724138, "grad_norm": 2.375, "learning_rate": 1.2194597773091837e-05, "loss": 0.009, "step": 2786 }, { "epoch": 96.10344827586206, "grad_norm": 0.3828125, "learning_rate": 1.2185873806958283e-05, "loss": 0.01, "step": 2787 }, { "epoch": 96.13793103448276, "grad_norm": 5.40625, "learning_rate": 1.2177149993926017e-05, "loss": 0.0257, "step": 2788 }, { "epoch": 96.17241379310344, "grad_norm": 0.1533203125, "learning_rate": 1.2168426338246924e-05, "loss": 0.0083, "step": 2789 }, { "epoch": 96.20689655172414, "grad_norm": 2.5625, "learning_rate": 1.2159702844172807e-05, "loss": 0.0169, "step": 2790 }, { "epoch": 96.24137931034483, "grad_norm": 2.265625, "learning_rate": 1.2150979515955392e-05, "loss": 0.0131, "step": 2791 }, { "epoch": 96.27586206896552, "grad_norm": 2.078125, "learning_rate": 1.2142256357846318e-05, "loss": 0.0352, "step": 2792 }, { "epoch": 96.3103448275862, "grad_norm": 4.90625, "learning_rate": 1.2133533374097148e-05, "loss": 0.028, "step": 2793 }, { "epoch": 96.34482758620689, "grad_norm": 2.484375, "learning_rate": 1.2124810568959361e-05, "loss": 0.0279, "step": 2794 }, { "epoch": 96.37931034482759, "grad_norm": 4.84375, "learning_rate": 1.211608794668434e-05, "loss": 0.047, "step": 2795 }, { "epoch": 96.41379310344827, "grad_norm": 1.1484375, "learning_rate": 1.2107365511523396e-05, "loss": 0.0156, "step": 2796 }, { "epoch": 96.44827586206897, "grad_norm": 0.47265625, "learning_rate": 1.2098643267727732e-05, "loss": 0.0066, "step": 2797 }, { "epoch": 96.48275862068965, "grad_norm": 2.71875, "learning_rate": 1.208992121954846e-05, "loss": 0.0209, "step": 2798 }, { "epoch": 96.51724137931035, "grad_norm": 3.453125, "learning_rate": 1.2081199371236607e-05, "loss": 0.0154, "step": 2799 }, { "epoch": 96.55172413793103, "grad_norm": 1.65625, "learning_rate": 1.2072477727043094e-05, "loss": 0.0151, "step": 2800 }, { "epoch": 96.58620689655173, "grad_norm": 2.890625, "learning_rate": 1.2063756291218742e-05, "loss": 0.0495, "step": 2801 }, { "epoch": 96.62068965517241, "grad_norm": 0.65234375, "learning_rate": 1.205503506801427e-05, "loss": 0.0036, "step": 2802 }, { "epoch": 96.65517241379311, "grad_norm": 0.890625, "learning_rate": 1.2046314061680304e-05, "loss": 0.014, "step": 2803 }, { "epoch": 96.6896551724138, "grad_norm": 2.125, "learning_rate": 1.2037593276467353e-05, "loss": 0.0156, "step": 2804 }, { "epoch": 96.72413793103448, "grad_norm": 0.7578125, "learning_rate": 1.2028872716625817e-05, "loss": 0.0059, "step": 2805 }, { "epoch": 96.75862068965517, "grad_norm": 0.640625, "learning_rate": 1.2020152386406001e-05, "loss": 0.0198, "step": 2806 }, { "epoch": 96.79310344827586, "grad_norm": 0.63671875, "learning_rate": 1.2011432290058082e-05, "loss": 0.0362, "step": 2807 }, { "epoch": 96.82758620689656, "grad_norm": 1.59375, "learning_rate": 1.2002712431832127e-05, "loss": 0.0087, "step": 2808 }, { "epoch": 96.86206896551724, "grad_norm": 2.59375, "learning_rate": 1.1993992815978096e-05, "loss": 0.0224, "step": 2809 }, { "epoch": 96.89655172413794, "grad_norm": 0.7109375, "learning_rate": 1.1985273446745818e-05, "loss": 0.0068, "step": 2810 }, { "epoch": 96.93103448275862, "grad_norm": 3.4375, "learning_rate": 1.1976554328385008e-05, "loss": 0.0137, "step": 2811 }, { "epoch": 96.96551724137932, "grad_norm": 2.046875, "learning_rate": 1.1967835465145256e-05, "loss": 0.0232, "step": 2812 }, { "epoch": 97.0, "grad_norm": 3.140625, "learning_rate": 1.195911686127604e-05, "loss": 0.0323, "step": 2813 }, { "epoch": 97.03448275862068, "grad_norm": 0.1337890625, "learning_rate": 1.1950398521026691e-05, "loss": 0.0204, "step": 2814 }, { "epoch": 97.06896551724138, "grad_norm": 0.71484375, "learning_rate": 1.1941680448646425e-05, "loss": 0.0227, "step": 2815 }, { "epoch": 97.10344827586206, "grad_norm": 4.34375, "learning_rate": 1.193296264838433e-05, "loss": 0.025, "step": 2816 }, { "epoch": 97.13793103448276, "grad_norm": 2.171875, "learning_rate": 1.1924245124489345e-05, "loss": 0.0168, "step": 2817 }, { "epoch": 97.17241379310344, "grad_norm": 0.88671875, "learning_rate": 1.1915527881210303e-05, "loss": 0.0182, "step": 2818 }, { "epoch": 97.20689655172414, "grad_norm": 2.734375, "learning_rate": 1.1906810922795863e-05, "loss": 0.0331, "step": 2819 }, { "epoch": 97.24137931034483, "grad_norm": 0.265625, "learning_rate": 1.1898094253494576e-05, "loss": 0.0017, "step": 2820 }, { "epoch": 97.27586206896552, "grad_norm": 2.15625, "learning_rate": 1.1889377877554835e-05, "loss": 0.0129, "step": 2821 }, { "epoch": 97.3103448275862, "grad_norm": 1.625, "learning_rate": 1.1880661799224901e-05, "loss": 0.0104, "step": 2822 }, { "epoch": 97.34482758620689, "grad_norm": 2.703125, "learning_rate": 1.1871946022752882e-05, "loss": 0.0144, "step": 2823 }, { "epoch": 97.37931034482759, "grad_norm": 1.0546875, "learning_rate": 1.186323055238674e-05, "loss": 0.0142, "step": 2824 }, { "epoch": 97.41379310344827, "grad_norm": 0.328125, "learning_rate": 1.1854515392374293e-05, "loss": 0.0185, "step": 2825 }, { "epoch": 97.44827586206897, "grad_norm": 0.328125, "learning_rate": 1.1845800546963202e-05, "loss": 0.0045, "step": 2826 }, { "epoch": 97.48275862068965, "grad_norm": 9.0, "learning_rate": 1.183708602040098e-05, "loss": 0.0273, "step": 2827 }, { "epoch": 97.51724137931035, "grad_norm": 4.90625, "learning_rate": 1.1828371816934974e-05, "loss": 0.0398, "step": 2828 }, { "epoch": 97.55172413793103, "grad_norm": 2.609375, "learning_rate": 1.1819657940812388e-05, "loss": 0.0228, "step": 2829 }, { "epoch": 97.58620689655173, "grad_norm": 2.0625, "learning_rate": 1.1810944396280257e-05, "loss": 0.0069, "step": 2830 }, { "epoch": 97.62068965517241, "grad_norm": 5.71875, "learning_rate": 1.1802231187585453e-05, "loss": 0.0432, "step": 2831 }, { "epoch": 97.65517241379311, "grad_norm": 0.87109375, "learning_rate": 1.1793518318974694e-05, "loss": 0.0084, "step": 2832 }, { "epoch": 97.6896551724138, "grad_norm": 2.578125, "learning_rate": 1.1784805794694524e-05, "loss": 0.022, "step": 2833 }, { "epoch": 97.72413793103448, "grad_norm": 1.25, "learning_rate": 1.177609361899132e-05, "loss": 0.023, "step": 2834 }, { "epoch": 97.75862068965517, "grad_norm": 0.98046875, "learning_rate": 1.1767381796111293e-05, "loss": 0.0329, "step": 2835 }, { "epoch": 97.79310344827586, "grad_norm": 1.5390625, "learning_rate": 1.1758670330300482e-05, "loss": 0.0409, "step": 2836 }, { "epoch": 97.82758620689656, "grad_norm": 0.7890625, "learning_rate": 1.1749959225804746e-05, "loss": 0.0044, "step": 2837 }, { "epoch": 97.86206896551724, "grad_norm": 3.296875, "learning_rate": 1.1741248486869764e-05, "loss": 0.0682, "step": 2838 }, { "epoch": 97.89655172413794, "grad_norm": 0.462890625, "learning_rate": 1.1732538117741058e-05, "loss": 0.0108, "step": 2839 }, { "epoch": 97.93103448275862, "grad_norm": 0.1171875, "learning_rate": 1.172382812266395e-05, "loss": 0.0237, "step": 2840 }, { "epoch": 97.96551724137932, "grad_norm": 0.72265625, "learning_rate": 1.1715118505883584e-05, "loss": 0.017, "step": 2841 }, { "epoch": 98.0, "grad_norm": 0.63671875, "learning_rate": 1.1706409271644925e-05, "loss": 0.007, "step": 2842 }, { "epoch": 98.03448275862068, "grad_norm": 1.71875, "learning_rate": 1.1697700424192743e-05, "loss": 0.0117, "step": 2843 }, { "epoch": 98.06896551724138, "grad_norm": 0.2158203125, "learning_rate": 1.1688991967771632e-05, "loss": 0.006, "step": 2844 }, { "epoch": 98.10344827586206, "grad_norm": 0.7109375, "learning_rate": 1.1680283906625985e-05, "loss": 0.0162, "step": 2845 }, { "epoch": 98.13793103448276, "grad_norm": 0.59765625, "learning_rate": 1.1671576245000002e-05, "loss": 0.0045, "step": 2846 }, { "epoch": 98.17241379310344, "grad_norm": 0.318359375, "learning_rate": 1.166286898713769e-05, "loss": 0.0134, "step": 2847 }, { "epoch": 98.20689655172414, "grad_norm": 2.421875, "learning_rate": 1.1654162137282867e-05, "loss": 0.0253, "step": 2848 }, { "epoch": 98.24137931034483, "grad_norm": 1.8515625, "learning_rate": 1.1645455699679143e-05, "loss": 0.0231, "step": 2849 }, { "epoch": 98.27586206896552, "grad_norm": 0.62890625, "learning_rate": 1.1636749678569928e-05, "loss": 0.0114, "step": 2850 }, { "epoch": 98.3103448275862, "grad_norm": 1.3359375, "learning_rate": 1.1628044078198434e-05, "loss": 0.0068, "step": 2851 }, { "epoch": 98.34482758620689, "grad_norm": 0.54296875, "learning_rate": 1.1619338902807664e-05, "loss": 0.0048, "step": 2852 }, { "epoch": 98.37931034482759, "grad_norm": 3.5625, "learning_rate": 1.1610634156640419e-05, "loss": 0.0141, "step": 2853 }, { "epoch": 98.41379310344827, "grad_norm": 7.21875, "learning_rate": 1.1601929843939274e-05, "loss": 0.0311, "step": 2854 }, { "epoch": 98.44827586206897, "grad_norm": 0.2412109375, "learning_rate": 1.1593225968946617e-05, "loss": 0.004, "step": 2855 }, { "epoch": 98.48275862068965, "grad_norm": 5.65625, "learning_rate": 1.1584522535904607e-05, "loss": 0.0309, "step": 2856 }, { "epoch": 98.51724137931035, "grad_norm": 1.75, "learning_rate": 1.1575819549055188e-05, "loss": 0.0225, "step": 2857 }, { "epoch": 98.55172413793103, "grad_norm": 1.8828125, "learning_rate": 1.1567117012640098e-05, "loss": 0.008, "step": 2858 }, { "epoch": 98.58620689655173, "grad_norm": 3.59375, "learning_rate": 1.1558414930900843e-05, "loss": 0.0441, "step": 2859 }, { "epoch": 98.62068965517241, "grad_norm": 1.234375, "learning_rate": 1.154971330807871e-05, "loss": 0.0091, "step": 2860 }, { "epoch": 98.65517241379311, "grad_norm": 4.375, "learning_rate": 1.1541012148414768e-05, "loss": 0.038, "step": 2861 }, { "epoch": 98.6896551724138, "grad_norm": 2.625, "learning_rate": 1.1532311456149856e-05, "loss": 0.0119, "step": 2862 }, { "epoch": 98.72413793103448, "grad_norm": 2.1875, "learning_rate": 1.1523611235524585e-05, "loss": 0.0143, "step": 2863 }, { "epoch": 98.75862068965517, "grad_norm": 2.015625, "learning_rate": 1.1514911490779333e-05, "loss": 0.0237, "step": 2864 }, { "epoch": 98.79310344827586, "grad_norm": 3.984375, "learning_rate": 1.1506212226154254e-05, "loss": 0.0548, "step": 2865 }, { "epoch": 98.82758620689656, "grad_norm": 0.28125, "learning_rate": 1.1497513445889265e-05, "loss": 0.0134, "step": 2866 }, { "epoch": 98.86206896551724, "grad_norm": 0.609375, "learning_rate": 1.1488815154224042e-05, "loss": 0.0085, "step": 2867 }, { "epoch": 98.89655172413794, "grad_norm": 1.1875, "learning_rate": 1.1480117355398032e-05, "loss": 0.0249, "step": 2868 }, { "epoch": 98.93103448275862, "grad_norm": 6.90625, "learning_rate": 1.147142005365043e-05, "loss": 0.032, "step": 2869 }, { "epoch": 98.96551724137932, "grad_norm": 0.4375, "learning_rate": 1.14627232532202e-05, "loss": 0.0075, "step": 2870 }, { "epoch": 99.0, "grad_norm": 0.369140625, "learning_rate": 1.1454026958346058e-05, "loss": 0.0069, "step": 2871 }, { "epoch": 99.03448275862068, "grad_norm": 0.796875, "learning_rate": 1.1445331173266468e-05, "loss": 0.0046, "step": 2872 }, { "epoch": 99.06896551724138, "grad_norm": 2.859375, "learning_rate": 1.143663590221965e-05, "loss": 0.0449, "step": 2873 }, { "epoch": 99.10344827586206, "grad_norm": 1.640625, "learning_rate": 1.1427941149443578e-05, "loss": 0.0111, "step": 2874 }, { "epoch": 99.13793103448276, "grad_norm": 0.51171875, "learning_rate": 1.1419246919175966e-05, "loss": 0.0047, "step": 2875 }, { "epoch": 99.17241379310344, "grad_norm": 2.34375, "learning_rate": 1.1410553215654272e-05, "loss": 0.0234, "step": 2876 }, { "epoch": 99.20689655172414, "grad_norm": 0.875, "learning_rate": 1.140186004311571e-05, "loss": 0.0075, "step": 2877 }, { "epoch": 99.24137931034483, "grad_norm": 0.38671875, "learning_rate": 1.1393167405797223e-05, "loss": 0.0153, "step": 2878 }, { "epoch": 99.27586206896552, "grad_norm": 0.35546875, "learning_rate": 1.1384475307935493e-05, "loss": 0.0139, "step": 2879 }, { "epoch": 99.3103448275862, "grad_norm": 4.875, "learning_rate": 1.1375783753766953e-05, "loss": 0.0473, "step": 2880 }, { "epoch": 99.34482758620689, "grad_norm": 0.1572265625, "learning_rate": 1.1367092747527752e-05, "loss": 0.0025, "step": 2881 }, { "epoch": 99.37931034482759, "grad_norm": 0.8984375, "learning_rate": 1.135840229345378e-05, "loss": 0.0052, "step": 2882 }, { "epoch": 99.41379310344827, "grad_norm": 0.6796875, "learning_rate": 1.1349712395780663e-05, "loss": 0.0061, "step": 2883 }, { "epoch": 99.44827586206897, "grad_norm": 1.34375, "learning_rate": 1.1341023058743749e-05, "loss": 0.014, "step": 2884 }, { "epoch": 99.48275862068965, "grad_norm": 1.8359375, "learning_rate": 1.133233428657812e-05, "loss": 0.0062, "step": 2885 }, { "epoch": 99.51724137931035, "grad_norm": 1.4296875, "learning_rate": 1.1323646083518572e-05, "loss": 0.0159, "step": 2886 }, { "epoch": 99.55172413793103, "grad_norm": 0.62109375, "learning_rate": 1.1314958453799634e-05, "loss": 0.0322, "step": 2887 }, { "epoch": 99.58620689655173, "grad_norm": 3.625, "learning_rate": 1.1306271401655547e-05, "loss": 0.0314, "step": 2888 }, { "epoch": 99.62068965517241, "grad_norm": 0.59765625, "learning_rate": 1.1297584931320284e-05, "loss": 0.0036, "step": 2889 }, { "epoch": 99.65517241379311, "grad_norm": 1.578125, "learning_rate": 1.1288899047027512e-05, "loss": 0.0168, "step": 2890 }, { "epoch": 99.6896551724138, "grad_norm": 0.412109375, "learning_rate": 1.1280213753010633e-05, "loss": 0.0061, "step": 2891 }, { "epoch": 99.72413793103448, "grad_norm": 1.328125, "learning_rate": 1.1271529053502755e-05, "loss": 0.0211, "step": 2892 }, { "epoch": 99.75862068965517, "grad_norm": 0.162109375, "learning_rate": 1.1262844952736688e-05, "loss": 0.0031, "step": 2893 }, { "epoch": 99.79310344827586, "grad_norm": 0.35546875, "learning_rate": 1.1254161454944965e-05, "loss": 0.0031, "step": 2894 }, { "epoch": 99.82758620689656, "grad_norm": 0.96484375, "learning_rate": 1.1245478564359812e-05, "loss": 0.0069, "step": 2895 }, { "epoch": 99.86206896551724, "grad_norm": 2.6875, "learning_rate": 1.123679628521317e-05, "loss": 0.0178, "step": 2896 }, { "epoch": 99.89655172413794, "grad_norm": 0.8984375, "learning_rate": 1.1228114621736673e-05, "loss": 0.0065, "step": 2897 }, { "epoch": 99.93103448275862, "grad_norm": 1.0546875, "learning_rate": 1.121943357816166e-05, "loss": 0.0066, "step": 2898 }, { "epoch": 99.96551724137932, "grad_norm": 1.2421875, "learning_rate": 1.1210753158719162e-05, "loss": 0.0055, "step": 2899 }, { "epoch": 100.0, "grad_norm": 1.203125, "learning_rate": 1.1202073367639911e-05, "loss": 0.0055, "step": 2900 }, { "epoch": 100.03448275862068, "grad_norm": 3.625, "learning_rate": 1.1193394209154334e-05, "loss": 0.0224, "step": 2901 }, { "epoch": 100.06896551724138, "grad_norm": 4.09375, "learning_rate": 1.1184715687492544e-05, "loss": 0.0556, "step": 2902 }, { "epoch": 100.10344827586206, "grad_norm": 0.283203125, "learning_rate": 1.1176037806884354e-05, "loss": 0.0153, "step": 2903 }, { "epoch": 100.13793103448276, "grad_norm": 0.765625, "learning_rate": 1.1167360571559252e-05, "loss": 0.0144, "step": 2904 }, { "epoch": 100.17241379310344, "grad_norm": 1.859375, "learning_rate": 1.1158683985746418e-05, "loss": 0.0118, "step": 2905 }, { "epoch": 100.20689655172414, "grad_norm": 6.625, "learning_rate": 1.1150008053674715e-05, "loss": 0.0411, "step": 2906 }, { "epoch": 100.24137931034483, "grad_norm": 2.640625, "learning_rate": 1.1141332779572689e-05, "loss": 0.0096, "step": 2907 }, { "epoch": 100.27586206896552, "grad_norm": 1.921875, "learning_rate": 1.113265816766856e-05, "loss": 0.026, "step": 2908 }, { "epoch": 100.3103448275862, "grad_norm": 1.390625, "learning_rate": 1.1123984222190225e-05, "loss": 0.0109, "step": 2909 }, { "epoch": 100.34482758620689, "grad_norm": 0.419921875, "learning_rate": 1.1115310947365268e-05, "loss": 0.0029, "step": 2910 }, { "epoch": 100.37931034482759, "grad_norm": 2.359375, "learning_rate": 1.1106638347420933e-05, "loss": 0.0201, "step": 2911 }, { "epoch": 100.41379310344827, "grad_norm": 0.20703125, "learning_rate": 1.1097966426584139e-05, "loss": 0.0169, "step": 2912 }, { "epoch": 100.44827586206897, "grad_norm": 0.4609375, "learning_rate": 1.108929518908148e-05, "loss": 0.0114, "step": 2913 }, { "epoch": 100.48275862068965, "grad_norm": 0.9375, "learning_rate": 1.1080624639139203e-05, "loss": 0.0034, "step": 2914 }, { "epoch": 100.51724137931035, "grad_norm": 0.66796875, "learning_rate": 1.1071954780983245e-05, "loss": 0.0552, "step": 2915 }, { "epoch": 100.55172413793103, "grad_norm": 1.9609375, "learning_rate": 1.106328561883917e-05, "loss": 0.017, "step": 2916 }, { "epoch": 100.58620689655173, "grad_norm": 1.2109375, "learning_rate": 1.1054617156932233e-05, "loss": 0.0171, "step": 2917 }, { "epoch": 100.62068965517241, "grad_norm": 2.984375, "learning_rate": 1.1045949399487334e-05, "loss": 0.0531, "step": 2918 }, { "epoch": 100.65517241379311, "grad_norm": 0.349609375, "learning_rate": 1.1037282350729033e-05, "loss": 0.003, "step": 2919 }, { "epoch": 100.6896551724138, "grad_norm": 1.171875, "learning_rate": 1.1028616014881546e-05, "loss": 0.0159, "step": 2920 }, { "epoch": 100.72413793103448, "grad_norm": 3.015625, "learning_rate": 1.1019950396168737e-05, "loss": 0.0342, "step": 2921 }, { "epoch": 100.75862068965517, "grad_norm": 1.7421875, "learning_rate": 1.1011285498814128e-05, "loss": 0.0135, "step": 2922 }, { "epoch": 100.79310344827586, "grad_norm": 2.109375, "learning_rate": 1.100262132704088e-05, "loss": 0.0213, "step": 2923 }, { "epoch": 100.82758620689656, "grad_norm": 1.265625, "learning_rate": 1.0993957885071812e-05, "loss": 0.0046, "step": 2924 }, { "epoch": 100.86206896551724, "grad_norm": 2.359375, "learning_rate": 1.0985295177129373e-05, "loss": 0.0122, "step": 2925 }, { "epoch": 100.89655172413794, "grad_norm": 1.5703125, "learning_rate": 1.097663320743566e-05, "loss": 0.0161, "step": 2926 }, { "epoch": 100.93103448275862, "grad_norm": 0.93359375, "learning_rate": 1.0967971980212417e-05, "loss": 0.029, "step": 2927 }, { "epoch": 100.96551724137932, "grad_norm": 1.390625, "learning_rate": 1.0959311499681018e-05, "loss": 0.0104, "step": 2928 }, { "epoch": 101.0, "grad_norm": 4.65625, "learning_rate": 1.0950651770062478e-05, "loss": 0.0384, "step": 2929 }, { "epoch": 101.03448275862068, "grad_norm": 1.0625, "learning_rate": 1.0941992795577447e-05, "loss": 0.0073, "step": 2930 }, { "epoch": 101.06896551724138, "grad_norm": 1.1484375, "learning_rate": 1.0933334580446195e-05, "loss": 0.0086, "step": 2931 }, { "epoch": 101.10344827586206, "grad_norm": 2.265625, "learning_rate": 1.092467712888864e-05, "loss": 0.0196, "step": 2932 }, { "epoch": 101.13793103448276, "grad_norm": 0.91015625, "learning_rate": 1.091602044512432e-05, "loss": 0.0075, "step": 2933 }, { "epoch": 101.17241379310344, "grad_norm": 0.76953125, "learning_rate": 1.0907364533372388e-05, "loss": 0.0149, "step": 2934 }, { "epoch": 101.20689655172414, "grad_norm": 0.265625, "learning_rate": 1.0898709397851634e-05, "loss": 0.0043, "step": 2935 }, { "epoch": 101.24137931034483, "grad_norm": 1.421875, "learning_rate": 1.0890055042780467e-05, "loss": 0.0082, "step": 2936 }, { "epoch": 101.27586206896552, "grad_norm": 0.490234375, "learning_rate": 1.0881401472376915e-05, "loss": 0.0059, "step": 2937 }, { "epoch": 101.3103448275862, "grad_norm": 2.734375, "learning_rate": 1.087274869085862e-05, "loss": 0.0071, "step": 2938 }, { "epoch": 101.34482758620689, "grad_norm": 1.1640625, "learning_rate": 1.0864096702442848e-05, "loss": 0.0055, "step": 2939 }, { "epoch": 101.37931034482759, "grad_norm": 4.1875, "learning_rate": 1.085544551134647e-05, "loss": 0.0183, "step": 2940 }, { "epoch": 101.41379310344827, "grad_norm": 1.171875, "learning_rate": 1.0846795121785969e-05, "loss": 0.0101, "step": 2941 }, { "epoch": 101.44827586206897, "grad_norm": 1.6171875, "learning_rate": 1.0838145537977447e-05, "loss": 0.0149, "step": 2942 }, { "epoch": 101.48275862068965, "grad_norm": 2.0625, "learning_rate": 1.0829496764136598e-05, "loss": 0.0282, "step": 2943 }, { "epoch": 101.51724137931035, "grad_norm": 2.4375, "learning_rate": 1.0820848804478733e-05, "loss": 0.0235, "step": 2944 }, { "epoch": 101.55172413793103, "grad_norm": 0.8125, "learning_rate": 1.0812201663218762e-05, "loss": 0.0037, "step": 2945 }, { "epoch": 101.58620689655173, "grad_norm": 3.734375, "learning_rate": 1.0803555344571197e-05, "loss": 0.0127, "step": 2946 }, { "epoch": 101.62068965517241, "grad_norm": 0.2421875, "learning_rate": 1.0794909852750145e-05, "loss": 0.027, "step": 2947 }, { "epoch": 101.65517241379311, "grad_norm": 1.5625, "learning_rate": 1.0786265191969322e-05, "loss": 0.0074, "step": 2948 }, { "epoch": 101.6896551724138, "grad_norm": 4.375, "learning_rate": 1.0777621366442026e-05, "loss": 0.0258, "step": 2949 }, { "epoch": 101.72413793103448, "grad_norm": 1.1640625, "learning_rate": 1.0768978380381148e-05, "loss": 0.0067, "step": 2950 }, { "epoch": 101.75862068965517, "grad_norm": 5.875, "learning_rate": 1.0760336237999187e-05, "loss": 0.0406, "step": 2951 }, { "epoch": 101.79310344827586, "grad_norm": 1.34375, "learning_rate": 1.0751694943508202e-05, "loss": 0.0228, "step": 2952 }, { "epoch": 101.82758620689656, "grad_norm": 1.5546875, "learning_rate": 1.0743054501119865e-05, "loss": 0.0152, "step": 2953 }, { "epoch": 101.86206896551724, "grad_norm": 0.6796875, "learning_rate": 1.0734414915045419e-05, "loss": 0.0048, "step": 2954 }, { "epoch": 101.89655172413794, "grad_norm": 2.375, "learning_rate": 1.0725776189495696e-05, "loss": 0.0239, "step": 2955 }, { "epoch": 101.93103448275862, "grad_norm": 3.15625, "learning_rate": 1.0717138328681106e-05, "loss": 0.0168, "step": 2956 }, { "epoch": 101.96551724137932, "grad_norm": 3.34375, "learning_rate": 1.0708501336811632e-05, "loss": 0.0199, "step": 2957 }, { "epoch": 102.0, "grad_norm": 5.90625, "learning_rate": 1.0699865218096849e-05, "loss": 0.0464, "step": 2958 }, { "epoch": 102.03448275862068, "grad_norm": 3.125, "learning_rate": 1.0691229976745887e-05, "loss": 0.013, "step": 2959 }, { "epoch": 102.06896551724138, "grad_norm": 1.0390625, "learning_rate": 1.0682595616967465e-05, "loss": 0.0037, "step": 2960 }, { "epoch": 102.10344827586206, "grad_norm": 0.1611328125, "learning_rate": 1.0673962142969857e-05, "loss": 0.0057, "step": 2961 }, { "epoch": 102.13793103448276, "grad_norm": 3.203125, "learning_rate": 1.0665329558960918e-05, "loss": 0.0411, "step": 2962 }, { "epoch": 102.17241379310344, "grad_norm": 2.078125, "learning_rate": 1.0656697869148067e-05, "loss": 0.0342, "step": 2963 }, { "epoch": 102.20689655172414, "grad_norm": 0.48828125, "learning_rate": 1.0648067077738275e-05, "loss": 0.0029, "step": 2964 }, { "epoch": 102.24137931034483, "grad_norm": 2.65625, "learning_rate": 1.0639437188938098e-05, "loss": 0.0249, "step": 2965 }, { "epoch": 102.27586206896552, "grad_norm": 0.765625, "learning_rate": 1.0630808206953632e-05, "loss": 0.0099, "step": 2966 }, { "epoch": 102.3103448275862, "grad_norm": 0.5078125, "learning_rate": 1.0622180135990537e-05, "loss": 0.0051, "step": 2967 }, { "epoch": 102.34482758620689, "grad_norm": 0.9453125, "learning_rate": 1.0613552980254032e-05, "loss": 0.0089, "step": 2968 }, { "epoch": 102.37931034482759, "grad_norm": 0.37890625, "learning_rate": 1.0604926743948892e-05, "loss": 0.0039, "step": 2969 }, { "epoch": 102.41379310344827, "grad_norm": 1.7890625, "learning_rate": 1.0596301431279432e-05, "loss": 0.0323, "step": 2970 }, { "epoch": 102.44827586206897, "grad_norm": 0.1953125, "learning_rate": 1.0587677046449526e-05, "loss": 0.0094, "step": 2971 }, { "epoch": 102.48275862068965, "grad_norm": 5.28125, "learning_rate": 1.0579053593662599e-05, "loss": 0.0355, "step": 2972 }, { "epoch": 102.51724137931035, "grad_norm": 2.5, "learning_rate": 1.057043107712161e-05, "loss": 0.0131, "step": 2973 }, { "epoch": 102.55172413793103, "grad_norm": 1.8984375, "learning_rate": 1.0561809501029078e-05, "loss": 0.0185, "step": 2974 }, { "epoch": 102.58620689655173, "grad_norm": 0.443359375, "learning_rate": 1.0553188869587049e-05, "loss": 0.0105, "step": 2975 }, { "epoch": 102.62068965517241, "grad_norm": 1.9921875, "learning_rate": 1.0544569186997112e-05, "loss": 0.0077, "step": 2976 }, { "epoch": 102.65517241379311, "grad_norm": 1.765625, "learning_rate": 1.0535950457460408e-05, "loss": 0.0587, "step": 2977 }, { "epoch": 102.6896551724138, "grad_norm": 1.734375, "learning_rate": 1.0527332685177583e-05, "loss": 0.0059, "step": 2978 }, { "epoch": 102.72413793103448, "grad_norm": 6.5, "learning_rate": 1.0518715874348844e-05, "loss": 0.0408, "step": 2979 }, { "epoch": 102.75862068965517, "grad_norm": 0.193359375, "learning_rate": 1.0510100029173917e-05, "loss": 0.0021, "step": 2980 }, { "epoch": 102.79310344827586, "grad_norm": 3.609375, "learning_rate": 1.0501485153852066e-05, "loss": 0.01, "step": 2981 }, { "epoch": 102.82758620689656, "grad_norm": 1.9375, "learning_rate": 1.0492871252582072e-05, "loss": 0.0107, "step": 2982 }, { "epoch": 102.86206896551724, "grad_norm": 0.8203125, "learning_rate": 1.0484258329562243e-05, "loss": 0.0251, "step": 2983 }, { "epoch": 102.89655172413794, "grad_norm": 0.86328125, "learning_rate": 1.047564638899042e-05, "loss": 0.0052, "step": 2984 }, { "epoch": 102.93103448275862, "grad_norm": 4.5625, "learning_rate": 1.0467035435063956e-05, "loss": 0.0189, "step": 2985 }, { "epoch": 102.96551724137932, "grad_norm": 0.32421875, "learning_rate": 1.0458425471979726e-05, "loss": 0.0036, "step": 2986 }, { "epoch": 103.0, "grad_norm": 0.396484375, "learning_rate": 1.0449816503934114e-05, "loss": 0.012, "step": 2987 }, { "epoch": 103.03448275862068, "grad_norm": 0.279296875, "learning_rate": 1.0441208535123035e-05, "loss": 0.0126, "step": 2988 }, { "epoch": 103.06896551724138, "grad_norm": 0.25, "learning_rate": 1.0432601569741906e-05, "loss": 0.0049, "step": 2989 }, { "epoch": 103.10344827586206, "grad_norm": 0.83984375, "learning_rate": 1.0423995611985653e-05, "loss": 0.0086, "step": 2990 }, { "epoch": 103.13793103448276, "grad_norm": 2.125, "learning_rate": 1.0415390666048723e-05, "loss": 0.0081, "step": 2991 }, { "epoch": 103.17241379310344, "grad_norm": 0.46875, "learning_rate": 1.0406786736125057e-05, "loss": 0.0333, "step": 2992 }, { "epoch": 103.20689655172414, "grad_norm": 2.21875, "learning_rate": 1.0398183826408107e-05, "loss": 0.0149, "step": 2993 }, { "epoch": 103.24137931034483, "grad_norm": 0.234375, "learning_rate": 1.0389581941090827e-05, "loss": 0.0039, "step": 2994 }, { "epoch": 103.27586206896552, "grad_norm": 6.71875, "learning_rate": 1.0380981084365675e-05, "loss": 0.0326, "step": 2995 }, { "epoch": 103.3103448275862, "grad_norm": 0.71875, "learning_rate": 1.0372381260424597e-05, "loss": 0.0041, "step": 2996 }, { "epoch": 103.34482758620689, "grad_norm": 3.03125, "learning_rate": 1.0363782473459044e-05, "loss": 0.017, "step": 2997 }, { "epoch": 103.37931034482759, "grad_norm": 1.2421875, "learning_rate": 1.0355184727659964e-05, "loss": 0.0322, "step": 2998 }, { "epoch": 103.41379310344827, "grad_norm": 4.0625, "learning_rate": 1.0346588027217793e-05, "loss": 0.0104, "step": 2999 }, { "epoch": 103.44827586206897, "grad_norm": 5.875, "learning_rate": 1.0337992376322456e-05, "loss": 0.0318, "step": 3000 } ], "logging_steps": 1, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 173, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 64, "trial_name": null, "trial_params": null }