{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9979310344827588, "eval_steps": 500, "global_step": 1449, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0020689655172413794, "grad_norm": 8.16248893737793, "learning_rate": 6.896551724137931e-08, "loss": 1.3047, "step": 1 }, { "epoch": 0.004137931034482759, "grad_norm": 8.501221656799316, "learning_rate": 1.3793103448275863e-07, "loss": 1.3364, "step": 2 }, { "epoch": 0.006206896551724138, "grad_norm": 8.676358222961426, "learning_rate": 2.0689655172413796e-07, "loss": 1.3572, "step": 3 }, { "epoch": 0.008275862068965517, "grad_norm": 8.630147933959961, "learning_rate": 2.7586206896551726e-07, "loss": 1.3387, "step": 4 }, { "epoch": 0.010344827586206896, "grad_norm": 8.223169326782227, "learning_rate": 3.4482758620689656e-07, "loss": 1.3156, "step": 5 }, { "epoch": 0.012413793103448275, "grad_norm": 8.330610275268555, "learning_rate": 4.137931034482759e-07, "loss": 1.2883, "step": 6 }, { "epoch": 0.014482758620689656, "grad_norm": 9.017132759094238, "learning_rate": 4.827586206896552e-07, "loss": 1.3781, "step": 7 }, { "epoch": 0.016551724137931035, "grad_norm": 8.214015007019043, "learning_rate": 5.517241379310345e-07, "loss": 1.2708, "step": 8 }, { "epoch": 0.018620689655172412, "grad_norm": 8.30260181427002, "learning_rate": 6.206896551724139e-07, "loss": 1.327, "step": 9 }, { "epoch": 0.020689655172413793, "grad_norm": 7.864175319671631, "learning_rate": 6.896551724137931e-07, "loss": 1.2873, "step": 10 }, { "epoch": 0.022758620689655173, "grad_norm": 7.549825191497803, "learning_rate": 7.586206896551725e-07, "loss": 1.2672, "step": 11 }, { "epoch": 0.02482758620689655, "grad_norm": 7.3428802490234375, "learning_rate": 8.275862068965518e-07, "loss": 1.2312, "step": 12 }, { "epoch": 0.02689655172413793, "grad_norm": 6.168758392333984, "learning_rate": 8.965517241379311e-07, "loss": 1.2331, "step": 13 }, { "epoch": 0.028965517241379312, "grad_norm": 6.219021320343018, "learning_rate": 9.655172413793103e-07, "loss": 1.2173, "step": 14 }, { "epoch": 0.03103448275862069, "grad_norm": 5.889226913452148, "learning_rate": 1.0344827586206898e-06, "loss": 1.2164, "step": 15 }, { "epoch": 0.03310344827586207, "grad_norm": 5.6153411865234375, "learning_rate": 1.103448275862069e-06, "loss": 1.2284, "step": 16 }, { "epoch": 0.03517241379310345, "grad_norm": 3.499351978302002, "learning_rate": 1.1724137931034483e-06, "loss": 1.1728, "step": 17 }, { "epoch": 0.037241379310344824, "grad_norm": 3.140091896057129, "learning_rate": 1.2413793103448277e-06, "loss": 1.1509, "step": 18 }, { "epoch": 0.039310344827586205, "grad_norm": 3.1030735969543457, "learning_rate": 1.3103448275862072e-06, "loss": 1.1557, "step": 19 }, { "epoch": 0.041379310344827586, "grad_norm": 2.808349370956421, "learning_rate": 1.3793103448275862e-06, "loss": 1.099, "step": 20 }, { "epoch": 0.043448275862068966, "grad_norm": 2.8193469047546387, "learning_rate": 1.4482758620689657e-06, "loss": 1.1728, "step": 21 }, { "epoch": 0.04551724137931035, "grad_norm": 2.7212672233581543, "learning_rate": 1.517241379310345e-06, "loss": 1.1511, "step": 22 }, { "epoch": 0.04758620689655172, "grad_norm": 2.5926826000213623, "learning_rate": 1.5862068965517244e-06, "loss": 1.1317, "step": 23 }, { "epoch": 0.0496551724137931, "grad_norm": 3.0901269912719727, "learning_rate": 1.6551724137931037e-06, "loss": 1.0843, "step": 24 }, { "epoch": 0.05172413793103448, "grad_norm": 3.495481252670288, "learning_rate": 1.724137931034483e-06, "loss": 1.1062, "step": 25 }, { "epoch": 0.05379310344827586, "grad_norm": 3.160388231277466, "learning_rate": 1.7931034482758622e-06, "loss": 1.0713, "step": 26 }, { "epoch": 0.05586206896551724, "grad_norm": 3.0356714725494385, "learning_rate": 1.8620689655172416e-06, "loss": 1.0513, "step": 27 }, { "epoch": 0.057931034482758624, "grad_norm": 2.9434399604797363, "learning_rate": 1.9310344827586207e-06, "loss": 1.0907, "step": 28 }, { "epoch": 0.06, "grad_norm": 2.5999321937561035, "learning_rate": 2.0000000000000003e-06, "loss": 1.0615, "step": 29 }, { "epoch": 0.06206896551724138, "grad_norm": 2.188593864440918, "learning_rate": 2.0689655172413796e-06, "loss": 1.0549, "step": 30 }, { "epoch": 0.06413793103448276, "grad_norm": 1.513140320777893, "learning_rate": 2.137931034482759e-06, "loss": 1.0179, "step": 31 }, { "epoch": 0.06620689655172414, "grad_norm": 1.3605451583862305, "learning_rate": 2.206896551724138e-06, "loss": 1.0188, "step": 32 }, { "epoch": 0.06827586206896552, "grad_norm": 1.7352983951568604, "learning_rate": 2.2758620689655173e-06, "loss": 1.0239, "step": 33 }, { "epoch": 0.0703448275862069, "grad_norm": 1.7656395435333252, "learning_rate": 2.3448275862068966e-06, "loss": 0.9941, "step": 34 }, { "epoch": 0.07241379310344828, "grad_norm": 1.684338092803955, "learning_rate": 2.4137931034482762e-06, "loss": 0.9742, "step": 35 }, { "epoch": 0.07448275862068965, "grad_norm": 1.4101715087890625, "learning_rate": 2.4827586206896555e-06, "loss": 0.9809, "step": 36 }, { "epoch": 0.07655172413793103, "grad_norm": 1.2714829444885254, "learning_rate": 2.5517241379310347e-06, "loss": 0.9693, "step": 37 }, { "epoch": 0.07862068965517241, "grad_norm": 1.1786231994628906, "learning_rate": 2.6206896551724144e-06, "loss": 0.9956, "step": 38 }, { "epoch": 0.08068965517241379, "grad_norm": 1.0718473196029663, "learning_rate": 2.6896551724137932e-06, "loss": 0.9836, "step": 39 }, { "epoch": 0.08275862068965517, "grad_norm": 0.9965543746948242, "learning_rate": 2.7586206896551725e-06, "loss": 0.9186, "step": 40 }, { "epoch": 0.08482758620689655, "grad_norm": 0.9652248620986938, "learning_rate": 2.827586206896552e-06, "loss": 0.9265, "step": 41 }, { "epoch": 0.08689655172413793, "grad_norm": 1.0196917057037354, "learning_rate": 2.8965517241379314e-06, "loss": 0.9483, "step": 42 }, { "epoch": 0.08896551724137931, "grad_norm": 1.2638320922851562, "learning_rate": 2.9655172413793102e-06, "loss": 0.9579, "step": 43 }, { "epoch": 0.0910344827586207, "grad_norm": 1.0733128786087036, "learning_rate": 3.03448275862069e-06, "loss": 0.9739, "step": 44 }, { "epoch": 0.09310344827586207, "grad_norm": 0.9054429531097412, "learning_rate": 3.103448275862069e-06, "loss": 0.9671, "step": 45 }, { "epoch": 0.09517241379310344, "grad_norm": 0.9127644896507263, "learning_rate": 3.172413793103449e-06, "loss": 0.9145, "step": 46 }, { "epoch": 0.09724137931034482, "grad_norm": 0.8835800886154175, "learning_rate": 3.2413793103448277e-06, "loss": 0.9169, "step": 47 }, { "epoch": 0.0993103448275862, "grad_norm": 1.0521382093429565, "learning_rate": 3.3103448275862073e-06, "loss": 0.9341, "step": 48 }, { "epoch": 0.10137931034482758, "grad_norm": 0.8984640836715698, "learning_rate": 3.3793103448275866e-06, "loss": 0.9007, "step": 49 }, { "epoch": 0.10344827586206896, "grad_norm": 0.8801516890525818, "learning_rate": 3.448275862068966e-06, "loss": 0.9447, "step": 50 }, { "epoch": 0.10551724137931034, "grad_norm": 0.873184859752655, "learning_rate": 3.517241379310345e-06, "loss": 0.9134, "step": 51 }, { "epoch": 0.10758620689655173, "grad_norm": 0.8689625263214111, "learning_rate": 3.5862068965517243e-06, "loss": 0.9404, "step": 52 }, { "epoch": 0.1096551724137931, "grad_norm": 0.8642359972000122, "learning_rate": 3.655172413793104e-06, "loss": 0.8982, "step": 53 }, { "epoch": 0.11172413793103449, "grad_norm": 0.9032179117202759, "learning_rate": 3.7241379310344832e-06, "loss": 0.8848, "step": 54 }, { "epoch": 0.11379310344827587, "grad_norm": 0.8974177837371826, "learning_rate": 3.793103448275862e-06, "loss": 0.9485, "step": 55 }, { "epoch": 0.11586206896551725, "grad_norm": 0.9621221423149109, "learning_rate": 3.862068965517241e-06, "loss": 0.8727, "step": 56 }, { "epoch": 0.11793103448275861, "grad_norm": 0.9366543889045715, "learning_rate": 3.931034482758621e-06, "loss": 0.926, "step": 57 }, { "epoch": 0.12, "grad_norm": 0.8639587163925171, "learning_rate": 4.000000000000001e-06, "loss": 0.8623, "step": 58 }, { "epoch": 0.12206896551724138, "grad_norm": 0.9308229684829712, "learning_rate": 4.0689655172413795e-06, "loss": 0.8973, "step": 59 }, { "epoch": 0.12413793103448276, "grad_norm": 0.8033000230789185, "learning_rate": 4.137931034482759e-06, "loss": 0.8861, "step": 60 }, { "epoch": 0.12620689655172415, "grad_norm": 0.9005990028381348, "learning_rate": 4.206896551724138e-06, "loss": 0.8823, "step": 61 }, { "epoch": 0.12827586206896552, "grad_norm": 0.9020888805389404, "learning_rate": 4.275862068965518e-06, "loss": 0.8914, "step": 62 }, { "epoch": 0.13034482758620689, "grad_norm": 0.8798072338104248, "learning_rate": 4.3448275862068965e-06, "loss": 0.9076, "step": 63 }, { "epoch": 0.13241379310344828, "grad_norm": 0.8254591226577759, "learning_rate": 4.413793103448276e-06, "loss": 0.9246, "step": 64 }, { "epoch": 0.13448275862068965, "grad_norm": 0.7802205085754395, "learning_rate": 4.482758620689656e-06, "loss": 0.8635, "step": 65 }, { "epoch": 0.13655172413793104, "grad_norm": 0.8493502736091614, "learning_rate": 4.551724137931035e-06, "loss": 0.8793, "step": 66 }, { "epoch": 0.1386206896551724, "grad_norm": 0.8410840034484863, "learning_rate": 4.620689655172414e-06, "loss": 0.8593, "step": 67 }, { "epoch": 0.1406896551724138, "grad_norm": 0.796180009841919, "learning_rate": 4.689655172413793e-06, "loss": 0.8797, "step": 68 }, { "epoch": 0.14275862068965517, "grad_norm": 0.7212627530097961, "learning_rate": 4.758620689655173e-06, "loss": 0.8661, "step": 69 }, { "epoch": 0.14482758620689656, "grad_norm": 0.8363891243934631, "learning_rate": 4.8275862068965525e-06, "loss": 0.8679, "step": 70 }, { "epoch": 0.14689655172413793, "grad_norm": 0.921846866607666, "learning_rate": 4.896551724137931e-06, "loss": 0.8643, "step": 71 }, { "epoch": 0.1489655172413793, "grad_norm": 0.7880343794822693, "learning_rate": 4.965517241379311e-06, "loss": 0.878, "step": 72 }, { "epoch": 0.1510344827586207, "grad_norm": 0.8128662705421448, "learning_rate": 5.03448275862069e-06, "loss": 0.8772, "step": 73 }, { "epoch": 0.15310344827586206, "grad_norm": 0.8667871952056885, "learning_rate": 5.1034482758620695e-06, "loss": 0.8731, "step": 74 }, { "epoch": 0.15517241379310345, "grad_norm": 0.8730148673057556, "learning_rate": 5.172413793103449e-06, "loss": 0.876, "step": 75 }, { "epoch": 0.15724137931034482, "grad_norm": 0.8205820918083191, "learning_rate": 5.241379310344829e-06, "loss": 0.8415, "step": 76 }, { "epoch": 0.15931034482758621, "grad_norm": 0.8799615502357483, "learning_rate": 5.310344827586207e-06, "loss": 0.8482, "step": 77 }, { "epoch": 0.16137931034482758, "grad_norm": 0.936702311038971, "learning_rate": 5.3793103448275865e-06, "loss": 0.8677, "step": 78 }, { "epoch": 0.16344827586206898, "grad_norm": 0.8674628734588623, "learning_rate": 5.448275862068966e-06, "loss": 0.8973, "step": 79 }, { "epoch": 0.16551724137931034, "grad_norm": 0.8501123785972595, "learning_rate": 5.517241379310345e-06, "loss": 0.8603, "step": 80 }, { "epoch": 0.16758620689655174, "grad_norm": 0.9434123039245605, "learning_rate": 5.586206896551725e-06, "loss": 0.8983, "step": 81 }, { "epoch": 0.1696551724137931, "grad_norm": 0.8477321267127991, "learning_rate": 5.655172413793104e-06, "loss": 0.8621, "step": 82 }, { "epoch": 0.17172413793103447, "grad_norm": 0.8743482232093811, "learning_rate": 5.724137931034483e-06, "loss": 0.8854, "step": 83 }, { "epoch": 0.17379310344827587, "grad_norm": 0.7993690967559814, "learning_rate": 5.793103448275863e-06, "loss": 0.8605, "step": 84 }, { "epoch": 0.17586206896551723, "grad_norm": 0.8073793649673462, "learning_rate": 5.862068965517242e-06, "loss": 0.8867, "step": 85 }, { "epoch": 0.17793103448275863, "grad_norm": 0.8699036240577698, "learning_rate": 5.9310344827586205e-06, "loss": 0.8752, "step": 86 }, { "epoch": 0.18, "grad_norm": 0.9213085770606995, "learning_rate": 6e-06, "loss": 0.8601, "step": 87 }, { "epoch": 0.1820689655172414, "grad_norm": 0.8272461295127869, "learning_rate": 6.06896551724138e-06, "loss": 0.8613, "step": 88 }, { "epoch": 0.18413793103448275, "grad_norm": 0.8762076497077942, "learning_rate": 6.1379310344827595e-06, "loss": 0.8755, "step": 89 }, { "epoch": 0.18620689655172415, "grad_norm": 0.8086516261100769, "learning_rate": 6.206896551724138e-06, "loss": 0.8779, "step": 90 }, { "epoch": 0.18827586206896552, "grad_norm": 0.8407052755355835, "learning_rate": 6.275862068965518e-06, "loss": 0.869, "step": 91 }, { "epoch": 0.19034482758620688, "grad_norm": 0.8412452340126038, "learning_rate": 6.344827586206898e-06, "loss": 0.8937, "step": 92 }, { "epoch": 0.19241379310344828, "grad_norm": 0.8644373416900635, "learning_rate": 6.413793103448276e-06, "loss": 0.8627, "step": 93 }, { "epoch": 0.19448275862068964, "grad_norm": 0.867426335811615, "learning_rate": 6.482758620689655e-06, "loss": 0.8714, "step": 94 }, { "epoch": 0.19655172413793104, "grad_norm": 0.8877378702163696, "learning_rate": 6.551724137931035e-06, "loss": 0.8603, "step": 95 }, { "epoch": 0.1986206896551724, "grad_norm": 0.7677062749862671, "learning_rate": 6.620689655172415e-06, "loss": 0.8273, "step": 96 }, { "epoch": 0.2006896551724138, "grad_norm": 0.8652222752571106, "learning_rate": 6.6896551724137935e-06, "loss": 0.8786, "step": 97 }, { "epoch": 0.20275862068965517, "grad_norm": 0.8587886691093445, "learning_rate": 6.758620689655173e-06, "loss": 0.8541, "step": 98 }, { "epoch": 0.20482758620689656, "grad_norm": 0.8325220942497253, "learning_rate": 6.827586206896553e-06, "loss": 0.8389, "step": 99 }, { "epoch": 0.20689655172413793, "grad_norm": 0.9126191735267639, "learning_rate": 6.896551724137932e-06, "loss": 0.8588, "step": 100 }, { "epoch": 0.20896551724137932, "grad_norm": 0.8659350872039795, "learning_rate": 6.9655172413793105e-06, "loss": 0.8649, "step": 101 }, { "epoch": 0.2110344827586207, "grad_norm": 0.80833899974823, "learning_rate": 7.03448275862069e-06, "loss": 0.8668, "step": 102 }, { "epoch": 0.21310344827586206, "grad_norm": 0.8443086743354797, "learning_rate": 7.103448275862069e-06, "loss": 0.852, "step": 103 }, { "epoch": 0.21517241379310345, "grad_norm": 0.9743650555610657, "learning_rate": 7.172413793103449e-06, "loss": 0.8727, "step": 104 }, { "epoch": 0.21724137931034482, "grad_norm": 0.8710004687309265, "learning_rate": 7.241379310344828e-06, "loss": 0.8605, "step": 105 }, { "epoch": 0.2193103448275862, "grad_norm": 0.9044391512870789, "learning_rate": 7.310344827586208e-06, "loss": 0.8395, "step": 106 }, { "epoch": 0.22137931034482758, "grad_norm": 0.7920845746994019, "learning_rate": 7.379310344827587e-06, "loss": 0.8361, "step": 107 }, { "epoch": 0.22344827586206897, "grad_norm": 0.9043503403663635, "learning_rate": 7.4482758620689665e-06, "loss": 0.8326, "step": 108 }, { "epoch": 0.22551724137931034, "grad_norm": 0.914540708065033, "learning_rate": 7.517241379310345e-06, "loss": 0.8592, "step": 109 }, { "epoch": 0.22758620689655173, "grad_norm": 0.889606773853302, "learning_rate": 7.586206896551724e-06, "loss": 0.871, "step": 110 }, { "epoch": 0.2296551724137931, "grad_norm": 0.8037165403366089, "learning_rate": 7.655172413793104e-06, "loss": 0.8059, "step": 111 }, { "epoch": 0.2317241379310345, "grad_norm": 0.9440395832061768, "learning_rate": 7.724137931034483e-06, "loss": 0.8511, "step": 112 }, { "epoch": 0.23379310344827586, "grad_norm": 0.8397315740585327, "learning_rate": 7.793103448275863e-06, "loss": 0.8363, "step": 113 }, { "epoch": 0.23586206896551723, "grad_norm": 0.9024421572685242, "learning_rate": 7.862068965517242e-06, "loss": 0.8276, "step": 114 }, { "epoch": 0.23793103448275862, "grad_norm": 0.9368509650230408, "learning_rate": 7.93103448275862e-06, "loss": 0.8896, "step": 115 }, { "epoch": 0.24, "grad_norm": 0.9322966933250427, "learning_rate": 8.000000000000001e-06, "loss": 0.8054, "step": 116 }, { "epoch": 0.24206896551724139, "grad_norm": 0.9048160910606384, "learning_rate": 8.06896551724138e-06, "loss": 0.8448, "step": 117 }, { "epoch": 0.24413793103448275, "grad_norm": 0.8558708429336548, "learning_rate": 8.137931034482759e-06, "loss": 0.8275, "step": 118 }, { "epoch": 0.24620689655172415, "grad_norm": 0.8424068093299866, "learning_rate": 8.206896551724138e-06, "loss": 0.8694, "step": 119 }, { "epoch": 0.2482758620689655, "grad_norm": 1.001757025718689, "learning_rate": 8.275862068965518e-06, "loss": 0.8453, "step": 120 }, { "epoch": 0.2503448275862069, "grad_norm": 0.9482073783874512, "learning_rate": 8.344827586206897e-06, "loss": 0.8696, "step": 121 }, { "epoch": 0.2524137931034483, "grad_norm": 0.9771032929420471, "learning_rate": 8.413793103448276e-06, "loss": 0.8373, "step": 122 }, { "epoch": 0.25448275862068964, "grad_norm": 1.0352799892425537, "learning_rate": 8.482758620689656e-06, "loss": 0.837, "step": 123 }, { "epoch": 0.25655172413793104, "grad_norm": 0.9529556035995483, "learning_rate": 8.551724137931035e-06, "loss": 0.8286, "step": 124 }, { "epoch": 0.25862068965517243, "grad_norm": 0.8518896698951721, "learning_rate": 8.620689655172414e-06, "loss": 0.847, "step": 125 }, { "epoch": 0.26068965517241377, "grad_norm": 0.8487024903297424, "learning_rate": 8.689655172413793e-06, "loss": 0.8257, "step": 126 }, { "epoch": 0.26275862068965516, "grad_norm": 1.0520756244659424, "learning_rate": 8.758620689655173e-06, "loss": 0.823, "step": 127 }, { "epoch": 0.26482758620689656, "grad_norm": 1.002064824104309, "learning_rate": 8.827586206896552e-06, "loss": 0.8618, "step": 128 }, { "epoch": 0.26689655172413795, "grad_norm": 0.9487592577934265, "learning_rate": 8.896551724137931e-06, "loss": 0.8344, "step": 129 }, { "epoch": 0.2689655172413793, "grad_norm": 0.9966208338737488, "learning_rate": 8.965517241379312e-06, "loss": 0.8372, "step": 130 }, { "epoch": 0.2710344827586207, "grad_norm": 0.8904616832733154, "learning_rate": 9.03448275862069e-06, "loss": 0.8547, "step": 131 }, { "epoch": 0.2731034482758621, "grad_norm": 0.8750537037849426, "learning_rate": 9.10344827586207e-06, "loss": 0.8423, "step": 132 }, { "epoch": 0.2751724137931034, "grad_norm": 1.0919158458709717, "learning_rate": 9.172413793103448e-06, "loss": 0.8295, "step": 133 }, { "epoch": 0.2772413793103448, "grad_norm": 0.8703680038452148, "learning_rate": 9.241379310344829e-06, "loss": 0.7951, "step": 134 }, { "epoch": 0.2793103448275862, "grad_norm": 0.8651219606399536, "learning_rate": 9.310344827586207e-06, "loss": 0.8656, "step": 135 }, { "epoch": 0.2813793103448276, "grad_norm": 1.1044590473175049, "learning_rate": 9.379310344827586e-06, "loss": 0.8359, "step": 136 }, { "epoch": 0.28344827586206894, "grad_norm": 0.8297178149223328, "learning_rate": 9.448275862068967e-06, "loss": 0.8055, "step": 137 }, { "epoch": 0.28551724137931034, "grad_norm": 0.8873213529586792, "learning_rate": 9.517241379310346e-06, "loss": 0.8346, "step": 138 }, { "epoch": 0.28758620689655173, "grad_norm": 0.9231070280075073, "learning_rate": 9.586206896551724e-06, "loss": 0.8647, "step": 139 }, { "epoch": 0.2896551724137931, "grad_norm": 0.9084100127220154, "learning_rate": 9.655172413793105e-06, "loss": 0.8342, "step": 140 }, { "epoch": 0.29172413793103447, "grad_norm": 0.9144700765609741, "learning_rate": 9.724137931034484e-06, "loss": 0.8468, "step": 141 }, { "epoch": 0.29379310344827586, "grad_norm": 0.9257712364196777, "learning_rate": 9.793103448275863e-06, "loss": 0.8192, "step": 142 }, { "epoch": 0.29586206896551726, "grad_norm": 0.8467687368392944, "learning_rate": 9.862068965517241e-06, "loss": 0.827, "step": 143 }, { "epoch": 0.2979310344827586, "grad_norm": 0.925663411617279, "learning_rate": 9.931034482758622e-06, "loss": 0.8436, "step": 144 }, { "epoch": 0.3, "grad_norm": 0.9307589530944824, "learning_rate": 1e-05, "loss": 0.7953, "step": 145 }, { "epoch": 0.3020689655172414, "grad_norm": 0.9580163955688477, "learning_rate": 9.999985489433723e-06, "loss": 0.829, "step": 146 }, { "epoch": 0.3041379310344828, "grad_norm": 0.8726505637168884, "learning_rate": 9.999941957819112e-06, "loss": 0.8571, "step": 147 }, { "epoch": 0.3062068965517241, "grad_norm": 0.934336245059967, "learning_rate": 9.999869405408833e-06, "loss": 0.8127, "step": 148 }, { "epoch": 0.3082758620689655, "grad_norm": 1.0289698839187622, "learning_rate": 9.999767832624e-06, "loss": 0.8592, "step": 149 }, { "epoch": 0.3103448275862069, "grad_norm": 0.8795868754386902, "learning_rate": 9.999637240054163e-06, "loss": 0.8207, "step": 150 }, { "epoch": 0.3124137931034483, "grad_norm": 0.8612985610961914, "learning_rate": 9.99947762845731e-06, "loss": 0.8548, "step": 151 }, { "epoch": 0.31448275862068964, "grad_norm": 0.90495365858078, "learning_rate": 9.999288998759865e-06, "loss": 0.8346, "step": 152 }, { "epoch": 0.31655172413793103, "grad_norm": 1.010977029800415, "learning_rate": 9.999071352056676e-06, "loss": 0.8504, "step": 153 }, { "epoch": 0.31862068965517243, "grad_norm": 0.9661422967910767, "learning_rate": 9.998824689611012e-06, "loss": 0.8264, "step": 154 }, { "epoch": 0.32068965517241377, "grad_norm": 1.0099589824676514, "learning_rate": 9.998549012854562e-06, "loss": 0.8054, "step": 155 }, { "epoch": 0.32275862068965516, "grad_norm": 0.9441735744476318, "learning_rate": 9.998244323387412e-06, "loss": 0.803, "step": 156 }, { "epoch": 0.32482758620689656, "grad_norm": 0.9308339953422546, "learning_rate": 9.99791062297805e-06, "loss": 0.8267, "step": 157 }, { "epoch": 0.32689655172413795, "grad_norm": 0.9326705932617188, "learning_rate": 9.997547913563352e-06, "loss": 0.8291, "step": 158 }, { "epoch": 0.3289655172413793, "grad_norm": 0.8982317447662354, "learning_rate": 9.997156197248564e-06, "loss": 0.837, "step": 159 }, { "epoch": 0.3310344827586207, "grad_norm": 0.93187016248703, "learning_rate": 9.996735476307292e-06, "loss": 0.812, "step": 160 }, { "epoch": 0.3331034482758621, "grad_norm": 0.9523342251777649, "learning_rate": 9.996285753181499e-06, "loss": 0.8545, "step": 161 }, { "epoch": 0.3351724137931035, "grad_norm": 0.8485278487205505, "learning_rate": 9.995807030481481e-06, "loss": 0.8405, "step": 162 }, { "epoch": 0.3372413793103448, "grad_norm": 0.9064662456512451, "learning_rate": 9.99529931098585e-06, "loss": 0.8164, "step": 163 }, { "epoch": 0.3393103448275862, "grad_norm": 0.8660606145858765, "learning_rate": 9.99476259764153e-06, "loss": 0.816, "step": 164 }, { "epoch": 0.3413793103448276, "grad_norm": 0.9558955430984497, "learning_rate": 9.994196893563722e-06, "loss": 0.8365, "step": 165 }, { "epoch": 0.34344827586206894, "grad_norm": 0.838666558265686, "learning_rate": 9.993602202035902e-06, "loss": 0.7915, "step": 166 }, { "epoch": 0.34551724137931034, "grad_norm": 0.8363468050956726, "learning_rate": 9.992978526509794e-06, "loss": 0.7789, "step": 167 }, { "epoch": 0.34758620689655173, "grad_norm": 0.8468877077102661, "learning_rate": 9.992325870605353e-06, "loss": 0.8306, "step": 168 }, { "epoch": 0.3496551724137931, "grad_norm": 0.8892214298248291, "learning_rate": 9.991644238110741e-06, "loss": 0.8433, "step": 169 }, { "epoch": 0.35172413793103446, "grad_norm": 0.9307257533073425, "learning_rate": 9.990933632982307e-06, "loss": 0.8402, "step": 170 }, { "epoch": 0.35379310344827586, "grad_norm": 0.8565515279769897, "learning_rate": 9.990194059344565e-06, "loss": 0.8403, "step": 171 }, { "epoch": 0.35586206896551725, "grad_norm": 0.89169842004776, "learning_rate": 9.989425521490169e-06, "loss": 0.8387, "step": 172 }, { "epoch": 0.35793103448275865, "grad_norm": 0.9352335333824158, "learning_rate": 9.988628023879883e-06, "loss": 0.8292, "step": 173 }, { "epoch": 0.36, "grad_norm": 0.8871222734451294, "learning_rate": 9.987801571142569e-06, "loss": 0.8245, "step": 174 }, { "epoch": 0.3620689655172414, "grad_norm": 0.7852333188056946, "learning_rate": 9.98694616807514e-06, "loss": 0.8523, "step": 175 }, { "epoch": 0.3641379310344828, "grad_norm": 0.763839840888977, "learning_rate": 9.986061819642554e-06, "loss": 0.8027, "step": 176 }, { "epoch": 0.3662068965517241, "grad_norm": 0.9269424080848694, "learning_rate": 9.985148530977767e-06, "loss": 0.8507, "step": 177 }, { "epoch": 0.3682758620689655, "grad_norm": 0.9032247066497803, "learning_rate": 9.984206307381714e-06, "loss": 0.82, "step": 178 }, { "epoch": 0.3703448275862069, "grad_norm": 0.8867712020874023, "learning_rate": 9.983235154323276e-06, "loss": 0.8395, "step": 179 }, { "epoch": 0.3724137931034483, "grad_norm": 0.8791627883911133, "learning_rate": 9.982235077439242e-06, "loss": 0.8677, "step": 180 }, { "epoch": 0.37448275862068964, "grad_norm": 0.9103829264640808, "learning_rate": 9.981206082534287e-06, "loss": 0.8136, "step": 181 }, { "epoch": 0.37655172413793103, "grad_norm": 0.8131744265556335, "learning_rate": 9.98014817558093e-06, "loss": 0.835, "step": 182 }, { "epoch": 0.3786206896551724, "grad_norm": 0.8920416235923767, "learning_rate": 9.979061362719502e-06, "loss": 0.7958, "step": 183 }, { "epoch": 0.38068965517241377, "grad_norm": 0.8333418369293213, "learning_rate": 9.977945650258111e-06, "loss": 0.8238, "step": 184 }, { "epoch": 0.38275862068965516, "grad_norm": 0.8242124915122986, "learning_rate": 9.976801044672608e-06, "loss": 0.8412, "step": 185 }, { "epoch": 0.38482758620689655, "grad_norm": 0.9200892448425293, "learning_rate": 9.975627552606538e-06, "loss": 0.8429, "step": 186 }, { "epoch": 0.38689655172413795, "grad_norm": 0.8656523823738098, "learning_rate": 9.974425180871118e-06, "loss": 0.8268, "step": 187 }, { "epoch": 0.3889655172413793, "grad_norm": 0.9208945631980896, "learning_rate": 9.973193936445185e-06, "loss": 0.8338, "step": 188 }, { "epoch": 0.3910344827586207, "grad_norm": 0.873414158821106, "learning_rate": 9.971933826475162e-06, "loss": 0.8236, "step": 189 }, { "epoch": 0.3931034482758621, "grad_norm": 0.9169439673423767, "learning_rate": 9.970644858275008e-06, "loss": 0.8242, "step": 190 }, { "epoch": 0.39517241379310347, "grad_norm": 0.8280132412910461, "learning_rate": 9.969327039326193e-06, "loss": 0.8269, "step": 191 }, { "epoch": 0.3972413793103448, "grad_norm": 0.8578395247459412, "learning_rate": 9.967980377277632e-06, "loss": 0.7664, "step": 192 }, { "epoch": 0.3993103448275862, "grad_norm": 0.8905231952667236, "learning_rate": 9.966604879945659e-06, "loss": 0.8239, "step": 193 }, { "epoch": 0.4013793103448276, "grad_norm": 0.8388824462890625, "learning_rate": 9.96520055531397e-06, "loss": 0.8128, "step": 194 }, { "epoch": 0.40344827586206894, "grad_norm": 0.7973628640174866, "learning_rate": 9.963767411533584e-06, "loss": 0.7949, "step": 195 }, { "epoch": 0.40551724137931033, "grad_norm": 0.8202701807022095, "learning_rate": 9.962305456922792e-06, "loss": 0.8213, "step": 196 }, { "epoch": 0.40758620689655173, "grad_norm": 0.7976661324501038, "learning_rate": 9.960814699967112e-06, "loss": 0.7915, "step": 197 }, { "epoch": 0.4096551724137931, "grad_norm": 0.7930517196655273, "learning_rate": 9.959295149319233e-06, "loss": 0.805, "step": 198 }, { "epoch": 0.41172413793103446, "grad_norm": 0.809988260269165, "learning_rate": 9.95774681379897e-06, "loss": 0.7896, "step": 199 }, { "epoch": 0.41379310344827586, "grad_norm": 0.8766646981239319, "learning_rate": 9.956169702393215e-06, "loss": 0.8189, "step": 200 }, { "epoch": 0.41586206896551725, "grad_norm": 0.9457526803016663, "learning_rate": 9.954563824255879e-06, "loss": 0.8189, "step": 201 }, { "epoch": 0.41793103448275865, "grad_norm": 0.9090128540992737, "learning_rate": 9.952929188707843e-06, "loss": 0.8614, "step": 202 }, { "epoch": 0.42, "grad_norm": 0.8124152421951294, "learning_rate": 9.951265805236903e-06, "loss": 0.8186, "step": 203 }, { "epoch": 0.4220689655172414, "grad_norm": 0.8770976066589355, "learning_rate": 9.94957368349771e-06, "loss": 0.7963, "step": 204 }, { "epoch": 0.4241379310344828, "grad_norm": 0.8936490416526794, "learning_rate": 9.947852833311725e-06, "loss": 0.8278, "step": 205 }, { "epoch": 0.4262068965517241, "grad_norm": 0.9065835475921631, "learning_rate": 9.946103264667151e-06, "loss": 0.8026, "step": 206 }, { "epoch": 0.4282758620689655, "grad_norm": 0.755096435546875, "learning_rate": 9.94432498771888e-06, "loss": 0.7987, "step": 207 }, { "epoch": 0.4303448275862069, "grad_norm": 0.8373960256576538, "learning_rate": 9.942518012788438e-06, "loss": 0.7967, "step": 208 }, { "epoch": 0.4324137931034483, "grad_norm": 0.8678641319274902, "learning_rate": 9.940682350363913e-06, "loss": 0.8212, "step": 209 }, { "epoch": 0.43448275862068964, "grad_norm": 0.8203259706497192, "learning_rate": 9.938818011099906e-06, "loss": 0.7922, "step": 210 }, { "epoch": 0.43655172413793103, "grad_norm": 0.8179094195365906, "learning_rate": 9.936925005817464e-06, "loss": 0.8271, "step": 211 }, { "epoch": 0.4386206896551724, "grad_norm": 0.8908324241638184, "learning_rate": 9.93500334550402e-06, "loss": 0.7867, "step": 212 }, { "epoch": 0.4406896551724138, "grad_norm": 0.8492280840873718, "learning_rate": 9.933053041313325e-06, "loss": 0.8225, "step": 213 }, { "epoch": 0.44275862068965516, "grad_norm": 0.8480848670005798, "learning_rate": 9.931074104565387e-06, "loss": 0.8418, "step": 214 }, { "epoch": 0.44482758620689655, "grad_norm": 0.857255756855011, "learning_rate": 9.929066546746403e-06, "loss": 0.8297, "step": 215 }, { "epoch": 0.44689655172413795, "grad_norm": 0.9532951712608337, "learning_rate": 9.927030379508691e-06, "loss": 0.8094, "step": 216 }, { "epoch": 0.4489655172413793, "grad_norm": 0.8437204360961914, "learning_rate": 9.924965614670629e-06, "loss": 0.7996, "step": 217 }, { "epoch": 0.4510344827586207, "grad_norm": 0.7982957363128662, "learning_rate": 9.922872264216581e-06, "loss": 0.7962, "step": 218 }, { "epoch": 0.4531034482758621, "grad_norm": 0.8589974045753479, "learning_rate": 9.920750340296823e-06, "loss": 0.8311, "step": 219 }, { "epoch": 0.45517241379310347, "grad_norm": 0.8524751663208008, "learning_rate": 9.918599855227487e-06, "loss": 0.8113, "step": 220 }, { "epoch": 0.4572413793103448, "grad_norm": 0.8594533205032349, "learning_rate": 9.916420821490474e-06, "loss": 0.8201, "step": 221 }, { "epoch": 0.4593103448275862, "grad_norm": 0.872061550617218, "learning_rate": 9.914213251733387e-06, "loss": 0.8014, "step": 222 }, { "epoch": 0.4613793103448276, "grad_norm": 0.8578385710716248, "learning_rate": 9.911977158769461e-06, "loss": 0.787, "step": 223 }, { "epoch": 0.463448275862069, "grad_norm": 0.9662230610847473, "learning_rate": 9.909712555577491e-06, "loss": 0.7924, "step": 224 }, { "epoch": 0.46551724137931033, "grad_norm": 0.909537672996521, "learning_rate": 9.90741945530174e-06, "loss": 0.8267, "step": 225 }, { "epoch": 0.4675862068965517, "grad_norm": 0.8204202651977539, "learning_rate": 9.905097871251887e-06, "loss": 0.8142, "step": 226 }, { "epoch": 0.4696551724137931, "grad_norm": 0.8849340081214905, "learning_rate": 9.90274781690293e-06, "loss": 0.8364, "step": 227 }, { "epoch": 0.47172413793103446, "grad_norm": 0.825469970703125, "learning_rate": 9.900369305895115e-06, "loss": 0.7874, "step": 228 }, { "epoch": 0.47379310344827585, "grad_norm": 0.799996018409729, "learning_rate": 9.89796235203386e-06, "loss": 0.8044, "step": 229 }, { "epoch": 0.47586206896551725, "grad_norm": 0.8811237812042236, "learning_rate": 9.895526969289671e-06, "loss": 0.8027, "step": 230 }, { "epoch": 0.47793103448275864, "grad_norm": 0.9074158072471619, "learning_rate": 9.89306317179806e-06, "loss": 0.8172, "step": 231 }, { "epoch": 0.48, "grad_norm": 0.7685100436210632, "learning_rate": 9.890570973859468e-06, "loss": 0.8144, "step": 232 }, { "epoch": 0.4820689655172414, "grad_norm": 0.8316488265991211, "learning_rate": 9.888050389939172e-06, "loss": 0.8188, "step": 233 }, { "epoch": 0.48413793103448277, "grad_norm": 0.8519396185874939, "learning_rate": 9.885501434667215e-06, "loss": 0.8242, "step": 234 }, { "epoch": 0.4862068965517241, "grad_norm": 0.8774333000183105, "learning_rate": 9.882924122838311e-06, "loss": 0.8083, "step": 235 }, { "epoch": 0.4882758620689655, "grad_norm": 0.8599486351013184, "learning_rate": 9.88031846941176e-06, "loss": 0.8024, "step": 236 }, { "epoch": 0.4903448275862069, "grad_norm": 0.8138340711593628, "learning_rate": 9.877684489511367e-06, "loss": 0.8242, "step": 237 }, { "epoch": 0.4924137931034483, "grad_norm": 0.9175668954849243, "learning_rate": 9.875022198425347e-06, "loss": 0.816, "step": 238 }, { "epoch": 0.49448275862068963, "grad_norm": 0.8907325267791748, "learning_rate": 9.872331611606238e-06, "loss": 0.7923, "step": 239 }, { "epoch": 0.496551724137931, "grad_norm": 0.8826651573181152, "learning_rate": 9.869612744670817e-06, "loss": 0.8359, "step": 240 }, { "epoch": 0.4986206896551724, "grad_norm": 0.866460919380188, "learning_rate": 9.866865613400008e-06, "loss": 0.7803, "step": 241 }, { "epoch": 0.5006896551724138, "grad_norm": 0.9463049173355103, "learning_rate": 9.864090233738776e-06, "loss": 0.8221, "step": 242 }, { "epoch": 0.5027586206896552, "grad_norm": 0.8751413226127625, "learning_rate": 9.861286621796056e-06, "loss": 0.8299, "step": 243 }, { "epoch": 0.5048275862068966, "grad_norm": 0.9759247303009033, "learning_rate": 9.858454793844648e-06, "loss": 0.8147, "step": 244 }, { "epoch": 0.506896551724138, "grad_norm": 0.899433434009552, "learning_rate": 9.855594766321122e-06, "loss": 0.8172, "step": 245 }, { "epoch": 0.5089655172413793, "grad_norm": 0.8528696298599243, "learning_rate": 9.852706555825727e-06, "loss": 0.7903, "step": 246 }, { "epoch": 0.5110344827586207, "grad_norm": 0.9641590714454651, "learning_rate": 9.849790179122287e-06, "loss": 0.7741, "step": 247 }, { "epoch": 0.5131034482758621, "grad_norm": 0.8884571194648743, "learning_rate": 9.846845653138117e-06, "loss": 0.8316, "step": 248 }, { "epoch": 0.5151724137931034, "grad_norm": 0.8643737435340881, "learning_rate": 9.843872994963912e-06, "loss": 0.7956, "step": 249 }, { "epoch": 0.5172413793103449, "grad_norm": 0.9500787854194641, "learning_rate": 9.84087222185365e-06, "loss": 0.8379, "step": 250 }, { "epoch": 0.5193103448275862, "grad_norm": 0.8723439574241638, "learning_rate": 9.837843351224504e-06, "loss": 0.8169, "step": 251 }, { "epoch": 0.5213793103448275, "grad_norm": 0.9550707936286926, "learning_rate": 9.834786400656719e-06, "loss": 0.8016, "step": 252 }, { "epoch": 0.523448275862069, "grad_norm": 0.9410514831542969, "learning_rate": 9.831701387893533e-06, "loss": 0.7718, "step": 253 }, { "epoch": 0.5255172413793103, "grad_norm": 0.8164445757865906, "learning_rate": 9.828588330841055e-06, "loss": 0.8035, "step": 254 }, { "epoch": 0.5275862068965518, "grad_norm": 0.8466964364051819, "learning_rate": 9.825447247568177e-06, "loss": 0.8199, "step": 255 }, { "epoch": 0.5296551724137931, "grad_norm": 1.039279818534851, "learning_rate": 9.822278156306455e-06, "loss": 0.8012, "step": 256 }, { "epoch": 0.5317241379310345, "grad_norm": 0.9037715196609497, "learning_rate": 9.819081075450014e-06, "loss": 0.8094, "step": 257 }, { "epoch": 0.5337931034482759, "grad_norm": 0.8636988401412964, "learning_rate": 9.815856023555437e-06, "loss": 0.8395, "step": 258 }, { "epoch": 0.5358620689655172, "grad_norm": 0.883048415184021, "learning_rate": 9.812603019341652e-06, "loss": 0.7844, "step": 259 }, { "epoch": 0.5379310344827586, "grad_norm": 0.861579954624176, "learning_rate": 9.809322081689834e-06, "loss": 0.8218, "step": 260 }, { "epoch": 0.54, "grad_norm": 0.9384899139404297, "learning_rate": 9.80601322964329e-06, "loss": 0.802, "step": 261 }, { "epoch": 0.5420689655172414, "grad_norm": 0.8394620418548584, "learning_rate": 9.802676482407343e-06, "loss": 0.7974, "step": 262 }, { "epoch": 0.5441379310344827, "grad_norm": 0.8985764384269714, "learning_rate": 9.799311859349235e-06, "loss": 0.8001, "step": 263 }, { "epoch": 0.5462068965517242, "grad_norm": 0.8577210307121277, "learning_rate": 9.795919379997993e-06, "loss": 0.8312, "step": 264 }, { "epoch": 0.5482758620689655, "grad_norm": 0.8467931151390076, "learning_rate": 9.792499064044343e-06, "loss": 0.8007, "step": 265 }, { "epoch": 0.5503448275862068, "grad_norm": 0.9455611705780029, "learning_rate": 9.789050931340572e-06, "loss": 0.8143, "step": 266 }, { "epoch": 0.5524137931034483, "grad_norm": 0.8457427620887756, "learning_rate": 9.785575001900418e-06, "loss": 0.7889, "step": 267 }, { "epoch": 0.5544827586206896, "grad_norm": 0.8870904445648193, "learning_rate": 9.78207129589897e-06, "loss": 0.823, "step": 268 }, { "epoch": 0.5565517241379311, "grad_norm": 0.8103102445602417, "learning_rate": 9.778539833672525e-06, "loss": 0.7962, "step": 269 }, { "epoch": 0.5586206896551724, "grad_norm": 0.8109989166259766, "learning_rate": 9.774980635718494e-06, "loss": 0.8107, "step": 270 }, { "epoch": 0.5606896551724138, "grad_norm": 0.8663826584815979, "learning_rate": 9.771393722695267e-06, "loss": 0.8158, "step": 271 }, { "epoch": 0.5627586206896552, "grad_norm": 1.5541844367980957, "learning_rate": 9.767779115422097e-06, "loss": 0.7898, "step": 272 }, { "epoch": 0.5648275862068965, "grad_norm": 0.8373553156852722, "learning_rate": 9.764136834878987e-06, "loss": 0.7963, "step": 273 }, { "epoch": 0.5668965517241379, "grad_norm": 0.7791042923927307, "learning_rate": 9.760466902206557e-06, "loss": 0.8085, "step": 274 }, { "epoch": 0.5689655172413793, "grad_norm": 0.8393211960792542, "learning_rate": 9.756769338705927e-06, "loss": 0.8136, "step": 275 }, { "epoch": 0.5710344827586207, "grad_norm": 0.811062216758728, "learning_rate": 9.753044165838594e-06, "loss": 0.8075, "step": 276 }, { "epoch": 0.573103448275862, "grad_norm": 0.8894556760787964, "learning_rate": 9.749291405226304e-06, "loss": 0.8371, "step": 277 }, { "epoch": 0.5751724137931035, "grad_norm": 0.8580628037452698, "learning_rate": 9.745511078650931e-06, "loss": 0.7695, "step": 278 }, { "epoch": 0.5772413793103448, "grad_norm": 0.7854070067405701, "learning_rate": 9.741703208054347e-06, "loss": 0.8076, "step": 279 }, { "epoch": 0.5793103448275863, "grad_norm": 0.8000266551971436, "learning_rate": 9.737867815538292e-06, "loss": 0.7894, "step": 280 }, { "epoch": 0.5813793103448276, "grad_norm": 0.8094033002853394, "learning_rate": 9.734004923364258e-06, "loss": 0.8035, "step": 281 }, { "epoch": 0.5834482758620689, "grad_norm": 0.8682005405426025, "learning_rate": 9.73011455395334e-06, "loss": 0.8219, "step": 282 }, { "epoch": 0.5855172413793104, "grad_norm": 0.8397322297096252, "learning_rate": 9.72619672988613e-06, "loss": 0.8241, "step": 283 }, { "epoch": 0.5875862068965517, "grad_norm": 0.8110212087631226, "learning_rate": 9.722251473902562e-06, "loss": 0.8099, "step": 284 }, { "epoch": 0.5896551724137931, "grad_norm": 0.8143454790115356, "learning_rate": 9.718278808901797e-06, "loss": 0.7744, "step": 285 }, { "epoch": 0.5917241379310345, "grad_norm": 0.8721611499786377, "learning_rate": 9.714278757942082e-06, "loss": 0.785, "step": 286 }, { "epoch": 0.5937931034482758, "grad_norm": 0.7908191084861755, "learning_rate": 9.710251344240619e-06, "loss": 0.7983, "step": 287 }, { "epoch": 0.5958620689655172, "grad_norm": 0.8809441924095154, "learning_rate": 9.70619659117343e-06, "loss": 0.8428, "step": 288 }, { "epoch": 0.5979310344827586, "grad_norm": 0.8129044771194458, "learning_rate": 9.702114522275216e-06, "loss": 0.8149, "step": 289 }, { "epoch": 0.6, "grad_norm": 0.826924204826355, "learning_rate": 9.698005161239237e-06, "loss": 0.8091, "step": 290 }, { "epoch": 0.6020689655172414, "grad_norm": 0.8344675302505493, "learning_rate": 9.69386853191715e-06, "loss": 0.7913, "step": 291 }, { "epoch": 0.6041379310344828, "grad_norm": 0.8673633933067322, "learning_rate": 9.689704658318888e-06, "loss": 0.8, "step": 292 }, { "epoch": 0.6062068965517241, "grad_norm": 0.7886486053466797, "learning_rate": 9.685513564612521e-06, "loss": 0.7967, "step": 293 }, { "epoch": 0.6082758620689656, "grad_norm": 0.8335657715797424, "learning_rate": 9.681295275124104e-06, "loss": 0.7843, "step": 294 }, { "epoch": 0.6103448275862069, "grad_norm": 0.934149980545044, "learning_rate": 9.67704981433754e-06, "loss": 0.7967, "step": 295 }, { "epoch": 0.6124137931034482, "grad_norm": 0.8753807544708252, "learning_rate": 9.672777206894452e-06, "loss": 0.8086, "step": 296 }, { "epoch": 0.6144827586206897, "grad_norm": 0.8328104615211487, "learning_rate": 9.668477477594021e-06, "loss": 0.8006, "step": 297 }, { "epoch": 0.616551724137931, "grad_norm": 0.8407275080680847, "learning_rate": 9.664150651392846e-06, "loss": 0.7968, "step": 298 }, { "epoch": 0.6186206896551724, "grad_norm": 0.8628553748130798, "learning_rate": 9.659796753404806e-06, "loss": 0.8256, "step": 299 }, { "epoch": 0.6206896551724138, "grad_norm": 0.7844734191894531, "learning_rate": 9.655415808900915e-06, "loss": 0.784, "step": 300 }, { "epoch": 0.6227586206896552, "grad_norm": 0.8278802633285522, "learning_rate": 9.651007843309164e-06, "loss": 0.7548, "step": 301 }, { "epoch": 0.6248275862068966, "grad_norm": 0.8672357201576233, "learning_rate": 9.646572882214385e-06, "loss": 0.8006, "step": 302 }, { "epoch": 0.6268965517241379, "grad_norm": 0.8474316000938416, "learning_rate": 9.642110951358097e-06, "loss": 0.8056, "step": 303 }, { "epoch": 0.6289655172413793, "grad_norm": 0.8918145895004272, "learning_rate": 9.637622076638358e-06, "loss": 0.8086, "step": 304 }, { "epoch": 0.6310344827586207, "grad_norm": 0.7701090574264526, "learning_rate": 9.633106284109612e-06, "loss": 0.7971, "step": 305 }, { "epoch": 0.6331034482758621, "grad_norm": 0.7959749102592468, "learning_rate": 9.62856359998254e-06, "loss": 0.7914, "step": 306 }, { "epoch": 0.6351724137931034, "grad_norm": 0.8288137912750244, "learning_rate": 9.623994050623917e-06, "loss": 0.8074, "step": 307 }, { "epoch": 0.6372413793103449, "grad_norm": 0.7820798754692078, "learning_rate": 9.619397662556434e-06, "loss": 0.764, "step": 308 }, { "epoch": 0.6393103448275862, "grad_norm": 0.8969067335128784, "learning_rate": 9.614774462458573e-06, "loss": 0.8464, "step": 309 }, { "epoch": 0.6413793103448275, "grad_norm": 0.8294548988342285, "learning_rate": 9.610124477164436e-06, "loss": 0.8576, "step": 310 }, { "epoch": 0.643448275862069, "grad_norm": 0.8388926982879639, "learning_rate": 9.605447733663586e-06, "loss": 0.7603, "step": 311 }, { "epoch": 0.6455172413793103, "grad_norm": 0.8787766098976135, "learning_rate": 9.600744259100905e-06, "loss": 0.7991, "step": 312 }, { "epoch": 0.6475862068965518, "grad_norm": 0.8679440021514893, "learning_rate": 9.596014080776424e-06, "loss": 0.7977, "step": 313 }, { "epoch": 0.6496551724137931, "grad_norm": 0.8179478049278259, "learning_rate": 9.59125722614517e-06, "loss": 0.8009, "step": 314 }, { "epoch": 0.6517241379310345, "grad_norm": 0.8390472531318665, "learning_rate": 9.586473722817002e-06, "loss": 0.8023, "step": 315 }, { "epoch": 0.6537931034482759, "grad_norm": 0.9145618677139282, "learning_rate": 9.581663598556461e-06, "loss": 0.7888, "step": 316 }, { "epoch": 0.6558620689655172, "grad_norm": 0.7960366010665894, "learning_rate": 9.576826881282595e-06, "loss": 0.8047, "step": 317 }, { "epoch": 0.6579310344827586, "grad_norm": 0.8886500000953674, "learning_rate": 9.571963599068807e-06, "loss": 0.7772, "step": 318 }, { "epoch": 0.66, "grad_norm": 0.8854326009750366, "learning_rate": 9.56707378014269e-06, "loss": 0.783, "step": 319 }, { "epoch": 0.6620689655172414, "grad_norm": 0.7884778380393982, "learning_rate": 9.56215745288586e-06, "loss": 0.7783, "step": 320 }, { "epoch": 0.6641379310344827, "grad_norm": 0.9254597425460815, "learning_rate": 9.557214645833792e-06, "loss": 0.7903, "step": 321 }, { "epoch": 0.6662068965517242, "grad_norm": 1.0196994543075562, "learning_rate": 9.55224538767566e-06, "loss": 0.7955, "step": 322 }, { "epoch": 0.6682758620689655, "grad_norm": 0.8419172167778015, "learning_rate": 9.547249707254166e-06, "loss": 0.8135, "step": 323 }, { "epoch": 0.670344827586207, "grad_norm": 0.85199373960495, "learning_rate": 9.542227633565364e-06, "loss": 0.7882, "step": 324 }, { "epoch": 0.6724137931034483, "grad_norm": 0.8458907008171082, "learning_rate": 9.537179195758513e-06, "loss": 0.7803, "step": 325 }, { "epoch": 0.6744827586206896, "grad_norm": 0.8774712681770325, "learning_rate": 9.532104423135885e-06, "loss": 0.8011, "step": 326 }, { "epoch": 0.6765517241379311, "grad_norm": 0.8695656657218933, "learning_rate": 9.527003345152616e-06, "loss": 0.792, "step": 327 }, { "epoch": 0.6786206896551724, "grad_norm": 0.799929141998291, "learning_rate": 9.521875991416514e-06, "loss": 0.7651, "step": 328 }, { "epoch": 0.6806896551724138, "grad_norm": 0.8962751626968384, "learning_rate": 9.516722391687903e-06, "loss": 0.7951, "step": 329 }, { "epoch": 0.6827586206896552, "grad_norm": 0.8643319010734558, "learning_rate": 9.51154257587944e-06, "loss": 0.7967, "step": 330 }, { "epoch": 0.6848275862068965, "grad_norm": 0.9759703874588013, "learning_rate": 9.506336574055954e-06, "loss": 0.7929, "step": 331 }, { "epoch": 0.6868965517241379, "grad_norm": 0.8432797193527222, "learning_rate": 9.501104416434255e-06, "loss": 0.7779, "step": 332 }, { "epoch": 0.6889655172413793, "grad_norm": 0.9403534531593323, "learning_rate": 9.495846133382973e-06, "loss": 0.8123, "step": 333 }, { "epoch": 0.6910344827586207, "grad_norm": 0.8679063320159912, "learning_rate": 9.490561755422374e-06, "loss": 0.7462, "step": 334 }, { "epoch": 0.6931034482758621, "grad_norm": 0.791249692440033, "learning_rate": 9.485251313224181e-06, "loss": 0.7982, "step": 335 }, { "epoch": 0.6951724137931035, "grad_norm": 0.830558180809021, "learning_rate": 9.479914837611409e-06, "loss": 0.7568, "step": 336 }, { "epoch": 0.6972413793103448, "grad_norm": 0.8038316369056702, "learning_rate": 9.474552359558167e-06, "loss": 0.7626, "step": 337 }, { "epoch": 0.6993103448275862, "grad_norm": 0.7749502658843994, "learning_rate": 9.469163910189494e-06, "loss": 0.79, "step": 338 }, { "epoch": 0.7013793103448276, "grad_norm": 0.9494272470474243, "learning_rate": 9.463749520781172e-06, "loss": 0.7803, "step": 339 }, { "epoch": 0.7034482758620689, "grad_norm": 0.8260956406593323, "learning_rate": 9.458309222759542e-06, "loss": 0.7729, "step": 340 }, { "epoch": 0.7055172413793104, "grad_norm": 0.8298391699790955, "learning_rate": 9.452843047701324e-06, "loss": 0.8141, "step": 341 }, { "epoch": 0.7075862068965517, "grad_norm": 0.9029043912887573, "learning_rate": 9.44735102733344e-06, "loss": 0.797, "step": 342 }, { "epoch": 0.7096551724137931, "grad_norm": 0.8290296196937561, "learning_rate": 9.441833193532817e-06, "loss": 0.7708, "step": 343 }, { "epoch": 0.7117241379310345, "grad_norm": 0.7795094847679138, "learning_rate": 9.436289578326215e-06, "loss": 0.8196, "step": 344 }, { "epoch": 0.7137931034482758, "grad_norm": 0.7970453500747681, "learning_rate": 9.43072021389003e-06, "loss": 0.822, "step": 345 }, { "epoch": 0.7158620689655173, "grad_norm": 0.8111122250556946, "learning_rate": 9.425125132550115e-06, "loss": 0.771, "step": 346 }, { "epoch": 0.7179310344827586, "grad_norm": 0.782451868057251, "learning_rate": 9.419504366781591e-06, "loss": 0.7865, "step": 347 }, { "epoch": 0.72, "grad_norm": 0.7692107558250427, "learning_rate": 9.413857949208655e-06, "loss": 0.7748, "step": 348 }, { "epoch": 0.7220689655172414, "grad_norm": 0.9149669408798218, "learning_rate": 9.408185912604395e-06, "loss": 0.7847, "step": 349 }, { "epoch": 0.7241379310344828, "grad_norm": 0.812156081199646, "learning_rate": 9.402488289890593e-06, "loss": 0.7921, "step": 350 }, { "epoch": 0.7262068965517241, "grad_norm": 0.8430172801017761, "learning_rate": 9.396765114137544e-06, "loss": 0.7752, "step": 351 }, { "epoch": 0.7282758620689656, "grad_norm": 0.8315995931625366, "learning_rate": 9.391016418563856e-06, "loss": 0.7867, "step": 352 }, { "epoch": 0.7303448275862069, "grad_norm": 0.8469542860984802, "learning_rate": 9.385242236536259e-06, "loss": 0.7573, "step": 353 }, { "epoch": 0.7324137931034482, "grad_norm": 0.8568265438079834, "learning_rate": 9.379442601569415e-06, "loss": 0.7873, "step": 354 }, { "epoch": 0.7344827586206897, "grad_norm": 0.8971392512321472, "learning_rate": 9.37361754732572e-06, "loss": 0.8196, "step": 355 }, { "epoch": 0.736551724137931, "grad_norm": 0.8264282941818237, "learning_rate": 9.367767107615106e-06, "loss": 0.7785, "step": 356 }, { "epoch": 0.7386206896551725, "grad_norm": 0.852152407169342, "learning_rate": 9.36189131639485e-06, "loss": 0.7655, "step": 357 }, { "epoch": 0.7406896551724138, "grad_norm": 0.7565860152244568, "learning_rate": 9.355990207769378e-06, "loss": 0.7672, "step": 358 }, { "epoch": 0.7427586206896551, "grad_norm": 0.8384962677955627, "learning_rate": 9.35006381599006e-06, "loss": 0.7768, "step": 359 }, { "epoch": 0.7448275862068966, "grad_norm": 0.806664764881134, "learning_rate": 9.344112175455013e-06, "loss": 0.7716, "step": 360 }, { "epoch": 0.7468965517241379, "grad_norm": 0.848978579044342, "learning_rate": 9.338135320708912e-06, "loss": 0.817, "step": 361 }, { "epoch": 0.7489655172413793, "grad_norm": 0.8164881467819214, "learning_rate": 9.332133286442772e-06, "loss": 0.7967, "step": 362 }, { "epoch": 0.7510344827586207, "grad_norm": 0.855451226234436, "learning_rate": 9.326106107493762e-06, "loss": 0.7519, "step": 363 }, { "epoch": 0.7531034482758621, "grad_norm": 0.8127675652503967, "learning_rate": 9.32005381884499e-06, "loss": 0.758, "step": 364 }, { "epoch": 0.7551724137931034, "grad_norm": 0.8386479020118713, "learning_rate": 9.313976455625316e-06, "loss": 0.7947, "step": 365 }, { "epoch": 0.7572413793103449, "grad_norm": 0.9732878804206848, "learning_rate": 9.307874053109127e-06, "loss": 0.7968, "step": 366 }, { "epoch": 0.7593103448275862, "grad_norm": 0.9192622900009155, "learning_rate": 9.301746646716152e-06, "loss": 0.7927, "step": 367 }, { "epoch": 0.7613793103448275, "grad_norm": 0.8056217432022095, "learning_rate": 9.295594272011246e-06, "loss": 0.7892, "step": 368 }, { "epoch": 0.763448275862069, "grad_norm": 0.8506579399108887, "learning_rate": 9.289416964704186e-06, "loss": 0.7746, "step": 369 }, { "epoch": 0.7655172413793103, "grad_norm": 0.8824191689491272, "learning_rate": 9.283214760649462e-06, "loss": 0.7749, "step": 370 }, { "epoch": 0.7675862068965518, "grad_norm": 0.894450306892395, "learning_rate": 9.27698769584607e-06, "loss": 0.7792, "step": 371 }, { "epoch": 0.7696551724137931, "grad_norm": 0.8521785140037537, "learning_rate": 9.270735806437306e-06, "loss": 0.8031, "step": 372 }, { "epoch": 0.7717241379310344, "grad_norm": 0.9518225789070129, "learning_rate": 9.26445912871055e-06, "loss": 0.8018, "step": 373 }, { "epoch": 0.7737931034482759, "grad_norm": 0.82342129945755, "learning_rate": 9.258157699097065e-06, "loss": 0.7745, "step": 374 }, { "epoch": 0.7758620689655172, "grad_norm": 0.9217233657836914, "learning_rate": 9.251831554171773e-06, "loss": 0.8179, "step": 375 }, { "epoch": 0.7779310344827586, "grad_norm": 0.9028921127319336, "learning_rate": 9.245480730653055e-06, "loss": 0.7961, "step": 376 }, { "epoch": 0.78, "grad_norm": 0.7794080376625061, "learning_rate": 9.239105265402525e-06, "loss": 0.7874, "step": 377 }, { "epoch": 0.7820689655172414, "grad_norm": 0.8010353446006775, "learning_rate": 9.232705195424832e-06, "loss": 0.7673, "step": 378 }, { "epoch": 0.7841379310344827, "grad_norm": 0.8443918824195862, "learning_rate": 9.226280557867428e-06, "loss": 0.7922, "step": 379 }, { "epoch": 0.7862068965517242, "grad_norm": 0.8193161487579346, "learning_rate": 9.219831390020367e-06, "loss": 0.7668, "step": 380 }, { "epoch": 0.7882758620689655, "grad_norm": 0.8224409818649292, "learning_rate": 9.213357729316077e-06, "loss": 0.7643, "step": 381 }, { "epoch": 0.7903448275862069, "grad_norm": 0.8491160273551941, "learning_rate": 9.206859613329155e-06, "loss": 0.7801, "step": 382 }, { "epoch": 0.7924137931034483, "grad_norm": 0.8366249203681946, "learning_rate": 9.200337079776136e-06, "loss": 0.7496, "step": 383 }, { "epoch": 0.7944827586206896, "grad_norm": 0.7923499345779419, "learning_rate": 9.193790166515282e-06, "loss": 0.7988, "step": 384 }, { "epoch": 0.7965517241379311, "grad_norm": 0.8000715970993042, "learning_rate": 9.187218911546363e-06, "loss": 0.7583, "step": 385 }, { "epoch": 0.7986206896551724, "grad_norm": 0.7873345017433167, "learning_rate": 9.180623353010428e-06, "loss": 0.7631, "step": 386 }, { "epoch": 0.8006896551724138, "grad_norm": 0.8444647789001465, "learning_rate": 9.174003529189592e-06, "loss": 0.7837, "step": 387 }, { "epoch": 0.8027586206896552, "grad_norm": 0.7419801354408264, "learning_rate": 9.167359478506816e-06, "loss": 0.7626, "step": 388 }, { "epoch": 0.8048275862068965, "grad_norm": 0.774566113948822, "learning_rate": 9.160691239525675e-06, "loss": 0.7728, "step": 389 }, { "epoch": 0.8068965517241379, "grad_norm": 0.730536699295044, "learning_rate": 9.153998850950133e-06, "loss": 0.7907, "step": 390 }, { "epoch": 0.8089655172413793, "grad_norm": 0.726324737071991, "learning_rate": 9.147282351624334e-06, "loss": 0.7827, "step": 391 }, { "epoch": 0.8110344827586207, "grad_norm": 0.7642649412155151, "learning_rate": 9.140541780532362e-06, "loss": 0.7417, "step": 392 }, { "epoch": 0.8131034482758621, "grad_norm": 0.8225309252738953, "learning_rate": 9.133777176798013e-06, "loss": 0.8105, "step": 393 }, { "epoch": 0.8151724137931035, "grad_norm": 0.7588528394699097, "learning_rate": 9.126988579684587e-06, "loss": 0.781, "step": 394 }, { "epoch": 0.8172413793103448, "grad_norm": 0.8489255309104919, "learning_rate": 9.120176028594632e-06, "loss": 0.7515, "step": 395 }, { "epoch": 0.8193103448275862, "grad_norm": 0.8431716561317444, "learning_rate": 9.113339563069742e-06, "loss": 0.7931, "step": 396 }, { "epoch": 0.8213793103448276, "grad_norm": 0.8133612871170044, "learning_rate": 9.106479222790312e-06, "loss": 0.7877, "step": 397 }, { "epoch": 0.8234482758620689, "grad_norm": 0.8307138085365295, "learning_rate": 9.099595047575308e-06, "loss": 0.7774, "step": 398 }, { "epoch": 0.8255172413793104, "grad_norm": 0.7466663122177124, "learning_rate": 9.092687077382043e-06, "loss": 0.7667, "step": 399 }, { "epoch": 0.8275862068965517, "grad_norm": 0.8137718439102173, "learning_rate": 9.08575535230594e-06, "loss": 0.8056, "step": 400 }, { "epoch": 0.829655172413793, "grad_norm": 0.8068400621414185, "learning_rate": 9.078799912580305e-06, "loss": 0.7588, "step": 401 }, { "epoch": 0.8317241379310345, "grad_norm": 0.8182150721549988, "learning_rate": 9.071820798576083e-06, "loss": 0.7748, "step": 402 }, { "epoch": 0.8337931034482758, "grad_norm": 0.8240687251091003, "learning_rate": 9.064818050801634e-06, "loss": 0.7854, "step": 403 }, { "epoch": 0.8358620689655173, "grad_norm": 0.8210726976394653, "learning_rate": 9.057791709902491e-06, "loss": 0.7809, "step": 404 }, { "epoch": 0.8379310344827586, "grad_norm": 0.8129264712333679, "learning_rate": 9.050741816661128e-06, "loss": 0.7747, "step": 405 }, { "epoch": 0.84, "grad_norm": 0.7455905079841614, "learning_rate": 9.04366841199672e-06, "loss": 0.777, "step": 406 }, { "epoch": 0.8420689655172414, "grad_norm": 0.8041906356811523, "learning_rate": 9.036571536964915e-06, "loss": 0.7634, "step": 407 }, { "epoch": 0.8441379310344828, "grad_norm": 0.7791020274162292, "learning_rate": 9.029451232757581e-06, "loss": 0.803, "step": 408 }, { "epoch": 0.8462068965517241, "grad_norm": 0.8423375487327576, "learning_rate": 9.022307540702576e-06, "loss": 0.7263, "step": 409 }, { "epoch": 0.8482758620689655, "grad_norm": 0.8119176030158997, "learning_rate": 9.015140502263506e-06, "loss": 0.7459, "step": 410 }, { "epoch": 0.8503448275862069, "grad_norm": 0.8209352493286133, "learning_rate": 9.007950159039488e-06, "loss": 0.7858, "step": 411 }, { "epoch": 0.8524137931034482, "grad_norm": 0.8128482699394226, "learning_rate": 9.0007365527649e-06, "loss": 0.7707, "step": 412 }, { "epoch": 0.8544827586206897, "grad_norm": 0.8830553293228149, "learning_rate": 8.993499725309148e-06, "loss": 0.7579, "step": 413 }, { "epoch": 0.856551724137931, "grad_norm": 0.8541165590286255, "learning_rate": 8.98623971867642e-06, "loss": 0.7638, "step": 414 }, { "epoch": 0.8586206896551725, "grad_norm": 0.7872908115386963, "learning_rate": 8.978956575005432e-06, "loss": 0.7967, "step": 415 }, { "epoch": 0.8606896551724138, "grad_norm": 0.7946335673332214, "learning_rate": 8.971650336569208e-06, "loss": 0.7769, "step": 416 }, { "epoch": 0.8627586206896551, "grad_norm": 0.8130462169647217, "learning_rate": 8.964321045774808e-06, "loss": 0.7889, "step": 417 }, { "epoch": 0.8648275862068966, "grad_norm": 0.7823083400726318, "learning_rate": 8.956968745163093e-06, "loss": 0.7814, "step": 418 }, { "epoch": 0.8668965517241379, "grad_norm": 0.8023502230644226, "learning_rate": 8.949593477408485e-06, "loss": 0.7647, "step": 419 }, { "epoch": 0.8689655172413793, "grad_norm": 0.7985740900039673, "learning_rate": 8.942195285318705e-06, "loss": 0.7794, "step": 420 }, { "epoch": 0.8710344827586207, "grad_norm": 0.8288792371749878, "learning_rate": 8.934774211834538e-06, "loss": 0.7891, "step": 421 }, { "epoch": 0.8731034482758621, "grad_norm": 0.8776261806488037, "learning_rate": 8.927330300029575e-06, "loss": 0.7723, "step": 422 }, { "epoch": 0.8751724137931034, "grad_norm": 0.8444452285766602, "learning_rate": 8.919863593109967e-06, "loss": 0.7812, "step": 423 }, { "epoch": 0.8772413793103448, "grad_norm": 0.8049442768096924, "learning_rate": 8.912374134414171e-06, "loss": 0.7765, "step": 424 }, { "epoch": 0.8793103448275862, "grad_norm": 0.8133780360221863, "learning_rate": 8.904861967412702e-06, "loss": 0.7289, "step": 425 }, { "epoch": 0.8813793103448276, "grad_norm": 0.7836434245109558, "learning_rate": 8.897327135707881e-06, "loss": 0.7887, "step": 426 }, { "epoch": 0.883448275862069, "grad_norm": 0.8233901858329773, "learning_rate": 8.889769683033573e-06, "loss": 0.7623, "step": 427 }, { "epoch": 0.8855172413793103, "grad_norm": 0.8158826231956482, "learning_rate": 8.882189653254951e-06, "loss": 0.7821, "step": 428 }, { "epoch": 0.8875862068965518, "grad_norm": 0.8711895942687988, "learning_rate": 8.874587090368221e-06, "loss": 0.7668, "step": 429 }, { "epoch": 0.8896551724137931, "grad_norm": 0.7956521511077881, "learning_rate": 8.86696203850038e-06, "loss": 0.762, "step": 430 }, { "epoch": 0.8917241379310344, "grad_norm": 0.7590909600257874, "learning_rate": 8.859314541908957e-06, "loss": 0.7269, "step": 431 }, { "epoch": 0.8937931034482759, "grad_norm": 0.74117511510849, "learning_rate": 8.851644644981756e-06, "loss": 0.7703, "step": 432 }, { "epoch": 0.8958620689655172, "grad_norm": 0.7806484699249268, "learning_rate": 8.843952392236595e-06, "loss": 0.7608, "step": 433 }, { "epoch": 0.8979310344827586, "grad_norm": 0.7991644740104675, "learning_rate": 8.83623782832105e-06, "loss": 0.7932, "step": 434 }, { "epoch": 0.9, "grad_norm": 0.7456468939781189, "learning_rate": 8.8285009980122e-06, "loss": 0.7638, "step": 435 }, { "epoch": 0.9020689655172414, "grad_norm": 0.814285933971405, "learning_rate": 8.820741946216357e-06, "loss": 0.7711, "step": 436 }, { "epoch": 0.9041379310344828, "grad_norm": 0.767855703830719, "learning_rate": 8.81296071796882e-06, "loss": 0.792, "step": 437 }, { "epoch": 0.9062068965517242, "grad_norm": 0.8761329054832458, "learning_rate": 8.805157358433595e-06, "loss": 0.7736, "step": 438 }, { "epoch": 0.9082758620689655, "grad_norm": 0.9907826781272888, "learning_rate": 8.797331912903152e-06, "loss": 0.7725, "step": 439 }, { "epoch": 0.9103448275862069, "grad_norm": 0.8171151280403137, "learning_rate": 8.789484426798148e-06, "loss": 0.7593, "step": 440 }, { "epoch": 0.9124137931034483, "grad_norm": 0.8003857731819153, "learning_rate": 8.78161494566717e-06, "loss": 0.7596, "step": 441 }, { "epoch": 0.9144827586206896, "grad_norm": 0.9617862105369568, "learning_rate": 8.77372351518647e-06, "loss": 0.7872, "step": 442 }, { "epoch": 0.9165517241379311, "grad_norm": 0.8046567440032959, "learning_rate": 8.765810181159696e-06, "loss": 0.7482, "step": 443 }, { "epoch": 0.9186206896551724, "grad_norm": 0.7760241031646729, "learning_rate": 8.757874989517631e-06, "loss": 0.7843, "step": 444 }, { "epoch": 0.9206896551724137, "grad_norm": 0.7372346520423889, "learning_rate": 8.74991798631793e-06, "loss": 0.7882, "step": 445 }, { "epoch": 0.9227586206896552, "grad_norm": 0.8651936650276184, "learning_rate": 8.741939217744835e-06, "loss": 0.7872, "step": 446 }, { "epoch": 0.9248275862068965, "grad_norm": 0.7748751640319824, "learning_rate": 8.733938730108928e-06, "loss": 0.7801, "step": 447 }, { "epoch": 0.926896551724138, "grad_norm": 0.8208494782447815, "learning_rate": 8.725916569846854e-06, "loss": 0.7878, "step": 448 }, { "epoch": 0.9289655172413793, "grad_norm": 0.7929314374923706, "learning_rate": 8.717872783521048e-06, "loss": 0.7709, "step": 449 }, { "epoch": 0.9310344827586207, "grad_norm": 0.7704197764396667, "learning_rate": 8.709807417819464e-06, "loss": 0.7916, "step": 450 }, { "epoch": 0.9331034482758621, "grad_norm": 0.8212770223617554, "learning_rate": 8.701720519555315e-06, "loss": 0.7711, "step": 451 }, { "epoch": 0.9351724137931035, "grad_norm": 0.8029863238334656, "learning_rate": 8.69361213566679e-06, "loss": 0.7774, "step": 452 }, { "epoch": 0.9372413793103448, "grad_norm": 0.7679287791252136, "learning_rate": 8.685482313216784e-06, "loss": 0.7461, "step": 453 }, { "epoch": 0.9393103448275862, "grad_norm": 0.771782398223877, "learning_rate": 8.67733109939263e-06, "loss": 0.7808, "step": 454 }, { "epoch": 0.9413793103448276, "grad_norm": 0.7877687811851501, "learning_rate": 8.669158541505817e-06, "loss": 0.7683, "step": 455 }, { "epoch": 0.9434482758620689, "grad_norm": 0.7960418462753296, "learning_rate": 8.660964686991725e-06, "loss": 0.7719, "step": 456 }, { "epoch": 0.9455172413793104, "grad_norm": 0.761481761932373, "learning_rate": 8.65274958340934e-06, "loss": 0.7635, "step": 457 }, { "epoch": 0.9475862068965517, "grad_norm": 0.846750795841217, "learning_rate": 8.644513278440984e-06, "loss": 0.7917, "step": 458 }, { "epoch": 0.9496551724137932, "grad_norm": 0.8491324782371521, "learning_rate": 8.636255819892037e-06, "loss": 0.7826, "step": 459 }, { "epoch": 0.9517241379310345, "grad_norm": 0.8026092648506165, "learning_rate": 8.627977255690658e-06, "loss": 0.7715, "step": 460 }, { "epoch": 0.9537931034482758, "grad_norm": 0.8143620491027832, "learning_rate": 8.61967763388751e-06, "loss": 0.7724, "step": 461 }, { "epoch": 0.9558620689655173, "grad_norm": 0.7983601689338684, "learning_rate": 8.611357002655475e-06, "loss": 0.8044, "step": 462 }, { "epoch": 0.9579310344827586, "grad_norm": 0.9372730851173401, "learning_rate": 8.603015410289387e-06, "loss": 0.7635, "step": 463 }, { "epoch": 0.96, "grad_norm": 0.809413731098175, "learning_rate": 8.594652905205731e-06, "loss": 0.7524, "step": 464 }, { "epoch": 0.9620689655172414, "grad_norm": 0.78413987159729, "learning_rate": 8.586269535942386e-06, "loss": 0.7623, "step": 465 }, { "epoch": 0.9641379310344828, "grad_norm": 0.8303465247154236, "learning_rate": 8.577865351158324e-06, "loss": 0.7832, "step": 466 }, { "epoch": 0.9662068965517241, "grad_norm": 0.9170937538146973, "learning_rate": 8.569440399633337e-06, "loss": 0.7817, "step": 467 }, { "epoch": 0.9682758620689655, "grad_norm": 0.835443377494812, "learning_rate": 8.560994730267751e-06, "loss": 0.7658, "step": 468 }, { "epoch": 0.9703448275862069, "grad_norm": 0.8764263987541199, "learning_rate": 8.552528392082147e-06, "loss": 0.7479, "step": 469 }, { "epoch": 0.9724137931034482, "grad_norm": 0.8065820336341858, "learning_rate": 8.544041434217067e-06, "loss": 0.7469, "step": 470 }, { "epoch": 0.9744827586206897, "grad_norm": 0.895577073097229, "learning_rate": 8.535533905932739e-06, "loss": 0.7716, "step": 471 }, { "epoch": 0.976551724137931, "grad_norm": 0.844038188457489, "learning_rate": 8.52700585660878e-06, "loss": 0.7586, "step": 472 }, { "epoch": 0.9786206896551725, "grad_norm": 0.8391221761703491, "learning_rate": 8.518457335743927e-06, "loss": 0.785, "step": 473 }, { "epoch": 0.9806896551724138, "grad_norm": 0.802529513835907, "learning_rate": 8.509888392955725e-06, "loss": 0.7786, "step": 474 }, { "epoch": 0.9827586206896551, "grad_norm": 0.793171763420105, "learning_rate": 8.50129907798026e-06, "loss": 0.7542, "step": 475 }, { "epoch": 0.9848275862068966, "grad_norm": 0.7531145811080933, "learning_rate": 8.492689440671865e-06, "loss": 0.7837, "step": 476 }, { "epoch": 0.9868965517241379, "grad_norm": 0.7774952054023743, "learning_rate": 8.484059531002822e-06, "loss": 0.7702, "step": 477 }, { "epoch": 0.9889655172413793, "grad_norm": 0.84196537733078, "learning_rate": 8.475409399063081e-06, "loss": 0.7628, "step": 478 }, { "epoch": 0.9910344827586207, "grad_norm": 0.760192334651947, "learning_rate": 8.46673909505997e-06, "loss": 0.7549, "step": 479 }, { "epoch": 0.993103448275862, "grad_norm": 0.7565990686416626, "learning_rate": 8.458048669317897e-06, "loss": 0.7473, "step": 480 }, { "epoch": 0.9951724137931034, "grad_norm": 0.792762041091919, "learning_rate": 8.44933817227806e-06, "loss": 0.7631, "step": 481 }, { "epoch": 0.9972413793103448, "grad_norm": 0.7413747906684875, "learning_rate": 8.440607654498156e-06, "loss": 0.778, "step": 482 }, { "epoch": 0.9993103448275862, "grad_norm": 0.7627434730529785, "learning_rate": 8.43185716665209e-06, "loss": 0.7737, "step": 483 }, { "epoch": 1.0013793103448276, "grad_norm": 0.8035737872123718, "learning_rate": 8.423086759529676e-06, "loss": 0.7007, "step": 484 }, { "epoch": 1.0034482758620689, "grad_norm": 0.7981570363044739, "learning_rate": 8.41429648403634e-06, "loss": 0.6635, "step": 485 }, { "epoch": 1.0055172413793103, "grad_norm": 0.7854894995689392, "learning_rate": 8.405486391192834e-06, "loss": 0.6601, "step": 486 }, { "epoch": 1.0075862068965518, "grad_norm": 0.7979158759117126, "learning_rate": 8.396656532134934e-06, "loss": 0.6369, "step": 487 }, { "epoch": 1.0096551724137932, "grad_norm": 0.8349688053131104, "learning_rate": 8.387806958113139e-06, "loss": 0.6569, "step": 488 }, { "epoch": 1.0117241379310344, "grad_norm": 0.8813797235488892, "learning_rate": 8.378937720492384e-06, "loss": 0.6697, "step": 489 }, { "epoch": 1.013793103448276, "grad_norm": 0.8171966671943665, "learning_rate": 8.370048870751732e-06, "loss": 0.6545, "step": 490 }, { "epoch": 1.0158620689655173, "grad_norm": 0.8515467047691345, "learning_rate": 8.361140460484081e-06, "loss": 0.6476, "step": 491 }, { "epoch": 1.0179310344827586, "grad_norm": 0.8771365284919739, "learning_rate": 8.352212541395859e-06, "loss": 0.6641, "step": 492 }, { "epoch": 1.02, "grad_norm": 0.8186706304550171, "learning_rate": 8.343265165306736e-06, "loss": 0.6558, "step": 493 }, { "epoch": 1.0220689655172415, "grad_norm": 0.821365237236023, "learning_rate": 8.334298384149305e-06, "loss": 0.6526, "step": 494 }, { "epoch": 1.0241379310344827, "grad_norm": 0.8323445320129395, "learning_rate": 8.325312249968795e-06, "loss": 0.6788, "step": 495 }, { "epoch": 1.0262068965517241, "grad_norm": 0.7858123183250427, "learning_rate": 8.316306814922769e-06, "loss": 0.6476, "step": 496 }, { "epoch": 1.0282758620689656, "grad_norm": 0.8636062741279602, "learning_rate": 8.307282131280805e-06, "loss": 0.6494, "step": 497 }, { "epoch": 1.0303448275862068, "grad_norm": 0.8002017736434937, "learning_rate": 8.298238251424216e-06, "loss": 0.6432, "step": 498 }, { "epoch": 1.0324137931034483, "grad_norm": 0.852461040019989, "learning_rate": 8.289175227845727e-06, "loss": 0.6602, "step": 499 }, { "epoch": 1.0344827586206897, "grad_norm": 0.782200813293457, "learning_rate": 8.280093113149182e-06, "loss": 0.6347, "step": 500 }, { "epoch": 1.036551724137931, "grad_norm": 0.7889745831489563, "learning_rate": 8.270991960049231e-06, "loss": 0.6626, "step": 501 }, { "epoch": 1.0386206896551724, "grad_norm": 0.7759472131729126, "learning_rate": 8.261871821371026e-06, "loss": 0.6374, "step": 502 }, { "epoch": 1.0406896551724139, "grad_norm": 0.7325034737586975, "learning_rate": 8.252732750049918e-06, "loss": 0.6531, "step": 503 }, { "epoch": 1.042758620689655, "grad_norm": 0.7706113457679749, "learning_rate": 8.24357479913115e-06, "loss": 0.6526, "step": 504 }, { "epoch": 1.0448275862068965, "grad_norm": 0.7855322360992432, "learning_rate": 8.234398021769541e-06, "loss": 0.6443, "step": 505 }, { "epoch": 1.046896551724138, "grad_norm": 0.7973263263702393, "learning_rate": 8.225202471229187e-06, "loss": 0.6422, "step": 506 }, { "epoch": 1.0489655172413792, "grad_norm": 0.7677010297775269, "learning_rate": 8.215988200883145e-06, "loss": 0.6378, "step": 507 }, { "epoch": 1.0510344827586207, "grad_norm": 0.8065165877342224, "learning_rate": 8.206755264213128e-06, "loss": 0.6291, "step": 508 }, { "epoch": 1.053103448275862, "grad_norm": 0.7429850101470947, "learning_rate": 8.19750371480919e-06, "loss": 0.6671, "step": 509 }, { "epoch": 1.0551724137931036, "grad_norm": 0.8575379252433777, "learning_rate": 8.188233606369422e-06, "loss": 0.6462, "step": 510 }, { "epoch": 1.0572413793103448, "grad_norm": 0.7762051224708557, "learning_rate": 8.178944992699632e-06, "loss": 0.6743, "step": 511 }, { "epoch": 1.0593103448275862, "grad_norm": 0.7541788220405579, "learning_rate": 8.169637927713037e-06, "loss": 0.6351, "step": 512 }, { "epoch": 1.0613793103448277, "grad_norm": 0.7777592539787292, "learning_rate": 8.160312465429952e-06, "loss": 0.6506, "step": 513 }, { "epoch": 1.063448275862069, "grad_norm": 0.7478598952293396, "learning_rate": 8.150968659977472e-06, "loss": 0.6781, "step": 514 }, { "epoch": 1.0655172413793104, "grad_norm": 0.7685233950614929, "learning_rate": 8.141606565589157e-06, "loss": 0.6802, "step": 515 }, { "epoch": 1.0675862068965518, "grad_norm": 0.7681663632392883, "learning_rate": 8.132226236604728e-06, "loss": 0.6421, "step": 516 }, { "epoch": 1.069655172413793, "grad_norm": 0.7492855787277222, "learning_rate": 8.122827727469737e-06, "loss": 0.6445, "step": 517 }, { "epoch": 1.0717241379310345, "grad_norm": 0.8009100556373596, "learning_rate": 8.113411092735262e-06, "loss": 0.6424, "step": 518 }, { "epoch": 1.073793103448276, "grad_norm": 0.7607418894767761, "learning_rate": 8.103976387057579e-06, "loss": 0.6456, "step": 519 }, { "epoch": 1.0758620689655172, "grad_norm": 0.805308997631073, "learning_rate": 8.094523665197862e-06, "loss": 0.6639, "step": 520 }, { "epoch": 1.0779310344827586, "grad_norm": 0.7429332137107849, "learning_rate": 8.085052982021849e-06, "loss": 0.644, "step": 521 }, { "epoch": 1.08, "grad_norm": 0.7760939598083496, "learning_rate": 8.075564392499527e-06, "loss": 0.6472, "step": 522 }, { "epoch": 1.0820689655172413, "grad_norm": 0.8237507343292236, "learning_rate": 8.066057951704821e-06, "loss": 0.643, "step": 523 }, { "epoch": 1.0841379310344827, "grad_norm": 0.7856370806694031, "learning_rate": 8.056533714815268e-06, "loss": 0.658, "step": 524 }, { "epoch": 1.0862068965517242, "grad_norm": 0.7537881731987, "learning_rate": 8.046991737111696e-06, "loss": 0.6441, "step": 525 }, { "epoch": 1.0882758620689654, "grad_norm": 0.7884009480476379, "learning_rate": 8.037432073977902e-06, "loss": 0.6253, "step": 526 }, { "epoch": 1.0903448275862069, "grad_norm": 0.9002606868743896, "learning_rate": 8.027854780900339e-06, "loss": 0.6652, "step": 527 }, { "epoch": 1.0924137931034483, "grad_norm": 0.7838542461395264, "learning_rate": 8.018259913467785e-06, "loss": 0.6783, "step": 528 }, { "epoch": 1.0944827586206896, "grad_norm": 0.74989253282547, "learning_rate": 8.008647527371022e-06, "loss": 0.655, "step": 529 }, { "epoch": 1.096551724137931, "grad_norm": 0.9094736576080322, "learning_rate": 7.999017678402518e-06, "loss": 0.6587, "step": 530 }, { "epoch": 1.0986206896551725, "grad_norm": 0.7564888000488281, "learning_rate": 7.989370422456098e-06, "loss": 0.6577, "step": 531 }, { "epoch": 1.1006896551724137, "grad_norm": 0.806409478187561, "learning_rate": 7.979705815526621e-06, "loss": 0.6666, "step": 532 }, { "epoch": 1.1027586206896551, "grad_norm": 0.7861296534538269, "learning_rate": 7.970023913709652e-06, "loss": 0.6849, "step": 533 }, { "epoch": 1.1048275862068966, "grad_norm": 0.7604372501373291, "learning_rate": 7.960324773201147e-06, "loss": 0.6363, "step": 534 }, { "epoch": 1.106896551724138, "grad_norm": 0.778635561466217, "learning_rate": 7.95060845029711e-06, "loss": 0.6551, "step": 535 }, { "epoch": 1.1089655172413793, "grad_norm": 0.8555791974067688, "learning_rate": 7.94087500139328e-06, "loss": 0.6587, "step": 536 }, { "epoch": 1.1110344827586207, "grad_norm": 0.7580198645591736, "learning_rate": 7.931124482984802e-06, "loss": 0.678, "step": 537 }, { "epoch": 1.1131034482758622, "grad_norm": 0.749212384223938, "learning_rate": 7.921356951665894e-06, "loss": 0.6235, "step": 538 }, { "epoch": 1.1151724137931034, "grad_norm": 0.7692828178405762, "learning_rate": 7.911572464129515e-06, "loss": 0.663, "step": 539 }, { "epoch": 1.1172413793103448, "grad_norm": 0.7994730472564697, "learning_rate": 7.901771077167051e-06, "loss": 0.6318, "step": 540 }, { "epoch": 1.1193103448275863, "grad_norm": 0.7523401975631714, "learning_rate": 7.891952847667973e-06, "loss": 0.6528, "step": 541 }, { "epoch": 1.1213793103448275, "grad_norm": 0.7888882160186768, "learning_rate": 7.882117832619506e-06, "loss": 0.6238, "step": 542 }, { "epoch": 1.123448275862069, "grad_norm": 0.7809697389602661, "learning_rate": 7.872266089106309e-06, "loss": 0.634, "step": 543 }, { "epoch": 1.1255172413793104, "grad_norm": 0.7669185400009155, "learning_rate": 7.862397674310127e-06, "loss": 0.6938, "step": 544 }, { "epoch": 1.1275862068965516, "grad_norm": 0.7955731749534607, "learning_rate": 7.85251264550948e-06, "loss": 0.6711, "step": 545 }, { "epoch": 1.129655172413793, "grad_norm": 0.7307600378990173, "learning_rate": 7.842611060079312e-06, "loss": 0.6459, "step": 546 }, { "epoch": 1.1317241379310345, "grad_norm": 0.8730869889259338, "learning_rate": 7.832692975490668e-06, "loss": 0.6805, "step": 547 }, { "epoch": 1.1337931034482758, "grad_norm": 0.7648000121116638, "learning_rate": 7.822758449310357e-06, "loss": 0.6502, "step": 548 }, { "epoch": 1.1358620689655172, "grad_norm": 0.7544475197792053, "learning_rate": 7.812807539200622e-06, "loss": 0.6807, "step": 549 }, { "epoch": 1.1379310344827587, "grad_norm": 0.7903164029121399, "learning_rate": 7.802840302918794e-06, "loss": 0.6667, "step": 550 }, { "epoch": 1.1400000000000001, "grad_norm": 0.7563683390617371, "learning_rate": 7.792856798316975e-06, "loss": 0.633, "step": 551 }, { "epoch": 1.1420689655172414, "grad_norm": 0.8001116514205933, "learning_rate": 7.782857083341685e-06, "loss": 0.6297, "step": 552 }, { "epoch": 1.1441379310344828, "grad_norm": 0.8235406279563904, "learning_rate": 7.772841216033534e-06, "loss": 0.6372, "step": 553 }, { "epoch": 1.1462068965517243, "grad_norm": 0.7986905574798584, "learning_rate": 7.762809254526884e-06, "loss": 0.6727, "step": 554 }, { "epoch": 1.1482758620689655, "grad_norm": 0.7706049680709839, "learning_rate": 7.752761257049517e-06, "loss": 0.6775, "step": 555 }, { "epoch": 1.150344827586207, "grad_norm": 0.8530354499816895, "learning_rate": 7.74269728192228e-06, "loss": 0.6551, "step": 556 }, { "epoch": 1.1524137931034484, "grad_norm": 0.8466417789459229, "learning_rate": 7.732617387558769e-06, "loss": 0.6418, "step": 557 }, { "epoch": 1.1544827586206896, "grad_norm": 0.7542173862457275, "learning_rate": 7.722521632464972e-06, "loss": 0.6017, "step": 558 }, { "epoch": 1.156551724137931, "grad_norm": 0.7655349373817444, "learning_rate": 7.712410075238936e-06, "loss": 0.6454, "step": 559 }, { "epoch": 1.1586206896551725, "grad_norm": 0.7444957494735718, "learning_rate": 7.702282774570434e-06, "loss": 0.653, "step": 560 }, { "epoch": 1.1606896551724137, "grad_norm": 0.7831370830535889, "learning_rate": 7.692139789240611e-06, "loss": 0.6443, "step": 561 }, { "epoch": 1.1627586206896552, "grad_norm": 0.7594817876815796, "learning_rate": 7.681981178121652e-06, "loss": 0.6422, "step": 562 }, { "epoch": 1.1648275862068966, "grad_norm": 0.7494958639144897, "learning_rate": 7.671807000176434e-06, "loss": 0.6217, "step": 563 }, { "epoch": 1.1668965517241379, "grad_norm": 0.862666130065918, "learning_rate": 7.661617314458194e-06, "loss": 0.6465, "step": 564 }, { "epoch": 1.1689655172413793, "grad_norm": 0.7856999635696411, "learning_rate": 7.651412180110176e-06, "loss": 0.6482, "step": 565 }, { "epoch": 1.1710344827586208, "grad_norm": 0.7493500709533691, "learning_rate": 7.641191656365288e-06, "loss": 0.6378, "step": 566 }, { "epoch": 1.173103448275862, "grad_norm": 0.7933719754219055, "learning_rate": 7.630955802545768e-06, "loss": 0.6481, "step": 567 }, { "epoch": 1.1751724137931034, "grad_norm": 0.7814990282058716, "learning_rate": 7.6207046780628295e-06, "loss": 0.6434, "step": 568 }, { "epoch": 1.177241379310345, "grad_norm": 0.7389763593673706, "learning_rate": 7.61043834241632e-06, "loss": 0.6241, "step": 569 }, { "epoch": 1.1793103448275861, "grad_norm": 0.7658818364143372, "learning_rate": 7.600156855194376e-06, "loss": 0.6446, "step": 570 }, { "epoch": 1.1813793103448276, "grad_norm": 0.8013659119606018, "learning_rate": 7.589860276073082e-06, "loss": 0.6253, "step": 571 }, { "epoch": 1.183448275862069, "grad_norm": 0.7773886919021606, "learning_rate": 7.579548664816112e-06, "loss": 0.6574, "step": 572 }, { "epoch": 1.1855172413793102, "grad_norm": 0.8565280437469482, "learning_rate": 7.569222081274396e-06, "loss": 0.6781, "step": 573 }, { "epoch": 1.1875862068965517, "grad_norm": 0.75011146068573, "learning_rate": 7.558880585385762e-06, "loss": 0.6628, "step": 574 }, { "epoch": 1.1896551724137931, "grad_norm": 0.854688823223114, "learning_rate": 7.548524237174595e-06, "loss": 0.6442, "step": 575 }, { "epoch": 1.1917241379310344, "grad_norm": 0.8342920541763306, "learning_rate": 7.538153096751486e-06, "loss": 0.6493, "step": 576 }, { "epoch": 1.1937931034482758, "grad_norm": 0.7875855565071106, "learning_rate": 7.527767224312883e-06, "loss": 0.6428, "step": 577 }, { "epoch": 1.1958620689655173, "grad_norm": 0.8200221061706543, "learning_rate": 7.517366680140743e-06, "loss": 0.6735, "step": 578 }, { "epoch": 1.1979310344827585, "grad_norm": 0.8583515882492065, "learning_rate": 7.50695152460218e-06, "loss": 0.6432, "step": 579 }, { "epoch": 1.2, "grad_norm": 0.7881141304969788, "learning_rate": 7.496521818149114e-06, "loss": 0.6455, "step": 580 }, { "epoch": 1.2020689655172414, "grad_norm": 0.8400844931602478, "learning_rate": 7.4860776213179264e-06, "loss": 0.6569, "step": 581 }, { "epoch": 1.2041379310344829, "grad_norm": 0.8359171748161316, "learning_rate": 7.4756189947291e-06, "loss": 0.637, "step": 582 }, { "epoch": 1.206206896551724, "grad_norm": 0.8451258540153503, "learning_rate": 7.465145999086874e-06, "loss": 0.6504, "step": 583 }, { "epoch": 1.2082758620689655, "grad_norm": 0.7891321778297424, "learning_rate": 7.454658695178885e-06, "loss": 0.6712, "step": 584 }, { "epoch": 1.210344827586207, "grad_norm": 0.8851990103721619, "learning_rate": 7.44415714387582e-06, "loss": 0.6334, "step": 585 }, { "epoch": 1.2124137931034482, "grad_norm": 0.9024987816810608, "learning_rate": 7.433641406131065e-06, "loss": 0.6604, "step": 586 }, { "epoch": 1.2144827586206897, "grad_norm": 0.7987050414085388, "learning_rate": 7.4231115429803405e-06, "loss": 0.6633, "step": 587 }, { "epoch": 1.2165517241379311, "grad_norm": 0.7416420578956604, "learning_rate": 7.41256761554136e-06, "loss": 0.6401, "step": 588 }, { "epoch": 1.2186206896551723, "grad_norm": 0.8680900931358337, "learning_rate": 7.4020096850134635e-06, "loss": 0.6553, "step": 589 }, { "epoch": 1.2206896551724138, "grad_norm": 0.8396124243736267, "learning_rate": 7.391437812677274e-06, "loss": 0.6374, "step": 590 }, { "epoch": 1.2227586206896552, "grad_norm": 0.7313915491104126, "learning_rate": 7.380852059894331e-06, "loss": 0.6733, "step": 591 }, { "epoch": 1.2248275862068965, "grad_norm": 0.8062239289283752, "learning_rate": 7.370252488106744e-06, "loss": 0.6697, "step": 592 }, { "epoch": 1.226896551724138, "grad_norm": 0.8707490563392639, "learning_rate": 7.359639158836828e-06, "loss": 0.6517, "step": 593 }, { "epoch": 1.2289655172413794, "grad_norm": 0.8169229030609131, "learning_rate": 7.349012133686748e-06, "loss": 0.6507, "step": 594 }, { "epoch": 1.2310344827586206, "grad_norm": 0.8145736455917358, "learning_rate": 7.3383714743381664e-06, "loss": 0.6594, "step": 595 }, { "epoch": 1.233103448275862, "grad_norm": 0.7559329271316528, "learning_rate": 7.327717242551882e-06, "loss": 0.6685, "step": 596 }, { "epoch": 1.2351724137931035, "grad_norm": 0.840366542339325, "learning_rate": 7.317049500167466e-06, "loss": 0.6152, "step": 597 }, { "epoch": 1.237241379310345, "grad_norm": 0.8220683932304382, "learning_rate": 7.306368309102914e-06, "loss": 0.6615, "step": 598 }, { "epoch": 1.2393103448275862, "grad_norm": 0.75614994764328, "learning_rate": 7.295673731354278e-06, "loss": 0.6451, "step": 599 }, { "epoch": 1.2413793103448276, "grad_norm": 0.8044620156288147, "learning_rate": 7.2849658289953084e-06, "loss": 0.6482, "step": 600 }, { "epoch": 1.243448275862069, "grad_norm": 0.7796791791915894, "learning_rate": 7.2742446641770985e-06, "loss": 0.6371, "step": 601 }, { "epoch": 1.2455172413793103, "grad_norm": 0.739811897277832, "learning_rate": 7.263510299127714e-06, "loss": 0.6489, "step": 602 }, { "epoch": 1.2475862068965518, "grad_norm": 0.7257277965545654, "learning_rate": 7.252762796151843e-06, "loss": 0.648, "step": 603 }, { "epoch": 1.2496551724137932, "grad_norm": 0.7809808850288391, "learning_rate": 7.242002217630428e-06, "loss": 0.6584, "step": 604 }, { "epoch": 1.2517241379310344, "grad_norm": 0.8291626572608948, "learning_rate": 7.231228626020303e-06, "loss": 0.652, "step": 605 }, { "epoch": 1.2537931034482759, "grad_norm": 0.7942583560943604, "learning_rate": 7.220442083853834e-06, "loss": 0.6538, "step": 606 }, { "epoch": 1.2558620689655173, "grad_norm": 0.7212340235710144, "learning_rate": 7.209642653738555e-06, "loss": 0.6351, "step": 607 }, { "epoch": 1.2579310344827586, "grad_norm": 0.8261105418205261, "learning_rate": 7.198830398356807e-06, "loss": 0.6725, "step": 608 }, { "epoch": 1.26, "grad_norm": 0.8113548755645752, "learning_rate": 7.188005380465365e-06, "loss": 0.6654, "step": 609 }, { "epoch": 1.2620689655172415, "grad_norm": 0.76350337266922, "learning_rate": 7.177167662895087e-06, "loss": 0.6775, "step": 610 }, { "epoch": 1.2641379310344827, "grad_norm": 0.6983400583267212, "learning_rate": 7.166317308550541e-06, "loss": 0.6494, "step": 611 }, { "epoch": 1.2662068965517241, "grad_norm": 0.7737719416618347, "learning_rate": 7.1554543804096425e-06, "loss": 0.6344, "step": 612 }, { "epoch": 1.2682758620689656, "grad_norm": 0.9658696055412292, "learning_rate": 7.144578941523283e-06, "loss": 0.6472, "step": 613 }, { "epoch": 1.2703448275862068, "grad_norm": 0.7466915845870972, "learning_rate": 7.1336910550149775e-06, "loss": 0.6659, "step": 614 }, { "epoch": 1.2724137931034483, "grad_norm": 0.7958390116691589, "learning_rate": 7.122790784080482e-06, "loss": 0.6792, "step": 615 }, { "epoch": 1.2744827586206897, "grad_norm": 0.8669654726982117, "learning_rate": 7.111878191987441e-06, "loss": 0.6508, "step": 616 }, { "epoch": 1.276551724137931, "grad_norm": 0.8021039366722107, "learning_rate": 7.10095334207501e-06, "loss": 0.6379, "step": 617 }, { "epoch": 1.2786206896551724, "grad_norm": 0.8026660680770874, "learning_rate": 7.090016297753492e-06, "loss": 0.6554, "step": 618 }, { "epoch": 1.2806896551724138, "grad_norm": 0.7808569669723511, "learning_rate": 7.079067122503969e-06, "loss": 0.6693, "step": 619 }, { "epoch": 1.282758620689655, "grad_norm": 0.8397783637046814, "learning_rate": 7.0681058798779335e-06, "loss": 0.6484, "step": 620 }, { "epoch": 1.2848275862068965, "grad_norm": 0.9213140606880188, "learning_rate": 7.057132633496924e-06, "loss": 0.6603, "step": 621 }, { "epoch": 1.286896551724138, "grad_norm": 0.7552454471588135, "learning_rate": 7.046147447052146e-06, "loss": 0.6536, "step": 622 }, { "epoch": 1.2889655172413792, "grad_norm": 0.7723549604415894, "learning_rate": 7.03515038430411e-06, "loss": 0.6549, "step": 623 }, { "epoch": 1.2910344827586206, "grad_norm": 0.965324878692627, "learning_rate": 7.024141509082259e-06, "loss": 0.6552, "step": 624 }, { "epoch": 1.293103448275862, "grad_norm": 0.8971583843231201, "learning_rate": 7.013120885284599e-06, "loss": 0.6667, "step": 625 }, { "epoch": 1.2951724137931033, "grad_norm": 0.8151547908782959, "learning_rate": 7.002088576877325e-06, "loss": 0.6451, "step": 626 }, { "epoch": 1.2972413793103448, "grad_norm": 0.7296470999717712, "learning_rate": 6.991044647894456e-06, "loss": 0.6517, "step": 627 }, { "epoch": 1.2993103448275862, "grad_norm": 0.9281187057495117, "learning_rate": 6.9799891624374565e-06, "loss": 0.6777, "step": 628 }, { "epoch": 1.3013793103448275, "grad_norm": 0.9756149053573608, "learning_rate": 6.968922184674868e-06, "loss": 0.6432, "step": 629 }, { "epoch": 1.303448275862069, "grad_norm": 0.7917487025260925, "learning_rate": 6.957843778841937e-06, "loss": 0.6461, "step": 630 }, { "epoch": 1.3055172413793104, "grad_norm": 0.7548885941505432, "learning_rate": 6.946754009240241e-06, "loss": 0.6323, "step": 631 }, { "epoch": 1.3075862068965518, "grad_norm": 0.8275778889656067, "learning_rate": 6.935652940237313e-06, "loss": 0.6651, "step": 632 }, { "epoch": 1.309655172413793, "grad_norm": 0.8038797974586487, "learning_rate": 6.924540636266272e-06, "loss": 0.6631, "step": 633 }, { "epoch": 1.3117241379310345, "grad_norm": 0.7313756346702576, "learning_rate": 6.913417161825449e-06, "loss": 0.6292, "step": 634 }, { "epoch": 1.313793103448276, "grad_norm": 0.7938227653503418, "learning_rate": 6.902282581478009e-06, "loss": 0.643, "step": 635 }, { "epoch": 1.3158620689655172, "grad_norm": 0.7737557291984558, "learning_rate": 6.891136959851576e-06, "loss": 0.6506, "step": 636 }, { "epoch": 1.3179310344827586, "grad_norm": 0.7778476476669312, "learning_rate": 6.879980361637865e-06, "loss": 0.6683, "step": 637 }, { "epoch": 1.32, "grad_norm": 0.8388012051582336, "learning_rate": 6.868812851592299e-06, "loss": 0.6508, "step": 638 }, { "epoch": 1.3220689655172415, "grad_norm": 0.8127637505531311, "learning_rate": 6.857634494533636e-06, "loss": 0.6521, "step": 639 }, { "epoch": 1.3241379310344827, "grad_norm": 0.718436062335968, "learning_rate": 6.846445355343591e-06, "loss": 0.6625, "step": 640 }, { "epoch": 1.3262068965517242, "grad_norm": 0.7375659346580505, "learning_rate": 6.835245498966461e-06, "loss": 0.6293, "step": 641 }, { "epoch": 1.3282758620689656, "grad_norm": 0.7901964783668518, "learning_rate": 6.824034990408753e-06, "loss": 0.6281, "step": 642 }, { "epoch": 1.3303448275862069, "grad_norm": 0.7931618094444275, "learning_rate": 6.8128138947387966e-06, "loss": 0.6634, "step": 643 }, { "epoch": 1.3324137931034483, "grad_norm": 0.7482447624206543, "learning_rate": 6.801582277086371e-06, "loss": 0.6446, "step": 644 }, { "epoch": 1.3344827586206898, "grad_norm": 0.8103066086769104, "learning_rate": 6.790340202642333e-06, "loss": 0.6436, "step": 645 }, { "epoch": 1.336551724137931, "grad_norm": 0.7419809699058533, "learning_rate": 6.7790877366582255e-06, "loss": 0.6632, "step": 646 }, { "epoch": 1.3386206896551724, "grad_norm": 0.7484923005104065, "learning_rate": 6.76782494444591e-06, "loss": 0.6609, "step": 647 }, { "epoch": 1.340689655172414, "grad_norm": 0.7377297282218933, "learning_rate": 6.756551891377185e-06, "loss": 0.6557, "step": 648 }, { "epoch": 1.3427586206896551, "grad_norm": 0.8004779815673828, "learning_rate": 6.7452686428834045e-06, "loss": 0.6413, "step": 649 }, { "epoch": 1.3448275862068966, "grad_norm": 0.7940952181816101, "learning_rate": 6.733975264455097e-06, "loss": 0.6604, "step": 650 }, { "epoch": 1.346896551724138, "grad_norm": 0.734728991985321, "learning_rate": 6.722671821641591e-06, "loss": 0.6445, "step": 651 }, { "epoch": 1.3489655172413793, "grad_norm": 0.811326801776886, "learning_rate": 6.711358380050628e-06, "loss": 0.6637, "step": 652 }, { "epoch": 1.3510344827586207, "grad_norm": 0.8057741522789001, "learning_rate": 6.700035005347983e-06, "loss": 0.6385, "step": 653 }, { "epoch": 1.3531034482758622, "grad_norm": 0.7723029851913452, "learning_rate": 6.68870176325709e-06, "loss": 0.6514, "step": 654 }, { "epoch": 1.3551724137931034, "grad_norm": 0.7846975326538086, "learning_rate": 6.677358719558655e-06, "loss": 0.6312, "step": 655 }, { "epoch": 1.3572413793103448, "grad_norm": 0.8255680203437805, "learning_rate": 6.666005940090271e-06, "loss": 0.6651, "step": 656 }, { "epoch": 1.3593103448275863, "grad_norm": 0.7784463763237, "learning_rate": 6.654643490746042e-06, "loss": 0.6654, "step": 657 }, { "epoch": 1.3613793103448275, "grad_norm": 0.8567957878112793, "learning_rate": 6.643271437476196e-06, "loss": 0.6493, "step": 658 }, { "epoch": 1.363448275862069, "grad_norm": 0.8170894980430603, "learning_rate": 6.6318898462867086e-06, "loss": 0.674, "step": 659 }, { "epoch": 1.3655172413793104, "grad_norm": 0.777654230594635, "learning_rate": 6.6204987832389115e-06, "loss": 0.7003, "step": 660 }, { "epoch": 1.3675862068965516, "grad_norm": 0.7541632056236267, "learning_rate": 6.609098314449116e-06, "loss": 0.6488, "step": 661 }, { "epoch": 1.369655172413793, "grad_norm": 0.806873083114624, "learning_rate": 6.597688506088224e-06, "loss": 0.6484, "step": 662 }, { "epoch": 1.3717241379310345, "grad_norm": 0.7729332447052002, "learning_rate": 6.586269424381349e-06, "loss": 0.63, "step": 663 }, { "epoch": 1.3737931034482758, "grad_norm": 0.7325103282928467, "learning_rate": 6.574841135607426e-06, "loss": 0.6377, "step": 664 }, { "epoch": 1.3758620689655172, "grad_norm": 0.7824835777282715, "learning_rate": 6.563403706098833e-06, "loss": 0.6493, "step": 665 }, { "epoch": 1.3779310344827587, "grad_norm": 0.8269271850585938, "learning_rate": 6.551957202241001e-06, "loss": 0.6763, "step": 666 }, { "epoch": 1.38, "grad_norm": 0.7548254728317261, "learning_rate": 6.540501690472032e-06, "loss": 0.6463, "step": 667 }, { "epoch": 1.3820689655172413, "grad_norm": 0.728628396987915, "learning_rate": 6.529037237282309e-06, "loss": 0.6571, "step": 668 }, { "epoch": 1.3841379310344828, "grad_norm": 0.7487698793411255, "learning_rate": 6.517563909214119e-06, "loss": 0.6819, "step": 669 }, { "epoch": 1.386206896551724, "grad_norm": 0.7777491807937622, "learning_rate": 6.5060817728612544e-06, "loss": 0.6384, "step": 670 }, { "epoch": 1.3882758620689655, "grad_norm": 0.7405235171318054, "learning_rate": 6.4945908948686355e-06, "loss": 0.6637, "step": 671 }, { "epoch": 1.390344827586207, "grad_norm": 0.730686366558075, "learning_rate": 6.4830913419319205e-06, "loss": 0.6264, "step": 672 }, { "epoch": 1.3924137931034481, "grad_norm": 0.7911974191665649, "learning_rate": 6.471583180797121e-06, "loss": 0.6251, "step": 673 }, { "epoch": 1.3944827586206896, "grad_norm": 0.7909592390060425, "learning_rate": 6.460066478260209e-06, "loss": 0.6639, "step": 674 }, { "epoch": 1.396551724137931, "grad_norm": 0.7824447154998779, "learning_rate": 6.448541301166737e-06, "loss": 0.6771, "step": 675 }, { "epoch": 1.3986206896551723, "grad_norm": 0.7654966711997986, "learning_rate": 6.437007716411441e-06, "loss": 0.6661, "step": 676 }, { "epoch": 1.4006896551724137, "grad_norm": 0.7122900485992432, "learning_rate": 6.4254657909378615e-06, "loss": 0.6724, "step": 677 }, { "epoch": 1.4027586206896552, "grad_norm": 0.7968724370002747, "learning_rate": 6.4139155917379445e-06, "loss": 0.6348, "step": 678 }, { "epoch": 1.4048275862068966, "grad_norm": 0.7387704849243164, "learning_rate": 6.402357185851669e-06, "loss": 0.6537, "step": 679 }, { "epoch": 1.4068965517241379, "grad_norm": 0.7359346747398376, "learning_rate": 6.3907906403666355e-06, "loss": 0.6649, "step": 680 }, { "epoch": 1.4089655172413793, "grad_norm": 0.7153935432434082, "learning_rate": 6.379216022417695e-06, "loss": 0.6574, "step": 681 }, { "epoch": 1.4110344827586208, "grad_norm": 0.7676519751548767, "learning_rate": 6.367633399186554e-06, "loss": 0.6811, "step": 682 }, { "epoch": 1.4131034482758622, "grad_norm": 0.8044560551643372, "learning_rate": 6.3560428379013795e-06, "loss": 0.6496, "step": 683 }, { "epoch": 1.4151724137931034, "grad_norm": 0.7096127867698669, "learning_rate": 6.344444405836414e-06, "loss": 0.6493, "step": 684 }, { "epoch": 1.4172413793103449, "grad_norm": 0.7820800542831421, "learning_rate": 6.332838170311586e-06, "loss": 0.6561, "step": 685 }, { "epoch": 1.4193103448275863, "grad_norm": 0.7546415328979492, "learning_rate": 6.321224198692115e-06, "loss": 0.672, "step": 686 }, { "epoch": 1.4213793103448276, "grad_norm": 0.7297517657279968, "learning_rate": 6.309602558388122e-06, "loss": 0.6697, "step": 687 }, { "epoch": 1.423448275862069, "grad_norm": 0.8077932000160217, "learning_rate": 6.297973316854241e-06, "loss": 0.6603, "step": 688 }, { "epoch": 1.4255172413793105, "grad_norm": 0.7859988808631897, "learning_rate": 6.286336541589224e-06, "loss": 0.6724, "step": 689 }, { "epoch": 1.4275862068965517, "grad_norm": 0.7754270434379578, "learning_rate": 6.27469230013555e-06, "loss": 0.6529, "step": 690 }, { "epoch": 1.4296551724137931, "grad_norm": 0.7747129797935486, "learning_rate": 6.263040660079031e-06, "loss": 0.635, "step": 691 }, { "epoch": 1.4317241379310346, "grad_norm": 0.7329117655754089, "learning_rate": 6.25138168904843e-06, "loss": 0.6489, "step": 692 }, { "epoch": 1.4337931034482758, "grad_norm": 0.7887305021286011, "learning_rate": 6.239715454715054e-06, "loss": 0.6317, "step": 693 }, { "epoch": 1.4358620689655173, "grad_norm": 0.7606728672981262, "learning_rate": 6.22804202479237e-06, "loss": 0.6604, "step": 694 }, { "epoch": 1.4379310344827587, "grad_norm": 0.7513334155082703, "learning_rate": 6.216361467035608e-06, "loss": 0.6437, "step": 695 }, { "epoch": 1.44, "grad_norm": 0.8091455698013306, "learning_rate": 6.204673849241372e-06, "loss": 0.6305, "step": 696 }, { "epoch": 1.4420689655172414, "grad_norm": 0.7892200946807861, "learning_rate": 6.192979239247243e-06, "loss": 0.6549, "step": 697 }, { "epoch": 1.4441379310344828, "grad_norm": 0.7235889434814453, "learning_rate": 6.181277704931386e-06, "loss": 0.651, "step": 698 }, { "epoch": 1.446206896551724, "grad_norm": 0.7446338534355164, "learning_rate": 6.169569314212157e-06, "loss": 0.636, "step": 699 }, { "epoch": 1.4482758620689655, "grad_norm": 0.7943323254585266, "learning_rate": 6.1578541350477076e-06, "loss": 0.6331, "step": 700 }, { "epoch": 1.450344827586207, "grad_norm": 0.6900318264961243, "learning_rate": 6.146132235435591e-06, "loss": 0.6569, "step": 701 }, { "epoch": 1.4524137931034482, "grad_norm": 0.7237994074821472, "learning_rate": 6.1344036834123695e-06, "loss": 0.6224, "step": 702 }, { "epoch": 1.4544827586206897, "grad_norm": 0.7555676102638245, "learning_rate": 6.1226685470532125e-06, "loss": 0.6633, "step": 703 }, { "epoch": 1.456551724137931, "grad_norm": 0.72216796875, "learning_rate": 6.1109268944715125e-06, "loss": 0.6551, "step": 704 }, { "epoch": 1.4586206896551723, "grad_norm": 0.7122531533241272, "learning_rate": 6.099178793818479e-06, "loss": 0.6363, "step": 705 }, { "epoch": 1.4606896551724138, "grad_norm": 0.7397927045822144, "learning_rate": 6.0874243132827505e-06, "loss": 0.6569, "step": 706 }, { "epoch": 1.4627586206896552, "grad_norm": 0.729145348072052, "learning_rate": 6.075663521089994e-06, "loss": 0.628, "step": 707 }, { "epoch": 1.4648275862068965, "grad_norm": 0.7200533151626587, "learning_rate": 6.06389648550251e-06, "loss": 0.6554, "step": 708 }, { "epoch": 1.466896551724138, "grad_norm": 0.7103410959243774, "learning_rate": 6.0521232748188416e-06, "loss": 0.6481, "step": 709 }, { "epoch": 1.4689655172413794, "grad_norm": 0.7074390053749084, "learning_rate": 6.040343957373367e-06, "loss": 0.6274, "step": 710 }, { "epoch": 1.4710344827586206, "grad_norm": 0.734455406665802, "learning_rate": 6.028558601535915e-06, "loss": 0.661, "step": 711 }, { "epoch": 1.473103448275862, "grad_norm": 0.7593621015548706, "learning_rate": 6.016767275711359e-06, "loss": 0.6339, "step": 712 }, { "epoch": 1.4751724137931035, "grad_norm": 0.7968884706497192, "learning_rate": 6.0049700483392256e-06, "loss": 0.6564, "step": 713 }, { "epoch": 1.4772413793103447, "grad_norm": 0.7559491991996765, "learning_rate": 5.993166987893294e-06, "loss": 0.6525, "step": 714 }, { "epoch": 1.4793103448275862, "grad_norm": 0.7332967519760132, "learning_rate": 5.981358162881202e-06, "loss": 0.6591, "step": 715 }, { "epoch": 1.4813793103448276, "grad_norm": 0.7570938467979431, "learning_rate": 5.969543641844044e-06, "loss": 0.6558, "step": 716 }, { "epoch": 1.4834482758620688, "grad_norm": 0.7297980189323425, "learning_rate": 5.957723493355977e-06, "loss": 0.6396, "step": 717 }, { "epoch": 1.4855172413793103, "grad_norm": 0.7381563782691956, "learning_rate": 5.945897786023817e-06, "loss": 0.6539, "step": 718 }, { "epoch": 1.4875862068965517, "grad_norm": 0.7127878665924072, "learning_rate": 5.9340665884866535e-06, "loss": 0.6587, "step": 719 }, { "epoch": 1.489655172413793, "grad_norm": 0.7142752408981323, "learning_rate": 5.922229969415432e-06, "loss": 0.6867, "step": 720 }, { "epoch": 1.4917241379310344, "grad_norm": 0.7629202604293823, "learning_rate": 5.910387997512573e-06, "loss": 0.6381, "step": 721 }, { "epoch": 1.4937931034482759, "grad_norm": 0.716674268245697, "learning_rate": 5.898540741511564e-06, "loss": 0.6406, "step": 722 }, { "epoch": 1.4958620689655173, "grad_norm": 0.732208251953125, "learning_rate": 5.8866882701765605e-06, "loss": 0.6366, "step": 723 }, { "epoch": 1.4979310344827585, "grad_norm": 0.7054344415664673, "learning_rate": 5.8748306523019925e-06, "loss": 0.6491, "step": 724 }, { "epoch": 1.5, "grad_norm": 0.7091671228408813, "learning_rate": 5.86296795671216e-06, "loss": 0.6448, "step": 725 }, { "epoch": 1.5020689655172412, "grad_norm": 0.7083839178085327, "learning_rate": 5.851100252260835e-06, "loss": 0.6302, "step": 726 }, { "epoch": 1.504137931034483, "grad_norm": 0.727642834186554, "learning_rate": 5.839227607830862e-06, "loss": 0.6375, "step": 727 }, { "epoch": 1.5062068965517241, "grad_norm": 0.7836138606071472, "learning_rate": 5.827350092333758e-06, "loss": 0.6429, "step": 728 }, { "epoch": 1.5082758620689654, "grad_norm": 0.7098393440246582, "learning_rate": 5.815467774709314e-06, "loss": 0.6579, "step": 729 }, { "epoch": 1.510344827586207, "grad_norm": 0.749321460723877, "learning_rate": 5.803580723925193e-06, "loss": 0.6361, "step": 730 }, { "epoch": 1.5124137931034483, "grad_norm": 0.7477707862854004, "learning_rate": 5.791689008976531e-06, "loss": 0.6525, "step": 731 }, { "epoch": 1.5144827586206897, "grad_norm": 0.7622563242912292, "learning_rate": 5.779792698885534e-06, "loss": 0.667, "step": 732 }, { "epoch": 1.5165517241379312, "grad_norm": 0.7400147318840027, "learning_rate": 5.767891862701081e-06, "loss": 0.618, "step": 733 }, { "epoch": 1.5186206896551724, "grad_norm": 0.7715649008750916, "learning_rate": 5.755986569498321e-06, "loss": 0.6501, "step": 734 }, { "epoch": 1.5206896551724138, "grad_norm": 0.7207207679748535, "learning_rate": 5.744076888378272e-06, "loss": 0.6616, "step": 735 }, { "epoch": 1.5227586206896553, "grad_norm": 0.7672625780105591, "learning_rate": 5.732162888467421e-06, "loss": 0.6117, "step": 736 }, { "epoch": 1.5248275862068965, "grad_norm": 0.8607432246208191, "learning_rate": 5.7202446389173225e-06, "loss": 0.6272, "step": 737 }, { "epoch": 1.526896551724138, "grad_norm": 0.8004416227340698, "learning_rate": 5.708322208904196e-06, "loss": 0.6538, "step": 738 }, { "epoch": 1.5289655172413794, "grad_norm": 0.7796547412872314, "learning_rate": 5.696395667628526e-06, "loss": 0.6301, "step": 739 }, { "epoch": 1.5310344827586206, "grad_norm": 0.7893610596656799, "learning_rate": 5.6844650843146595e-06, "loss": 0.6299, "step": 740 }, { "epoch": 1.533103448275862, "grad_norm": 0.7845308184623718, "learning_rate": 5.672530528210405e-06, "loss": 0.6635, "step": 741 }, { "epoch": 1.5351724137931035, "grad_norm": 0.7558560371398926, "learning_rate": 5.660592068586629e-06, "loss": 0.6637, "step": 742 }, { "epoch": 1.5372413793103448, "grad_norm": 0.7075406908988953, "learning_rate": 5.648649774736855e-06, "loss": 0.6433, "step": 743 }, { "epoch": 1.5393103448275862, "grad_norm": 0.7514592409133911, "learning_rate": 5.6367037159768625e-06, "loss": 0.6547, "step": 744 }, { "epoch": 1.5413793103448277, "grad_norm": 0.7782887816429138, "learning_rate": 5.624753961644281e-06, "loss": 0.6541, "step": 745 }, { "epoch": 1.543448275862069, "grad_norm": 0.7897704243659973, "learning_rate": 5.612800581098193e-06, "loss": 0.6582, "step": 746 }, { "epoch": 1.5455172413793103, "grad_norm": 0.7592941522598267, "learning_rate": 5.600843643718728e-06, "loss": 0.6309, "step": 747 }, { "epoch": 1.5475862068965518, "grad_norm": 0.7635563015937805, "learning_rate": 5.588883218906653e-06, "loss": 0.6407, "step": 748 }, { "epoch": 1.549655172413793, "grad_norm": 0.7806128859519958, "learning_rate": 5.57691937608299e-06, "loss": 0.659, "step": 749 }, { "epoch": 1.5517241379310345, "grad_norm": 0.7472797632217407, "learning_rate": 5.564952184688588e-06, "loss": 0.661, "step": 750 }, { "epoch": 1.553793103448276, "grad_norm": 0.7442753314971924, "learning_rate": 5.552981714183738e-06, "loss": 0.6582, "step": 751 }, { "epoch": 1.5558620689655172, "grad_norm": 0.7324591279029846, "learning_rate": 5.5410080340477634e-06, "loss": 0.6454, "step": 752 }, { "epoch": 1.5579310344827586, "grad_norm": 0.7217804789543152, "learning_rate": 5.529031213778615e-06, "loss": 0.6441, "step": 753 }, { "epoch": 1.56, "grad_norm": 0.7129271626472473, "learning_rate": 5.517051322892468e-06, "loss": 0.6642, "step": 754 }, { "epoch": 1.5620689655172413, "grad_norm": 0.6992490291595459, "learning_rate": 5.505068430923327e-06, "loss": 0.6382, "step": 755 }, { "epoch": 1.5641379310344827, "grad_norm": 0.7824325561523438, "learning_rate": 5.4930826074226085e-06, "loss": 0.6657, "step": 756 }, { "epoch": 1.5662068965517242, "grad_norm": 0.7302783727645874, "learning_rate": 5.481093921958749e-06, "loss": 0.6643, "step": 757 }, { "epoch": 1.5682758620689654, "grad_norm": 0.7289795875549316, "learning_rate": 5.469102444116791e-06, "loss": 0.6429, "step": 758 }, { "epoch": 1.5703448275862069, "grad_norm": 0.7278138995170593, "learning_rate": 5.457108243497993e-06, "loss": 0.6518, "step": 759 }, { "epoch": 1.5724137931034483, "grad_norm": 0.7361022233963013, "learning_rate": 5.445111389719408e-06, "loss": 0.6445, "step": 760 }, { "epoch": 1.5744827586206895, "grad_norm": 0.7398816347122192, "learning_rate": 5.433111952413496e-06, "loss": 0.6601, "step": 761 }, { "epoch": 1.576551724137931, "grad_norm": 0.7518286108970642, "learning_rate": 5.421110001227705e-06, "loss": 0.6119, "step": 762 }, { "epoch": 1.5786206896551724, "grad_norm": 0.766607403755188, "learning_rate": 5.409105605824082e-06, "loss": 0.6555, "step": 763 }, { "epoch": 1.5806896551724137, "grad_norm": 0.7912081480026245, "learning_rate": 5.3970988358788565e-06, "loss": 0.6652, "step": 764 }, { "epoch": 1.5827586206896553, "grad_norm": 0.7829515337944031, "learning_rate": 5.385089761082039e-06, "loss": 0.6272, "step": 765 }, { "epoch": 1.5848275862068966, "grad_norm": 0.7310935854911804, "learning_rate": 5.3730784511370204e-06, "loss": 0.6714, "step": 766 }, { "epoch": 1.5868965517241378, "grad_norm": 0.7284607291221619, "learning_rate": 5.361064975760166e-06, "loss": 0.6563, "step": 767 }, { "epoch": 1.5889655172413795, "grad_norm": 0.7712138891220093, "learning_rate": 5.349049404680407e-06, "loss": 0.6611, "step": 768 }, { "epoch": 1.5910344827586207, "grad_norm": 0.7651541233062744, "learning_rate": 5.3370318076388405e-06, "loss": 0.6278, "step": 769 }, { "epoch": 1.593103448275862, "grad_norm": 0.7919555902481079, "learning_rate": 5.32501225438832e-06, "loss": 0.6497, "step": 770 }, { "epoch": 1.5951724137931036, "grad_norm": 0.6936086416244507, "learning_rate": 5.3129908146930565e-06, "loss": 0.6327, "step": 771 }, { "epoch": 1.5972413793103448, "grad_norm": 0.7541401386260986, "learning_rate": 5.300967558328208e-06, "loss": 0.6141, "step": 772 }, { "epoch": 1.599310344827586, "grad_norm": 0.785559892654419, "learning_rate": 5.288942555079479e-06, "loss": 0.6763, "step": 773 }, { "epoch": 1.6013793103448277, "grad_norm": 0.7877768874168396, "learning_rate": 5.2769158747427115e-06, "loss": 0.6647, "step": 774 }, { "epoch": 1.603448275862069, "grad_norm": 0.7559152245521545, "learning_rate": 5.264887587123483e-06, "loss": 0.6589, "step": 775 }, { "epoch": 1.6055172413793104, "grad_norm": 0.7801457643508911, "learning_rate": 5.2528577620366986e-06, "loss": 0.6643, "step": 776 }, { "epoch": 1.6075862068965519, "grad_norm": 0.7516634464263916, "learning_rate": 5.240826469306187e-06, "loss": 0.6471, "step": 777 }, { "epoch": 1.609655172413793, "grad_norm": 0.7641648054122925, "learning_rate": 5.228793778764297e-06, "loss": 0.6796, "step": 778 }, { "epoch": 1.6117241379310345, "grad_norm": 0.7582610845565796, "learning_rate": 5.216759760251493e-06, "loss": 0.6602, "step": 779 }, { "epoch": 1.613793103448276, "grad_norm": 0.7532828450202942, "learning_rate": 5.204724483615941e-06, "loss": 0.6594, "step": 780 }, { "epoch": 1.6158620689655172, "grad_norm": 0.774940550327301, "learning_rate": 5.1926880187131134e-06, "loss": 0.656, "step": 781 }, { "epoch": 1.6179310344827587, "grad_norm": 0.7041785717010498, "learning_rate": 5.18065043540538e-06, "loss": 0.645, "step": 782 }, { "epoch": 1.62, "grad_norm": 0.7526170611381531, "learning_rate": 5.168611803561599e-06, "loss": 0.658, "step": 783 }, { "epoch": 1.6220689655172413, "grad_norm": 0.8174700736999512, "learning_rate": 5.156572193056718e-06, "loss": 0.6502, "step": 784 }, { "epoch": 1.6241379310344828, "grad_norm": 0.7578688263893127, "learning_rate": 5.144531673771364e-06, "loss": 0.6664, "step": 785 }, { "epoch": 1.6262068965517242, "grad_norm": 0.7004210948944092, "learning_rate": 5.132490315591437e-06, "loss": 0.6351, "step": 786 }, { "epoch": 1.6282758620689655, "grad_norm": 0.7185120582580566, "learning_rate": 5.1204481884077075e-06, "loss": 0.6148, "step": 787 }, { "epoch": 1.630344827586207, "grad_norm": 0.8190487623214722, "learning_rate": 5.10840536211541e-06, "loss": 0.6241, "step": 788 }, { "epoch": 1.6324137931034484, "grad_norm": 0.8276094198226929, "learning_rate": 5.096361906613836e-06, "loss": 0.6241, "step": 789 }, { "epoch": 1.6344827586206896, "grad_norm": 0.7432799339294434, "learning_rate": 5.084317891805928e-06, "loss": 0.6408, "step": 790 }, { "epoch": 1.636551724137931, "grad_norm": 0.7277430891990662, "learning_rate": 5.072273387597877e-06, "loss": 0.6406, "step": 791 }, { "epoch": 1.6386206896551725, "grad_norm": 0.7360126972198486, "learning_rate": 5.0602284638987145e-06, "loss": 0.6422, "step": 792 }, { "epoch": 1.6406896551724137, "grad_norm": 0.8584643006324768, "learning_rate": 5.048183190619904e-06, "loss": 0.6329, "step": 793 }, { "epoch": 1.6427586206896552, "grad_norm": 0.7649242281913757, "learning_rate": 5.036137637674943e-06, "loss": 0.6356, "step": 794 }, { "epoch": 1.6448275862068966, "grad_norm": 0.7129114270210266, "learning_rate": 5.0240918749789455e-06, "loss": 0.6334, "step": 795 }, { "epoch": 1.6468965517241378, "grad_norm": 0.7862198352813721, "learning_rate": 5.01204597244825e-06, "loss": 0.675, "step": 796 }, { "epoch": 1.6489655172413793, "grad_norm": 0.7846953272819519, "learning_rate": 5e-06, "loss": 0.6573, "step": 797 }, { "epoch": 1.6510344827586207, "grad_norm": 0.7969672083854675, "learning_rate": 4.987954027551751e-06, "loss": 0.652, "step": 798 }, { "epoch": 1.653103448275862, "grad_norm": 0.6976673007011414, "learning_rate": 4.975908125021055e-06, "loss": 0.6251, "step": 799 }, { "epoch": 1.6551724137931034, "grad_norm": 0.7470307946205139, "learning_rate": 4.963862362325058e-06, "loss": 0.6489, "step": 800 }, { "epoch": 1.6572413793103449, "grad_norm": 0.7507964372634888, "learning_rate": 4.951816809380098e-06, "loss": 0.6355, "step": 801 }, { "epoch": 1.659310344827586, "grad_norm": 0.7489534616470337, "learning_rate": 4.939771536101286e-06, "loss": 0.6555, "step": 802 }, { "epoch": 1.6613793103448276, "grad_norm": 0.7056592106819153, "learning_rate": 4.9277266124021245e-06, "loss": 0.6497, "step": 803 }, { "epoch": 1.663448275862069, "grad_norm": 0.7045999765396118, "learning_rate": 4.915682108194073e-06, "loss": 0.6547, "step": 804 }, { "epoch": 1.6655172413793102, "grad_norm": 0.7271073460578918, "learning_rate": 4.903638093386167e-06, "loss": 0.6267, "step": 805 }, { "epoch": 1.6675862068965517, "grad_norm": 0.7735942602157593, "learning_rate": 4.891594637884591e-06, "loss": 0.6369, "step": 806 }, { "epoch": 1.6696551724137931, "grad_norm": 0.7303586006164551, "learning_rate": 4.879551811592295e-06, "loss": 0.6438, "step": 807 }, { "epoch": 1.6717241379310344, "grad_norm": 0.734630286693573, "learning_rate": 4.867509684408564e-06, "loss": 0.6273, "step": 808 }, { "epoch": 1.6737931034482758, "grad_norm": 0.7351275682449341, "learning_rate": 4.855468326228638e-06, "loss": 0.6628, "step": 809 }, { "epoch": 1.6758620689655173, "grad_norm": 0.7179361581802368, "learning_rate": 4.843427806943283e-06, "loss": 0.6422, "step": 810 }, { "epoch": 1.6779310344827585, "grad_norm": 0.6784632802009583, "learning_rate": 4.831388196438402e-06, "loss": 0.6407, "step": 811 }, { "epoch": 1.6800000000000002, "grad_norm": 0.73552006483078, "learning_rate": 4.819349564594622e-06, "loss": 0.6448, "step": 812 }, { "epoch": 1.6820689655172414, "grad_norm": 0.7421280145645142, "learning_rate": 4.807311981286888e-06, "loss": 0.6621, "step": 813 }, { "epoch": 1.6841379310344826, "grad_norm": 0.6813828349113464, "learning_rate": 4.79527551638406e-06, "loss": 0.6442, "step": 814 }, { "epoch": 1.6862068965517243, "grad_norm": 0.7029781937599182, "learning_rate": 4.783240239748509e-06, "loss": 0.6351, "step": 815 }, { "epoch": 1.6882758620689655, "grad_norm": 0.6841586828231812, "learning_rate": 4.7712062212357045e-06, "loss": 0.6373, "step": 816 }, { "epoch": 1.6903448275862067, "grad_norm": 0.7366838455200195, "learning_rate": 4.7591735306938144e-06, "loss": 0.6733, "step": 817 }, { "epoch": 1.6924137931034484, "grad_norm": 0.7882569432258606, "learning_rate": 4.747142237963305e-06, "loss": 0.6266, "step": 818 }, { "epoch": 1.6944827586206896, "grad_norm": 0.6938716173171997, "learning_rate": 4.7351124128765185e-06, "loss": 0.6318, "step": 819 }, { "epoch": 1.6965517241379309, "grad_norm": 0.7524465918540955, "learning_rate": 4.723084125257291e-06, "loss": 0.6437, "step": 820 }, { "epoch": 1.6986206896551725, "grad_norm": 0.7383887767791748, "learning_rate": 4.711057444920522e-06, "loss": 0.6525, "step": 821 }, { "epoch": 1.7006896551724138, "grad_norm": 0.7824906706809998, "learning_rate": 4.699032441671794e-06, "loss": 0.6458, "step": 822 }, { "epoch": 1.7027586206896552, "grad_norm": 0.7252950072288513, "learning_rate": 4.687009185306945e-06, "loss": 0.6536, "step": 823 }, { "epoch": 1.7048275862068967, "grad_norm": 0.7240720391273499, "learning_rate": 4.67498774561168e-06, "loss": 0.6583, "step": 824 }, { "epoch": 1.706896551724138, "grad_norm": 0.739566445350647, "learning_rate": 4.662968192361161e-06, "loss": 0.6758, "step": 825 }, { "epoch": 1.7089655172413794, "grad_norm": 0.7385093569755554, "learning_rate": 4.650950595319593e-06, "loss": 0.6273, "step": 826 }, { "epoch": 1.7110344827586208, "grad_norm": 0.7286804914474487, "learning_rate": 4.6389350242398354e-06, "loss": 0.6436, "step": 827 }, { "epoch": 1.713103448275862, "grad_norm": 0.7011599540710449, "learning_rate": 4.6269215488629795e-06, "loss": 0.6532, "step": 828 }, { "epoch": 1.7151724137931035, "grad_norm": 0.7191659212112427, "learning_rate": 4.614910238917963e-06, "loss": 0.6544, "step": 829 }, { "epoch": 1.717241379310345, "grad_norm": 0.7047889232635498, "learning_rate": 4.602901164121145e-06, "loss": 0.6421, "step": 830 }, { "epoch": 1.7193103448275862, "grad_norm": 0.7209073901176453, "learning_rate": 4.59089439417592e-06, "loss": 0.6289, "step": 831 }, { "epoch": 1.7213793103448276, "grad_norm": 0.7176074385643005, "learning_rate": 4.578889998772296e-06, "loss": 0.642, "step": 832 }, { "epoch": 1.723448275862069, "grad_norm": 0.7749189138412476, "learning_rate": 4.5668880475865074e-06, "loss": 0.6755, "step": 833 }, { "epoch": 1.7255172413793103, "grad_norm": 0.689206600189209, "learning_rate": 4.554888610280593e-06, "loss": 0.6552, "step": 834 }, { "epoch": 1.7275862068965517, "grad_norm": 0.7584705352783203, "learning_rate": 4.542891756502008e-06, "loss": 0.6109, "step": 835 }, { "epoch": 1.7296551724137932, "grad_norm": 0.7560999393463135, "learning_rate": 4.5308975558832095e-06, "loss": 0.6619, "step": 836 }, { "epoch": 1.7317241379310344, "grad_norm": 0.8325402140617371, "learning_rate": 4.518906078041252e-06, "loss": 0.6344, "step": 837 }, { "epoch": 1.7337931034482759, "grad_norm": 0.7222752571105957, "learning_rate": 4.506917392577393e-06, "loss": 0.6255, "step": 838 }, { "epoch": 1.7358620689655173, "grad_norm": 0.7850416898727417, "learning_rate": 4.494931569076674e-06, "loss": 0.6213, "step": 839 }, { "epoch": 1.7379310344827585, "grad_norm": 0.7065498232841492, "learning_rate": 4.482948677107533e-06, "loss": 0.6146, "step": 840 }, { "epoch": 1.74, "grad_norm": 0.6791179776191711, "learning_rate": 4.4709687862213866e-06, "loss": 0.6461, "step": 841 }, { "epoch": 1.7420689655172414, "grad_norm": 0.750182032585144, "learning_rate": 4.458991965952238e-06, "loss": 0.652, "step": 842 }, { "epoch": 1.7441379310344827, "grad_norm": 0.700097382068634, "learning_rate": 4.447018285816263e-06, "loss": 0.6437, "step": 843 }, { "epoch": 1.7462068965517241, "grad_norm": 0.7411333322525024, "learning_rate": 4.435047815311414e-06, "loss": 0.6502, "step": 844 }, { "epoch": 1.7482758620689656, "grad_norm": 0.7353593707084656, "learning_rate": 4.423080623917012e-06, "loss": 0.6502, "step": 845 }, { "epoch": 1.7503448275862068, "grad_norm": 0.8049408793449402, "learning_rate": 4.411116781093348e-06, "loss": 0.6482, "step": 846 }, { "epoch": 1.7524137931034482, "grad_norm": 0.7221590876579285, "learning_rate": 4.399156356281274e-06, "loss": 0.6549, "step": 847 }, { "epoch": 1.7544827586206897, "grad_norm": 0.7125415802001953, "learning_rate": 4.3871994189018075e-06, "loss": 0.6409, "step": 848 }, { "epoch": 1.756551724137931, "grad_norm": 0.7120684385299683, "learning_rate": 4.3752460383557195e-06, "loss": 0.6476, "step": 849 }, { "epoch": 1.7586206896551724, "grad_norm": 0.7340549230575562, "learning_rate": 4.363296284023139e-06, "loss": 0.6606, "step": 850 }, { "epoch": 1.7606896551724138, "grad_norm": 0.8672581911087036, "learning_rate": 4.351350225263147e-06, "loss": 0.6447, "step": 851 }, { "epoch": 1.762758620689655, "grad_norm": 0.7489596009254456, "learning_rate": 4.339407931413372e-06, "loss": 0.634, "step": 852 }, { "epoch": 1.7648275862068965, "grad_norm": 0.7057939171791077, "learning_rate": 4.327469471789597e-06, "loss": 0.6683, "step": 853 }, { "epoch": 1.766896551724138, "grad_norm": 0.7291073799133301, "learning_rate": 4.315534915685341e-06, "loss": 0.6766, "step": 854 }, { "epoch": 1.7689655172413792, "grad_norm": 0.7649169564247131, "learning_rate": 4.303604332371476e-06, "loss": 0.649, "step": 855 }, { "epoch": 1.7710344827586209, "grad_norm": 0.7476722002029419, "learning_rate": 4.291677791095805e-06, "loss": 0.6111, "step": 856 }, { "epoch": 1.773103448275862, "grad_norm": 0.7485122680664062, "learning_rate": 4.27975536108268e-06, "loss": 0.6375, "step": 857 }, { "epoch": 1.7751724137931033, "grad_norm": 0.7042410373687744, "learning_rate": 4.26783711153258e-06, "loss": 0.6406, "step": 858 }, { "epoch": 1.777241379310345, "grad_norm": 0.7430243492126465, "learning_rate": 4.2559231116217305e-06, "loss": 0.6511, "step": 859 }, { "epoch": 1.7793103448275862, "grad_norm": 0.6917579174041748, "learning_rate": 4.244013430501681e-06, "loss": 0.6316, "step": 860 }, { "epoch": 1.7813793103448274, "grad_norm": 0.6834431290626526, "learning_rate": 4.2321081372989195e-06, "loss": 0.6413, "step": 861 }, { "epoch": 1.783448275862069, "grad_norm": 0.6975676417350769, "learning_rate": 4.220207301114467e-06, "loss": 0.6331, "step": 862 }, { "epoch": 1.7855172413793103, "grad_norm": 0.7284821271896362, "learning_rate": 4.208310991023469e-06, "loss": 0.6431, "step": 863 }, { "epoch": 1.7875862068965516, "grad_norm": 0.742372453212738, "learning_rate": 4.1964192760748085e-06, "loss": 0.6307, "step": 864 }, { "epoch": 1.7896551724137932, "grad_norm": 0.7629566788673401, "learning_rate": 4.184532225290687e-06, "loss": 0.6186, "step": 865 }, { "epoch": 1.7917241379310345, "grad_norm": 0.7036967277526855, "learning_rate": 4.172649907666244e-06, "loss": 0.6212, "step": 866 }, { "epoch": 1.793793103448276, "grad_norm": 0.6951998472213745, "learning_rate": 4.16077239216914e-06, "loss": 0.6023, "step": 867 }, { "epoch": 1.7958620689655174, "grad_norm": 0.728022575378418, "learning_rate": 4.148899747739168e-06, "loss": 0.6444, "step": 868 }, { "epoch": 1.7979310344827586, "grad_norm": 0.6999719142913818, "learning_rate": 4.137032043287841e-06, "loss": 0.6368, "step": 869 }, { "epoch": 1.8, "grad_norm": 0.7263539433479309, "learning_rate": 4.125169347698009e-06, "loss": 0.6584, "step": 870 }, { "epoch": 1.8020689655172415, "grad_norm": 0.7751326560974121, "learning_rate": 4.11331172982344e-06, "loss": 0.6245, "step": 871 }, { "epoch": 1.8041379310344827, "grad_norm": 0.7183523178100586, "learning_rate": 4.101459258488438e-06, "loss": 0.5938, "step": 872 }, { "epoch": 1.8062068965517242, "grad_norm": 0.7222279906272888, "learning_rate": 4.089612002487428e-06, "loss": 0.6162, "step": 873 }, { "epoch": 1.8082758620689656, "grad_norm": 0.7456408143043518, "learning_rate": 4.077770030584569e-06, "loss": 0.6602, "step": 874 }, { "epoch": 1.8103448275862069, "grad_norm": 0.7470499873161316, "learning_rate": 4.065933411513349e-06, "loss": 0.653, "step": 875 }, { "epoch": 1.8124137931034483, "grad_norm": 0.7037544846534729, "learning_rate": 4.054102213976184e-06, "loss": 0.6216, "step": 876 }, { "epoch": 1.8144827586206898, "grad_norm": 0.7381490468978882, "learning_rate": 4.042276506644024e-06, "loss": 0.6422, "step": 877 }, { "epoch": 1.816551724137931, "grad_norm": 0.7278012037277222, "learning_rate": 4.030456358155957e-06, "loss": 0.6475, "step": 878 }, { "epoch": 1.8186206896551724, "grad_norm": 0.7335582971572876, "learning_rate": 4.018641837118799e-06, "loss": 0.643, "step": 879 }, { "epoch": 1.8206896551724139, "grad_norm": 0.6921178698539734, "learning_rate": 4.006833012106707e-06, "loss": 0.6241, "step": 880 }, { "epoch": 1.822758620689655, "grad_norm": 0.7078073024749756, "learning_rate": 3.995029951660777e-06, "loss": 0.6295, "step": 881 }, { "epoch": 1.8248275862068966, "grad_norm": 0.6960477232933044, "learning_rate": 3.983232724288642e-06, "loss": 0.6596, "step": 882 }, { "epoch": 1.826896551724138, "grad_norm": 0.7017378211021423, "learning_rate": 3.971441398464088e-06, "loss": 0.611, "step": 883 }, { "epoch": 1.8289655172413792, "grad_norm": 0.729068398475647, "learning_rate": 3.959656042626634e-06, "loss": 0.6184, "step": 884 }, { "epoch": 1.8310344827586207, "grad_norm": 0.7174490690231323, "learning_rate": 3.94787672518116e-06, "loss": 0.6383, "step": 885 }, { "epoch": 1.8331034482758621, "grad_norm": 0.777277946472168, "learning_rate": 3.9361035144974905e-06, "loss": 0.6508, "step": 886 }, { "epoch": 1.8351724137931034, "grad_norm": 0.7446042895317078, "learning_rate": 3.924336478910007e-06, "loss": 0.6588, "step": 887 }, { "epoch": 1.8372413793103448, "grad_norm": 0.7238491773605347, "learning_rate": 3.912575686717251e-06, "loss": 0.6394, "step": 888 }, { "epoch": 1.8393103448275863, "grad_norm": 0.7243683338165283, "learning_rate": 3.900821206181521e-06, "loss": 0.6529, "step": 889 }, { "epoch": 1.8413793103448275, "grad_norm": 0.719829261302948, "learning_rate": 3.889073105528489e-06, "loss": 0.6535, "step": 890 }, { "epoch": 1.843448275862069, "grad_norm": 0.7592584490776062, "learning_rate": 3.8773314529467875e-06, "loss": 0.6391, "step": 891 }, { "epoch": 1.8455172413793104, "grad_norm": 0.7423802018165588, "learning_rate": 3.865596316587633e-06, "loss": 0.643, "step": 892 }, { "epoch": 1.8475862068965516, "grad_norm": 0.7308037281036377, "learning_rate": 3.853867764564409e-06, "loss": 0.6329, "step": 893 }, { "epoch": 1.849655172413793, "grad_norm": 0.7450445294380188, "learning_rate": 3.842145864952295e-06, "loss": 0.6516, "step": 894 }, { "epoch": 1.8517241379310345, "grad_norm": 0.8020148277282715, "learning_rate": 3.830430685787844e-06, "loss": 0.6378, "step": 895 }, { "epoch": 1.8537931034482757, "grad_norm": 0.6937201023101807, "learning_rate": 3.818722295068616e-06, "loss": 0.646, "step": 896 }, { "epoch": 1.8558620689655172, "grad_norm": 0.7302286028862, "learning_rate": 3.8070207607527587e-06, "loss": 0.6489, "step": 897 }, { "epoch": 1.8579310344827586, "grad_norm": 0.7561166286468506, "learning_rate": 3.7953261507586282e-06, "loss": 0.6249, "step": 898 }, { "epoch": 1.8599999999999999, "grad_norm": 0.6970481276512146, "learning_rate": 3.7836385329643933e-06, "loss": 0.6355, "step": 899 }, { "epoch": 1.8620689655172413, "grad_norm": 0.7228543162345886, "learning_rate": 3.77195797520763e-06, "loss": 0.6435, "step": 900 }, { "epoch": 1.8641379310344828, "grad_norm": 0.7981168627738953, "learning_rate": 3.760284545284947e-06, "loss": 0.6294, "step": 901 }, { "epoch": 1.866206896551724, "grad_norm": 0.758095383644104, "learning_rate": 3.7486183109515696e-06, "loss": 0.6652, "step": 902 }, { "epoch": 1.8682758620689657, "grad_norm": 0.7205570936203003, "learning_rate": 3.7369593399209704e-06, "loss": 0.6236, "step": 903 }, { "epoch": 1.870344827586207, "grad_norm": 0.7436397075653076, "learning_rate": 3.725307699864452e-06, "loss": 0.639, "step": 904 }, { "epoch": 1.8724137931034481, "grad_norm": 0.7406452298164368, "learning_rate": 3.7136634584107787e-06, "loss": 0.6274, "step": 905 }, { "epoch": 1.8744827586206898, "grad_norm": 0.7269127368927002, "learning_rate": 3.7020266831457598e-06, "loss": 0.6544, "step": 906 }, { "epoch": 1.876551724137931, "grad_norm": 0.732607901096344, "learning_rate": 3.6903974416118786e-06, "loss": 0.6784, "step": 907 }, { "epoch": 1.8786206896551723, "grad_norm": 0.7445371150970459, "learning_rate": 3.6787758013078863e-06, "loss": 0.6298, "step": 908 }, { "epoch": 1.880689655172414, "grad_norm": 0.7168945670127869, "learning_rate": 3.6671618296884147e-06, "loss": 0.6331, "step": 909 }, { "epoch": 1.8827586206896552, "grad_norm": 0.7478596568107605, "learning_rate": 3.655555594163587e-06, "loss": 0.6396, "step": 910 }, { "epoch": 1.8848275862068964, "grad_norm": 0.7174217700958252, "learning_rate": 3.6439571620986213e-06, "loss": 0.6619, "step": 911 }, { "epoch": 1.886896551724138, "grad_norm": 0.7665561437606812, "learning_rate": 3.6323666008134465e-06, "loss": 0.6513, "step": 912 }, { "epoch": 1.8889655172413793, "grad_norm": 0.6955576539039612, "learning_rate": 3.620783977582305e-06, "loss": 0.6409, "step": 913 }, { "epoch": 1.8910344827586207, "grad_norm": 0.6954272389411926, "learning_rate": 3.6092093596333654e-06, "loss": 0.6423, "step": 914 }, { "epoch": 1.8931034482758622, "grad_norm": 0.744321882724762, "learning_rate": 3.5976428141483326e-06, "loss": 0.6142, "step": 915 }, { "epoch": 1.8951724137931034, "grad_norm": 0.7438650131225586, "learning_rate": 3.586084408262056e-06, "loss": 0.6602, "step": 916 }, { "epoch": 1.8972413793103449, "grad_norm": 0.6955951452255249, "learning_rate": 3.5745342090621406e-06, "loss": 0.6311, "step": 917 }, { "epoch": 1.8993103448275863, "grad_norm": 0.6816698312759399, "learning_rate": 3.5629922835885618e-06, "loss": 0.6514, "step": 918 }, { "epoch": 1.9013793103448275, "grad_norm": 0.6610486507415771, "learning_rate": 3.551458698833265e-06, "loss": 0.6401, "step": 919 }, { "epoch": 1.903448275862069, "grad_norm": 0.7027138471603394, "learning_rate": 3.5399335217397936e-06, "loss": 0.6452, "step": 920 }, { "epoch": 1.9055172413793104, "grad_norm": 0.715492844581604, "learning_rate": 3.528416819202881e-06, "loss": 0.6386, "step": 921 }, { "epoch": 1.9075862068965517, "grad_norm": 0.7160024642944336, "learning_rate": 3.5169086580680824e-06, "loss": 0.6416, "step": 922 }, { "epoch": 1.9096551724137931, "grad_norm": 0.7222967743873596, "learning_rate": 3.5054091051313666e-06, "loss": 0.6377, "step": 923 }, { "epoch": 1.9117241379310346, "grad_norm": 0.6950160264968872, "learning_rate": 3.493918227138746e-06, "loss": 0.6458, "step": 924 }, { "epoch": 1.9137931034482758, "grad_norm": 0.6922501921653748, "learning_rate": 3.4824360907858824e-06, "loss": 0.6528, "step": 925 }, { "epoch": 1.9158620689655173, "grad_norm": 0.7411875128746033, "learning_rate": 3.47096276271769e-06, "loss": 0.6178, "step": 926 }, { "epoch": 1.9179310344827587, "grad_norm": 0.6832643151283264, "learning_rate": 3.45949830952797e-06, "loss": 0.6365, "step": 927 }, { "epoch": 1.92, "grad_norm": 0.7130672931671143, "learning_rate": 3.4480427977590004e-06, "loss": 0.6319, "step": 928 }, { "epoch": 1.9220689655172414, "grad_norm": 0.6737676858901978, "learning_rate": 3.43659629390117e-06, "loss": 0.6644, "step": 929 }, { "epoch": 1.9241379310344828, "grad_norm": 0.707264244556427, "learning_rate": 3.4251588643925757e-06, "loss": 0.659, "step": 930 }, { "epoch": 1.926206896551724, "grad_norm": 0.687591016292572, "learning_rate": 3.4137305756186543e-06, "loss": 0.6175, "step": 931 }, { "epoch": 1.9282758620689655, "grad_norm": 0.685309112071991, "learning_rate": 3.4023114939117773e-06, "loss": 0.6384, "step": 932 }, { "epoch": 1.930344827586207, "grad_norm": 0.7224867343902588, "learning_rate": 3.390901685550887e-06, "loss": 0.6648, "step": 933 }, { "epoch": 1.9324137931034482, "grad_norm": 0.7063190340995789, "learning_rate": 3.3795012167610898e-06, "loss": 0.6494, "step": 934 }, { "epoch": 1.9344827586206896, "grad_norm": 0.7511458992958069, "learning_rate": 3.3681101537132914e-06, "loss": 0.6272, "step": 935 }, { "epoch": 1.936551724137931, "grad_norm": 0.6883670091629028, "learning_rate": 3.356728562523805e-06, "loss": 0.6276, "step": 936 }, { "epoch": 1.9386206896551723, "grad_norm": 0.7098174691200256, "learning_rate": 3.3453565092539586e-06, "loss": 0.6675, "step": 937 }, { "epoch": 1.9406896551724138, "grad_norm": 0.7062398791313171, "learning_rate": 3.33399405990973e-06, "loss": 0.6408, "step": 938 }, { "epoch": 1.9427586206896552, "grad_norm": 0.7228072881698608, "learning_rate": 3.3226412804413444e-06, "loss": 0.6422, "step": 939 }, { "epoch": 1.9448275862068964, "grad_norm": 0.7137162685394287, "learning_rate": 3.3112982367429105e-06, "loss": 0.6474, "step": 940 }, { "epoch": 1.946896551724138, "grad_norm": 0.6957323551177979, "learning_rate": 3.299964994652017e-06, "loss": 0.6492, "step": 941 }, { "epoch": 1.9489655172413793, "grad_norm": 0.6673961281776428, "learning_rate": 3.288641619949374e-06, "loss": 0.6512, "step": 942 }, { "epoch": 1.9510344827586206, "grad_norm": 0.7092306017875671, "learning_rate": 3.2773281783584104e-06, "loss": 0.6623, "step": 943 }, { "epoch": 1.953103448275862, "grad_norm": 0.7391819953918457, "learning_rate": 3.2660247355449037e-06, "loss": 0.6698, "step": 944 }, { "epoch": 1.9551724137931035, "grad_norm": 0.7239782214164734, "learning_rate": 3.254731357116597e-06, "loss": 0.6404, "step": 945 }, { "epoch": 1.9572413793103447, "grad_norm": 0.735502302646637, "learning_rate": 3.2434481086228163e-06, "loss": 0.6278, "step": 946 }, { "epoch": 1.9593103448275864, "grad_norm": 0.7158029079437256, "learning_rate": 3.232175055554091e-06, "loss": 0.6441, "step": 947 }, { "epoch": 1.9613793103448276, "grad_norm": 0.7179998755455017, "learning_rate": 3.2209122633417766e-06, "loss": 0.6464, "step": 948 }, { "epoch": 1.9634482758620688, "grad_norm": 0.6865692734718323, "learning_rate": 3.209659797357669e-06, "loss": 0.6373, "step": 949 }, { "epoch": 1.9655172413793105, "grad_norm": 0.7228394150733948, "learning_rate": 3.1984177229136287e-06, "loss": 0.6583, "step": 950 }, { "epoch": 1.9675862068965517, "grad_norm": 0.7167955636978149, "learning_rate": 3.187186105261205e-06, "loss": 0.6276, "step": 951 }, { "epoch": 1.969655172413793, "grad_norm": 0.6937686204910278, "learning_rate": 3.1759650095912474e-06, "loss": 0.6203, "step": 952 }, { "epoch": 1.9717241379310346, "grad_norm": 0.6669635772705078, "learning_rate": 3.16475450103354e-06, "loss": 0.6447, "step": 953 }, { "epoch": 1.9737931034482759, "grad_norm": 0.7243427634239197, "learning_rate": 3.1535546446564107e-06, "loss": 0.6375, "step": 954 }, { "epoch": 1.975862068965517, "grad_norm": 0.7239643931388855, "learning_rate": 3.1423655054663665e-06, "loss": 0.666, "step": 955 }, { "epoch": 1.9779310344827588, "grad_norm": 0.7242691516876221, "learning_rate": 3.1311871484077017e-06, "loss": 0.6427, "step": 956 }, { "epoch": 1.98, "grad_norm": 0.7019519209861755, "learning_rate": 3.1200196383621363e-06, "loss": 0.637, "step": 957 }, { "epoch": 1.9820689655172414, "grad_norm": 0.7041625380516052, "learning_rate": 3.1088630401484245e-06, "loss": 0.6297, "step": 958 }, { "epoch": 1.9841379310344829, "grad_norm": 0.7110829949378967, "learning_rate": 3.0977174185219937e-06, "loss": 0.6353, "step": 959 }, { "epoch": 1.986206896551724, "grad_norm": 0.6561233401298523, "learning_rate": 3.0865828381745515e-06, "loss": 0.6184, "step": 960 }, { "epoch": 1.9882758620689656, "grad_norm": 0.7260603904724121, "learning_rate": 3.0754593637337276e-06, "loss": 0.6435, "step": 961 }, { "epoch": 1.990344827586207, "grad_norm": 0.7066130638122559, "learning_rate": 3.0643470597626885e-06, "loss": 0.6555, "step": 962 }, { "epoch": 1.9924137931034482, "grad_norm": 0.6790770888328552, "learning_rate": 3.05324599075976e-06, "loss": 0.6216, "step": 963 }, { "epoch": 1.9944827586206897, "grad_norm": 0.7200824022293091, "learning_rate": 3.042156221158064e-06, "loss": 0.6475, "step": 964 }, { "epoch": 1.9965517241379311, "grad_norm": 0.6905136108398438, "learning_rate": 3.0310778153251325e-06, "loss": 0.6133, "step": 965 }, { "epoch": 1.9986206896551724, "grad_norm": 0.7127056121826172, "learning_rate": 3.020010837562546e-06, "loss": 0.667, "step": 966 }, { "epoch": 2.0006896551724136, "grad_norm": 0.6826462149620056, "learning_rate": 3.0089553521055455e-06, "loss": 0.6117, "step": 967 }, { "epoch": 2.0027586206896553, "grad_norm": 0.803715169429779, "learning_rate": 2.9979114231226776e-06, "loss": 0.5496, "step": 968 }, { "epoch": 2.0048275862068965, "grad_norm": 0.7920373678207397, "learning_rate": 2.986879114715403e-06, "loss": 0.5323, "step": 969 }, { "epoch": 2.0068965517241377, "grad_norm": 0.7448180913925171, "learning_rate": 2.9758584909177434e-06, "loss": 0.5318, "step": 970 }, { "epoch": 2.0089655172413794, "grad_norm": 0.7098000049591064, "learning_rate": 2.964849615695892e-06, "loss": 0.5341, "step": 971 }, { "epoch": 2.0110344827586206, "grad_norm": 0.7263830304145813, "learning_rate": 2.9538525529478557e-06, "loss": 0.5605, "step": 972 }, { "epoch": 2.013103448275862, "grad_norm": 0.7626384496688843, "learning_rate": 2.9428673665030772e-06, "loss": 0.5407, "step": 973 }, { "epoch": 2.0151724137931035, "grad_norm": 0.8168633580207825, "learning_rate": 2.9318941201220664e-06, "loss": 0.5628, "step": 974 }, { "epoch": 2.0172413793103448, "grad_norm": 0.8351486325263977, "learning_rate": 2.9209328774960327e-06, "loss": 0.5295, "step": 975 }, { "epoch": 2.0193103448275864, "grad_norm": 0.7826777100563049, "learning_rate": 2.90998370224651e-06, "loss": 0.5141, "step": 976 }, { "epoch": 2.0213793103448277, "grad_norm": 0.7777988314628601, "learning_rate": 2.899046657924992e-06, "loss": 0.5577, "step": 977 }, { "epoch": 2.023448275862069, "grad_norm": 0.751868724822998, "learning_rate": 2.8881218080125605e-06, "loss": 0.5402, "step": 978 }, { "epoch": 2.0255172413793106, "grad_norm": 0.6992483735084534, "learning_rate": 2.8772092159195193e-06, "loss": 0.5323, "step": 979 }, { "epoch": 2.027586206896552, "grad_norm": 0.7283136248588562, "learning_rate": 2.8663089449850246e-06, "loss": 0.5364, "step": 980 }, { "epoch": 2.029655172413793, "grad_norm": 0.7618082761764526, "learning_rate": 2.8554210584767188e-06, "loss": 0.5388, "step": 981 }, { "epoch": 2.0317241379310347, "grad_norm": 0.7302858829498291, "learning_rate": 2.844545619590361e-06, "loss": 0.5247, "step": 982 }, { "epoch": 2.033793103448276, "grad_norm": 0.706373393535614, "learning_rate": 2.8336826914494607e-06, "loss": 0.5138, "step": 983 }, { "epoch": 2.035862068965517, "grad_norm": 0.6692239046096802, "learning_rate": 2.822832337104915e-06, "loss": 0.5363, "step": 984 }, { "epoch": 2.037931034482759, "grad_norm": 0.7576499581336975, "learning_rate": 2.8119946195346375e-06, "loss": 0.5389, "step": 985 }, { "epoch": 2.04, "grad_norm": 0.7035110592842102, "learning_rate": 2.8011696016431967e-06, "loss": 0.5333, "step": 986 }, { "epoch": 2.0420689655172413, "grad_norm": 0.7431889176368713, "learning_rate": 2.790357346261445e-06, "loss": 0.5228, "step": 987 }, { "epoch": 2.044137931034483, "grad_norm": 0.7341644167900085, "learning_rate": 2.7795579161461685e-06, "loss": 0.5153, "step": 988 }, { "epoch": 2.046206896551724, "grad_norm": 0.7432145476341248, "learning_rate": 2.7687713739796972e-06, "loss": 0.5299, "step": 989 }, { "epoch": 2.0482758620689654, "grad_norm": 0.6737946271896362, "learning_rate": 2.7579977823695727e-06, "loss": 0.5301, "step": 990 }, { "epoch": 2.050344827586207, "grad_norm": 0.7434407472610474, "learning_rate": 2.7472372038481575e-06, "loss": 0.5442, "step": 991 }, { "epoch": 2.0524137931034483, "grad_norm": 0.7228389382362366, "learning_rate": 2.7364897008722867e-06, "loss": 0.5392, "step": 992 }, { "epoch": 2.0544827586206895, "grad_norm": 0.6981167793273926, "learning_rate": 2.725755335822903e-06, "loss": 0.5503, "step": 993 }, { "epoch": 2.056551724137931, "grad_norm": 0.6900709867477417, "learning_rate": 2.7150341710046924e-06, "loss": 0.5448, "step": 994 }, { "epoch": 2.0586206896551724, "grad_norm": 0.6905665993690491, "learning_rate": 2.704326268645724e-06, "loss": 0.5304, "step": 995 }, { "epoch": 2.0606896551724136, "grad_norm": 0.720930814743042, "learning_rate": 2.6936316908970874e-06, "loss": 0.5304, "step": 996 }, { "epoch": 2.0627586206896553, "grad_norm": 0.697242796421051, "learning_rate": 2.6829504998325352e-06, "loss": 0.5209, "step": 997 }, { "epoch": 2.0648275862068965, "grad_norm": 0.6550352573394775, "learning_rate": 2.67228275744812e-06, "loss": 0.5319, "step": 998 }, { "epoch": 2.0668965517241378, "grad_norm": 0.7028206586837769, "learning_rate": 2.661628525661835e-06, "loss": 0.5253, "step": 999 }, { "epoch": 2.0689655172413794, "grad_norm": 0.7251700162887573, "learning_rate": 2.6509878663132514e-06, "loss": 0.5589, "step": 1000 }, { "epoch": 2.0710344827586207, "grad_norm": 0.67062908411026, "learning_rate": 2.6403608411631744e-06, "loss": 0.5462, "step": 1001 }, { "epoch": 2.073103448275862, "grad_norm": 0.7188374996185303, "learning_rate": 2.629747511893255e-06, "loss": 0.5201, "step": 1002 }, { "epoch": 2.0751724137931036, "grad_norm": 0.6892343163490295, "learning_rate": 2.61914794010567e-06, "loss": 0.5357, "step": 1003 }, { "epoch": 2.077241379310345, "grad_norm": 0.6836119294166565, "learning_rate": 2.6085621873227264e-06, "loss": 0.5123, "step": 1004 }, { "epoch": 2.079310344827586, "grad_norm": 0.6911861896514893, "learning_rate": 2.5979903149865386e-06, "loss": 0.546, "step": 1005 }, { "epoch": 2.0813793103448277, "grad_norm": 0.6947477459907532, "learning_rate": 2.5874323844586414e-06, "loss": 0.5408, "step": 1006 }, { "epoch": 2.083448275862069, "grad_norm": 0.6671912670135498, "learning_rate": 2.5768884570196615e-06, "loss": 0.5256, "step": 1007 }, { "epoch": 2.08551724137931, "grad_norm": 0.6696575284004211, "learning_rate": 2.566358593868936e-06, "loss": 0.5186, "step": 1008 }, { "epoch": 2.087586206896552, "grad_norm": 0.6655759215354919, "learning_rate": 2.555842856124182e-06, "loss": 0.5686, "step": 1009 }, { "epoch": 2.089655172413793, "grad_norm": 0.685197114944458, "learning_rate": 2.5453413048211175e-06, "loss": 0.4912, "step": 1010 }, { "epoch": 2.0917241379310343, "grad_norm": 0.7344222664833069, "learning_rate": 2.5348540009131283e-06, "loss": 0.5489, "step": 1011 }, { "epoch": 2.093793103448276, "grad_norm": 0.7047103047370911, "learning_rate": 2.5243810052709006e-06, "loss": 0.5366, "step": 1012 }, { "epoch": 2.095862068965517, "grad_norm": 0.6715806722640991, "learning_rate": 2.513922378682075e-06, "loss": 0.5337, "step": 1013 }, { "epoch": 2.0979310344827584, "grad_norm": 0.6645084619522095, "learning_rate": 2.5034781818508867e-06, "loss": 0.5451, "step": 1014 }, { "epoch": 2.1, "grad_norm": 0.6680448651313782, "learning_rate": 2.4930484753978224e-06, "loss": 0.5141, "step": 1015 }, { "epoch": 2.1020689655172413, "grad_norm": 0.6710419058799744, "learning_rate": 2.482633319859259e-06, "loss": 0.5233, "step": 1016 }, { "epoch": 2.1041379310344825, "grad_norm": 0.6609877943992615, "learning_rate": 2.472232775687119e-06, "loss": 0.5343, "step": 1017 }, { "epoch": 2.106206896551724, "grad_norm": 0.6360195279121399, "learning_rate": 2.4618469032485164e-06, "loss": 0.5549, "step": 1018 }, { "epoch": 2.1082758620689654, "grad_norm": 0.6452751159667969, "learning_rate": 2.4514757628254075e-06, "loss": 0.5369, "step": 1019 }, { "epoch": 2.110344827586207, "grad_norm": 0.6557055115699768, "learning_rate": 2.441119414614241e-06, "loss": 0.5277, "step": 1020 }, { "epoch": 2.1124137931034483, "grad_norm": 0.7097801566123962, "learning_rate": 2.4307779187256064e-06, "loss": 0.5358, "step": 1021 }, { "epoch": 2.1144827586206896, "grad_norm": 0.6572842597961426, "learning_rate": 2.4204513351838897e-06, "loss": 0.5394, "step": 1022 }, { "epoch": 2.1165517241379312, "grad_norm": 0.6733007431030273, "learning_rate": 2.4101397239269202e-06, "loss": 0.5579, "step": 1023 }, { "epoch": 2.1186206896551725, "grad_norm": 0.6480697393417358, "learning_rate": 2.3998431448056237e-06, "loss": 0.5635, "step": 1024 }, { "epoch": 2.1206896551724137, "grad_norm": 0.6844390034675598, "learning_rate": 2.389561657583681e-06, "loss": 0.5477, "step": 1025 }, { "epoch": 2.1227586206896554, "grad_norm": 0.6834397912025452, "learning_rate": 2.3792953219371713e-06, "loss": 0.5252, "step": 1026 }, { "epoch": 2.1248275862068966, "grad_norm": 0.666690468788147, "learning_rate": 2.3690441974542322e-06, "loss": 0.574, "step": 1027 }, { "epoch": 2.126896551724138, "grad_norm": 0.6561058759689331, "learning_rate": 2.358808343634712e-06, "loss": 0.5291, "step": 1028 }, { "epoch": 2.1289655172413795, "grad_norm": 0.6525347828865051, "learning_rate": 2.3485878198898253e-06, "loss": 0.5626, "step": 1029 }, { "epoch": 2.1310344827586207, "grad_norm": 0.6866719126701355, "learning_rate": 2.3383826855418064e-06, "loss": 0.524, "step": 1030 }, { "epoch": 2.133103448275862, "grad_norm": 0.6798598170280457, "learning_rate": 2.3281929998235664e-06, "loss": 0.5184, "step": 1031 }, { "epoch": 2.1351724137931036, "grad_norm": 0.6799730658531189, "learning_rate": 2.31801882187835e-06, "loss": 0.5086, "step": 1032 }, { "epoch": 2.137241379310345, "grad_norm": 0.6736025214195251, "learning_rate": 2.30786021075939e-06, "loss": 0.5333, "step": 1033 }, { "epoch": 2.139310344827586, "grad_norm": 0.6704842448234558, "learning_rate": 2.297717225429567e-06, "loss": 0.5065, "step": 1034 }, { "epoch": 2.1413793103448278, "grad_norm": 0.6423225998878479, "learning_rate": 2.287589924761065e-06, "loss": 0.5292, "step": 1035 }, { "epoch": 2.143448275862069, "grad_norm": 0.6814275979995728, "learning_rate": 2.2774783675350303e-06, "loss": 0.5418, "step": 1036 }, { "epoch": 2.14551724137931, "grad_norm": 0.6510114073753357, "learning_rate": 2.2673826124412314e-06, "loss": 0.5175, "step": 1037 }, { "epoch": 2.147586206896552, "grad_norm": 0.6611030697822571, "learning_rate": 2.257302718077721e-06, "loss": 0.5526, "step": 1038 }, { "epoch": 2.149655172413793, "grad_norm": 0.6714975833892822, "learning_rate": 2.247238742950483e-06, "loss": 0.524, "step": 1039 }, { "epoch": 2.1517241379310343, "grad_norm": 0.6750874519348145, "learning_rate": 2.2371907454731168e-06, "loss": 0.5529, "step": 1040 }, { "epoch": 2.153793103448276, "grad_norm": 0.6794918179512024, "learning_rate": 2.2271587839664673e-06, "loss": 0.5034, "step": 1041 }, { "epoch": 2.1558620689655172, "grad_norm": 0.6698225140571594, "learning_rate": 2.217142916658318e-06, "loss": 0.531, "step": 1042 }, { "epoch": 2.1579310344827585, "grad_norm": 0.6470540761947632, "learning_rate": 2.2071432016830257e-06, "loss": 0.5631, "step": 1043 }, { "epoch": 2.16, "grad_norm": 0.7159327268600464, "learning_rate": 2.197159697081208e-06, "loss": 0.5476, "step": 1044 }, { "epoch": 2.1620689655172414, "grad_norm": 0.7387550473213196, "learning_rate": 2.18719246079938e-06, "loss": 0.5019, "step": 1045 }, { "epoch": 2.1641379310344826, "grad_norm": 0.7351087331771851, "learning_rate": 2.1772415506896447e-06, "loss": 0.5607, "step": 1046 }, { "epoch": 2.1662068965517243, "grad_norm": 0.6578730940818787, "learning_rate": 2.1673070245093324e-06, "loss": 0.5494, "step": 1047 }, { "epoch": 2.1682758620689655, "grad_norm": 0.675528347492218, "learning_rate": 2.157388939920689e-06, "loss": 0.526, "step": 1048 }, { "epoch": 2.1703448275862067, "grad_norm": 0.6719648838043213, "learning_rate": 2.1474873544905204e-06, "loss": 0.5334, "step": 1049 }, { "epoch": 2.1724137931034484, "grad_norm": 0.6997095346450806, "learning_rate": 2.137602325689873e-06, "loss": 0.5274, "step": 1050 }, { "epoch": 2.1744827586206896, "grad_norm": 0.7052441835403442, "learning_rate": 2.127733910893693e-06, "loss": 0.5378, "step": 1051 }, { "epoch": 2.176551724137931, "grad_norm": 0.6952463388442993, "learning_rate": 2.1178821673804943e-06, "loss": 0.5244, "step": 1052 }, { "epoch": 2.1786206896551725, "grad_norm": 0.664283037185669, "learning_rate": 2.108047152332028e-06, "loss": 0.5335, "step": 1053 }, { "epoch": 2.1806896551724138, "grad_norm": 0.6841456294059753, "learning_rate": 2.0982289228329495e-06, "loss": 0.5367, "step": 1054 }, { "epoch": 2.182758620689655, "grad_norm": 0.6873970031738281, "learning_rate": 2.0884275358704863e-06, "loss": 0.5242, "step": 1055 }, { "epoch": 2.1848275862068967, "grad_norm": 0.704255998134613, "learning_rate": 2.078643048334109e-06, "loss": 0.5369, "step": 1056 }, { "epoch": 2.186896551724138, "grad_norm": 0.6614420413970947, "learning_rate": 2.0688755170152e-06, "loss": 0.5397, "step": 1057 }, { "epoch": 2.188965517241379, "grad_norm": 0.65423583984375, "learning_rate": 2.0591249986067214e-06, "loss": 0.5374, "step": 1058 }, { "epoch": 2.191034482758621, "grad_norm": 0.6417624950408936, "learning_rate": 2.0493915497028933e-06, "loss": 0.5238, "step": 1059 }, { "epoch": 2.193103448275862, "grad_norm": 0.6377779841423035, "learning_rate": 2.039675226798854e-06, "loss": 0.5377, "step": 1060 }, { "epoch": 2.1951724137931032, "grad_norm": 0.6852798461914062, "learning_rate": 2.029976086290347e-06, "loss": 0.5386, "step": 1061 }, { "epoch": 2.197241379310345, "grad_norm": 0.6332745552062988, "learning_rate": 2.0202941844733792e-06, "loss": 0.5541, "step": 1062 }, { "epoch": 2.199310344827586, "grad_norm": 0.6756484508514404, "learning_rate": 2.0106295775439018e-06, "loss": 0.5092, "step": 1063 }, { "epoch": 2.2013793103448274, "grad_norm": 0.6440528035163879, "learning_rate": 2.0009823215974815e-06, "loss": 0.5256, "step": 1064 }, { "epoch": 2.203448275862069, "grad_norm": 0.6707040071487427, "learning_rate": 1.9913524726289784e-06, "loss": 0.5317, "step": 1065 }, { "epoch": 2.2055172413793103, "grad_norm": 0.6765915155410767, "learning_rate": 1.9817400865322166e-06, "loss": 0.5319, "step": 1066 }, { "epoch": 2.2075862068965515, "grad_norm": 0.6627142429351807, "learning_rate": 1.972145219099662e-06, "loss": 0.5246, "step": 1067 }, { "epoch": 2.209655172413793, "grad_norm": 0.650766909122467, "learning_rate": 1.962567926022099e-06, "loss": 0.541, "step": 1068 }, { "epoch": 2.2117241379310344, "grad_norm": 0.6753216981887817, "learning_rate": 1.9530082628883058e-06, "loss": 0.5472, "step": 1069 }, { "epoch": 2.213793103448276, "grad_norm": 0.6764883399009705, "learning_rate": 1.9434662851847325e-06, "loss": 0.522, "step": 1070 }, { "epoch": 2.2158620689655173, "grad_norm": 0.6282169818878174, "learning_rate": 1.9339420482951794e-06, "loss": 0.5127, "step": 1071 }, { "epoch": 2.2179310344827585, "grad_norm": 0.6872683763504028, "learning_rate": 1.9244356075004743e-06, "loss": 0.5661, "step": 1072 }, { "epoch": 2.22, "grad_norm": 0.6766654849052429, "learning_rate": 1.914947017978153e-06, "loss": 0.543, "step": 1073 }, { "epoch": 2.2220689655172414, "grad_norm": 0.6761453747749329, "learning_rate": 1.9054763348021372e-06, "loss": 0.5139, "step": 1074 }, { "epoch": 2.2241379310344827, "grad_norm": 0.6696473360061646, "learning_rate": 1.896023612942422e-06, "loss": 0.5224, "step": 1075 }, { "epoch": 2.2262068965517243, "grad_norm": 0.7067750096321106, "learning_rate": 1.8865889072647386e-06, "loss": 0.5376, "step": 1076 }, { "epoch": 2.2282758620689656, "grad_norm": 0.654654324054718, "learning_rate": 1.8771722725302644e-06, "loss": 0.5438, "step": 1077 }, { "epoch": 2.230344827586207, "grad_norm": 0.6717188358306885, "learning_rate": 1.8677737633952715e-06, "loss": 0.5354, "step": 1078 }, { "epoch": 2.2324137931034485, "grad_norm": 0.6855020523071289, "learning_rate": 1.8583934344108446e-06, "loss": 0.5294, "step": 1079 }, { "epoch": 2.2344827586206897, "grad_norm": 0.6623315811157227, "learning_rate": 1.8490313400225296e-06, "loss": 0.5317, "step": 1080 }, { "epoch": 2.236551724137931, "grad_norm": 0.687870979309082, "learning_rate": 1.8396875345700498e-06, "loss": 0.5539, "step": 1081 }, { "epoch": 2.2386206896551726, "grad_norm": 0.6914966106414795, "learning_rate": 1.830362072286963e-06, "loss": 0.5537, "step": 1082 }, { "epoch": 2.240689655172414, "grad_norm": 0.7108622193336487, "learning_rate": 1.8210550073003701e-06, "loss": 0.5252, "step": 1083 }, { "epoch": 2.242758620689655, "grad_norm": 0.6858078241348267, "learning_rate": 1.8117663936305785e-06, "loss": 0.5542, "step": 1084 }, { "epoch": 2.2448275862068967, "grad_norm": 0.6860451698303223, "learning_rate": 1.8024962851908106e-06, "loss": 0.5246, "step": 1085 }, { "epoch": 2.246896551724138, "grad_norm": 0.6590988636016846, "learning_rate": 1.7932447357868737e-06, "loss": 0.5328, "step": 1086 }, { "epoch": 2.248965517241379, "grad_norm": 0.7076303958892822, "learning_rate": 1.7840117991168566e-06, "loss": 0.5205, "step": 1087 }, { "epoch": 2.251034482758621, "grad_norm": 0.6729241013526917, "learning_rate": 1.7747975287708147e-06, "loss": 0.5322, "step": 1088 }, { "epoch": 2.253103448275862, "grad_norm": 0.6530643701553345, "learning_rate": 1.7656019782304602e-06, "loss": 0.552, "step": 1089 }, { "epoch": 2.2551724137931033, "grad_norm": 0.6714941263198853, "learning_rate": 1.7564252008688514e-06, "loss": 0.5344, "step": 1090 }, { "epoch": 2.257241379310345, "grad_norm": 0.6543527841567993, "learning_rate": 1.7472672499500837e-06, "loss": 0.5282, "step": 1091 }, { "epoch": 2.259310344827586, "grad_norm": 0.6821870803833008, "learning_rate": 1.7381281786289767e-06, "loss": 0.5009, "step": 1092 }, { "epoch": 2.2613793103448274, "grad_norm": 0.6596633791923523, "learning_rate": 1.729008039950772e-06, "loss": 0.546, "step": 1093 }, { "epoch": 2.263448275862069, "grad_norm": 0.6522555351257324, "learning_rate": 1.71990688685082e-06, "loss": 0.5464, "step": 1094 }, { "epoch": 2.2655172413793103, "grad_norm": 0.668368399143219, "learning_rate": 1.7108247721542725e-06, "loss": 0.528, "step": 1095 }, { "epoch": 2.2675862068965515, "grad_norm": 0.697405219078064, "learning_rate": 1.701761748575786e-06, "loss": 0.5657, "step": 1096 }, { "epoch": 2.269655172413793, "grad_norm": 0.6927202939987183, "learning_rate": 1.6927178687191953e-06, "loss": 0.518, "step": 1097 }, { "epoch": 2.2717241379310344, "grad_norm": 0.658198356628418, "learning_rate": 1.6836931850772331e-06, "loss": 0.5482, "step": 1098 }, { "epoch": 2.2737931034482757, "grad_norm": 0.688456654548645, "learning_rate": 1.6746877500312054e-06, "loss": 0.5114, "step": 1099 }, { "epoch": 2.2758620689655173, "grad_norm": 0.6654470562934875, "learning_rate": 1.6657016158506966e-06, "loss": 0.5417, "step": 1100 }, { "epoch": 2.2779310344827586, "grad_norm": 0.6720593571662903, "learning_rate": 1.656734834693266e-06, "loss": 0.5243, "step": 1101 }, { "epoch": 2.2800000000000002, "grad_norm": 0.6629576086997986, "learning_rate": 1.6477874586041415e-06, "loss": 0.5244, "step": 1102 }, { "epoch": 2.2820689655172415, "grad_norm": 0.6911601424217224, "learning_rate": 1.6388595395159207e-06, "loss": 0.5541, "step": 1103 }, { "epoch": 2.2841379310344827, "grad_norm": 0.6672130823135376, "learning_rate": 1.6299511292482689e-06, "loss": 0.5315, "step": 1104 }, { "epoch": 2.2862068965517244, "grad_norm": 0.688179612159729, "learning_rate": 1.621062279507617e-06, "loss": 0.526, "step": 1105 }, { "epoch": 2.2882758620689656, "grad_norm": 0.6778724193572998, "learning_rate": 1.612193041886862e-06, "loss": 0.5029, "step": 1106 }, { "epoch": 2.290344827586207, "grad_norm": 0.6525766253471375, "learning_rate": 1.6033434678650684e-06, "loss": 0.5367, "step": 1107 }, { "epoch": 2.2924137931034485, "grad_norm": 0.6853109002113342, "learning_rate": 1.5945136088071677e-06, "loss": 0.5411, "step": 1108 }, { "epoch": 2.2944827586206897, "grad_norm": 0.6768192648887634, "learning_rate": 1.5857035159636625e-06, "loss": 0.5474, "step": 1109 }, { "epoch": 2.296551724137931, "grad_norm": 0.6869329214096069, "learning_rate": 1.5769132404703274e-06, "loss": 0.5156, "step": 1110 }, { "epoch": 2.2986206896551726, "grad_norm": 0.641927182674408, "learning_rate": 1.5681428333479104e-06, "loss": 0.5612, "step": 1111 }, { "epoch": 2.300689655172414, "grad_norm": 0.6849982738494873, "learning_rate": 1.5593923455018462e-06, "loss": 0.524, "step": 1112 }, { "epoch": 2.302758620689655, "grad_norm": 0.6639160513877869, "learning_rate": 1.550661827721941e-06, "loss": 0.5219, "step": 1113 }, { "epoch": 2.3048275862068968, "grad_norm": 0.6571563482284546, "learning_rate": 1.5419513306821055e-06, "loss": 0.5213, "step": 1114 }, { "epoch": 2.306896551724138, "grad_norm": 0.6656675338745117, "learning_rate": 1.5332609049400298e-06, "loss": 0.5142, "step": 1115 }, { "epoch": 2.308965517241379, "grad_norm": 0.6509329080581665, "learning_rate": 1.5245906009369211e-06, "loss": 0.5379, "step": 1116 }, { "epoch": 2.311034482758621, "grad_norm": 0.6804161667823792, "learning_rate": 1.5159404689971797e-06, "loss": 0.5348, "step": 1117 }, { "epoch": 2.313103448275862, "grad_norm": 0.6578646898269653, "learning_rate": 1.5073105593281385e-06, "loss": 0.5218, "step": 1118 }, { "epoch": 2.3151724137931033, "grad_norm": 0.6653922200202942, "learning_rate": 1.4987009220197406e-06, "loss": 0.5304, "step": 1119 }, { "epoch": 2.317241379310345, "grad_norm": 0.6492714881896973, "learning_rate": 1.4901116070442772e-06, "loss": 0.5218, "step": 1120 }, { "epoch": 2.3193103448275862, "grad_norm": 0.7024757862091064, "learning_rate": 1.4815426642560753e-06, "loss": 0.545, "step": 1121 }, { "epoch": 2.3213793103448275, "grad_norm": 0.6717721223831177, "learning_rate": 1.47299414339122e-06, "loss": 0.5376, "step": 1122 }, { "epoch": 2.323448275862069, "grad_norm": 0.6341143250465393, "learning_rate": 1.4644660940672628e-06, "loss": 0.5158, "step": 1123 }, { "epoch": 2.3255172413793104, "grad_norm": 0.6935365200042725, "learning_rate": 1.455958565782934e-06, "loss": 0.537, "step": 1124 }, { "epoch": 2.3275862068965516, "grad_norm": 0.6960786581039429, "learning_rate": 1.4474716079178541e-06, "loss": 0.5547, "step": 1125 }, { "epoch": 2.3296551724137933, "grad_norm": 0.6549028754234314, "learning_rate": 1.4390052697322499e-06, "loss": 0.544, "step": 1126 }, { "epoch": 2.3317241379310345, "grad_norm": 0.6705878973007202, "learning_rate": 1.430559600366665e-06, "loss": 0.5502, "step": 1127 }, { "epoch": 2.3337931034482757, "grad_norm": 0.6721364855766296, "learning_rate": 1.4221346488416765e-06, "loss": 0.5532, "step": 1128 }, { "epoch": 2.3358620689655174, "grad_norm": 0.6675882935523987, "learning_rate": 1.4137304640576161e-06, "loss": 0.5174, "step": 1129 }, { "epoch": 2.3379310344827586, "grad_norm": 0.650718629360199, "learning_rate": 1.4053470947942694e-06, "loss": 0.5565, "step": 1130 }, { "epoch": 2.34, "grad_norm": 0.6801978945732117, "learning_rate": 1.3969845897106166e-06, "loss": 0.5276, "step": 1131 }, { "epoch": 2.3420689655172415, "grad_norm": 0.6504865288734436, "learning_rate": 1.3886429973445253e-06, "loss": 0.5126, "step": 1132 }, { "epoch": 2.3441379310344828, "grad_norm": 0.6334866881370544, "learning_rate": 1.3803223661124938e-06, "loss": 0.5036, "step": 1133 }, { "epoch": 2.346206896551724, "grad_norm": 0.6434235572814941, "learning_rate": 1.3720227443093437e-06, "loss": 0.5119, "step": 1134 }, { "epoch": 2.3482758620689657, "grad_norm": 0.6684521436691284, "learning_rate": 1.3637441801079666e-06, "loss": 0.526, "step": 1135 }, { "epoch": 2.350344827586207, "grad_norm": 0.6425202488899231, "learning_rate": 1.3554867215590174e-06, "loss": 0.5348, "step": 1136 }, { "epoch": 2.352413793103448, "grad_norm": 0.6454555988311768, "learning_rate": 1.3472504165906614e-06, "loss": 0.5457, "step": 1137 }, { "epoch": 2.35448275862069, "grad_norm": 0.6798628568649292, "learning_rate": 1.3390353130082756e-06, "loss": 0.5436, "step": 1138 }, { "epoch": 2.356551724137931, "grad_norm": 0.6718437671661377, "learning_rate": 1.330841458494183e-06, "loss": 0.5232, "step": 1139 }, { "epoch": 2.3586206896551722, "grad_norm": 0.6361545324325562, "learning_rate": 1.3226689006073712e-06, "loss": 0.5313, "step": 1140 }, { "epoch": 2.360689655172414, "grad_norm": 0.6867586374282837, "learning_rate": 1.3145176867832165e-06, "loss": 0.5452, "step": 1141 }, { "epoch": 2.362758620689655, "grad_norm": 0.665239691734314, "learning_rate": 1.3063878643332112e-06, "loss": 0.5136, "step": 1142 }, { "epoch": 2.3648275862068964, "grad_norm": 0.6792328953742981, "learning_rate": 1.2982794804446858e-06, "loss": 0.5325, "step": 1143 }, { "epoch": 2.366896551724138, "grad_norm": 0.6773557066917419, "learning_rate": 1.2901925821805372e-06, "loss": 0.5112, "step": 1144 }, { "epoch": 2.3689655172413793, "grad_norm": 0.6913496851921082, "learning_rate": 1.2821272164789544e-06, "loss": 0.5547, "step": 1145 }, { "epoch": 2.3710344827586205, "grad_norm": 0.6805727481842041, "learning_rate": 1.2740834301531468e-06, "loss": 0.5473, "step": 1146 }, { "epoch": 2.373103448275862, "grad_norm": 0.7125365734100342, "learning_rate": 1.266061269891073e-06, "loss": 0.5248, "step": 1147 }, { "epoch": 2.3751724137931034, "grad_norm": 0.6634067893028259, "learning_rate": 1.2580607822551677e-06, "loss": 0.5287, "step": 1148 }, { "epoch": 2.3772413793103446, "grad_norm": 0.6796189546585083, "learning_rate": 1.2500820136820735e-06, "loss": 0.532, "step": 1149 }, { "epoch": 2.3793103448275863, "grad_norm": 0.6692973971366882, "learning_rate": 1.2421250104823683e-06, "loss": 0.5265, "step": 1150 }, { "epoch": 2.3813793103448275, "grad_norm": 0.6606705784797668, "learning_rate": 1.2341898188403068e-06, "loss": 0.5184, "step": 1151 }, { "epoch": 2.3834482758620688, "grad_norm": 0.6459177136421204, "learning_rate": 1.2262764848135316e-06, "loss": 0.5578, "step": 1152 }, { "epoch": 2.3855172413793104, "grad_norm": 0.67464679479599, "learning_rate": 1.2183850543328313e-06, "loss": 0.5367, "step": 1153 }, { "epoch": 2.3875862068965517, "grad_norm": 0.6794815063476562, "learning_rate": 1.2105155732018532e-06, "loss": 0.5458, "step": 1154 }, { "epoch": 2.389655172413793, "grad_norm": 0.6938785910606384, "learning_rate": 1.2026680870968494e-06, "loss": 0.5217, "step": 1155 }, { "epoch": 2.3917241379310346, "grad_norm": 0.6906960010528564, "learning_rate": 1.194842641566406e-06, "loss": 0.5315, "step": 1156 }, { "epoch": 2.393793103448276, "grad_norm": 0.6642972826957703, "learning_rate": 1.187039282031182e-06, "loss": 0.5315, "step": 1157 }, { "epoch": 2.395862068965517, "grad_norm": 0.6678175926208496, "learning_rate": 1.179258053783644e-06, "loss": 0.5007, "step": 1158 }, { "epoch": 2.3979310344827587, "grad_norm": 0.6204706430435181, "learning_rate": 1.171499001987802e-06, "loss": 0.5617, "step": 1159 }, { "epoch": 2.4, "grad_norm": 0.6954528093338013, "learning_rate": 1.163762171678951e-06, "loss": 0.5211, "step": 1160 }, { "epoch": 2.402068965517241, "grad_norm": 0.6939405798912048, "learning_rate": 1.156047607763407e-06, "loss": 0.5264, "step": 1161 }, { "epoch": 2.404137931034483, "grad_norm": 0.7006513476371765, "learning_rate": 1.1483553550182453e-06, "loss": 0.5488, "step": 1162 }, { "epoch": 2.406206896551724, "grad_norm": 0.67670738697052, "learning_rate": 1.1406854580910426e-06, "loss": 0.5304, "step": 1163 }, { "epoch": 2.4082758620689657, "grad_norm": 0.681532084941864, "learning_rate": 1.1330379614996223e-06, "loss": 0.5307, "step": 1164 }, { "epoch": 2.410344827586207, "grad_norm": 0.671879768371582, "learning_rate": 1.1254129096317807e-06, "loss": 0.5423, "step": 1165 }, { "epoch": 2.412413793103448, "grad_norm": 0.6575496196746826, "learning_rate": 1.1178103467450512e-06, "loss": 0.5261, "step": 1166 }, { "epoch": 2.41448275862069, "grad_norm": 0.692674458026886, "learning_rate": 1.110230316966427e-06, "loss": 0.522, "step": 1167 }, { "epoch": 2.416551724137931, "grad_norm": 0.6568799018859863, "learning_rate": 1.102672864292122e-06, "loss": 0.5187, "step": 1168 }, { "epoch": 2.4186206896551723, "grad_norm": 0.6846016049385071, "learning_rate": 1.095138032587298e-06, "loss": 0.5518, "step": 1169 }, { "epoch": 2.420689655172414, "grad_norm": 0.6592535376548767, "learning_rate": 1.0876258655858307e-06, "loss": 0.5241, "step": 1170 }, { "epoch": 2.422758620689655, "grad_norm": 0.6726316809654236, "learning_rate": 1.0801364068900334e-06, "loss": 0.5153, "step": 1171 }, { "epoch": 2.4248275862068964, "grad_norm": 0.661979615688324, "learning_rate": 1.0726696999704266e-06, "loss": 0.5321, "step": 1172 }, { "epoch": 2.426896551724138, "grad_norm": 0.6790592074394226, "learning_rate": 1.0652257881654625e-06, "loss": 0.5361, "step": 1173 }, { "epoch": 2.4289655172413793, "grad_norm": 0.6957859396934509, "learning_rate": 1.057804714681296e-06, "loss": 0.5129, "step": 1174 }, { "epoch": 2.4310344827586206, "grad_norm": 0.6732125282287598, "learning_rate": 1.0504065225915166e-06, "loss": 0.56, "step": 1175 }, { "epoch": 2.4331034482758622, "grad_norm": 0.6757326126098633, "learning_rate": 1.0430312548369076e-06, "loss": 0.5469, "step": 1176 }, { "epoch": 2.4351724137931035, "grad_norm": 0.673953115940094, "learning_rate": 1.0356789542251939e-06, "loss": 0.507, "step": 1177 }, { "epoch": 2.4372413793103447, "grad_norm": 0.6575935482978821, "learning_rate": 1.028349663430792e-06, "loss": 0.5134, "step": 1178 }, { "epoch": 2.4393103448275864, "grad_norm": 0.673143208026886, "learning_rate": 1.0210434249945677e-06, "loss": 0.5443, "step": 1179 }, { "epoch": 2.4413793103448276, "grad_norm": 0.6751227974891663, "learning_rate": 1.0137602813235824e-06, "loss": 0.5535, "step": 1180 }, { "epoch": 2.443448275862069, "grad_norm": 0.6402212977409363, "learning_rate": 1.0065002746908532e-06, "loss": 0.5517, "step": 1181 }, { "epoch": 2.4455172413793105, "grad_norm": 0.6959084868431091, "learning_rate": 9.992634472351016e-07, "loss": 0.5104, "step": 1182 }, { "epoch": 2.4475862068965517, "grad_norm": 0.673514187335968, "learning_rate": 9.92049840960514e-07, "loss": 0.5485, "step": 1183 }, { "epoch": 2.449655172413793, "grad_norm": 0.64295893907547, "learning_rate": 9.84859497736495e-07, "loss": 0.5202, "step": 1184 }, { "epoch": 2.4517241379310346, "grad_norm": 0.669701337814331, "learning_rate": 9.776924592974257e-07, "loss": 0.5201, "step": 1185 }, { "epoch": 2.453793103448276, "grad_norm": 0.6700237393379211, "learning_rate": 9.705487672424202e-07, "loss": 0.5437, "step": 1186 }, { "epoch": 2.455862068965517, "grad_norm": 0.6907758712768555, "learning_rate": 9.63428463035085e-07, "loss": 0.511, "step": 1187 }, { "epoch": 2.4579310344827587, "grad_norm": 0.6732699275016785, "learning_rate": 9.563315880032798e-07, "loss": 0.5418, "step": 1188 }, { "epoch": 2.46, "grad_norm": 0.7099267840385437, "learning_rate": 9.492581833388736e-07, "loss": 0.5403, "step": 1189 }, { "epoch": 2.462068965517241, "grad_norm": 0.6354034543037415, "learning_rate": 9.422082900975105e-07, "loss": 0.5227, "step": 1190 }, { "epoch": 2.464137931034483, "grad_norm": 0.6575520038604736, "learning_rate": 9.351819491983671e-07, "loss": 0.5327, "step": 1191 }, { "epoch": 2.466206896551724, "grad_norm": 0.6583318114280701, "learning_rate": 9.281792014239171e-07, "loss": 0.5292, "step": 1192 }, { "epoch": 2.4682758620689658, "grad_norm": 0.6778759360313416, "learning_rate": 9.212000874196953e-07, "loss": 0.5091, "step": 1193 }, { "epoch": 2.470344827586207, "grad_norm": 0.6768158674240112, "learning_rate": 9.142446476940598e-07, "loss": 0.5043, "step": 1194 }, { "epoch": 2.472413793103448, "grad_norm": 0.6922417879104614, "learning_rate": 9.073129226179589e-07, "loss": 0.5382, "step": 1195 }, { "epoch": 2.47448275862069, "grad_norm": 0.6937352418899536, "learning_rate": 9.004049524246944e-07, "loss": 0.5334, "step": 1196 }, { "epoch": 2.476551724137931, "grad_norm": 0.6940251588821411, "learning_rate": 8.935207772096904e-07, "loss": 0.5235, "step": 1197 }, { "epoch": 2.4786206896551723, "grad_norm": 0.6610797643661499, "learning_rate": 8.86660436930259e-07, "loss": 0.5369, "step": 1198 }, { "epoch": 2.480689655172414, "grad_norm": 0.6743729114532471, "learning_rate": 8.798239714053692e-07, "loss": 0.5494, "step": 1199 }, { "epoch": 2.4827586206896552, "grad_norm": 0.6906112432479858, "learning_rate": 8.730114203154144e-07, "loss": 0.5221, "step": 1200 }, { "epoch": 2.4848275862068965, "grad_norm": 0.6822516322135925, "learning_rate": 8.662228232019876e-07, "loss": 0.5364, "step": 1201 }, { "epoch": 2.486896551724138, "grad_norm": 0.6591139435768127, "learning_rate": 8.594582194676393e-07, "loss": 0.5412, "step": 1202 }, { "epoch": 2.4889655172413794, "grad_norm": 0.694092333316803, "learning_rate": 8.527176483756671e-07, "loss": 0.5233, "step": 1203 }, { "epoch": 2.4910344827586206, "grad_norm": 0.6754273772239685, "learning_rate": 8.460011490498671e-07, "loss": 0.5501, "step": 1204 }, { "epoch": 2.4931034482758623, "grad_norm": 0.670056164264679, "learning_rate": 8.393087604743283e-07, "loss": 0.528, "step": 1205 }, { "epoch": 2.4951724137931035, "grad_norm": 0.6645668148994446, "learning_rate": 8.326405214931838e-07, "loss": 0.5586, "step": 1206 }, { "epoch": 2.4972413793103447, "grad_norm": 0.6548720002174377, "learning_rate": 8.259964708104091e-07, "loss": 0.5316, "step": 1207 }, { "epoch": 2.4993103448275864, "grad_norm": 0.6840012073516846, "learning_rate": 8.193766469895737e-07, "loss": 0.5274, "step": 1208 }, { "epoch": 2.5013793103448276, "grad_norm": 0.655719518661499, "learning_rate": 8.127810884536402e-07, "loss": 0.521, "step": 1209 }, { "epoch": 2.503448275862069, "grad_norm": 0.6610667109489441, "learning_rate": 8.062098334847185e-07, "loss": 0.5472, "step": 1210 }, { "epoch": 2.5055172413793105, "grad_norm": 0.6627329587936401, "learning_rate": 7.996629202238654e-07, "loss": 0.5101, "step": 1211 }, { "epoch": 2.5075862068965518, "grad_norm": 0.6512289047241211, "learning_rate": 7.931403866708464e-07, "loss": 0.5263, "step": 1212 }, { "epoch": 2.509655172413793, "grad_norm": 0.6700037121772766, "learning_rate": 7.866422706839239e-07, "loss": 0.5265, "step": 1213 }, { "epoch": 2.5117241379310347, "grad_norm": 0.649949312210083, "learning_rate": 7.801686099796358e-07, "loss": 0.5439, "step": 1214 }, { "epoch": 2.513793103448276, "grad_norm": 0.6628941893577576, "learning_rate": 7.737194421325739e-07, "loss": 0.5112, "step": 1215 }, { "epoch": 2.515862068965517, "grad_norm": 0.6837835311889648, "learning_rate": 7.672948045751699e-07, "loss": 0.5497, "step": 1216 }, { "epoch": 2.517931034482759, "grad_norm": 0.6798170804977417, "learning_rate": 7.60894734597476e-07, "loss": 0.5439, "step": 1217 }, { "epoch": 2.52, "grad_norm": 0.6628232002258301, "learning_rate": 7.545192693469467e-07, "loss": 0.5435, "step": 1218 }, { "epoch": 2.5220689655172412, "grad_norm": 0.6693230867385864, "learning_rate": 7.481684458282273e-07, "loss": 0.5377, "step": 1219 }, { "epoch": 2.524137931034483, "grad_norm": 0.6691813468933105, "learning_rate": 7.418423009029363e-07, "loss": 0.5463, "step": 1220 }, { "epoch": 2.526206896551724, "grad_norm": 0.6763612031936646, "learning_rate": 7.355408712894508e-07, "loss": 0.5616, "step": 1221 }, { "epoch": 2.5282758620689654, "grad_norm": 0.6607294082641602, "learning_rate": 7.292641935626966e-07, "loss": 0.5381, "step": 1222 }, { "epoch": 2.530344827586207, "grad_norm": 0.6794742941856384, "learning_rate": 7.23012304153931e-07, "loss": 0.5427, "step": 1223 }, { "epoch": 2.5324137931034483, "grad_norm": 0.6626706719398499, "learning_rate": 7.167852393505393e-07, "loss": 0.534, "step": 1224 }, { "epoch": 2.5344827586206895, "grad_norm": 0.6796131730079651, "learning_rate": 7.105830352958143e-07, "loss": 0.5493, "step": 1225 }, { "epoch": 2.536551724137931, "grad_norm": 0.6909650564193726, "learning_rate": 7.04405727988754e-07, "loss": 0.5359, "step": 1226 }, { "epoch": 2.5386206896551724, "grad_norm": 0.6646261811256409, "learning_rate": 6.982533532838481e-07, "loss": 0.4803, "step": 1227 }, { "epoch": 2.5406896551724136, "grad_norm": 0.6714354753494263, "learning_rate": 6.921259468908742e-07, "loss": 0.5642, "step": 1228 }, { "epoch": 2.5427586206896553, "grad_norm": 0.6736488342285156, "learning_rate": 6.86023544374686e-07, "loss": 0.5513, "step": 1229 }, { "epoch": 2.5448275862068965, "grad_norm": 0.7024829387664795, "learning_rate": 6.799461811550101e-07, "loss": 0.5382, "step": 1230 }, { "epoch": 2.5468965517241378, "grad_norm": 0.670295000076294, "learning_rate": 6.738938925062399e-07, "loss": 0.5301, "step": 1231 }, { "epoch": 2.5489655172413794, "grad_norm": 0.6387397050857544, "learning_rate": 6.678667135572293e-07, "loss": 0.5093, "step": 1232 }, { "epoch": 2.5510344827586207, "grad_norm": 0.6766871809959412, "learning_rate": 6.618646792910893e-07, "loss": 0.5519, "step": 1233 }, { "epoch": 2.553103448275862, "grad_norm": 0.7046573758125305, "learning_rate": 6.558878245449879e-07, "loss": 0.5218, "step": 1234 }, { "epoch": 2.5551724137931036, "grad_norm": 0.6664929389953613, "learning_rate": 6.499361840099421e-07, "loss": 0.522, "step": 1235 }, { "epoch": 2.557241379310345, "grad_norm": 0.6669551730155945, "learning_rate": 6.440097922306232e-07, "loss": 0.565, "step": 1236 }, { "epoch": 2.559310344827586, "grad_norm": 0.6854833364486694, "learning_rate": 6.381086836051498e-07, "loss": 0.5561, "step": 1237 }, { "epoch": 2.5613793103448277, "grad_norm": 0.6877058148384094, "learning_rate": 6.322328923848958e-07, "loss": 0.5313, "step": 1238 }, { "epoch": 2.563448275862069, "grad_norm": 0.6529080867767334, "learning_rate": 6.263824526742807e-07, "loss": 0.5318, "step": 1239 }, { "epoch": 2.56551724137931, "grad_norm": 0.6818994879722595, "learning_rate": 6.205573984305857e-07, "loss": 0.5387, "step": 1240 }, { "epoch": 2.567586206896552, "grad_norm": 0.6354886889457703, "learning_rate": 6.147577634637413e-07, "loss": 0.5351, "step": 1241 }, { "epoch": 2.569655172413793, "grad_norm": 0.6590231657028198, "learning_rate": 6.089835814361461e-07, "loss": 0.5288, "step": 1242 }, { "epoch": 2.5717241379310343, "grad_norm": 0.6300837993621826, "learning_rate": 6.03234885862457e-07, "loss": 0.554, "step": 1243 }, { "epoch": 2.573793103448276, "grad_norm": 0.64389568567276, "learning_rate": 5.975117101094091e-07, "loss": 0.5548, "step": 1244 }, { "epoch": 2.575862068965517, "grad_norm": 0.65228271484375, "learning_rate": 5.918140873956063e-07, "loss": 0.5418, "step": 1245 }, { "epoch": 2.5779310344827584, "grad_norm": 0.6716015934944153, "learning_rate": 5.861420507913462e-07, "loss": 0.5506, "step": 1246 }, { "epoch": 2.58, "grad_norm": 0.6640960574150085, "learning_rate": 5.804956332184092e-07, "loss": 0.5215, "step": 1247 }, { "epoch": 2.5820689655172413, "grad_norm": 0.6824872493743896, "learning_rate": 5.748748674498855e-07, "loss": 0.546, "step": 1248 }, { "epoch": 2.5841379310344825, "grad_norm": 0.6688941121101379, "learning_rate": 5.692797861099719e-07, "loss": 0.5166, "step": 1249 }, { "epoch": 2.586206896551724, "grad_norm": 0.6445797085762024, "learning_rate": 5.637104216737871e-07, "loss": 0.5286, "step": 1250 }, { "epoch": 2.5882758620689654, "grad_norm": 0.6584656238555908, "learning_rate": 5.581668064671847e-07, "loss": 0.5364, "step": 1251 }, { "epoch": 2.5903448275862067, "grad_norm": 0.6681441068649292, "learning_rate": 5.526489726665624e-07, "loss": 0.5203, "step": 1252 }, { "epoch": 2.5924137931034483, "grad_norm": 0.6537905931472778, "learning_rate": 5.471569522986775e-07, "loss": 0.5203, "step": 1253 }, { "epoch": 2.5944827586206896, "grad_norm": 0.613910973072052, "learning_rate": 5.416907772404606e-07, "loss": 0.5105, "step": 1254 }, { "epoch": 2.596551724137931, "grad_norm": 0.6802027225494385, "learning_rate": 5.3625047921883e-07, "loss": 0.5231, "step": 1255 }, { "epoch": 2.5986206896551725, "grad_norm": 0.6352335214614868, "learning_rate": 5.30836089810507e-07, "loss": 0.5125, "step": 1256 }, { "epoch": 2.6006896551724137, "grad_norm": 0.6671028137207031, "learning_rate": 5.254476404418341e-07, "loss": 0.5221, "step": 1257 }, { "epoch": 2.602758620689655, "grad_norm": 0.6607826948165894, "learning_rate": 5.200851623885922e-07, "loss": 0.5384, "step": 1258 }, { "epoch": 2.6048275862068966, "grad_norm": 0.645964503288269, "learning_rate": 5.147486867758201e-07, "loss": 0.5306, "step": 1259 }, { "epoch": 2.606896551724138, "grad_norm": 0.6596983671188354, "learning_rate": 5.094382445776281e-07, "loss": 0.5464, "step": 1260 }, { "epoch": 2.608965517241379, "grad_norm": 0.678046464920044, "learning_rate": 5.041538666170282e-07, "loss": 0.524, "step": 1261 }, { "epoch": 2.6110344827586207, "grad_norm": 0.6556622385978699, "learning_rate": 4.988955835657456e-07, "loss": 0.5395, "step": 1262 }, { "epoch": 2.613103448275862, "grad_norm": 0.6770464181900024, "learning_rate": 4.93663425944047e-07, "loss": 0.5394, "step": 1263 }, { "epoch": 2.6151724137931036, "grad_norm": 0.6753197312355042, "learning_rate": 4.884574241205603e-07, "loss": 0.5239, "step": 1264 }, { "epoch": 2.617241379310345, "grad_norm": 0.655596911907196, "learning_rate": 4.832776083120983e-07, "loss": 0.5148, "step": 1265 }, { "epoch": 2.619310344827586, "grad_norm": 0.6836827397346497, "learning_rate": 4.781240085834865e-07, "loss": 0.5453, "step": 1266 }, { "epoch": 2.6213793103448277, "grad_norm": 0.680099368095398, "learning_rate": 4.729966548473841e-07, "loss": 0.5148, "step": 1267 }, { "epoch": 2.623448275862069, "grad_norm": 0.6692262887954712, "learning_rate": 4.678955768641147e-07, "loss": 0.554, "step": 1268 }, { "epoch": 2.62551724137931, "grad_norm": 0.6534875631332397, "learning_rate": 4.628208042414889e-07, "loss": 0.5548, "step": 1269 }, { "epoch": 2.627586206896552, "grad_norm": 0.617273211479187, "learning_rate": 4.5777236643463787e-07, "loss": 0.5305, "step": 1270 }, { "epoch": 2.629655172413793, "grad_norm": 0.6691027283668518, "learning_rate": 4.5275029274583715e-07, "loss": 0.5292, "step": 1271 }, { "epoch": 2.6317241379310343, "grad_norm": 0.6685481071472168, "learning_rate": 4.477546123243409e-07, "loss": 0.5306, "step": 1272 }, { "epoch": 2.633793103448276, "grad_norm": 0.6642303466796875, "learning_rate": 4.4278535416620914e-07, "loss": 0.5276, "step": 1273 }, { "epoch": 2.6358620689655172, "grad_norm": 0.6658126711845398, "learning_rate": 4.3784254711414064e-07, "loss": 0.5447, "step": 1274 }, { "epoch": 2.637931034482759, "grad_norm": 0.6844912767410278, "learning_rate": 4.329262198573109e-07, "loss": 0.5419, "step": 1275 }, { "epoch": 2.64, "grad_norm": 0.6531012654304504, "learning_rate": 4.280364009311927e-07, "loss": 0.5283, "step": 1276 }, { "epoch": 2.6420689655172414, "grad_norm": 0.6809141635894775, "learning_rate": 4.231731187174065e-07, "loss": 0.5486, "step": 1277 }, { "epoch": 2.644137931034483, "grad_norm": 0.6723445653915405, "learning_rate": 4.183364014435398e-07, "loss": 0.5441, "step": 1278 }, { "epoch": 2.6462068965517243, "grad_norm": 0.6648197174072266, "learning_rate": 4.1352627718299953e-07, "loss": 0.5506, "step": 1279 }, { "epoch": 2.6482758620689655, "grad_norm": 0.668096661567688, "learning_rate": 4.087427738548322e-07, "loss": 0.5363, "step": 1280 }, { "epoch": 2.650344827586207, "grad_norm": 0.6981197595596313, "learning_rate": 4.0398591922357787e-07, "loss": 0.5673, "step": 1281 }, { "epoch": 2.6524137931034484, "grad_norm": 0.6475323438644409, "learning_rate": 3.9925574089909603e-07, "loss": 0.5402, "step": 1282 }, { "epoch": 2.6544827586206896, "grad_norm": 0.640293300151825, "learning_rate": 3.945522663364154e-07, "loss": 0.5302, "step": 1283 }, { "epoch": 2.6565517241379313, "grad_norm": 0.6663315296173096, "learning_rate": 3.898755228355661e-07, "loss": 0.5379, "step": 1284 }, { "epoch": 2.6586206896551725, "grad_norm": 0.6679863929748535, "learning_rate": 3.852255375414271e-07, "loss": 0.5605, "step": 1285 }, { "epoch": 2.6606896551724137, "grad_norm": 0.6763351559638977, "learning_rate": 3.8060233744356634e-07, "loss": 0.5102, "step": 1286 }, { "epoch": 2.6627586206896554, "grad_norm": 0.6907029151916504, "learning_rate": 3.760059493760848e-07, "loss": 0.522, "step": 1287 }, { "epoch": 2.6648275862068966, "grad_norm": 0.6595637202262878, "learning_rate": 3.7143640001745916e-07, "loss": 0.549, "step": 1288 }, { "epoch": 2.666896551724138, "grad_norm": 0.6693307161331177, "learning_rate": 3.6689371589039013e-07, "loss": 0.5257, "step": 1289 }, { "epoch": 2.6689655172413795, "grad_norm": 0.6373291611671448, "learning_rate": 3.6237792336164424e-07, "loss": 0.5478, "step": 1290 }, { "epoch": 2.6710344827586208, "grad_norm": 0.6235068440437317, "learning_rate": 3.5788904864190346e-07, "loss": 0.5484, "step": 1291 }, { "epoch": 2.673103448275862, "grad_norm": 0.6607393026351929, "learning_rate": 3.534271177856163e-07, "loss": 0.5082, "step": 1292 }, { "epoch": 2.6751724137931037, "grad_norm": 0.6282607316970825, "learning_rate": 3.489921566908372e-07, "loss": 0.5366, "step": 1293 }, { "epoch": 2.677241379310345, "grad_norm": 0.6705973148345947, "learning_rate": 3.445841910990877e-07, "loss": 0.5382, "step": 1294 }, { "epoch": 2.679310344827586, "grad_norm": 0.6531069278717041, "learning_rate": 3.4020324659519487e-07, "loss": 0.5556, "step": 1295 }, { "epoch": 2.681379310344828, "grad_norm": 0.6508034467697144, "learning_rate": 3.358493486071568e-07, "loss": 0.5256, "step": 1296 }, { "epoch": 2.683448275862069, "grad_norm": 0.6649227142333984, "learning_rate": 3.315225224059809e-07, "loss": 0.5104, "step": 1297 }, { "epoch": 2.6855172413793102, "grad_norm": 0.6882157325744629, "learning_rate": 3.2722279310554825e-07, "loss": 0.542, "step": 1298 }, { "epoch": 2.687586206896552, "grad_norm": 0.6619060039520264, "learning_rate": 3.2295018566246007e-07, "loss": 0.5379, "step": 1299 }, { "epoch": 2.689655172413793, "grad_norm": 0.6553718447685242, "learning_rate": 3.1870472487589845e-07, "loss": 0.5122, "step": 1300 }, { "epoch": 2.6917241379310344, "grad_norm": 0.691162645816803, "learning_rate": 3.1448643538748045e-07, "loss": 0.5102, "step": 1301 }, { "epoch": 2.693793103448276, "grad_norm": 0.6377774477005005, "learning_rate": 3.1029534168111217e-07, "loss": 0.524, "step": 1302 }, { "epoch": 2.6958620689655173, "grad_norm": 0.6641302704811096, "learning_rate": 3.06131468082852e-07, "loss": 0.5126, "step": 1303 }, { "epoch": 2.6979310344827585, "grad_norm": 0.6569216251373291, "learning_rate": 3.0199483876076465e-07, "loss": 0.5347, "step": 1304 }, { "epoch": 2.7, "grad_norm": 0.6541246175765991, "learning_rate": 2.9788547772478416e-07, "loss": 0.5384, "step": 1305 }, { "epoch": 2.7020689655172414, "grad_norm": 0.6962401866912842, "learning_rate": 2.938034088265723e-07, "loss": 0.5321, "step": 1306 }, { "epoch": 2.7041379310344826, "grad_norm": 0.6459115147590637, "learning_rate": 2.897486557593826e-07, "loss": 0.5477, "step": 1307 }, { "epoch": 2.7062068965517243, "grad_norm": 0.6690937280654907, "learning_rate": 2.8572124205791916e-07, "loss": 0.5361, "step": 1308 }, { "epoch": 2.7082758620689655, "grad_norm": 0.651241660118103, "learning_rate": 2.817211910982037e-07, "loss": 0.5231, "step": 1309 }, { "epoch": 2.7103448275862068, "grad_norm": 0.6507266759872437, "learning_rate": 2.7774852609743887e-07, "loss": 0.5331, "step": 1310 }, { "epoch": 2.7124137931034484, "grad_norm": 0.6396605968475342, "learning_rate": 2.7380327011387096e-07, "loss": 0.5418, "step": 1311 }, { "epoch": 2.7144827586206897, "grad_norm": 0.635365903377533, "learning_rate": 2.698854460466599e-07, "loss": 0.5258, "step": 1312 }, { "epoch": 2.716551724137931, "grad_norm": 0.6772063374519348, "learning_rate": 2.6599507663574387e-07, "loss": 0.5274, "step": 1313 }, { "epoch": 2.7186206896551726, "grad_norm": 0.6762843132019043, "learning_rate": 2.621321844617092e-07, "loss": 0.5279, "step": 1314 }, { "epoch": 2.720689655172414, "grad_norm": 0.6602565050125122, "learning_rate": 2.582967919456547e-07, "loss": 0.5343, "step": 1315 }, { "epoch": 2.722758620689655, "grad_norm": 0.6536930799484253, "learning_rate": 2.544889213490698e-07, "loss": 0.5146, "step": 1316 }, { "epoch": 2.7248275862068967, "grad_norm": 0.6757822632789612, "learning_rate": 2.5070859477369645e-07, "loss": 0.5247, "step": 1317 }, { "epoch": 2.726896551724138, "grad_norm": 0.6500410437583923, "learning_rate": 2.46955834161407e-07, "loss": 0.5383, "step": 1318 }, { "epoch": 2.728965517241379, "grad_norm": 0.6390625834465027, "learning_rate": 2.432306612940738e-07, "loss": 0.5254, "step": 1319 }, { "epoch": 2.731034482758621, "grad_norm": 0.6645777821540833, "learning_rate": 2.39533097793444e-07, "loss": 0.4968, "step": 1320 }, { "epoch": 2.733103448275862, "grad_norm": 0.6638719439506531, "learning_rate": 2.3586316512101416e-07, "loss": 0.5147, "step": 1321 }, { "epoch": 2.7351724137931033, "grad_norm": 0.6468044519424438, "learning_rate": 2.3222088457790448e-07, "loss": 0.5321, "step": 1322 }, { "epoch": 2.737241379310345, "grad_norm": 0.6644803285598755, "learning_rate": 2.286062773047354e-07, "loss": 0.5462, "step": 1323 }, { "epoch": 2.739310344827586, "grad_norm": 0.6668086647987366, "learning_rate": 2.2501936428150794e-07, "loss": 0.5093, "step": 1324 }, { "epoch": 2.7413793103448274, "grad_norm": 0.6860660910606384, "learning_rate": 2.2146016632747624e-07, "loss": 0.5485, "step": 1325 }, { "epoch": 2.743448275862069, "grad_norm": 0.6771222949028015, "learning_rate": 2.1792870410103195e-07, "loss": 0.5049, "step": 1326 }, { "epoch": 2.7455172413793103, "grad_norm": 0.6570666432380676, "learning_rate": 2.1442499809958273e-07, "loss": 0.5263, "step": 1327 }, { "epoch": 2.7475862068965515, "grad_norm": 0.6926378011703491, "learning_rate": 2.109490686594301e-07, "loss": 0.5532, "step": 1328 }, { "epoch": 2.749655172413793, "grad_norm": 0.6724755764007568, "learning_rate": 2.0750093595565735e-07, "loss": 0.5282, "step": 1329 }, { "epoch": 2.7517241379310344, "grad_norm": 0.6550016403198242, "learning_rate": 2.0408062000200623e-07, "loss": 0.5475, "step": 1330 }, { "epoch": 2.7537931034482757, "grad_norm": 0.6652045249938965, "learning_rate": 2.006881406507677e-07, "loss": 0.5517, "step": 1331 }, { "epoch": 2.7558620689655173, "grad_norm": 0.6777829527854919, "learning_rate": 1.9732351759265743e-07, "loss": 0.5755, "step": 1332 }, { "epoch": 2.7579310344827586, "grad_norm": 0.632814347743988, "learning_rate": 1.9398677035671222e-07, "loss": 0.5004, "step": 1333 }, { "epoch": 2.76, "grad_norm": 0.6533367037773132, "learning_rate": 1.9067791831016647e-07, "loss": 0.5314, "step": 1334 }, { "epoch": 2.7620689655172415, "grad_norm": 0.6759463548660278, "learning_rate": 1.8739698065834977e-07, "loss": 0.5277, "step": 1335 }, { "epoch": 2.7641379310344827, "grad_norm": 0.6170943975448608, "learning_rate": 1.841439764445646e-07, "loss": 0.5207, "step": 1336 }, { "epoch": 2.766206896551724, "grad_norm": 0.6548056602478027, "learning_rate": 1.8091892454998595e-07, "loss": 0.5539, "step": 1337 }, { "epoch": 2.7682758620689656, "grad_norm": 0.6568596363067627, "learning_rate": 1.7772184369354517e-07, "loss": 0.5267, "step": 1338 }, { "epoch": 2.770344827586207, "grad_norm": 0.6602894067764282, "learning_rate": 1.7455275243182413e-07, "loss": 0.5589, "step": 1339 }, { "epoch": 2.772413793103448, "grad_norm": 0.6518042087554932, "learning_rate": 1.714116691589457e-07, "loss": 0.518, "step": 1340 }, { "epoch": 2.7744827586206897, "grad_norm": 0.6620248556137085, "learning_rate": 1.682986121064689e-07, "loss": 0.5521, "step": 1341 }, { "epoch": 2.776551724137931, "grad_norm": 0.6624969840049744, "learning_rate": 1.6521359934328185e-07, "loss": 0.5467, "step": 1342 }, { "epoch": 2.778620689655172, "grad_norm": 0.6587342619895935, "learning_rate": 1.6215664877549774e-07, "loss": 0.5063, "step": 1343 }, { "epoch": 2.780689655172414, "grad_norm": 0.6892203092575073, "learning_rate": 1.5912777814635018e-07, "loss": 0.5463, "step": 1344 }, { "epoch": 2.782758620689655, "grad_norm": 0.6598497033119202, "learning_rate": 1.561270050360897e-07, "loss": 0.5265, "step": 1345 }, { "epoch": 2.7848275862068963, "grad_norm": 0.6603181958198547, "learning_rate": 1.5315434686188345e-07, "loss": 0.5467, "step": 1346 }, { "epoch": 2.786896551724138, "grad_norm": 0.6583660840988159, "learning_rate": 1.5020982087771362e-07, "loss": 0.5537, "step": 1347 }, { "epoch": 2.788965517241379, "grad_norm": 0.6598078012466431, "learning_rate": 1.472934441742746e-07, "loss": 0.5563, "step": 1348 }, { "epoch": 2.7910344827586204, "grad_norm": 0.6481250524520874, "learning_rate": 1.444052336788787e-07, "loss": 0.5573, "step": 1349 }, { "epoch": 2.793103448275862, "grad_norm": 0.6612368226051331, "learning_rate": 1.4154520615535185e-07, "loss": 0.5604, "step": 1350 }, { "epoch": 2.7951724137931033, "grad_norm": 0.6412969827651978, "learning_rate": 1.3871337820394404e-07, "loss": 0.5592, "step": 1351 }, { "epoch": 2.7972413793103446, "grad_norm": 0.6546990275382996, "learning_rate": 1.359097662612252e-07, "loss": 0.5295, "step": 1352 }, { "epoch": 2.7993103448275862, "grad_norm": 0.6505655646324158, "learning_rate": 1.33134386599994e-07, "loss": 0.5208, "step": 1353 }, { "epoch": 2.8013793103448275, "grad_norm": 0.6737064123153687, "learning_rate": 1.303872553291824e-07, "loss": 0.529, "step": 1354 }, { "epoch": 2.803448275862069, "grad_norm": 0.6469187140464783, "learning_rate": 1.2766838839376294e-07, "loss": 0.5435, "step": 1355 }, { "epoch": 2.8055172413793104, "grad_norm": 0.6671730279922485, "learning_rate": 1.24977801574655e-07, "loss": 0.5453, "step": 1356 }, { "epoch": 2.8075862068965516, "grad_norm": 0.656061053276062, "learning_rate": 1.2231551048863421e-07, "loss": 0.5404, "step": 1357 }, { "epoch": 2.8096551724137933, "grad_norm": 0.6444647312164307, "learning_rate": 1.196815305882404e-07, "loss": 0.5321, "step": 1358 }, { "epoch": 2.8117241379310345, "grad_norm": 0.6776153445243835, "learning_rate": 1.1707587716169033e-07, "loss": 0.5226, "step": 1359 }, { "epoch": 2.8137931034482757, "grad_norm": 0.6640626192092896, "learning_rate": 1.1449856533278624e-07, "loss": 0.5381, "step": 1360 }, { "epoch": 2.8158620689655174, "grad_norm": 0.6596142053604126, "learning_rate": 1.1194961006082972e-07, "loss": 0.52, "step": 1361 }, { "epoch": 2.8179310344827586, "grad_norm": 0.6379115581512451, "learning_rate": 1.0942902614053453e-07, "loss": 0.5294, "step": 1362 }, { "epoch": 2.82, "grad_norm": 0.6417571902275085, "learning_rate": 1.0693682820194062e-07, "loss": 0.56, "step": 1363 }, { "epoch": 2.8220689655172415, "grad_norm": 0.672452986240387, "learning_rate": 1.0447303071032977e-07, "loss": 0.5091, "step": 1364 }, { "epoch": 2.8241379310344827, "grad_norm": 0.6543784737586975, "learning_rate": 1.0203764796614057e-07, "loss": 0.5518, "step": 1365 }, { "epoch": 2.8262068965517244, "grad_norm": 0.6682656407356262, "learning_rate": 9.963069410488635e-08, "loss": 0.5146, "step": 1366 }, { "epoch": 2.8282758620689656, "grad_norm": 0.6736637949943542, "learning_rate": 9.725218309707129e-08, "loss": 0.5245, "step": 1367 }, { "epoch": 2.830344827586207, "grad_norm": 0.6793366074562073, "learning_rate": 9.490212874811388e-08, "loss": 0.5301, "step": 1368 }, { "epoch": 2.8324137931034485, "grad_norm": 0.6533759832382202, "learning_rate": 9.258054469825972e-08, "loss": 0.5245, "step": 1369 }, { "epoch": 2.8344827586206898, "grad_norm": 0.6463536024093628, "learning_rate": 9.028744442251103e-08, "loss": 0.5174, "step": 1370 }, { "epoch": 2.836551724137931, "grad_norm": 0.691880464553833, "learning_rate": 8.802284123053894e-08, "loss": 0.5345, "step": 1371 }, { "epoch": 2.8386206896551727, "grad_norm": 0.6536058187484741, "learning_rate": 8.578674826661527e-08, "loss": 0.5371, "step": 1372 }, { "epoch": 2.840689655172414, "grad_norm": 0.658385157585144, "learning_rate": 8.357917850952802e-08, "loss": 0.5064, "step": 1373 }, { "epoch": 2.842758620689655, "grad_norm": 0.662955105304718, "learning_rate": 8.140014477251379e-08, "loss": 0.5247, "step": 1374 }, { "epoch": 2.844827586206897, "grad_norm": 0.6408557295799255, "learning_rate": 7.924965970317722e-08, "loss": 0.5348, "step": 1375 }, { "epoch": 2.846896551724138, "grad_norm": 0.6421958208084106, "learning_rate": 7.712773578342047e-08, "loss": 0.5583, "step": 1376 }, { "epoch": 2.8489655172413793, "grad_norm": 0.6565936803817749, "learning_rate": 7.503438532937169e-08, "loss": 0.5502, "step": 1377 }, { "epoch": 2.851034482758621, "grad_norm": 0.6656866073608398, "learning_rate": 7.296962049131051e-08, "loss": 0.5256, "step": 1378 }, { "epoch": 2.853103448275862, "grad_norm": 0.662927508354187, "learning_rate": 7.093345325359935e-08, "loss": 0.5399, "step": 1379 }, { "epoch": 2.8551724137931034, "grad_norm": 0.664005696773529, "learning_rate": 6.892589543461392e-08, "loss": 0.5233, "step": 1380 }, { "epoch": 2.857241379310345, "grad_norm": 0.6524354219436646, "learning_rate": 6.694695868667556e-08, "loss": 0.5318, "step": 1381 }, { "epoch": 2.8593103448275863, "grad_norm": 0.6665657758712769, "learning_rate": 6.49966544959807e-08, "loss": 0.5241, "step": 1382 }, { "epoch": 2.8613793103448275, "grad_norm": 0.6510589122772217, "learning_rate": 6.307499418253705e-08, "loss": 0.5293, "step": 1383 }, { "epoch": 2.863448275862069, "grad_norm": 0.6608887910842896, "learning_rate": 6.118198890009586e-08, "loss": 0.5257, "step": 1384 }, { "epoch": 2.8655172413793104, "grad_norm": 0.6739917993545532, "learning_rate": 5.9317649636088656e-08, "loss": 0.5544, "step": 1385 }, { "epoch": 2.8675862068965516, "grad_norm": 0.6483108401298523, "learning_rate": 5.748198721156284e-08, "loss": 0.5297, "step": 1386 }, { "epoch": 2.8696551724137933, "grad_norm": 0.653653621673584, "learning_rate": 5.5675012281119486e-08, "loss": 0.5445, "step": 1387 }, { "epoch": 2.8717241379310345, "grad_norm": 0.6576898097991943, "learning_rate": 5.389673533284956e-08, "loss": 0.4969, "step": 1388 }, { "epoch": 2.8737931034482758, "grad_norm": 0.6513156294822693, "learning_rate": 5.214716668827558e-08, "loss": 0.547, "step": 1389 }, { "epoch": 2.8758620689655174, "grad_norm": 0.6595351099967957, "learning_rate": 5.042631650229057e-08, "loss": 0.5567, "step": 1390 }, { "epoch": 2.8779310344827587, "grad_norm": 0.644473671913147, "learning_rate": 4.8734194763098706e-08, "loss": 0.5005, "step": 1391 }, { "epoch": 2.88, "grad_norm": 0.6628778576850891, "learning_rate": 4.7070811292157513e-08, "loss": 0.5162, "step": 1392 }, { "epoch": 2.8820689655172416, "grad_norm": 0.6727180480957031, "learning_rate": 4.543617574412185e-08, "loss": 0.5441, "step": 1393 }, { "epoch": 2.884137931034483, "grad_norm": 0.666273832321167, "learning_rate": 4.383029760678614e-08, "loss": 0.5409, "step": 1394 }, { "epoch": 2.886206896551724, "grad_norm": 0.6508522629737854, "learning_rate": 4.2253186201031135e-08, "loss": 0.5193, "step": 1395 }, { "epoch": 2.8882758620689657, "grad_norm": 0.6973995566368103, "learning_rate": 4.07048506807689e-08, "loss": 0.5415, "step": 1396 }, { "epoch": 2.890344827586207, "grad_norm": 0.6525006890296936, "learning_rate": 3.9185300032889005e-08, "loss": 0.5379, "step": 1397 }, { "epoch": 2.892413793103448, "grad_norm": 0.677098274230957, "learning_rate": 3.7694543077208546e-08, "loss": 0.5413, "step": 1398 }, { "epoch": 2.89448275862069, "grad_norm": 0.6358975172042847, "learning_rate": 3.6232588466417195e-08, "loss": 0.5449, "step": 1399 }, { "epoch": 2.896551724137931, "grad_norm": 0.6577171087265015, "learning_rate": 3.479944468603169e-08, "loss": 0.5363, "step": 1400 }, { "epoch": 2.8986206896551723, "grad_norm": 0.6604806780815125, "learning_rate": 3.339512005434309e-08, "loss": 0.5307, "step": 1401 }, { "epoch": 2.900689655172414, "grad_norm": 0.6712777614593506, "learning_rate": 3.2019622722369024e-08, "loss": 0.537, "step": 1402 }, { "epoch": 2.902758620689655, "grad_norm": 0.6645734310150146, "learning_rate": 3.0672960673808205e-08, "loss": 0.5775, "step": 1403 }, { "epoch": 2.9048275862068964, "grad_norm": 0.6652960777282715, "learning_rate": 2.93551417249921e-08, "loss": 0.5579, "step": 1404 }, { "epoch": 2.906896551724138, "grad_norm": 0.65793377161026, "learning_rate": 2.8066173524839978e-08, "loss": 0.5572, "step": 1405 }, { "epoch": 2.9089655172413793, "grad_norm": 0.6762325167655945, "learning_rate": 2.6806063554815632e-08, "loss": 0.5502, "step": 1406 }, { "epoch": 2.9110344827586205, "grad_norm": 0.6445053219795227, "learning_rate": 2.5574819128882933e-08, "loss": 0.5287, "step": 1407 }, { "epoch": 2.913103448275862, "grad_norm": 0.651478111743927, "learning_rate": 2.4372447393462562e-08, "loss": 0.5344, "step": 1408 }, { "epoch": 2.9151724137931034, "grad_norm": 0.6526092290878296, "learning_rate": 2.319895532739369e-08, "loss": 0.5081, "step": 1409 }, { "epoch": 2.9172413793103447, "grad_norm": 0.6735363602638245, "learning_rate": 2.205434974188847e-08, "loss": 0.5468, "step": 1410 }, { "epoch": 2.9193103448275863, "grad_norm": 0.6453359127044678, "learning_rate": 2.093863728049872e-08, "loss": 0.5245, "step": 1411 }, { "epoch": 2.9213793103448276, "grad_norm": 0.6538481116294861, "learning_rate": 1.9851824419070965e-08, "loss": 0.5291, "step": 1412 }, { "epoch": 2.923448275862069, "grad_norm": 0.6372977495193481, "learning_rate": 1.8793917465713686e-08, "loss": 0.5489, "step": 1413 }, { "epoch": 2.9255172413793105, "grad_norm": 0.6539329290390015, "learning_rate": 1.7764922560759014e-08, "loss": 0.5307, "step": 1414 }, { "epoch": 2.9275862068965517, "grad_norm": 0.6570875644683838, "learning_rate": 1.676484567672554e-08, "loss": 0.5509, "step": 1415 }, { "epoch": 2.929655172413793, "grad_norm": 0.6732608079910278, "learning_rate": 1.5793692618286115e-08, "loss": 0.5501, "step": 1416 }, { "epoch": 2.9317241379310346, "grad_norm": 0.638278067111969, "learning_rate": 1.4851469022234e-08, "loss": 0.5377, "step": 1417 }, { "epoch": 2.933793103448276, "grad_norm": 0.6890416741371155, "learning_rate": 1.3938180357447318e-08, "loss": 0.503, "step": 1418 }, { "epoch": 2.935862068965517, "grad_norm": 0.6768400073051453, "learning_rate": 1.3053831924861315e-08, "loss": 0.531, "step": 1419 }, { "epoch": 2.9379310344827587, "grad_norm": 0.6668310165405273, "learning_rate": 1.2198428857433941e-08, "loss": 0.5394, "step": 1420 }, { "epoch": 2.94, "grad_norm": 0.6531161665916443, "learning_rate": 1.137197612011809e-08, "loss": 0.526, "step": 1421 }, { "epoch": 2.942068965517241, "grad_norm": 0.6525026559829712, "learning_rate": 1.0574478509833286e-08, "loss": 0.5175, "step": 1422 }, { "epoch": 2.944137931034483, "grad_norm": 0.6503228545188904, "learning_rate": 9.805940655436274e-09, "loss": 0.534, "step": 1423 }, { "epoch": 2.946206896551724, "grad_norm": 0.6437545418739319, "learning_rate": 9.066367017694366e-09, "loss": 0.5318, "step": 1424 }, { "epoch": 2.9482758620689653, "grad_norm": 0.6803710460662842, "learning_rate": 8.35576188926046e-09, "loss": 0.5196, "step": 1425 }, { "epoch": 2.950344827586207, "grad_norm": 0.6403658390045166, "learning_rate": 7.674129394648623e-09, "loss": 0.5252, "step": 1426 }, { "epoch": 2.952413793103448, "grad_norm": 0.6625465750694275, "learning_rate": 7.0214734902074314e-09, "loss": 0.5389, "step": 1427 }, { "epoch": 2.9544827586206894, "grad_norm": 0.6770764589309692, "learning_rate": 6.3977979640994504e-09, "loss": 0.5674, "step": 1428 }, { "epoch": 2.956551724137931, "grad_norm": 0.6394649147987366, "learning_rate": 5.803106436279571e-09, "loss": 0.5134, "step": 1429 }, { "epoch": 2.9586206896551723, "grad_norm": 0.6759665608406067, "learning_rate": 5.237402358471144e-09, "loss": 0.5349, "step": 1430 }, { "epoch": 2.9606896551724136, "grad_norm": 0.6301043629646301, "learning_rate": 4.700689014149329e-09, "loss": 0.5539, "step": 1431 }, { "epoch": 2.9627586206896552, "grad_norm": 0.6656786799430847, "learning_rate": 4.192969518519441e-09, "loss": 0.534, "step": 1432 }, { "epoch": 2.9648275862068965, "grad_norm": 0.681606113910675, "learning_rate": 3.71424681850141e-09, "loss": 0.5136, "step": 1433 }, { "epoch": 2.9668965517241377, "grad_norm": 0.6622211933135986, "learning_rate": 3.2645236927092426e-09, "loss": 0.5406, "step": 1434 }, { "epoch": 2.9689655172413794, "grad_norm": 0.6423860192298889, "learning_rate": 2.8438027514382517e-09, "loss": 0.5265, "step": 1435 }, { "epoch": 2.9710344827586206, "grad_norm": 0.6584990620613098, "learning_rate": 2.452086436648404e-09, "loss": 0.5342, "step": 1436 }, { "epoch": 2.973103448275862, "grad_norm": 0.6501452922821045, "learning_rate": 2.0893770219493347e-09, "loss": 0.544, "step": 1437 }, { "epoch": 2.9751724137931035, "grad_norm": 0.6575037240982056, "learning_rate": 1.755676612588686e-09, "loss": 0.5459, "step": 1438 }, { "epoch": 2.9772413793103447, "grad_norm": 0.6507884860038757, "learning_rate": 1.450987145439342e-09, "loss": 0.5337, "step": 1439 }, { "epoch": 2.979310344827586, "grad_norm": 0.6674343347549438, "learning_rate": 1.1753103889883267e-09, "loss": 0.5298, "step": 1440 }, { "epoch": 2.9813793103448276, "grad_norm": 0.6635849475860596, "learning_rate": 9.286479433257e-10, "loss": 0.5505, "step": 1441 }, { "epoch": 2.983448275862069, "grad_norm": 0.6534397006034851, "learning_rate": 7.110012401362332e-10, "loss": 0.5287, "step": 1442 }, { "epoch": 2.98551724137931, "grad_norm": 0.6629782319068909, "learning_rate": 5.2237154268997e-10, "loss": 0.5412, "step": 1443 }, { "epoch": 2.9875862068965517, "grad_norm": 0.628498375415802, "learning_rate": 3.627599458377873e-10, "loss": 0.5274, "step": 1444 }, { "epoch": 2.989655172413793, "grad_norm": 0.6589900851249695, "learning_rate": 2.321673760002918e-10, "loss": 0.5285, "step": 1445 }, { "epoch": 2.9917241379310346, "grad_norm": 0.6807800531387329, "learning_rate": 1.305945911672657e-10, "loss": 0.5263, "step": 1446 }, { "epoch": 2.993793103448276, "grad_norm": 0.6564441919326782, "learning_rate": 5.804218088933944e-11, "loss": 0.5142, "step": 1447 }, { "epoch": 2.995862068965517, "grad_norm": 0.6432551145553589, "learning_rate": 1.4510566277992077e-11, "loss": 0.5098, "step": 1448 }, { "epoch": 2.9979310344827588, "grad_norm": 0.6673936247825623, "learning_rate": 0.0, "loss": 0.5185, "step": 1449 }, { "epoch": 2.9979310344827588, "step": 1449, "total_flos": 7.092133536927252e+17, "train_loss": 0.6742002179780608, "train_runtime": 25544.3965, "train_samples_per_second": 5.449, "train_steps_per_second": 0.057 } ], "logging_steps": 1, "max_steps": 1449, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.092133536927252e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }