{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 939, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003194888178913738, "grad_norm": 6.409920785382119, "learning_rate": 2.1276595744680852e-07, "loss": 0.9383, "step": 1 }, { "epoch": 0.006389776357827476, "grad_norm": 5.882817659413674, "learning_rate": 4.2553191489361704e-07, "loss": 0.8501, "step": 2 }, { "epoch": 0.009584664536741214, "grad_norm": 5.949158200484977, "learning_rate": 6.382978723404255e-07, "loss": 0.8872, "step": 3 }, { "epoch": 0.012779552715654952, "grad_norm": 5.705459169654703, "learning_rate": 8.510638297872341e-07, "loss": 0.8123, "step": 4 }, { "epoch": 0.01597444089456869, "grad_norm": 6.165398865418825, "learning_rate": 1.0638297872340427e-06, "loss": 0.8867, "step": 5 }, { "epoch": 0.019169329073482427, "grad_norm": 5.901621861739117, "learning_rate": 1.276595744680851e-06, "loss": 0.848, "step": 6 }, { "epoch": 0.022364217252396165, "grad_norm": 5.729605975390999, "learning_rate": 1.4893617021276596e-06, "loss": 0.8766, "step": 7 }, { "epoch": 0.025559105431309903, "grad_norm": 4.7109637488813965, "learning_rate": 1.7021276595744682e-06, "loss": 0.8249, "step": 8 }, { "epoch": 0.02875399361022364, "grad_norm": 4.440852197395593, "learning_rate": 1.9148936170212767e-06, "loss": 0.8056, "step": 9 }, { "epoch": 0.03194888178913738, "grad_norm": 4.068255335955712, "learning_rate": 2.1276595744680853e-06, "loss": 0.7646, "step": 10 }, { "epoch": 0.03514376996805112, "grad_norm": 2.9662994476369344, "learning_rate": 2.340425531914894e-06, "loss": 0.7915, "step": 11 }, { "epoch": 0.038338658146964855, "grad_norm": 2.5528675082475676, "learning_rate": 2.553191489361702e-06, "loss": 0.8249, "step": 12 }, { "epoch": 0.04153354632587859, "grad_norm": 2.442136937760188, "learning_rate": 2.765957446808511e-06, "loss": 0.767, "step": 13 }, { "epoch": 0.04472843450479233, "grad_norm": 3.009485060078495, "learning_rate": 2.978723404255319e-06, "loss": 0.7684, "step": 14 }, { "epoch": 0.04792332268370607, "grad_norm": 3.530505323803651, "learning_rate": 3.191489361702128e-06, "loss": 0.7301, "step": 15 }, { "epoch": 0.051118210862619806, "grad_norm": 3.5912586753687012, "learning_rate": 3.4042553191489363e-06, "loss": 0.7483, "step": 16 }, { "epoch": 0.054313099041533544, "grad_norm": 3.242695851912524, "learning_rate": 3.6170212765957453e-06, "loss": 0.7003, "step": 17 }, { "epoch": 0.05750798722044728, "grad_norm": 3.1020257848632498, "learning_rate": 3.8297872340425535e-06, "loss": 0.7508, "step": 18 }, { "epoch": 0.06070287539936102, "grad_norm": 3.1595177932395546, "learning_rate": 4.042553191489362e-06, "loss": 0.7397, "step": 19 }, { "epoch": 0.06389776357827476, "grad_norm": 2.2887672112178294, "learning_rate": 4.255319148936171e-06, "loss": 0.669, "step": 20 }, { "epoch": 0.0670926517571885, "grad_norm": 1.8739235547973887, "learning_rate": 4.468085106382979e-06, "loss": 0.6473, "step": 21 }, { "epoch": 0.07028753993610223, "grad_norm": 1.6043319479062044, "learning_rate": 4.680851063829788e-06, "loss": 0.5916, "step": 22 }, { "epoch": 0.07348242811501597, "grad_norm": 1.4903043423735904, "learning_rate": 4.893617021276596e-06, "loss": 0.7035, "step": 23 }, { "epoch": 0.07667731629392971, "grad_norm": 1.5284966886424023, "learning_rate": 5.106382978723404e-06, "loss": 0.6841, "step": 24 }, { "epoch": 0.07987220447284345, "grad_norm": 1.7400544873250572, "learning_rate": 5.319148936170213e-06, "loss": 0.608, "step": 25 }, { "epoch": 0.08306709265175719, "grad_norm": 1.823355346961419, "learning_rate": 5.531914893617022e-06, "loss": 0.6812, "step": 26 }, { "epoch": 0.08626198083067092, "grad_norm": 1.5434873639853426, "learning_rate": 5.744680851063831e-06, "loss": 0.6243, "step": 27 }, { "epoch": 0.08945686900958466, "grad_norm": 1.4366139254965609, "learning_rate": 5.957446808510638e-06, "loss": 0.6559, "step": 28 }, { "epoch": 0.0926517571884984, "grad_norm": 1.2885186500127617, "learning_rate": 6.170212765957447e-06, "loss": 0.6615, "step": 29 }, { "epoch": 0.09584664536741214, "grad_norm": 1.24950215759162, "learning_rate": 6.382978723404256e-06, "loss": 0.6502, "step": 30 }, { "epoch": 0.09904153354632587, "grad_norm": 1.3498220957297518, "learning_rate": 6.595744680851064e-06, "loss": 0.6197, "step": 31 }, { "epoch": 0.10223642172523961, "grad_norm": 1.3715708826482595, "learning_rate": 6.808510638297873e-06, "loss": 0.648, "step": 32 }, { "epoch": 0.10543130990415335, "grad_norm": 1.2724124105045853, "learning_rate": 7.021276595744682e-06, "loss": 0.5946, "step": 33 }, { "epoch": 0.10862619808306709, "grad_norm": 1.0470319859101942, "learning_rate": 7.234042553191491e-06, "loss": 0.6183, "step": 34 }, { "epoch": 0.11182108626198083, "grad_norm": 1.0168611398932235, "learning_rate": 7.446808510638298e-06, "loss": 0.6104, "step": 35 }, { "epoch": 0.11501597444089456, "grad_norm": 1.0363638467793455, "learning_rate": 7.659574468085107e-06, "loss": 0.5497, "step": 36 }, { "epoch": 0.1182108626198083, "grad_norm": 1.0065437324012412, "learning_rate": 7.872340425531916e-06, "loss": 0.6082, "step": 37 }, { "epoch": 0.12140575079872204, "grad_norm": 0.9397900918794712, "learning_rate": 8.085106382978723e-06, "loss": 0.607, "step": 38 }, { "epoch": 0.12460063897763578, "grad_norm": 0.8971589534820209, "learning_rate": 8.297872340425532e-06, "loss": 0.5706, "step": 39 }, { "epoch": 0.12779552715654952, "grad_norm": 0.9733107673169904, "learning_rate": 8.510638297872341e-06, "loss": 0.5445, "step": 40 }, { "epoch": 0.13099041533546327, "grad_norm": 1.0242616537914584, "learning_rate": 8.72340425531915e-06, "loss": 0.5692, "step": 41 }, { "epoch": 0.134185303514377, "grad_norm": 0.9024153216202099, "learning_rate": 8.936170212765958e-06, "loss": 0.5916, "step": 42 }, { "epoch": 0.13738019169329074, "grad_norm": 0.9635264772129024, "learning_rate": 9.148936170212767e-06, "loss": 0.5842, "step": 43 }, { "epoch": 0.14057507987220447, "grad_norm": 1.0056120995653648, "learning_rate": 9.361702127659576e-06, "loss": 0.5588, "step": 44 }, { "epoch": 0.14376996805111822, "grad_norm": 0.9275308523019801, "learning_rate": 9.574468085106385e-06, "loss": 0.5985, "step": 45 }, { "epoch": 0.14696485623003194, "grad_norm": 1.0395462293521542, "learning_rate": 9.787234042553192e-06, "loss": 0.5751, "step": 46 }, { "epoch": 0.1501597444089457, "grad_norm": 0.8523118843095494, "learning_rate": 1e-05, "loss": 0.6098, "step": 47 }, { "epoch": 0.15335463258785942, "grad_norm": 0.8942183665281777, "learning_rate": 1.0212765957446808e-05, "loss": 0.5898, "step": 48 }, { "epoch": 0.15654952076677317, "grad_norm": 0.8818105787041829, "learning_rate": 1.0425531914893619e-05, "loss": 0.5829, "step": 49 }, { "epoch": 0.1597444089456869, "grad_norm": 1.0300441845112693, "learning_rate": 1.0638297872340426e-05, "loss": 0.5514, "step": 50 }, { "epoch": 0.16293929712460065, "grad_norm": 1.0423440565604452, "learning_rate": 1.0851063829787233e-05, "loss": 0.593, "step": 51 }, { "epoch": 0.16613418530351437, "grad_norm": 0.949481155031613, "learning_rate": 1.1063829787234044e-05, "loss": 0.6092, "step": 52 }, { "epoch": 0.16932907348242812, "grad_norm": 0.9369849481377671, "learning_rate": 1.1276595744680851e-05, "loss": 0.5577, "step": 53 }, { "epoch": 0.17252396166134185, "grad_norm": 0.9618399687304353, "learning_rate": 1.1489361702127662e-05, "loss": 0.551, "step": 54 }, { "epoch": 0.1757188498402556, "grad_norm": 0.9655092891280169, "learning_rate": 1.170212765957447e-05, "loss": 0.5664, "step": 55 }, { "epoch": 0.17891373801916932, "grad_norm": 1.0520390138780007, "learning_rate": 1.1914893617021277e-05, "loss": 0.5341, "step": 56 }, { "epoch": 0.18210862619808307, "grad_norm": 0.969457321374093, "learning_rate": 1.2127659574468087e-05, "loss": 0.5643, "step": 57 }, { "epoch": 0.1853035143769968, "grad_norm": 0.8852786189340646, "learning_rate": 1.2340425531914895e-05, "loss": 0.543, "step": 58 }, { "epoch": 0.18849840255591055, "grad_norm": 1.1488496953158756, "learning_rate": 1.2553191489361702e-05, "loss": 0.5653, "step": 59 }, { "epoch": 0.19169329073482427, "grad_norm": 1.1296682096329418, "learning_rate": 1.2765957446808513e-05, "loss": 0.5565, "step": 60 }, { "epoch": 0.19488817891373802, "grad_norm": 0.8854908232020008, "learning_rate": 1.297872340425532e-05, "loss": 0.5551, "step": 61 }, { "epoch": 0.19808306709265175, "grad_norm": 0.8324763984138673, "learning_rate": 1.3191489361702127e-05, "loss": 0.5148, "step": 62 }, { "epoch": 0.2012779552715655, "grad_norm": 1.1990838238795918, "learning_rate": 1.3404255319148938e-05, "loss": 0.5541, "step": 63 }, { "epoch": 0.20447284345047922, "grad_norm": 0.9277279636520797, "learning_rate": 1.3617021276595745e-05, "loss": 0.5484, "step": 64 }, { "epoch": 0.20766773162939298, "grad_norm": 1.0648177480107957, "learning_rate": 1.3829787234042556e-05, "loss": 0.5834, "step": 65 }, { "epoch": 0.2108626198083067, "grad_norm": 1.1696287149009121, "learning_rate": 1.4042553191489363e-05, "loss": 0.5202, "step": 66 }, { "epoch": 0.21405750798722045, "grad_norm": 1.0169526918813456, "learning_rate": 1.425531914893617e-05, "loss": 0.5276, "step": 67 }, { "epoch": 0.21725239616613418, "grad_norm": 1.0568993802356397, "learning_rate": 1.4468085106382981e-05, "loss": 0.5184, "step": 68 }, { "epoch": 0.22044728434504793, "grad_norm": 1.1725904166905932, "learning_rate": 1.4680851063829789e-05, "loss": 0.5715, "step": 69 }, { "epoch": 0.22364217252396165, "grad_norm": 1.0117321152787573, "learning_rate": 1.4893617021276596e-05, "loss": 0.5911, "step": 70 }, { "epoch": 0.2268370607028754, "grad_norm": 0.8952457851305932, "learning_rate": 1.5106382978723407e-05, "loss": 0.5053, "step": 71 }, { "epoch": 0.23003194888178913, "grad_norm": 1.0847433664961181, "learning_rate": 1.5319148936170214e-05, "loss": 0.5301, "step": 72 }, { "epoch": 0.23322683706070288, "grad_norm": 0.9381541668113672, "learning_rate": 1.5531914893617023e-05, "loss": 0.5325, "step": 73 }, { "epoch": 0.2364217252396166, "grad_norm": 0.9870753843267154, "learning_rate": 1.5744680851063832e-05, "loss": 0.5132, "step": 74 }, { "epoch": 0.23961661341853036, "grad_norm": 0.8860262179526839, "learning_rate": 1.595744680851064e-05, "loss": 0.5175, "step": 75 }, { "epoch": 0.24281150159744408, "grad_norm": 0.838264095104911, "learning_rate": 1.6170212765957446e-05, "loss": 0.5411, "step": 76 }, { "epoch": 0.24600638977635783, "grad_norm": 0.9067720513160165, "learning_rate": 1.6382978723404255e-05, "loss": 0.4906, "step": 77 }, { "epoch": 0.24920127795527156, "grad_norm": 0.9230169949949911, "learning_rate": 1.6595744680851064e-05, "loss": 0.5305, "step": 78 }, { "epoch": 0.2523961661341853, "grad_norm": 1.0482361365213537, "learning_rate": 1.6808510638297873e-05, "loss": 0.5442, "step": 79 }, { "epoch": 0.25559105431309903, "grad_norm": 0.8237515236418056, "learning_rate": 1.7021276595744682e-05, "loss": 0.5029, "step": 80 }, { "epoch": 0.25878594249201275, "grad_norm": 0.8569463159961227, "learning_rate": 1.723404255319149e-05, "loss": 0.5536, "step": 81 }, { "epoch": 0.26198083067092653, "grad_norm": 0.9656094274228123, "learning_rate": 1.74468085106383e-05, "loss": 0.5666, "step": 82 }, { "epoch": 0.26517571884984026, "grad_norm": 0.879784737531998, "learning_rate": 1.765957446808511e-05, "loss": 0.5171, "step": 83 }, { "epoch": 0.268370607028754, "grad_norm": 0.8745385454305299, "learning_rate": 1.7872340425531915e-05, "loss": 0.4754, "step": 84 }, { "epoch": 0.2715654952076677, "grad_norm": 0.8610209256691467, "learning_rate": 1.8085106382978724e-05, "loss": 0.5713, "step": 85 }, { "epoch": 0.2747603833865815, "grad_norm": 1.0200117897200043, "learning_rate": 1.8297872340425533e-05, "loss": 0.549, "step": 86 }, { "epoch": 0.2779552715654952, "grad_norm": 1.0381714037531296, "learning_rate": 1.8510638297872342e-05, "loss": 0.5522, "step": 87 }, { "epoch": 0.28115015974440893, "grad_norm": 0.9237569605348899, "learning_rate": 1.872340425531915e-05, "loss": 0.5825, "step": 88 }, { "epoch": 0.28434504792332266, "grad_norm": 1.0109855714809834, "learning_rate": 1.893617021276596e-05, "loss": 0.4851, "step": 89 }, { "epoch": 0.28753993610223644, "grad_norm": 0.9947036818367249, "learning_rate": 1.914893617021277e-05, "loss": 0.5195, "step": 90 }, { "epoch": 0.29073482428115016, "grad_norm": 1.188164995511023, "learning_rate": 1.9361702127659575e-05, "loss": 0.5328, "step": 91 }, { "epoch": 0.2939297124600639, "grad_norm": 0.9985112774386741, "learning_rate": 1.9574468085106384e-05, "loss": 0.4985, "step": 92 }, { "epoch": 0.2971246006389776, "grad_norm": 1.147610921973416, "learning_rate": 1.9787234042553193e-05, "loss": 0.5353, "step": 93 }, { "epoch": 0.3003194888178914, "grad_norm": 1.2244624011374305, "learning_rate": 2e-05, "loss": 0.5853, "step": 94 }, { "epoch": 0.3035143769968051, "grad_norm": 1.0237954246117738, "learning_rate": 1.9999930887622756e-05, "loss": 0.5223, "step": 95 }, { "epoch": 0.30670926517571884, "grad_norm": 0.9347818625754574, "learning_rate": 1.9999723551446323e-05, "loss": 0.5499, "step": 96 }, { "epoch": 0.30990415335463256, "grad_norm": 0.9931906754170975, "learning_rate": 1.9999377994336602e-05, "loss": 0.5153, "step": 97 }, { "epoch": 0.31309904153354634, "grad_norm": 1.100664016646544, "learning_rate": 1.9998894221070052e-05, "loss": 0.5421, "step": 98 }, { "epoch": 0.31629392971246006, "grad_norm": 1.072898943505642, "learning_rate": 1.9998272238333606e-05, "loss": 0.522, "step": 99 }, { "epoch": 0.3194888178913738, "grad_norm": 0.8788562431890847, "learning_rate": 1.9997512054724616e-05, "loss": 0.5768, "step": 100 }, { "epoch": 0.3226837060702875, "grad_norm": 1.005114294076789, "learning_rate": 1.99966136807507e-05, "loss": 0.5366, "step": 101 }, { "epoch": 0.3258785942492013, "grad_norm": 0.9818208632299064, "learning_rate": 1.99955771288296e-05, "loss": 0.5629, "step": 102 }, { "epoch": 0.329073482428115, "grad_norm": 0.9550602655061554, "learning_rate": 1.9994402413289044e-05, "loss": 0.5057, "step": 103 }, { "epoch": 0.33226837060702874, "grad_norm": 0.8574474170220133, "learning_rate": 1.99930895503665e-05, "loss": 0.5436, "step": 104 }, { "epoch": 0.3354632587859425, "grad_norm": 0.8630661623280808, "learning_rate": 1.9991638558208987e-05, "loss": 0.5043, "step": 105 }, { "epoch": 0.33865814696485624, "grad_norm": 0.9861864624707406, "learning_rate": 1.9990049456872805e-05, "loss": 0.4998, "step": 106 }, { "epoch": 0.34185303514376997, "grad_norm": 1.1731144874278645, "learning_rate": 1.998832226832327e-05, "loss": 0.5621, "step": 107 }, { "epoch": 0.3450479233226837, "grad_norm": 0.9198118564044715, "learning_rate": 1.9986457016434402e-05, "loss": 0.5209, "step": 108 }, { "epoch": 0.34824281150159747, "grad_norm": 0.9818139368666934, "learning_rate": 1.99844537269886e-05, "loss": 0.5304, "step": 109 }, { "epoch": 0.3514376996805112, "grad_norm": 0.871284930323514, "learning_rate": 1.9982312427676286e-05, "loss": 0.5295, "step": 110 }, { "epoch": 0.3546325878594249, "grad_norm": 1.1442536730263613, "learning_rate": 1.9980033148095514e-05, "loss": 0.6233, "step": 111 }, { "epoch": 0.35782747603833864, "grad_norm": 1.0303042499178714, "learning_rate": 1.997761591975157e-05, "loss": 0.5829, "step": 112 }, { "epoch": 0.3610223642172524, "grad_norm": 0.9699824744336917, "learning_rate": 1.9975060776056538e-05, "loss": 0.5688, "step": 113 }, { "epoch": 0.36421725239616615, "grad_norm": 0.9059578869436408, "learning_rate": 1.9972367752328824e-05, "loss": 0.6128, "step": 114 }, { "epoch": 0.36741214057507987, "grad_norm": 1.0030044320881064, "learning_rate": 1.9969536885792685e-05, "loss": 0.491, "step": 115 }, { "epoch": 0.3706070287539936, "grad_norm": 0.880432268699565, "learning_rate": 1.9966568215577704e-05, "loss": 0.525, "step": 116 }, { "epoch": 0.3738019169329074, "grad_norm": 0.9288360039901852, "learning_rate": 1.9963461782718247e-05, "loss": 0.4981, "step": 117 }, { "epoch": 0.3769968051118211, "grad_norm": 0.8685285833946184, "learning_rate": 1.996021763015291e-05, "loss": 0.5174, "step": 118 }, { "epoch": 0.3801916932907348, "grad_norm": 0.9456704365811888, "learning_rate": 1.9956835802723916e-05, "loss": 0.5116, "step": 119 }, { "epoch": 0.38338658146964855, "grad_norm": 0.8672982876474767, "learning_rate": 1.995331634717649e-05, "loss": 0.5191, "step": 120 }, { "epoch": 0.3865814696485623, "grad_norm": 1.2439896447102605, "learning_rate": 1.9949659312158212e-05, "loss": 0.5418, "step": 121 }, { "epoch": 0.38977635782747605, "grad_norm": 1.1971242384225746, "learning_rate": 1.9945864748218368e-05, "loss": 0.5135, "step": 122 }, { "epoch": 0.3929712460063898, "grad_norm": 0.8834188032057455, "learning_rate": 1.994193270780722e-05, "loss": 0.555, "step": 123 }, { "epoch": 0.3961661341853035, "grad_norm": 0.9505972320015716, "learning_rate": 1.9937863245275303e-05, "loss": 0.5506, "step": 124 }, { "epoch": 0.3993610223642173, "grad_norm": 1.2212739220097744, "learning_rate": 1.9933656416872664e-05, "loss": 0.5404, "step": 125 }, { "epoch": 0.402555910543131, "grad_norm": 0.9304190644575443, "learning_rate": 1.992931228074808e-05, "loss": 0.6051, "step": 126 }, { "epoch": 0.4057507987220447, "grad_norm": 0.9409095688366127, "learning_rate": 1.9924830896948268e-05, "loss": 0.5515, "step": 127 }, { "epoch": 0.40894568690095845, "grad_norm": 0.8699906408491971, "learning_rate": 1.9920212327417044e-05, "loss": 0.5022, "step": 128 }, { "epoch": 0.41214057507987223, "grad_norm": 0.9946776707428178, "learning_rate": 1.991545663599448e-05, "loss": 0.5309, "step": 129 }, { "epoch": 0.41533546325878595, "grad_norm": 1.01014181570919, "learning_rate": 1.9910563888415996e-05, "loss": 0.5681, "step": 130 }, { "epoch": 0.4185303514376997, "grad_norm": 0.93629937085659, "learning_rate": 1.990553415231148e-05, "loss": 0.5576, "step": 131 }, { "epoch": 0.4217252396166134, "grad_norm": 1.1184865559025652, "learning_rate": 1.9900367497204328e-05, "loss": 0.5088, "step": 132 }, { "epoch": 0.4249201277955272, "grad_norm": 1.0388424246571961, "learning_rate": 1.9895063994510512e-05, "loss": 0.5269, "step": 133 }, { "epoch": 0.4281150159744409, "grad_norm": 0.9588973054432186, "learning_rate": 1.9889623717537564e-05, "loss": 0.5914, "step": 134 }, { "epoch": 0.43130990415335463, "grad_norm": 1.0047992755730388, "learning_rate": 1.988404674148358e-05, "loss": 0.5335, "step": 135 }, { "epoch": 0.43450479233226835, "grad_norm": 0.9513706884109058, "learning_rate": 1.9878333143436172e-05, "loss": 0.5252, "step": 136 }, { "epoch": 0.43769968051118213, "grad_norm": 0.9491047212617997, "learning_rate": 1.987248300237141e-05, "loss": 0.5289, "step": 137 }, { "epoch": 0.44089456869009586, "grad_norm": 1.033338903728736, "learning_rate": 1.9866496399152732e-05, "loss": 0.548, "step": 138 }, { "epoch": 0.4440894568690096, "grad_norm": 0.981330967452213, "learning_rate": 1.9860373416529804e-05, "loss": 0.5533, "step": 139 }, { "epoch": 0.4472843450479233, "grad_norm": 0.9771678046126266, "learning_rate": 1.9854114139137408e-05, "loss": 0.5165, "step": 140 }, { "epoch": 0.4504792332268371, "grad_norm": 0.9318509073079462, "learning_rate": 1.984771865349425e-05, "loss": 0.5897, "step": 141 }, { "epoch": 0.4536741214057508, "grad_norm": 0.9749289154723085, "learning_rate": 1.9841187048001774e-05, "loss": 0.5209, "step": 142 }, { "epoch": 0.45686900958466453, "grad_norm": 0.9585929892181612, "learning_rate": 1.9834519412942938e-05, "loss": 0.5318, "step": 143 }, { "epoch": 0.46006389776357826, "grad_norm": 0.9990483753288364, "learning_rate": 1.9827715840480962e-05, "loss": 0.5166, "step": 144 }, { "epoch": 0.46325878594249204, "grad_norm": 1.0335238778784028, "learning_rate": 1.982077642465806e-05, "loss": 0.5423, "step": 145 }, { "epoch": 0.46645367412140576, "grad_norm": 1.0867942922933842, "learning_rate": 1.9813701261394136e-05, "loss": 0.525, "step": 146 }, { "epoch": 0.4696485623003195, "grad_norm": 1.116159290481366, "learning_rate": 1.9806490448485463e-05, "loss": 0.5217, "step": 147 }, { "epoch": 0.4728434504792332, "grad_norm": 1.0276431277812408, "learning_rate": 1.9799144085603323e-05, "loss": 0.5613, "step": 148 }, { "epoch": 0.476038338658147, "grad_norm": 0.9607572461751404, "learning_rate": 1.9791662274292638e-05, "loss": 0.5696, "step": 149 }, { "epoch": 0.4792332268370607, "grad_norm": 0.7743742651857493, "learning_rate": 1.978404511797056e-05, "loss": 0.4712, "step": 150 }, { "epoch": 0.48242811501597443, "grad_norm": 0.9379781740009565, "learning_rate": 1.9776292721925046e-05, "loss": 0.5456, "step": 151 }, { "epoch": 0.48562300319488816, "grad_norm": 0.9500162315992806, "learning_rate": 1.97684051933134e-05, "loss": 0.5093, "step": 152 }, { "epoch": 0.48881789137380194, "grad_norm": 0.9463427069451845, "learning_rate": 1.9760382641160792e-05, "loss": 0.523, "step": 153 }, { "epoch": 0.49201277955271566, "grad_norm": 0.964902547990394, "learning_rate": 1.9752225176358757e-05, "loss": 0.5154, "step": 154 }, { "epoch": 0.4952076677316294, "grad_norm": 0.9266397117812072, "learning_rate": 1.974393291166364e-05, "loss": 0.5242, "step": 155 }, { "epoch": 0.4984025559105431, "grad_norm": 1.0373423802821105, "learning_rate": 1.973550596169508e-05, "loss": 0.5526, "step": 156 }, { "epoch": 0.5015974440894568, "grad_norm": 0.8331184703908079, "learning_rate": 1.9726944442934378e-05, "loss": 0.4775, "step": 157 }, { "epoch": 0.5047923322683706, "grad_norm": 0.8424437954994451, "learning_rate": 1.971824847372292e-05, "loss": 0.4946, "step": 158 }, { "epoch": 0.5079872204472844, "grad_norm": 0.8924532125572432, "learning_rate": 1.9709418174260523e-05, "loss": 0.5213, "step": 159 }, { "epoch": 0.5111821086261981, "grad_norm": 1.0238137951842978, "learning_rate": 1.9700453666603786e-05, "loss": 0.504, "step": 160 }, { "epoch": 0.5143769968051118, "grad_norm": 0.8696826053287715, "learning_rate": 1.96913550746644e-05, "loss": 0.5285, "step": 161 }, { "epoch": 0.5175718849840255, "grad_norm": 1.1419625361492167, "learning_rate": 1.9682122524207426e-05, "loss": 0.5258, "step": 162 }, { "epoch": 0.5207667731629393, "grad_norm": 1.0717562011005641, "learning_rate": 1.9672756142849564e-05, "loss": 0.5622, "step": 163 }, { "epoch": 0.5239616613418531, "grad_norm": 1.082362120758369, "learning_rate": 1.9663256060057395e-05, "loss": 0.5459, "step": 164 }, { "epoch": 0.5271565495207667, "grad_norm": 0.9058526491846522, "learning_rate": 1.965362240714557e-05, "loss": 0.535, "step": 165 }, { "epoch": 0.5303514376996805, "grad_norm": 1.0636353754052472, "learning_rate": 1.9643855317275036e-05, "loss": 0.5187, "step": 166 }, { "epoch": 0.5335463258785943, "grad_norm": 0.8841768134611658, "learning_rate": 1.963395492545114e-05, "loss": 0.4926, "step": 167 }, { "epoch": 0.536741214057508, "grad_norm": 0.8454615826601668, "learning_rate": 1.9623921368521814e-05, "loss": 0.4925, "step": 168 }, { "epoch": 0.5399361022364217, "grad_norm": 0.8525089279579146, "learning_rate": 1.961375478517564e-05, "loss": 0.554, "step": 169 }, { "epoch": 0.5431309904153354, "grad_norm": 0.9507938815359402, "learning_rate": 1.9603455315939976e-05, "loss": 0.5381, "step": 170 }, { "epoch": 0.5463258785942492, "grad_norm": 0.8663401759485011, "learning_rate": 1.9593023103178984e-05, "loss": 0.5398, "step": 171 }, { "epoch": 0.549520766773163, "grad_norm": 0.9927086656087859, "learning_rate": 1.9582458291091664e-05, "loss": 0.5363, "step": 172 }, { "epoch": 0.5527156549520766, "grad_norm": 0.8142362736099577, "learning_rate": 1.9571761025709874e-05, "loss": 0.5441, "step": 173 }, { "epoch": 0.5559105431309904, "grad_norm": 0.9363468107783036, "learning_rate": 1.95609314548963e-05, "loss": 0.5771, "step": 174 }, { "epoch": 0.5591054313099042, "grad_norm": 0.7489334767030862, "learning_rate": 1.954996972834242e-05, "loss": 0.5208, "step": 175 }, { "epoch": 0.5623003194888179, "grad_norm": 0.934302740617694, "learning_rate": 1.953887599756643e-05, "loss": 0.5107, "step": 176 }, { "epoch": 0.5654952076677316, "grad_norm": 0.8012783490329485, "learning_rate": 1.9527650415911154e-05, "loss": 0.5377, "step": 177 }, { "epoch": 0.5686900958466453, "grad_norm": 0.9181068406934503, "learning_rate": 1.9516293138541914e-05, "loss": 0.5111, "step": 178 }, { "epoch": 0.5718849840255591, "grad_norm": 0.8305500501619189, "learning_rate": 1.95048043224444e-05, "loss": 0.5088, "step": 179 }, { "epoch": 0.5750798722044729, "grad_norm": 1.066440074003196, "learning_rate": 1.9493184126422494e-05, "loss": 0.5453, "step": 180 }, { "epoch": 0.5782747603833865, "grad_norm": 1.107675123138524, "learning_rate": 1.948143271109606e-05, "loss": 0.5425, "step": 181 }, { "epoch": 0.5814696485623003, "grad_norm": 0.8978276499968547, "learning_rate": 1.946955023889876e-05, "loss": 0.527, "step": 182 }, { "epoch": 0.5846645367412141, "grad_norm": 0.7949008147878583, "learning_rate": 1.9457536874075765e-05, "loss": 0.5247, "step": 183 }, { "epoch": 0.5878594249201278, "grad_norm": 0.974636844868348, "learning_rate": 1.9445392782681523e-05, "loss": 0.5075, "step": 184 }, { "epoch": 0.5910543130990416, "grad_norm": 0.9163909881555694, "learning_rate": 1.9433118132577432e-05, "loss": 0.5242, "step": 185 }, { "epoch": 0.5942492012779552, "grad_norm": 0.9615667703817601, "learning_rate": 1.9420713093429548e-05, "loss": 0.4878, "step": 186 }, { "epoch": 0.597444089456869, "grad_norm": 0.8959730227876819, "learning_rate": 1.9408177836706215e-05, "loss": 0.5006, "step": 187 }, { "epoch": 0.6006389776357828, "grad_norm": 0.9791864357100734, "learning_rate": 1.939551253567571e-05, "loss": 0.508, "step": 188 }, { "epoch": 0.6038338658146964, "grad_norm": 0.8656035894165334, "learning_rate": 1.9382717365403854e-05, "loss": 0.4921, "step": 189 }, { "epoch": 0.6070287539936102, "grad_norm": 1.0378802554109634, "learning_rate": 1.9369792502751568e-05, "loss": 0.5269, "step": 190 }, { "epoch": 0.610223642172524, "grad_norm": 0.843711222780129, "learning_rate": 1.9356738126372448e-05, "loss": 0.4818, "step": 191 }, { "epoch": 0.6134185303514377, "grad_norm": 0.9506506682319595, "learning_rate": 1.9343554416710293e-05, "loss": 0.5222, "step": 192 }, { "epoch": 0.6166134185303515, "grad_norm": 0.9158618299618471, "learning_rate": 1.9330241555996606e-05, "loss": 0.5458, "step": 193 }, { "epoch": 0.6198083067092651, "grad_norm": 0.9450209037071783, "learning_rate": 1.9316799728248074e-05, "loss": 0.5328, "step": 194 }, { "epoch": 0.6230031948881789, "grad_norm": 0.9060814081186569, "learning_rate": 1.930322911926404e-05, "loss": 0.5213, "step": 195 }, { "epoch": 0.6261980830670927, "grad_norm": 0.9414520961878969, "learning_rate": 1.92895299166239e-05, "loss": 0.5231, "step": 196 }, { "epoch": 0.6293929712460063, "grad_norm": 0.7921738327634565, "learning_rate": 1.9275702309684557e-05, "loss": 0.5229, "step": 197 }, { "epoch": 0.6325878594249201, "grad_norm": 0.8314517204939224, "learning_rate": 1.9261746489577767e-05, "loss": 0.4838, "step": 198 }, { "epoch": 0.6357827476038339, "grad_norm": 0.849639036650939, "learning_rate": 1.924766264920751e-05, "loss": 0.4933, "step": 199 }, { "epoch": 0.6389776357827476, "grad_norm": 0.847794453627695, "learning_rate": 1.923345098324732e-05, "loss": 0.4876, "step": 200 }, { "epoch": 0.6421725239616614, "grad_norm": 0.9533248917368539, "learning_rate": 1.92191116881376e-05, "loss": 0.527, "step": 201 }, { "epoch": 0.645367412140575, "grad_norm": 0.8517039461818795, "learning_rate": 1.9204644962082915e-05, "loss": 0.54, "step": 202 }, { "epoch": 0.6485623003194888, "grad_norm": 0.9145485201406112, "learning_rate": 1.9190051005049218e-05, "loss": 0.4843, "step": 203 }, { "epoch": 0.6517571884984026, "grad_norm": 1.0820771145546035, "learning_rate": 1.917533001876113e-05, "loss": 0.5369, "step": 204 }, { "epoch": 0.6549520766773163, "grad_norm": 0.9383437313008715, "learning_rate": 1.9160482206699124e-05, "loss": 0.4931, "step": 205 }, { "epoch": 0.65814696485623, "grad_norm": 0.9268071001516084, "learning_rate": 1.9145507774096706e-05, "loss": 0.5258, "step": 206 }, { "epoch": 0.6613418530351438, "grad_norm": 1.0204489467118907, "learning_rate": 1.9130406927937616e-05, "loss": 0.4991, "step": 207 }, { "epoch": 0.6645367412140575, "grad_norm": 0.9779602496725159, "learning_rate": 1.9115179876952925e-05, "loss": 0.5631, "step": 208 }, { "epoch": 0.6677316293929713, "grad_norm": 0.8699335339360664, "learning_rate": 1.9099826831618168e-05, "loss": 0.5526, "step": 209 }, { "epoch": 0.670926517571885, "grad_norm": 0.8695595019031475, "learning_rate": 1.9084348004150442e-05, "loss": 0.4992, "step": 210 }, { "epoch": 0.6741214057507987, "grad_norm": 0.9392019689943826, "learning_rate": 1.9068743608505454e-05, "loss": 0.521, "step": 211 }, { "epoch": 0.6773162939297125, "grad_norm": 0.8143307536090024, "learning_rate": 1.9053013860374588e-05, "loss": 0.5414, "step": 212 }, { "epoch": 0.6805111821086262, "grad_norm": 0.9222591688630268, "learning_rate": 1.9037158977181894e-05, "loss": 0.5494, "step": 213 }, { "epoch": 0.6837060702875399, "grad_norm": 0.9812040866307158, "learning_rate": 1.9021179178081107e-05, "loss": 0.5621, "step": 214 }, { "epoch": 0.6869009584664537, "grad_norm": 0.8094811465097489, "learning_rate": 1.900507468395261e-05, "loss": 0.4696, "step": 215 }, { "epoch": 0.6900958466453674, "grad_norm": 0.9772505283003321, "learning_rate": 1.8988845717400375e-05, "loss": 0.5201, "step": 216 }, { "epoch": 0.6932907348242812, "grad_norm": 0.9431810275346089, "learning_rate": 1.89724925027489e-05, "loss": 0.5601, "step": 217 }, { "epoch": 0.6964856230031949, "grad_norm": 0.9561884071389894, "learning_rate": 1.8956015266040086e-05, "loss": 0.5517, "step": 218 }, { "epoch": 0.6996805111821086, "grad_norm": 1.058918826761409, "learning_rate": 1.8939414235030137e-05, "loss": 0.5498, "step": 219 }, { "epoch": 0.7028753993610224, "grad_norm": 0.953107551204602, "learning_rate": 1.8922689639186387e-05, "loss": 0.5442, "step": 220 }, { "epoch": 0.7060702875399361, "grad_norm": 1.0633099150216911, "learning_rate": 1.890584170968417e-05, "loss": 0.516, "step": 221 }, { "epoch": 0.7092651757188498, "grad_norm": 0.8612205651102657, "learning_rate": 1.888887067940356e-05, "loss": 0.5438, "step": 222 }, { "epoch": 0.7124600638977636, "grad_norm": 0.9596989186750442, "learning_rate": 1.8871776782926216e-05, "loss": 0.487, "step": 223 }, { "epoch": 0.7156549520766773, "grad_norm": 0.8744406461422123, "learning_rate": 1.8854560256532098e-05, "loss": 0.4993, "step": 224 }, { "epoch": 0.7188498402555911, "grad_norm": 0.8321794065646121, "learning_rate": 1.8837221338196228e-05, "loss": 0.5049, "step": 225 }, { "epoch": 0.7220447284345048, "grad_norm": 0.983227338821679, "learning_rate": 1.8819760267585368e-05, "loss": 0.5522, "step": 226 }, { "epoch": 0.7252396166134185, "grad_norm": 0.8335553459135192, "learning_rate": 1.880217728605474e-05, "loss": 0.5468, "step": 227 }, { "epoch": 0.7284345047923323, "grad_norm": 0.8749858950751359, "learning_rate": 1.8784472636644677e-05, "loss": 0.5146, "step": 228 }, { "epoch": 0.731629392971246, "grad_norm": 0.9530182539257116, "learning_rate": 1.8766646564077265e-05, "loss": 0.52, "step": 229 }, { "epoch": 0.7348242811501597, "grad_norm": 0.8423941851636887, "learning_rate": 1.8748699314752947e-05, "loss": 0.5076, "step": 230 }, { "epoch": 0.7380191693290735, "grad_norm": 1.0674983177020294, "learning_rate": 1.8730631136747138e-05, "loss": 0.5162, "step": 231 }, { "epoch": 0.7412140575079872, "grad_norm": 0.7547725014119157, "learning_rate": 1.871244227980679e-05, "loss": 0.5258, "step": 232 }, { "epoch": 0.744408945686901, "grad_norm": 0.8478932101355969, "learning_rate": 1.8694132995346924e-05, "loss": 0.5008, "step": 233 }, { "epoch": 0.7476038338658147, "grad_norm": 0.7976275751937845, "learning_rate": 1.8675703536447178e-05, "loss": 0.4979, "step": 234 }, { "epoch": 0.7507987220447284, "grad_norm": 0.9815318332469554, "learning_rate": 1.8657154157848302e-05, "loss": 0.5145, "step": 235 }, { "epoch": 0.7539936102236422, "grad_norm": 0.7607009059819897, "learning_rate": 1.8638485115948612e-05, "loss": 0.4847, "step": 236 }, { "epoch": 0.7571884984025559, "grad_norm": 0.99749471399467, "learning_rate": 1.8619696668800494e-05, "loss": 0.5308, "step": 237 }, { "epoch": 0.7603833865814696, "grad_norm": 0.8240339314330095, "learning_rate": 1.860078907610679e-05, "loss": 0.5046, "step": 238 }, { "epoch": 0.7635782747603834, "grad_norm": 1.1284168756143547, "learning_rate": 1.858176259921724e-05, "loss": 0.5171, "step": 239 }, { "epoch": 0.7667731629392971, "grad_norm": 0.8803422345999297, "learning_rate": 1.856261750112485e-05, "loss": 0.4918, "step": 240 }, { "epoch": 0.7699680511182109, "grad_norm": 0.982879274763817, "learning_rate": 1.8543354046462275e-05, "loss": 0.5022, "step": 241 }, { "epoch": 0.7731629392971247, "grad_norm": 0.9891187325137296, "learning_rate": 1.8523972501498136e-05, "loss": 0.48, "step": 242 }, { "epoch": 0.7763578274760383, "grad_norm": 0.9467043035719503, "learning_rate": 1.850447313413337e-05, "loss": 0.4905, "step": 243 }, { "epoch": 0.7795527156549521, "grad_norm": 1.0150262185499397, "learning_rate": 1.8484856213897496e-05, "loss": 0.5968, "step": 244 }, { "epoch": 0.7827476038338658, "grad_norm": 0.8248018852173341, "learning_rate": 1.846512201194492e-05, "loss": 0.5219, "step": 245 }, { "epoch": 0.7859424920127795, "grad_norm": 0.8832290166642166, "learning_rate": 1.844527080105116e-05, "loss": 0.5187, "step": 246 }, { "epoch": 0.7891373801916933, "grad_norm": 1.009035400264542, "learning_rate": 1.8425302855609088e-05, "loss": 0.5627, "step": 247 }, { "epoch": 0.792332268370607, "grad_norm": 0.8594597717171026, "learning_rate": 1.8405218451625144e-05, "loss": 0.514, "step": 248 }, { "epoch": 0.7955271565495208, "grad_norm": 0.7078775094856737, "learning_rate": 1.8385017866715507e-05, "loss": 0.4998, "step": 249 }, { "epoch": 0.7987220447284346, "grad_norm": 0.8371833475631262, "learning_rate": 1.8364701380102267e-05, "loss": 0.5127, "step": 250 }, { "epoch": 0.8019169329073482, "grad_norm": 0.787045292407146, "learning_rate": 1.8344269272609567e-05, "loss": 0.4652, "step": 251 }, { "epoch": 0.805111821086262, "grad_norm": 0.7958452197463651, "learning_rate": 1.83237218266597e-05, "loss": 0.5054, "step": 252 }, { "epoch": 0.8083067092651757, "grad_norm": 0.8631229364890092, "learning_rate": 1.830305932626924e-05, "loss": 0.4939, "step": 253 }, { "epoch": 0.8115015974440895, "grad_norm": 0.8057179814750826, "learning_rate": 1.8282282057045087e-05, "loss": 0.4639, "step": 254 }, { "epoch": 0.8146964856230032, "grad_norm": 0.8064404110602067, "learning_rate": 1.8261390306180544e-05, "loss": 0.4844, "step": 255 }, { "epoch": 0.8178913738019169, "grad_norm": 0.7816773313940623, "learning_rate": 1.8240384362451316e-05, "loss": 0.5076, "step": 256 }, { "epoch": 0.8210862619808307, "grad_norm": 0.9362930862782485, "learning_rate": 1.8219264516211543e-05, "loss": 0.4893, "step": 257 }, { "epoch": 0.8242811501597445, "grad_norm": 0.9138184401157835, "learning_rate": 1.819803105938979e-05, "loss": 0.4813, "step": 258 }, { "epoch": 0.8274760383386581, "grad_norm": 0.9494662952500172, "learning_rate": 1.8176684285484985e-05, "loss": 0.4997, "step": 259 }, { "epoch": 0.8306709265175719, "grad_norm": 0.8364808839715582, "learning_rate": 1.815522448956239e-05, "loss": 0.5245, "step": 260 }, { "epoch": 0.8338658146964856, "grad_norm": 1.0245391927888499, "learning_rate": 1.8133651968249503e-05, "loss": 0.511, "step": 261 }, { "epoch": 0.8370607028753994, "grad_norm": 0.8892995493967196, "learning_rate": 1.8111967019731977e-05, "loss": 0.5461, "step": 262 }, { "epoch": 0.8402555910543131, "grad_norm": 0.8390501367891469, "learning_rate": 1.8090169943749477e-05, "loss": 0.533, "step": 263 }, { "epoch": 0.8434504792332268, "grad_norm": 0.9936383454596885, "learning_rate": 1.8068261041591548e-05, "loss": 0.5595, "step": 264 }, { "epoch": 0.8466453674121406, "grad_norm": 0.8843876722914896, "learning_rate": 1.8046240616093452e-05, "loss": 0.5091, "step": 265 }, { "epoch": 0.8498402555910544, "grad_norm": 0.8462745657107433, "learning_rate": 1.802410897163199e-05, "loss": 0.4691, "step": 266 }, { "epoch": 0.853035143769968, "grad_norm": 0.8659375070431975, "learning_rate": 1.800186641412126e-05, "loss": 0.495, "step": 267 }, { "epoch": 0.8562300319488818, "grad_norm": 0.9944074080731445, "learning_rate": 1.7979513251008475e-05, "loss": 0.4873, "step": 268 }, { "epoch": 0.8594249201277955, "grad_norm": 0.9728054270611681, "learning_rate": 1.7957049791269684e-05, "loss": 0.5081, "step": 269 }, { "epoch": 0.8626198083067093, "grad_norm": 0.8692424606646845, "learning_rate": 1.7934476345405508e-05, "loss": 0.5499, "step": 270 }, { "epoch": 0.865814696485623, "grad_norm": 1.1943727229261953, "learning_rate": 1.791179322543684e-05, "loss": 0.5376, "step": 271 }, { "epoch": 0.8690095846645367, "grad_norm": 1.0506722546815361, "learning_rate": 1.788900074490056e-05, "loss": 0.5016, "step": 272 }, { "epoch": 0.8722044728434505, "grad_norm": 0.9965124501430621, "learning_rate": 1.7866099218845164e-05, "loss": 0.4912, "step": 273 }, { "epoch": 0.8753993610223643, "grad_norm": 1.0256398553570618, "learning_rate": 1.7843088963826437e-05, "loss": 0.5407, "step": 274 }, { "epoch": 0.8785942492012779, "grad_norm": 0.8102238962254681, "learning_rate": 1.7819970297903056e-05, "loss": 0.5202, "step": 275 }, { "epoch": 0.8817891373801917, "grad_norm": 1.1591745272192227, "learning_rate": 1.7796743540632226e-05, "loss": 0.5356, "step": 276 }, { "epoch": 0.8849840255591054, "grad_norm": 1.0153401381137175, "learning_rate": 1.7773409013065222e-05, "loss": 0.5088, "step": 277 }, { "epoch": 0.8881789137380192, "grad_norm": 0.8212195883494724, "learning_rate": 1.7749967037742976e-05, "loss": 0.4674, "step": 278 }, { "epoch": 0.8913738019169329, "grad_norm": 1.0157100696911838, "learning_rate": 1.772641793869162e-05, "loss": 0.5141, "step": 279 }, { "epoch": 0.8945686900958466, "grad_norm": 0.9808941118590042, "learning_rate": 1.7702762041418e-05, "loss": 0.5436, "step": 280 }, { "epoch": 0.8977635782747604, "grad_norm": 0.9347484669225299, "learning_rate": 1.7678999672905167e-05, "loss": 0.5469, "step": 281 }, { "epoch": 0.9009584664536742, "grad_norm": 0.7877300604114963, "learning_rate": 1.7655131161607887e-05, "loss": 0.477, "step": 282 }, { "epoch": 0.9041533546325878, "grad_norm": 0.9228991931025834, "learning_rate": 1.7631156837448063e-05, "loss": 0.4949, "step": 283 }, { "epoch": 0.9073482428115016, "grad_norm": 0.8738629943355437, "learning_rate": 1.7607077031810204e-05, "loss": 0.4885, "step": 284 }, { "epoch": 0.9105431309904153, "grad_norm": 0.9228825501229833, "learning_rate": 1.7582892077536832e-05, "loss": 0.5402, "step": 285 }, { "epoch": 0.9137380191693291, "grad_norm": 0.7810257376041579, "learning_rate": 1.755860230892389e-05, "loss": 0.5139, "step": 286 }, { "epoch": 0.9169329073482428, "grad_norm": 0.8073649232141846, "learning_rate": 1.75342080617161e-05, "loss": 0.5388, "step": 287 }, { "epoch": 0.9201277955271565, "grad_norm": 0.865996445702188, "learning_rate": 1.750970967310235e-05, "loss": 0.5097, "step": 288 }, { "epoch": 0.9233226837060703, "grad_norm": 0.7542868702308059, "learning_rate": 1.7485107481711014e-05, "loss": 0.4671, "step": 289 }, { "epoch": 0.9265175718849841, "grad_norm": 1.0625341412401543, "learning_rate": 1.7460401827605277e-05, "loss": 0.5505, "step": 290 }, { "epoch": 0.9297124600638977, "grad_norm": 0.8122297211204726, "learning_rate": 1.7435593052278437e-05, "loss": 0.4982, "step": 291 }, { "epoch": 0.9329073482428115, "grad_norm": 0.946574361770195, "learning_rate": 1.741068149864918e-05, "loss": 0.5131, "step": 292 }, { "epoch": 0.9361022364217252, "grad_norm": 1.0835158989817582, "learning_rate": 1.7385667511056853e-05, "loss": 0.5345, "step": 293 }, { "epoch": 0.939297124600639, "grad_norm": 0.8670067651200516, "learning_rate": 1.7360551435256673e-05, "loss": 0.508, "step": 294 }, { "epoch": 0.9424920127795527, "grad_norm": 0.9387855819886006, "learning_rate": 1.733533361841499e-05, "loss": 0.5033, "step": 295 }, { "epoch": 0.9456869009584664, "grad_norm": 0.9542224381022377, "learning_rate": 1.7310014409104456e-05, "loss": 0.5782, "step": 296 }, { "epoch": 0.9488817891373802, "grad_norm": 0.87126390248402, "learning_rate": 1.728459415729922e-05, "loss": 0.4991, "step": 297 }, { "epoch": 0.952076677316294, "grad_norm": 0.8979696763884574, "learning_rate": 1.7259073214370085e-05, "loss": 0.5262, "step": 298 }, { "epoch": 0.9552715654952076, "grad_norm": 0.8097687637292601, "learning_rate": 1.7233451933079663e-05, "loss": 0.5198, "step": 299 }, { "epoch": 0.9584664536741214, "grad_norm": 0.904220596008717, "learning_rate": 1.720773066757748e-05, "loss": 0.5044, "step": 300 }, { "epoch": 0.9616613418530351, "grad_norm": 0.9163146597366306, "learning_rate": 1.7181909773395108e-05, "loss": 0.4913, "step": 301 }, { "epoch": 0.9648562300319489, "grad_norm": 0.8142143382005672, "learning_rate": 1.715598960744121e-05, "loss": 0.5032, "step": 302 }, { "epoch": 0.9680511182108626, "grad_norm": 0.8738304555552013, "learning_rate": 1.7129970527996654e-05, "loss": 0.4819, "step": 303 }, { "epoch": 0.9712460063897763, "grad_norm": 0.7808301311984981, "learning_rate": 1.7103852894709517e-05, "loss": 0.4706, "step": 304 }, { "epoch": 0.9744408945686901, "grad_norm": 0.9343948345313137, "learning_rate": 1.707763706859015e-05, "loss": 0.4925, "step": 305 }, { "epoch": 0.9776357827476039, "grad_norm": 0.8406849688228994, "learning_rate": 1.7051323412006167e-05, "loss": 0.4857, "step": 306 }, { "epoch": 0.9808306709265175, "grad_norm": 0.8030422977982409, "learning_rate": 1.7024912288677434e-05, "loss": 0.5001, "step": 307 }, { "epoch": 0.9840255591054313, "grad_norm": 0.7637836848154631, "learning_rate": 1.6998404063671064e-05, "loss": 0.5012, "step": 308 }, { "epoch": 0.987220447284345, "grad_norm": 0.8696703757631259, "learning_rate": 1.6971799103396332e-05, "loss": 0.512, "step": 309 }, { "epoch": 0.9904153354632588, "grad_norm": 0.926358777744064, "learning_rate": 1.6945097775599667e-05, "loss": 0.4864, "step": 310 }, { "epoch": 0.9936102236421726, "grad_norm": 0.7337895110591559, "learning_rate": 1.69183004493595e-05, "loss": 0.4719, "step": 311 }, { "epoch": 0.9968051118210862, "grad_norm": 0.9274672714820749, "learning_rate": 1.6891407495081228e-05, "loss": 0.5792, "step": 312 }, { "epoch": 1.0, "grad_norm": 0.8821002895164746, "learning_rate": 1.686441928449204e-05, "loss": 0.5718, "step": 313 }, { "epoch": 1.0031948881789137, "grad_norm": 0.9737433871203176, "learning_rate": 1.6837336190635824e-05, "loss": 0.4178, "step": 314 }, { "epoch": 1.0063897763578276, "grad_norm": 0.8184630714297582, "learning_rate": 1.6810158587867973e-05, "loss": 0.423, "step": 315 }, { "epoch": 1.0095846645367412, "grad_norm": 0.7010267787641967, "learning_rate": 1.6782886851850237e-05, "loss": 0.3532, "step": 316 }, { "epoch": 1.012779552715655, "grad_norm": 0.7552621709700076, "learning_rate": 1.675552135954552e-05, "loss": 0.3665, "step": 317 }, { "epoch": 1.0159744408945688, "grad_norm": 0.8245979183281297, "learning_rate": 1.672806248921266e-05, "loss": 0.3836, "step": 318 }, { "epoch": 1.0191693290734825, "grad_norm": 0.8683334721491652, "learning_rate": 1.6700510620401223e-05, "loss": 0.3768, "step": 319 }, { "epoch": 1.0223642172523961, "grad_norm": 0.9751233496086866, "learning_rate": 1.6672866133946244e-05, "loss": 0.4059, "step": 320 }, { "epoch": 1.0255591054313098, "grad_norm": 0.838948811266222, "learning_rate": 1.664512941196295e-05, "loss": 0.4057, "step": 321 }, { "epoch": 1.0287539936102237, "grad_norm": 0.9443479654525763, "learning_rate": 1.6617300837841502e-05, "loss": 0.4334, "step": 322 }, { "epoch": 1.0319488817891374, "grad_norm": 0.8658369811689361, "learning_rate": 1.6589380796241687e-05, "loss": 0.3821, "step": 323 }, { "epoch": 1.035143769968051, "grad_norm": 0.8860556738851113, "learning_rate": 1.6561369673087588e-05, "loss": 0.3899, "step": 324 }, { "epoch": 1.038338658146965, "grad_norm": 0.6682829096727095, "learning_rate": 1.653326785556227e-05, "loss": 0.3522, "step": 325 }, { "epoch": 1.0415335463258786, "grad_norm": 0.7410215144498764, "learning_rate": 1.650507573210242e-05, "loss": 0.3838, "step": 326 }, { "epoch": 1.0447284345047922, "grad_norm": 0.9609136708860833, "learning_rate": 1.6476793692392966e-05, "loss": 0.4566, "step": 327 }, { "epoch": 1.0479233226837061, "grad_norm": 0.6908849716070313, "learning_rate": 1.6448422127361707e-05, "loss": 0.3501, "step": 328 }, { "epoch": 1.0511182108626198, "grad_norm": 0.837313532223091, "learning_rate": 1.641996142917391e-05, "loss": 0.382, "step": 329 }, { "epoch": 1.0543130990415335, "grad_norm": 0.834950559417814, "learning_rate": 1.6391411991226873e-05, "loss": 0.4044, "step": 330 }, { "epoch": 1.0575079872204474, "grad_norm": 0.8560681611142513, "learning_rate": 1.6362774208144504e-05, "loss": 0.435, "step": 331 }, { "epoch": 1.060702875399361, "grad_norm": 0.8646105738148943, "learning_rate": 1.6334048475771855e-05, "loss": 0.3865, "step": 332 }, { "epoch": 1.0638977635782747, "grad_norm": 0.9206127193709772, "learning_rate": 1.6305235191169654e-05, "loss": 0.4109, "step": 333 }, { "epoch": 1.0670926517571886, "grad_norm": 0.7719035201516966, "learning_rate": 1.6276334752608823e-05, "loss": 0.356, "step": 334 }, { "epoch": 1.0702875399361023, "grad_norm": 1.1225193958448179, "learning_rate": 1.6247347559564962e-05, "loss": 0.3787, "step": 335 }, { "epoch": 1.073482428115016, "grad_norm": 0.8380968241340917, "learning_rate": 1.621827401271284e-05, "loss": 0.3143, "step": 336 }, { "epoch": 1.0766773162939298, "grad_norm": 0.8303781325962523, "learning_rate": 1.6189114513920838e-05, "loss": 0.384, "step": 337 }, { "epoch": 1.0798722044728435, "grad_norm": 1.129732310070753, "learning_rate": 1.6159869466245416e-05, "loss": 0.3807, "step": 338 }, { "epoch": 1.0830670926517572, "grad_norm": 0.7401369861082813, "learning_rate": 1.613053927392553e-05, "loss": 0.4185, "step": 339 }, { "epoch": 1.0862619808306708, "grad_norm": 0.795100265885747, "learning_rate": 1.610112434237704e-05, "loss": 0.4263, "step": 340 }, { "epoch": 1.0894568690095847, "grad_norm": 0.9130074370141869, "learning_rate": 1.6071625078187113e-05, "loss": 0.4352, "step": 341 }, { "epoch": 1.0926517571884984, "grad_norm": 0.7978114041961164, "learning_rate": 1.604204188910861e-05, "loss": 0.3718, "step": 342 }, { "epoch": 1.095846645367412, "grad_norm": 0.7906733557235865, "learning_rate": 1.601237518405443e-05, "loss": 0.4199, "step": 343 }, { "epoch": 1.099041533546326, "grad_norm": 0.7427188221038984, "learning_rate": 1.5982625373091877e-05, "loss": 0.3778, "step": 344 }, { "epoch": 1.1022364217252396, "grad_norm": 0.8721565333693361, "learning_rate": 1.5952792867436984e-05, "loss": 0.3701, "step": 345 }, { "epoch": 1.1054313099041533, "grad_norm": 0.8525717279830611, "learning_rate": 1.5922878079448827e-05, "loss": 0.4079, "step": 346 }, { "epoch": 1.1086261980830672, "grad_norm": 0.9085260822651806, "learning_rate": 1.5892881422623826e-05, "loss": 0.3416, "step": 347 }, { "epoch": 1.1118210862619808, "grad_norm": 0.844899253921564, "learning_rate": 1.5862803311590043e-05, "loss": 0.3841, "step": 348 }, { "epoch": 1.1150159744408945, "grad_norm": 0.8533360961376679, "learning_rate": 1.5832644162101417e-05, "loss": 0.4002, "step": 349 }, { "epoch": 1.1182108626198084, "grad_norm": 0.8830687311970081, "learning_rate": 1.580240439103206e-05, "loss": 0.3471, "step": 350 }, { "epoch": 1.121405750798722, "grad_norm": 0.7529407391712328, "learning_rate": 1.577208441637046e-05, "loss": 0.3819, "step": 351 }, { "epoch": 1.1246006389776357, "grad_norm": 0.926454792722682, "learning_rate": 1.5741684657213726e-05, "loss": 0.373, "step": 352 }, { "epoch": 1.1277955271565494, "grad_norm": 0.8653791591805564, "learning_rate": 1.571120553376178e-05, "loss": 0.3994, "step": 353 }, { "epoch": 1.1309904153354633, "grad_norm": 0.6366340699520703, "learning_rate": 1.568064746731156e-05, "loss": 0.3566, "step": 354 }, { "epoch": 1.134185303514377, "grad_norm": 0.8133151809880976, "learning_rate": 1.565001088025119e-05, "loss": 0.3836, "step": 355 }, { "epoch": 1.1373801916932909, "grad_norm": 0.795384552885173, "learning_rate": 1.5619296196054135e-05, "loss": 0.3711, "step": 356 }, { "epoch": 1.1405750798722045, "grad_norm": 0.7607341893849747, "learning_rate": 1.5588503839273373e-05, "loss": 0.4241, "step": 357 }, { "epoch": 1.1437699680511182, "grad_norm": 0.7704534797439848, "learning_rate": 1.5557634235535496e-05, "loss": 0.3941, "step": 358 }, { "epoch": 1.1469648562300319, "grad_norm": 0.7128626541107346, "learning_rate": 1.552668781153484e-05, "loss": 0.341, "step": 359 }, { "epoch": 1.1501597444089458, "grad_norm": 0.763750359311604, "learning_rate": 1.5495664995027594e-05, "loss": 0.3727, "step": 360 }, { "epoch": 1.1533546325878594, "grad_norm": 0.7417308830642706, "learning_rate": 1.5464566214825876e-05, "loss": 0.4052, "step": 361 }, { "epoch": 1.156549520766773, "grad_norm": 0.7993216873695975, "learning_rate": 1.543339190079182e-05, "loss": 0.3959, "step": 362 }, { "epoch": 1.159744408945687, "grad_norm": 0.7715779361907723, "learning_rate": 1.5402142483831607e-05, "loss": 0.403, "step": 363 }, { "epoch": 1.1629392971246006, "grad_norm": 0.8331596188661798, "learning_rate": 1.5370818395889536e-05, "loss": 0.3796, "step": 364 }, { "epoch": 1.1661341853035143, "grad_norm": 0.8303258308472176, "learning_rate": 1.533942006994205e-05, "loss": 0.3828, "step": 365 }, { "epoch": 1.1693290734824282, "grad_norm": 0.7308405138430362, "learning_rate": 1.530794793999173e-05, "loss": 0.3739, "step": 366 }, { "epoch": 1.1725239616613419, "grad_norm": 0.6670562893393245, "learning_rate": 1.527640244106133e-05, "loss": 0.4019, "step": 367 }, { "epoch": 1.1757188498402555, "grad_norm": 0.8948726572245147, "learning_rate": 1.5244784009187724e-05, "loss": 0.3774, "step": 368 }, { "epoch": 1.1789137380191694, "grad_norm": 0.7894907155797702, "learning_rate": 1.521309308141592e-05, "loss": 0.3486, "step": 369 }, { "epoch": 1.182108626198083, "grad_norm": 0.7556853574798593, "learning_rate": 1.518133009579298e-05, "loss": 0.3935, "step": 370 }, { "epoch": 1.1853035143769968, "grad_norm": 0.8278844793998316, "learning_rate": 1.5149495491362e-05, "loss": 0.4155, "step": 371 }, { "epoch": 1.1884984025559104, "grad_norm": 0.8208958496385201, "learning_rate": 1.5117589708156013e-05, "loss": 0.3603, "step": 372 }, { "epoch": 1.1916932907348243, "grad_norm": 0.7236738730838786, "learning_rate": 1.5085613187191926e-05, "loss": 0.3986, "step": 373 }, { "epoch": 1.194888178913738, "grad_norm": 0.8755456163986371, "learning_rate": 1.5053566370464416e-05, "loss": 0.3945, "step": 374 }, { "epoch": 1.1980830670926517, "grad_norm": 0.7911158266689877, "learning_rate": 1.502144970093982e-05, "loss": 0.3812, "step": 375 }, { "epoch": 1.2012779552715656, "grad_norm": 0.7551583697235154, "learning_rate": 1.4989263622550017e-05, "loss": 0.3637, "step": 376 }, { "epoch": 1.2044728434504792, "grad_norm": 0.7703326368866282, "learning_rate": 1.4957008580186276e-05, "loss": 0.3958, "step": 377 }, { "epoch": 1.207667731629393, "grad_norm": 0.8019207400958424, "learning_rate": 1.4924685019693142e-05, "loss": 0.3753, "step": 378 }, { "epoch": 1.2108626198083068, "grad_norm": 0.7447320316646009, "learning_rate": 1.4892293387862221e-05, "loss": 0.3678, "step": 379 }, { "epoch": 1.2140575079872205, "grad_norm": 0.8236833165393697, "learning_rate": 1.485983413242606e-05, "loss": 0.4608, "step": 380 }, { "epoch": 1.2172523961661341, "grad_norm": 0.809350983162518, "learning_rate": 1.4827307702051919e-05, "loss": 0.4427, "step": 381 }, { "epoch": 1.220447284345048, "grad_norm": 0.6761221914505472, "learning_rate": 1.4794714546335578e-05, "loss": 0.4402, "step": 382 }, { "epoch": 1.2236421725239617, "grad_norm": 0.7020134281897162, "learning_rate": 1.4762055115795136e-05, "loss": 0.3388, "step": 383 }, { "epoch": 1.2268370607028753, "grad_norm": 0.8020099653178805, "learning_rate": 1.472932986186477e-05, "loss": 0.4078, "step": 384 }, { "epoch": 1.230031948881789, "grad_norm": 0.8929243606671824, "learning_rate": 1.4696539236888495e-05, "loss": 0.4397, "step": 385 }, { "epoch": 1.233226837060703, "grad_norm": 0.854936507274674, "learning_rate": 1.4663683694113924e-05, "loss": 0.4025, "step": 386 }, { "epoch": 1.2364217252396166, "grad_norm": 0.8116806635693596, "learning_rate": 1.463076368768599e-05, "loss": 0.357, "step": 387 }, { "epoch": 1.2396166134185305, "grad_norm": 0.7525728570276254, "learning_rate": 1.4597779672640668e-05, "loss": 0.4191, "step": 388 }, { "epoch": 1.2428115015974441, "grad_norm": 0.8072471116119602, "learning_rate": 1.4564732104898702e-05, "loss": 0.3903, "step": 389 }, { "epoch": 1.2460063897763578, "grad_norm": 0.7161886277690318, "learning_rate": 1.4531621441259285e-05, "loss": 0.4142, "step": 390 }, { "epoch": 1.2492012779552715, "grad_norm": 0.7547658320106545, "learning_rate": 1.4498448139393752e-05, "loss": 0.398, "step": 391 }, { "epoch": 1.2523961661341854, "grad_norm": 0.7773591617963772, "learning_rate": 1.4465212657839254e-05, "loss": 0.4327, "step": 392 }, { "epoch": 1.255591054313099, "grad_norm": 0.6904639860069021, "learning_rate": 1.4431915455992416e-05, "loss": 0.3591, "step": 393 }, { "epoch": 1.2587859424920127, "grad_norm": 0.7049006865325205, "learning_rate": 1.4398556994102996e-05, "loss": 0.4282, "step": 394 }, { "epoch": 1.2619808306709266, "grad_norm": 0.7175700995734923, "learning_rate": 1.4365137733267514e-05, "loss": 0.3281, "step": 395 }, { "epoch": 1.2651757188498403, "grad_norm": 0.8073941319432364, "learning_rate": 1.433165813542288e-05, "loss": 0.3913, "step": 396 }, { "epoch": 1.268370607028754, "grad_norm": 0.7003414727566679, "learning_rate": 1.4298118663340018e-05, "loss": 0.3638, "step": 397 }, { "epoch": 1.2715654952076676, "grad_norm": 0.7435882291411996, "learning_rate": 1.4264519780617452e-05, "loss": 0.3795, "step": 398 }, { "epoch": 1.2747603833865815, "grad_norm": 0.9193494825848232, "learning_rate": 1.4230861951674914e-05, "loss": 0.3923, "step": 399 }, { "epoch": 1.2779552715654952, "grad_norm": 0.7316900213493632, "learning_rate": 1.4197145641746923e-05, "loss": 0.4167, "step": 400 }, { "epoch": 1.281150159744409, "grad_norm": 0.7980774035403849, "learning_rate": 1.4163371316876347e-05, "loss": 0.3917, "step": 401 }, { "epoch": 1.2843450479233227, "grad_norm": 0.8480589998659835, "learning_rate": 1.412953944390795e-05, "loss": 0.4241, "step": 402 }, { "epoch": 1.2875399361022364, "grad_norm": 0.7315237373460427, "learning_rate": 1.4095650490481978e-05, "loss": 0.4207, "step": 403 }, { "epoch": 1.29073482428115, "grad_norm": 0.6375027271180014, "learning_rate": 1.4061704925027653e-05, "loss": 0.4, "step": 404 }, { "epoch": 1.293929712460064, "grad_norm": 0.8134150199474265, "learning_rate": 1.4027703216756718e-05, "loss": 0.3669, "step": 405 }, { "epoch": 1.2971246006389776, "grad_norm": 0.7898379126973761, "learning_rate": 1.3993645835656955e-05, "loss": 0.4186, "step": 406 }, { "epoch": 1.3003194888178915, "grad_norm": 0.7703707080336293, "learning_rate": 1.3959533252485678e-05, "loss": 0.3739, "step": 407 }, { "epoch": 1.3035143769968052, "grad_norm": 0.7611127075634048, "learning_rate": 1.3925365938763227e-05, "loss": 0.3894, "step": 408 }, { "epoch": 1.3067092651757188, "grad_norm": 0.5958132128301956, "learning_rate": 1.3891144366766457e-05, "loss": 0.3959, "step": 409 }, { "epoch": 1.3099041533546325, "grad_norm": 0.6334352951724201, "learning_rate": 1.3856869009522212e-05, "loss": 0.3619, "step": 410 }, { "epoch": 1.3130990415335464, "grad_norm": 0.7130135068724587, "learning_rate": 1.382254034080077e-05, "loss": 0.386, "step": 411 }, { "epoch": 1.31629392971246, "grad_norm": 0.7032314829984485, "learning_rate": 1.3788158835109313e-05, "loss": 0.3893, "step": 412 }, { "epoch": 1.3194888178913737, "grad_norm": 0.7185920347681584, "learning_rate": 1.3753724967685363e-05, "loss": 0.3601, "step": 413 }, { "epoch": 1.3226837060702876, "grad_norm": 0.6332079804609233, "learning_rate": 1.3719239214490203e-05, "loss": 0.4075, "step": 414 }, { "epoch": 1.3258785942492013, "grad_norm": 0.7404658527045683, "learning_rate": 1.368470205220231e-05, "loss": 0.3993, "step": 415 }, { "epoch": 1.329073482428115, "grad_norm": 0.7103673991700076, "learning_rate": 1.3650113958210764e-05, "loss": 0.4013, "step": 416 }, { "epoch": 1.3322683706070286, "grad_norm": 0.7999109629095718, "learning_rate": 1.3615475410608647e-05, "loss": 0.3968, "step": 417 }, { "epoch": 1.3354632587859425, "grad_norm": 0.7066086147191345, "learning_rate": 1.3580786888186428e-05, "loss": 0.4061, "step": 418 }, { "epoch": 1.3386581469648562, "grad_norm": 0.8014839647186196, "learning_rate": 1.3546048870425356e-05, "loss": 0.419, "step": 419 }, { "epoch": 1.34185303514377, "grad_norm": 0.5901005264351044, "learning_rate": 1.3511261837490837e-05, "loss": 0.4084, "step": 420 }, { "epoch": 1.3450479233226837, "grad_norm": 0.739391619088848, "learning_rate": 1.3476426270225768e-05, "loss": 0.3988, "step": 421 }, { "epoch": 1.3482428115015974, "grad_norm": 0.8316348210677322, "learning_rate": 1.344154265014393e-05, "loss": 0.3769, "step": 422 }, { "epoch": 1.351437699680511, "grad_norm": 0.7970493555353015, "learning_rate": 1.3406611459423306e-05, "loss": 0.3849, "step": 423 }, { "epoch": 1.354632587859425, "grad_norm": 0.7564976916720336, "learning_rate": 1.3371633180899417e-05, "loss": 0.4197, "step": 424 }, { "epoch": 1.3578274760383386, "grad_norm": 0.961478015601037, "learning_rate": 1.3336608298058662e-05, "loss": 0.3626, "step": 425 }, { "epoch": 1.3610223642172525, "grad_norm": 0.6797153190498834, "learning_rate": 1.3301537295031626e-05, "loss": 0.3978, "step": 426 }, { "epoch": 1.3642172523961662, "grad_norm": 0.7404534217433801, "learning_rate": 1.326642065658638e-05, "loss": 0.3743, "step": 427 }, { "epoch": 1.3674121405750799, "grad_norm": 0.7466684080396095, "learning_rate": 1.3231258868121806e-05, "loss": 0.3951, "step": 428 }, { "epoch": 1.3706070287539935, "grad_norm": 0.804922013834555, "learning_rate": 1.3196052415660856e-05, "loss": 0.3805, "step": 429 }, { "epoch": 1.3738019169329074, "grad_norm": 0.7133222438264706, "learning_rate": 1.3160801785843857e-05, "loss": 0.365, "step": 430 }, { "epoch": 1.376996805111821, "grad_norm": 0.6606882521977088, "learning_rate": 1.3125507465921775e-05, "loss": 0.3718, "step": 431 }, { "epoch": 1.3801916932907348, "grad_norm": 0.8048485138423717, "learning_rate": 1.3090169943749475e-05, "loss": 0.348, "step": 432 }, { "epoch": 1.3833865814696487, "grad_norm": 0.7057236132502058, "learning_rate": 1.3054789707778998e-05, "loss": 0.3768, "step": 433 }, { "epoch": 1.3865814696485623, "grad_norm": 0.7427223594984532, "learning_rate": 1.3019367247052781e-05, "loss": 0.4046, "step": 434 }, { "epoch": 1.389776357827476, "grad_norm": 0.777982348919025, "learning_rate": 1.2983903051196922e-05, "loss": 0.4084, "step": 435 }, { "epoch": 1.3929712460063897, "grad_norm": 0.7920793454584819, "learning_rate": 1.2948397610414393e-05, "loss": 0.3788, "step": 436 }, { "epoch": 1.3961661341853036, "grad_norm": 0.8369304632263217, "learning_rate": 1.291285141547828e-05, "loss": 0.3867, "step": 437 }, { "epoch": 1.3993610223642172, "grad_norm": 0.7008958369408834, "learning_rate": 1.287726495772499e-05, "loss": 0.357, "step": 438 }, { "epoch": 1.4025559105431311, "grad_norm": 0.880961196911694, "learning_rate": 1.2841638729047463e-05, "loss": 0.3871, "step": 439 }, { "epoch": 1.4057507987220448, "grad_norm": 0.7018004421621855, "learning_rate": 1.2805973221888366e-05, "loss": 0.392, "step": 440 }, { "epoch": 1.4089456869009584, "grad_norm": 0.6646094458865319, "learning_rate": 1.2770268929233298e-05, "loss": 0.4033, "step": 441 }, { "epoch": 1.4121405750798721, "grad_norm": 0.7343716165782984, "learning_rate": 1.273452634460397e-05, "loss": 0.3523, "step": 442 }, { "epoch": 1.415335463258786, "grad_norm": 0.6140404980778273, "learning_rate": 1.2698745962051379e-05, "loss": 0.3827, "step": 443 }, { "epoch": 1.4185303514376997, "grad_norm": 0.7817854886271103, "learning_rate": 1.2662928276148985e-05, "loss": 0.3811, "step": 444 }, { "epoch": 1.4217252396166133, "grad_norm": 0.7446921251394508, "learning_rate": 1.262707378198587e-05, "loss": 0.347, "step": 445 }, { "epoch": 1.4249201277955272, "grad_norm": 0.7985988097138982, "learning_rate": 1.25911829751599e-05, "loss": 0.4242, "step": 446 }, { "epoch": 1.428115015974441, "grad_norm": 0.6490635913633626, "learning_rate": 1.2555256351770873e-05, "loss": 0.3819, "step": 447 }, { "epoch": 1.4313099041533546, "grad_norm": 0.8958688000041772, "learning_rate": 1.2519294408413655e-05, "loss": 0.4268, "step": 448 }, { "epoch": 1.4345047923322682, "grad_norm": 0.7760938214202414, "learning_rate": 1.2483297642171332e-05, "loss": 0.3781, "step": 449 }, { "epoch": 1.4376996805111821, "grad_norm": 0.6672706606455567, "learning_rate": 1.2447266550608315e-05, "loss": 0.3916, "step": 450 }, { "epoch": 1.4408945686900958, "grad_norm": 0.783741926506166, "learning_rate": 1.2411201631763483e-05, "loss": 0.3712, "step": 451 }, { "epoch": 1.4440894568690097, "grad_norm": 0.9120231546467329, "learning_rate": 1.2375103384143297e-05, "loss": 0.3636, "step": 452 }, { "epoch": 1.4472843450479234, "grad_norm": 0.7973665501327444, "learning_rate": 1.2338972306714889e-05, "loss": 0.3368, "step": 453 }, { "epoch": 1.450479233226837, "grad_norm": 0.6655599307050887, "learning_rate": 1.23028088988992e-05, "loss": 0.3629, "step": 454 }, { "epoch": 1.4536741214057507, "grad_norm": 0.7334761926748368, "learning_rate": 1.2266613660564042e-05, "loss": 0.4479, "step": 455 }, { "epoch": 1.4568690095846646, "grad_norm": 0.7535020493818221, "learning_rate": 1.2230387092017204e-05, "loss": 0.393, "step": 456 }, { "epoch": 1.4600638977635783, "grad_norm": 0.7490052786483912, "learning_rate": 1.2194129693999549e-05, "loss": 0.389, "step": 457 }, { "epoch": 1.4632587859424921, "grad_norm": 0.7487324907432512, "learning_rate": 1.2157841967678064e-05, "loss": 0.3936, "step": 458 }, { "epoch": 1.4664536741214058, "grad_norm": 0.6713166876605019, "learning_rate": 1.2121524414638958e-05, "loss": 0.4003, "step": 459 }, { "epoch": 1.4696485623003195, "grad_norm": 0.6898492721518805, "learning_rate": 1.2085177536880717e-05, "loss": 0.3879, "step": 460 }, { "epoch": 1.4728434504792332, "grad_norm": 0.6341478496791327, "learning_rate": 1.2048801836807162e-05, "loss": 0.3638, "step": 461 }, { "epoch": 1.476038338658147, "grad_norm": 0.7101904301148986, "learning_rate": 1.2012397817220522e-05, "loss": 0.3399, "step": 462 }, { "epoch": 1.4792332268370607, "grad_norm": 0.8232525976178354, "learning_rate": 1.1975965981314459e-05, "loss": 0.4104, "step": 463 }, { "epoch": 1.4824281150159744, "grad_norm": 0.7043019842589794, "learning_rate": 1.1939506832667129e-05, "loss": 0.4181, "step": 464 }, { "epoch": 1.4856230031948883, "grad_norm": 0.7083596672114337, "learning_rate": 1.190302087523422e-05, "loss": 0.3815, "step": 465 }, { "epoch": 1.488817891373802, "grad_norm": 0.6847992152093964, "learning_rate": 1.186650861334199e-05, "loss": 0.3866, "step": 466 }, { "epoch": 1.4920127795527156, "grad_norm": 0.6177262873949391, "learning_rate": 1.182997055168027e-05, "loss": 0.3975, "step": 467 }, { "epoch": 1.4952076677316293, "grad_norm": 0.7517416052088943, "learning_rate": 1.179340719529553e-05, "loss": 0.3694, "step": 468 }, { "epoch": 1.4984025559105432, "grad_norm": 0.7414131237861542, "learning_rate": 1.1756819049583861e-05, "loss": 0.4014, "step": 469 }, { "epoch": 1.5015974440894568, "grad_norm": 0.7428769885232025, "learning_rate": 1.1720206620284011e-05, "loss": 0.377, "step": 470 }, { "epoch": 1.5047923322683707, "grad_norm": 0.7208343616237635, "learning_rate": 1.1683570413470384e-05, "loss": 0.3618, "step": 471 }, { "epoch": 1.5079872204472844, "grad_norm": 0.6254200530335278, "learning_rate": 1.1646910935546055e-05, "loss": 0.354, "step": 472 }, { "epoch": 1.511182108626198, "grad_norm": 0.7137137920366021, "learning_rate": 1.1610228693235748e-05, "loss": 0.391, "step": 473 }, { "epoch": 1.5143769968051117, "grad_norm": 0.7835829374466873, "learning_rate": 1.1573524193578863e-05, "loss": 0.3794, "step": 474 }, { "epoch": 1.5175718849840254, "grad_norm": 0.6785632963125671, "learning_rate": 1.1536797943922442e-05, "loss": 0.3203, "step": 475 }, { "epoch": 1.5207667731629393, "grad_norm": 0.6482994241562808, "learning_rate": 1.1500050451914171e-05, "loss": 0.3735, "step": 476 }, { "epoch": 1.5239616613418532, "grad_norm": 0.6552308867215125, "learning_rate": 1.1463282225495358e-05, "loss": 0.3885, "step": 477 }, { "epoch": 1.5271565495207668, "grad_norm": 0.6081556542356011, "learning_rate": 1.1426493772893907e-05, "loss": 0.3824, "step": 478 }, { "epoch": 1.5303514376996805, "grad_norm": 0.6578350633788698, "learning_rate": 1.1389685602617302e-05, "loss": 0.3937, "step": 479 }, { "epoch": 1.5335463258785942, "grad_norm": 0.6548231568961034, "learning_rate": 1.1352858223445571e-05, "loss": 0.4007, "step": 480 }, { "epoch": 1.5367412140575079, "grad_norm": 0.6863453973626347, "learning_rate": 1.1316012144424265e-05, "loss": 0.421, "step": 481 }, { "epoch": 1.5399361022364217, "grad_norm": 0.6349909711548734, "learning_rate": 1.1279147874857397e-05, "loss": 0.4273, "step": 482 }, { "epoch": 1.5431309904153354, "grad_norm": 0.6681450715179307, "learning_rate": 1.1242265924300433e-05, "loss": 0.3429, "step": 483 }, { "epoch": 1.5463258785942493, "grad_norm": 0.6888043347361502, "learning_rate": 1.1205366802553231e-05, "loss": 0.3975, "step": 484 }, { "epoch": 1.549520766773163, "grad_norm": 0.7223078250216108, "learning_rate": 1.1168451019652995e-05, "loss": 0.3613, "step": 485 }, { "epoch": 1.5527156549520766, "grad_norm": 0.8636641582227061, "learning_rate": 1.1131519085867223e-05, "loss": 0.3823, "step": 486 }, { "epoch": 1.5559105431309903, "grad_norm": 0.6802538435332052, "learning_rate": 1.1094571511686669e-05, "loss": 0.3921, "step": 487 }, { "epoch": 1.5591054313099042, "grad_norm": 0.6821928369769188, "learning_rate": 1.1057608807818267e-05, "loss": 0.3705, "step": 488 }, { "epoch": 1.5623003194888179, "grad_norm": 0.6868604612664317, "learning_rate": 1.1020631485178084e-05, "loss": 0.4362, "step": 489 }, { "epoch": 1.5654952076677318, "grad_norm": 0.6814388248892118, "learning_rate": 1.0983640054884252e-05, "loss": 0.3445, "step": 490 }, { "epoch": 1.5686900958466454, "grad_norm": 0.6238777532494983, "learning_rate": 1.0946635028249916e-05, "loss": 0.3926, "step": 491 }, { "epoch": 1.571884984025559, "grad_norm": 0.7288697316229533, "learning_rate": 1.0909616916776138e-05, "loss": 0.325, "step": 492 }, { "epoch": 1.5750798722044728, "grad_norm": 0.5919468997525036, "learning_rate": 1.0872586232144861e-05, "loss": 0.4112, "step": 493 }, { "epoch": 1.5782747603833864, "grad_norm": 0.6404385296704492, "learning_rate": 1.0835543486211815e-05, "loss": 0.3608, "step": 494 }, { "epoch": 1.5814696485623003, "grad_norm": 0.6699847382159787, "learning_rate": 1.0798489190999441e-05, "loss": 0.3565, "step": 495 }, { "epoch": 1.5846645367412142, "grad_norm": 0.6843597801238872, "learning_rate": 1.076142385868983e-05, "loss": 0.3909, "step": 496 }, { "epoch": 1.5878594249201279, "grad_norm": 0.634709992907837, "learning_rate": 1.0724348001617626e-05, "loss": 0.3873, "step": 497 }, { "epoch": 1.5910543130990416, "grad_norm": 0.6548440586247501, "learning_rate": 1.0687262132262952e-05, "loss": 0.3983, "step": 498 }, { "epoch": 1.5942492012779552, "grad_norm": 0.642604066379002, "learning_rate": 1.065016676324433e-05, "loss": 0.3682, "step": 499 }, { "epoch": 1.5974440894568689, "grad_norm": 0.7073578469704601, "learning_rate": 1.061306240731158e-05, "loss": 0.4251, "step": 500 }, { "epoch": 1.6006389776357828, "grad_norm": 0.777600887075542, "learning_rate": 1.057594957733876e-05, "loss": 0.4026, "step": 501 }, { "epoch": 1.6038338658146964, "grad_norm": 0.7531810084419379, "learning_rate": 1.0538828786317046e-05, "loss": 0.3423, "step": 502 }, { "epoch": 1.6070287539936103, "grad_norm": 0.6079896810219849, "learning_rate": 1.0501700547347662e-05, "loss": 0.3497, "step": 503 }, { "epoch": 1.610223642172524, "grad_norm": 0.7036526411507877, "learning_rate": 1.0464565373634784e-05, "loss": 0.376, "step": 504 }, { "epoch": 1.6134185303514377, "grad_norm": 0.6979697325962348, "learning_rate": 1.0427423778478428e-05, "loss": 0.3836, "step": 505 }, { "epoch": 1.6166134185303513, "grad_norm": 0.7415189360097981, "learning_rate": 1.039027627526739e-05, "loss": 0.4158, "step": 506 }, { "epoch": 1.619808306709265, "grad_norm": 0.6308217438942585, "learning_rate": 1.035312337747212e-05, "loss": 0.3668, "step": 507 }, { "epoch": 1.623003194888179, "grad_norm": 0.7294035513207259, "learning_rate": 1.0315965598637634e-05, "loss": 0.3341, "step": 508 }, { "epoch": 1.6261980830670928, "grad_norm": 0.6726328452330528, "learning_rate": 1.0278803452376416e-05, "loss": 0.353, "step": 509 }, { "epoch": 1.6293929712460065, "grad_norm": 0.7190369852430928, "learning_rate": 1.0241637452361323e-05, "loss": 0.3465, "step": 510 }, { "epoch": 1.6325878594249201, "grad_norm": 0.7705034557107943, "learning_rate": 1.0204468112318479e-05, "loss": 0.4002, "step": 511 }, { "epoch": 1.6357827476038338, "grad_norm": 0.7767754657361531, "learning_rate": 1.016729594602017e-05, "loss": 0.3735, "step": 512 }, { "epoch": 1.6389776357827475, "grad_norm": 0.7826578684750594, "learning_rate": 1.0130121467277755e-05, "loss": 0.4192, "step": 513 }, { "epoch": 1.6421725239616614, "grad_norm": 0.7359999647065305, "learning_rate": 1.0092945189934558e-05, "loss": 0.3698, "step": 514 }, { "epoch": 1.645367412140575, "grad_norm": 0.596609232423867, "learning_rate": 1.0055767627858748e-05, "loss": 0.3738, "step": 515 }, { "epoch": 1.648562300319489, "grad_norm": 0.6849486475833665, "learning_rate": 1.0018589294936273e-05, "loss": 0.3703, "step": 516 }, { "epoch": 1.6517571884984026, "grad_norm": 0.7070173939461314, "learning_rate": 9.981410705063728e-06, "loss": 0.3497, "step": 517 }, { "epoch": 1.6549520766773163, "grad_norm": 0.6510068381611003, "learning_rate": 9.944232372141252e-06, "loss": 0.3779, "step": 518 }, { "epoch": 1.65814696485623, "grad_norm": 0.6408548186573009, "learning_rate": 9.907054810065446e-06, "loss": 0.3545, "step": 519 }, { "epoch": 1.6613418530351438, "grad_norm": 0.6992267985738108, "learning_rate": 9.869878532722246e-06, "loss": 0.3819, "step": 520 }, { "epoch": 1.6645367412140575, "grad_norm": 0.8229340777031916, "learning_rate": 9.832704053979828e-06, "loss": 0.3974, "step": 521 }, { "epoch": 1.6677316293929714, "grad_norm": 0.6624955649595131, "learning_rate": 9.795531887681523e-06, "loss": 0.367, "step": 522 }, { "epoch": 1.670926517571885, "grad_norm": 0.7223508220259945, "learning_rate": 9.75836254763868e-06, "loss": 0.3762, "step": 523 }, { "epoch": 1.6741214057507987, "grad_norm": 0.666821717668502, "learning_rate": 9.721196547623585e-06, "loss": 0.4124, "step": 524 }, { "epoch": 1.6773162939297124, "grad_norm": 0.6667393357744014, "learning_rate": 9.68403440136237e-06, "loss": 0.3474, "step": 525 }, { "epoch": 1.680511182108626, "grad_norm": 0.7990772639121522, "learning_rate": 9.646876622527886e-06, "loss": 0.421, "step": 526 }, { "epoch": 1.68370607028754, "grad_norm": 0.602666552687939, "learning_rate": 9.609723724732611e-06, "loss": 0.3308, "step": 527 }, { "epoch": 1.6869009584664538, "grad_norm": 0.6268003913810553, "learning_rate": 9.572576221521574e-06, "loss": 0.416, "step": 528 }, { "epoch": 1.6900958466453675, "grad_norm": 0.5993146844292886, "learning_rate": 9.535434626365221e-06, "loss": 0.3952, "step": 529 }, { "epoch": 1.6932907348242812, "grad_norm": 0.600874293925157, "learning_rate": 9.49829945265234e-06, "loss": 0.3887, "step": 530 }, { "epoch": 1.6964856230031948, "grad_norm": 0.8734215117818372, "learning_rate": 9.461171213682957e-06, "loss": 0.386, "step": 531 }, { "epoch": 1.6996805111821085, "grad_norm": 0.722628235693675, "learning_rate": 9.424050422661243e-06, "loss": 0.4161, "step": 532 }, { "epoch": 1.7028753993610224, "grad_norm": 0.6988954063267449, "learning_rate": 9.386937592688422e-06, "loss": 0.4119, "step": 533 }, { "epoch": 1.706070287539936, "grad_norm": 0.7384635614942465, "learning_rate": 9.349833236755675e-06, "loss": 0.4044, "step": 534 }, { "epoch": 1.70926517571885, "grad_norm": 0.7882261202758374, "learning_rate": 9.31273786773705e-06, "loss": 0.3701, "step": 535 }, { "epoch": 1.7124600638977636, "grad_norm": 0.7035288193657737, "learning_rate": 9.275651998382377e-06, "loss": 0.3825, "step": 536 }, { "epoch": 1.7156549520766773, "grad_norm": 0.6439779352342753, "learning_rate": 9.238576141310172e-06, "loss": 0.3596, "step": 537 }, { "epoch": 1.718849840255591, "grad_norm": 0.6669213876269842, "learning_rate": 9.201510809000562e-06, "loss": 0.3482, "step": 538 }, { "epoch": 1.7220447284345048, "grad_norm": 0.6933389154618239, "learning_rate": 9.164456513788186e-06, "loss": 0.3781, "step": 539 }, { "epoch": 1.7252396166134185, "grad_norm": 0.7376867128140121, "learning_rate": 9.12741376785514e-06, "loss": 0.3964, "step": 540 }, { "epoch": 1.7284345047923324, "grad_norm": 0.6705508648989672, "learning_rate": 9.090383083223866e-06, "loss": 0.4017, "step": 541 }, { "epoch": 1.731629392971246, "grad_norm": 0.6576975605420414, "learning_rate": 9.053364971750087e-06, "loss": 0.3975, "step": 542 }, { "epoch": 1.7348242811501597, "grad_norm": 0.6410586797875636, "learning_rate": 9.01635994511575e-06, "loss": 0.4352, "step": 543 }, { "epoch": 1.7380191693290734, "grad_norm": 0.660772708577865, "learning_rate": 8.979368514821917e-06, "loss": 0.364, "step": 544 }, { "epoch": 1.741214057507987, "grad_norm": 0.7048979114742526, "learning_rate": 8.942391192181735e-06, "loss": 0.4064, "step": 545 }, { "epoch": 1.744408945686901, "grad_norm": 0.6113060002775187, "learning_rate": 8.905428488313335e-06, "loss": 0.3934, "step": 546 }, { "epoch": 1.7476038338658149, "grad_norm": 0.7545453314822455, "learning_rate": 8.868480914132777e-06, "loss": 0.4098, "step": 547 }, { "epoch": 1.7507987220447285, "grad_norm": 0.6514600293747569, "learning_rate": 8.831548980347009e-06, "loss": 0.4287, "step": 548 }, { "epoch": 1.7539936102236422, "grad_norm": 0.6646813541705668, "learning_rate": 8.79463319744677e-06, "loss": 0.3431, "step": 549 }, { "epoch": 1.7571884984025559, "grad_norm": 0.6154846894981999, "learning_rate": 8.757734075699567e-06, "loss": 0.4097, "step": 550 }, { "epoch": 1.7603833865814695, "grad_norm": 0.6954784339750693, "learning_rate": 8.720852125142606e-06, "loss": 0.3757, "step": 551 }, { "epoch": 1.7635782747603834, "grad_norm": 0.745806458726026, "learning_rate": 8.683987855575742e-06, "loss": 0.4016, "step": 552 }, { "epoch": 1.766773162939297, "grad_norm": 0.6265426574548529, "learning_rate": 8.64714177655443e-06, "loss": 0.3306, "step": 553 }, { "epoch": 1.769968051118211, "grad_norm": 0.6529854645903413, "learning_rate": 8.610314397382701e-06, "loss": 0.373, "step": 554 }, { "epoch": 1.7731629392971247, "grad_norm": 0.7241000354204435, "learning_rate": 8.573506227106098e-06, "loss": 0.435, "step": 555 }, { "epoch": 1.7763578274760383, "grad_norm": 0.6613396360810041, "learning_rate": 8.536717774504644e-06, "loss": 0.37, "step": 556 }, { "epoch": 1.779552715654952, "grad_norm": 0.7090691178398771, "learning_rate": 8.49994954808583e-06, "loss": 0.3259, "step": 557 }, { "epoch": 1.7827476038338657, "grad_norm": 0.7010800080078065, "learning_rate": 8.463202056077562e-06, "loss": 0.3763, "step": 558 }, { "epoch": 1.7859424920127795, "grad_norm": 0.8135213458217521, "learning_rate": 8.426475806421139e-06, "loss": 0.3633, "step": 559 }, { "epoch": 1.7891373801916934, "grad_norm": 0.8057923307232946, "learning_rate": 8.389771306764255e-06, "loss": 0.387, "step": 560 }, { "epoch": 1.792332268370607, "grad_norm": 0.6868563818667783, "learning_rate": 8.353089064453949e-06, "loss": 0.3919, "step": 561 }, { "epoch": 1.7955271565495208, "grad_norm": 0.729148092530284, "learning_rate": 8.316429586529616e-06, "loss": 0.3739, "step": 562 }, { "epoch": 1.7987220447284344, "grad_norm": 0.7454589082195145, "learning_rate": 8.27979337971599e-06, "loss": 0.366, "step": 563 }, { "epoch": 1.8019169329073481, "grad_norm": 0.7363609782622037, "learning_rate": 8.243180950416142e-06, "loss": 0.3746, "step": 564 }, { "epoch": 1.805111821086262, "grad_norm": 0.7051907706488193, "learning_rate": 8.206592804704473e-06, "loss": 0.4004, "step": 565 }, { "epoch": 1.8083067092651757, "grad_norm": 0.7077685360891058, "learning_rate": 8.170029448319732e-06, "loss": 0.3605, "step": 566 }, { "epoch": 1.8115015974440896, "grad_norm": 0.630064546716315, "learning_rate": 8.133491386658016e-06, "loss": 0.3956, "step": 567 }, { "epoch": 1.8146964856230032, "grad_norm": 0.7209999797137789, "learning_rate": 8.096979124765781e-06, "loss": 0.4296, "step": 568 }, { "epoch": 1.817891373801917, "grad_norm": 0.7309587998391873, "learning_rate": 8.060493167332874e-06, "loss": 0.3941, "step": 569 }, { "epoch": 1.8210862619808306, "grad_norm": 0.6946638921204976, "learning_rate": 8.024034018685546e-06, "loss": 0.3368, "step": 570 }, { "epoch": 1.8242811501597445, "grad_norm": 0.6290744849288465, "learning_rate": 7.98760218277948e-06, "loss": 0.3967, "step": 571 }, { "epoch": 1.8274760383386581, "grad_norm": 0.6756197738312434, "learning_rate": 7.95119816319284e-06, "loss": 0.4374, "step": 572 }, { "epoch": 1.830670926517572, "grad_norm": 0.6195041582890459, "learning_rate": 7.914822463119285e-06, "loss": 0.3764, "step": 573 }, { "epoch": 1.8338658146964857, "grad_norm": 0.6520016022036852, "learning_rate": 7.878475585361045e-06, "loss": 0.3697, "step": 574 }, { "epoch": 1.8370607028753994, "grad_norm": 0.6673365035852971, "learning_rate": 7.84215803232194e-06, "loss": 0.3751, "step": 575 }, { "epoch": 1.840255591054313, "grad_norm": 0.7467414573410552, "learning_rate": 7.805870306000453e-06, "loss": 0.4025, "step": 576 }, { "epoch": 1.8434504792332267, "grad_norm": 0.7254888570735092, "learning_rate": 7.769612907982798e-06, "loss": 0.3572, "step": 577 }, { "epoch": 1.8466453674121406, "grad_norm": 0.6517647297304884, "learning_rate": 7.733386339435965e-06, "loss": 0.3919, "step": 578 }, { "epoch": 1.8498402555910545, "grad_norm": 0.5750405762377315, "learning_rate": 7.697191101100802e-06, "loss": 0.3996, "step": 579 }, { "epoch": 1.8530351437699681, "grad_norm": 0.6982903562614436, "learning_rate": 7.661027693285113e-06, "loss": 0.3773, "step": 580 }, { "epoch": 1.8562300319488818, "grad_norm": 0.6536016012178709, "learning_rate": 7.624896615856709e-06, "loss": 0.3519, "step": 581 }, { "epoch": 1.8594249201277955, "grad_norm": 0.6044514365984781, "learning_rate": 7.588798368236517e-06, "loss": 0.3914, "step": 582 }, { "epoch": 1.8626198083067091, "grad_norm": 0.6005793640293874, "learning_rate": 7.5527334493916894e-06, "loss": 0.3374, "step": 583 }, { "epoch": 1.865814696485623, "grad_norm": 0.6837392480104575, "learning_rate": 7.516702357828672e-06, "loss": 0.3906, "step": 584 }, { "epoch": 1.8690095846645367, "grad_norm": 0.6869443137528136, "learning_rate": 7.480705591586343e-06, "loss": 0.3648, "step": 585 }, { "epoch": 1.8722044728434506, "grad_norm": 0.6640758656121128, "learning_rate": 7.444743648229129e-06, "loss": 0.3795, "step": 586 }, { "epoch": 1.8753993610223643, "grad_norm": 0.711636064138404, "learning_rate": 7.408817024840103e-06, "loss": 0.3569, "step": 587 }, { "epoch": 1.878594249201278, "grad_norm": 0.6951157527099808, "learning_rate": 7.372926218014131e-06, "loss": 0.3535, "step": 588 }, { "epoch": 1.8817891373801916, "grad_norm": 0.6706065636049294, "learning_rate": 7.337071723851018e-06, "loss": 0.3598, "step": 589 }, { "epoch": 1.8849840255591053, "grad_norm": 0.8241961745024526, "learning_rate": 7.301254037948624e-06, "loss": 0.4376, "step": 590 }, { "epoch": 1.8881789137380192, "grad_norm": 0.7234855502758178, "learning_rate": 7.26547365539603e-06, "loss": 0.3529, "step": 591 }, { "epoch": 1.891373801916933, "grad_norm": 0.6238866827009744, "learning_rate": 7.2297310707667036e-06, "loss": 0.4163, "step": 592 }, { "epoch": 1.8945686900958467, "grad_norm": 0.5736271877912065, "learning_rate": 7.194026778111637e-06, "loss": 0.3587, "step": 593 }, { "epoch": 1.8977635782747604, "grad_norm": 0.6572875943312029, "learning_rate": 7.1583612709525405e-06, "loss": 0.3783, "step": 594 }, { "epoch": 1.900958466453674, "grad_norm": 0.6593416011341751, "learning_rate": 7.122735042275013e-06, "loss": 0.3675, "step": 595 }, { "epoch": 1.9041533546325877, "grad_norm": 0.6249202723608295, "learning_rate": 7.0871485845217235e-06, "loss": 0.4192, "step": 596 }, { "epoch": 1.9073482428115016, "grad_norm": 0.6174171894639658, "learning_rate": 7.05160238958561e-06, "loss": 0.3545, "step": 597 }, { "epoch": 1.9105431309904153, "grad_norm": 0.5978066146036624, "learning_rate": 7.016096948803082e-06, "loss": 0.3882, "step": 598 }, { "epoch": 1.9137380191693292, "grad_norm": 0.7508662460345255, "learning_rate": 6.980632752947221e-06, "loss": 0.4201, "step": 599 }, { "epoch": 1.9169329073482428, "grad_norm": 0.6674809760382143, "learning_rate": 6.945210292221003e-06, "loss": 0.4007, "step": 600 }, { "epoch": 1.9201277955271565, "grad_norm": 0.6845401841035883, "learning_rate": 6.909830056250527e-06, "loss": 0.4242, "step": 601 }, { "epoch": 1.9233226837060702, "grad_norm": 0.6747837702166705, "learning_rate": 6.87449253407823e-06, "loss": 0.3484, "step": 602 }, { "epoch": 1.926517571884984, "grad_norm": 0.6572349129459626, "learning_rate": 6.839198214156146e-06, "loss": 0.399, "step": 603 }, { "epoch": 1.9297124600638977, "grad_norm": 0.667673745198207, "learning_rate": 6.803947584339148e-06, "loss": 0.3812, "step": 604 }, { "epoch": 1.9329073482428116, "grad_norm": 0.7997277616197571, "learning_rate": 6.768741131878196e-06, "loss": 0.3719, "step": 605 }, { "epoch": 1.9361022364217253, "grad_norm": 0.6607956272467509, "learning_rate": 6.733579343413621e-06, "loss": 0.4025, "step": 606 }, { "epoch": 1.939297124600639, "grad_norm": 0.6424153208610056, "learning_rate": 6.69846270496838e-06, "loss": 0.4088, "step": 607 }, { "epoch": 1.9424920127795526, "grad_norm": 0.6841836944354628, "learning_rate": 6.66339170194134e-06, "loss": 0.3638, "step": 608 }, { "epoch": 1.9456869009584663, "grad_norm": 0.7327498988996838, "learning_rate": 6.628366819100586e-06, "loss": 0.3958, "step": 609 }, { "epoch": 1.9488817891373802, "grad_norm": 0.6623970255054498, "learning_rate": 6.593388540576699e-06, "loss": 0.4044, "step": 610 }, { "epoch": 1.952076677316294, "grad_norm": 0.6762762317320167, "learning_rate": 6.55845734985607e-06, "loss": 0.341, "step": 611 }, { "epoch": 1.9552715654952078, "grad_norm": 0.7352041630492657, "learning_rate": 6.523573729774234e-06, "loss": 0.419, "step": 612 }, { "epoch": 1.9584664536741214, "grad_norm": 0.6937804334119967, "learning_rate": 6.48873816250917e-06, "loss": 0.3757, "step": 613 }, { "epoch": 1.961661341853035, "grad_norm": 0.7897600638634236, "learning_rate": 6.453951129574644e-06, "loss": 0.4072, "step": 614 }, { "epoch": 1.9648562300319488, "grad_norm": 0.760236193668744, "learning_rate": 6.419213111813576e-06, "loss": 0.3435, "step": 615 }, { "epoch": 1.9680511182108626, "grad_norm": 0.8111151020455108, "learning_rate": 6.384524589391358e-06, "loss": 0.3664, "step": 616 }, { "epoch": 1.9712460063897763, "grad_norm": 0.6283684271490623, "learning_rate": 6.349886041789236e-06, "loss": 0.3853, "step": 617 }, { "epoch": 1.9744408945686902, "grad_norm": 0.6193900473349803, "learning_rate": 6.315297947797691e-06, "loss": 0.3598, "step": 618 }, { "epoch": 1.9776357827476039, "grad_norm": 0.8051890922129095, "learning_rate": 6.280760785509802e-06, "loss": 0.3793, "step": 619 }, { "epoch": 1.9808306709265175, "grad_norm": 0.7239793820151423, "learning_rate": 6.24627503231464e-06, "loss": 0.3721, "step": 620 }, { "epoch": 1.9840255591054312, "grad_norm": 0.7063077062824192, "learning_rate": 6.211841164890689e-06, "loss": 0.3797, "step": 621 }, { "epoch": 1.9872204472843449, "grad_norm": 0.7372976490512757, "learning_rate": 6.177459659199237e-06, "loss": 0.3738, "step": 622 }, { "epoch": 1.9904153354632588, "grad_norm": 0.6839285813717391, "learning_rate": 6.1431309904777915e-06, "loss": 0.3894, "step": 623 }, { "epoch": 1.9936102236421727, "grad_norm": 0.6817876012906897, "learning_rate": 6.108855633233546e-06, "loss": 0.4159, "step": 624 }, { "epoch": 1.9968051118210863, "grad_norm": 0.6386452889929469, "learning_rate": 6.074634061236777e-06, "loss": 0.3517, "step": 625 }, { "epoch": 2.0, "grad_norm": 0.638099341863494, "learning_rate": 6.040466747514324e-06, "loss": 0.3387, "step": 626 }, { "epoch": 2.0031948881789137, "grad_norm": 0.8271347304416454, "learning_rate": 6.006354164343047e-06, "loss": 0.2879, "step": 627 }, { "epoch": 2.0063897763578273, "grad_norm": 0.646769849130275, "learning_rate": 5.972296783243281e-06, "loss": 0.2889, "step": 628 }, { "epoch": 2.009584664536741, "grad_norm": 0.6621296057507008, "learning_rate": 5.93829507497235e-06, "loss": 0.3002, "step": 629 }, { "epoch": 2.012779552715655, "grad_norm": 0.6556200090741601, "learning_rate": 5.904349509518024e-06, "loss": 0.2477, "step": 630 }, { "epoch": 2.015974440894569, "grad_norm": 0.9911955110119478, "learning_rate": 5.87046055609205e-06, "loss": 0.3116, "step": 631 }, { "epoch": 2.0191693290734825, "grad_norm": 0.80574892593569, "learning_rate": 5.836628683123659e-06, "loss": 0.2621, "step": 632 }, { "epoch": 2.022364217252396, "grad_norm": 0.6910559059978796, "learning_rate": 5.8028543582530805e-06, "loss": 0.2855, "step": 633 }, { "epoch": 2.02555910543131, "grad_norm": 0.5968749973323358, "learning_rate": 5.769138048325087e-06, "loss": 0.2414, "step": 634 }, { "epoch": 2.0287539936102235, "grad_norm": 0.6536748958061217, "learning_rate": 5.735480219382554e-06, "loss": 0.2973, "step": 635 }, { "epoch": 2.0319488817891376, "grad_norm": 0.6989686634311535, "learning_rate": 5.701881336659992e-06, "loss": 0.268, "step": 636 }, { "epoch": 2.0351437699680512, "grad_norm": 0.6569035525910143, "learning_rate": 5.668341864577125e-06, "loss": 0.3044, "step": 637 }, { "epoch": 2.038338658146965, "grad_norm": 0.602669314331145, "learning_rate": 5.634862266732492e-06, "loss": 0.281, "step": 638 }, { "epoch": 2.0415335463258786, "grad_norm": 0.5728367364913933, "learning_rate": 5.601443005897012e-06, "loss": 0.332, "step": 639 }, { "epoch": 2.0447284345047922, "grad_norm": 0.6368451330277513, "learning_rate": 5.5680845440075885e-06, "loss": 0.2893, "step": 640 }, { "epoch": 2.047923322683706, "grad_norm": 0.5734258778625324, "learning_rate": 5.534787342160752e-06, "loss": 0.2975, "step": 641 }, { "epoch": 2.0511182108626196, "grad_norm": 0.6266002126502631, "learning_rate": 5.501551860606251e-06, "loss": 0.2898, "step": 642 }, { "epoch": 2.0543130990415337, "grad_norm": 0.5878513634227986, "learning_rate": 5.468378558740719e-06, "loss": 0.2666, "step": 643 }, { "epoch": 2.0575079872204474, "grad_norm": 0.5702013097999747, "learning_rate": 5.435267895101303e-06, "loss": 0.2692, "step": 644 }, { "epoch": 2.060702875399361, "grad_norm": 0.602338703074719, "learning_rate": 5.402220327359335e-06, "loss": 0.2721, "step": 645 }, { "epoch": 2.0638977635782747, "grad_norm": 0.5811139388678999, "learning_rate": 5.369236312314017e-06, "loss": 0.2928, "step": 646 }, { "epoch": 2.0670926517571884, "grad_norm": 0.5882325787521038, "learning_rate": 5.336316305886078e-06, "loss": 0.2966, "step": 647 }, { "epoch": 2.070287539936102, "grad_norm": 0.6412192942110215, "learning_rate": 5.303460763111508e-06, "loss": 0.2946, "step": 648 }, { "epoch": 2.073482428115016, "grad_norm": 0.6529865737566748, "learning_rate": 5.270670138135234e-06, "loss": 0.2574, "step": 649 }, { "epoch": 2.07667731629393, "grad_norm": 0.5986874454585268, "learning_rate": 5.237944884204864e-06, "loss": 0.2461, "step": 650 }, { "epoch": 2.0798722044728435, "grad_norm": 0.5492039524576837, "learning_rate": 5.205285453664424e-06, "loss": 0.3039, "step": 651 }, { "epoch": 2.083067092651757, "grad_norm": 0.6487742195433347, "learning_rate": 5.1726922979480805e-06, "loss": 0.2899, "step": 652 }, { "epoch": 2.086261980830671, "grad_norm": 0.6995349782327247, "learning_rate": 5.14016586757394e-06, "loss": 0.288, "step": 653 }, { "epoch": 2.0894568690095845, "grad_norm": 0.5774907524747939, "learning_rate": 5.107706612137776e-06, "loss": 0.2571, "step": 654 }, { "epoch": 2.0926517571884986, "grad_norm": 0.602454561129501, "learning_rate": 5.075314980306861e-06, "loss": 0.2588, "step": 655 }, { "epoch": 2.0958466453674123, "grad_norm": 0.615167046308701, "learning_rate": 5.042991419813723e-06, "loss": 0.2361, "step": 656 }, { "epoch": 2.099041533546326, "grad_norm": 0.5655289081350507, "learning_rate": 5.010736377449983e-06, "loss": 0.2763, "step": 657 }, { "epoch": 2.1022364217252396, "grad_norm": 0.5867220298505782, "learning_rate": 4.978550299060179e-06, "loss": 0.256, "step": 658 }, { "epoch": 2.1054313099041533, "grad_norm": 0.5617761964475673, "learning_rate": 4.946433629535585e-06, "loss": 0.3115, "step": 659 }, { "epoch": 2.108626198083067, "grad_norm": 0.530087482930863, "learning_rate": 4.914386812808073e-06, "loss": 0.2481, "step": 660 }, { "epoch": 2.1118210862619806, "grad_norm": 0.5821201498478065, "learning_rate": 4.882410291843989e-06, "loss": 0.2687, "step": 661 }, { "epoch": 2.1150159744408947, "grad_norm": 0.6009245895614106, "learning_rate": 4.850504508638004e-06, "loss": 0.2672, "step": 662 }, { "epoch": 2.1182108626198084, "grad_norm": 0.5812729760109654, "learning_rate": 4.818669904207021e-06, "loss": 0.2669, "step": 663 }, { "epoch": 2.121405750798722, "grad_norm": 0.5562495650685109, "learning_rate": 4.786906918584083e-06, "loss": 0.2708, "step": 664 }, { "epoch": 2.1246006389776357, "grad_norm": 0.565127621348711, "learning_rate": 4.755215990812277e-06, "loss": 0.2521, "step": 665 }, { "epoch": 2.1277955271565494, "grad_norm": 0.5406508398417975, "learning_rate": 4.7235975589386715e-06, "loss": 0.2579, "step": 666 }, { "epoch": 2.130990415335463, "grad_norm": 0.5579616048617105, "learning_rate": 4.692052060008271e-06, "loss": 0.3169, "step": 667 }, { "epoch": 2.134185303514377, "grad_norm": 0.565592556878981, "learning_rate": 4.6605799300579546e-06, "loss": 0.2653, "step": 668 }, { "epoch": 2.137380191693291, "grad_norm": 0.5683205240290914, "learning_rate": 4.629181604110464e-06, "loss": 0.2718, "step": 669 }, { "epoch": 2.1405750798722045, "grad_norm": 0.6460113832000962, "learning_rate": 4.597857516168397e-06, "loss": 0.2682, "step": 670 }, { "epoch": 2.143769968051118, "grad_norm": 0.5862339654865363, "learning_rate": 4.566608099208185e-06, "loss": 0.3328, "step": 671 }, { "epoch": 2.146964856230032, "grad_norm": 0.5435510212641139, "learning_rate": 4.535433785174124e-06, "loss": 0.2685, "step": 672 }, { "epoch": 2.1501597444089455, "grad_norm": 0.5399021750624424, "learning_rate": 4.50433500497241e-06, "loss": 0.2739, "step": 673 }, { "epoch": 2.1533546325878596, "grad_norm": 0.5286896610464358, "learning_rate": 4.4733121884651665e-06, "loss": 0.3251, "step": 674 }, { "epoch": 2.1565495207667733, "grad_norm": 0.5432882677437948, "learning_rate": 4.442365764464509e-06, "loss": 0.3201, "step": 675 }, { "epoch": 2.159744408945687, "grad_norm": 0.5748779573239078, "learning_rate": 4.411496160726632e-06, "loss": 0.2604, "step": 676 }, { "epoch": 2.1629392971246006, "grad_norm": 0.5033583313339888, "learning_rate": 4.38070380394587e-06, "loss": 0.2686, "step": 677 }, { "epoch": 2.1661341853035143, "grad_norm": 0.5985918860143106, "learning_rate": 4.349989119748815e-06, "loss": 0.2199, "step": 678 }, { "epoch": 2.169329073482428, "grad_norm": 0.5377363065642814, "learning_rate": 4.319352532688444e-06, "loss": 0.2533, "step": 679 }, { "epoch": 2.1725239616613417, "grad_norm": 0.5208201227205966, "learning_rate": 4.288794466238224e-06, "loss": 0.2494, "step": 680 }, { "epoch": 2.1757188498402558, "grad_norm": 0.546873580650872, "learning_rate": 4.2583153427862765e-06, "loss": 0.2587, "step": 681 }, { "epoch": 2.1789137380191694, "grad_norm": 0.6235527130870171, "learning_rate": 4.227915583629543e-06, "loss": 0.2825, "step": 682 }, { "epoch": 2.182108626198083, "grad_norm": 0.5268102602354202, "learning_rate": 4.197595608967941e-06, "loss": 0.275, "step": 683 }, { "epoch": 2.1853035143769968, "grad_norm": 0.5401897921628086, "learning_rate": 4.167355837898585e-06, "loss": 0.3187, "step": 684 }, { "epoch": 2.1884984025559104, "grad_norm": 0.5995690052883824, "learning_rate": 4.137196688409962e-06, "loss": 0.2842, "step": 685 }, { "epoch": 2.191693290734824, "grad_norm": 0.530440581581717, "learning_rate": 4.107118577376172e-06, "loss": 0.2655, "step": 686 }, { "epoch": 2.194888178913738, "grad_norm": 0.5468145774158746, "learning_rate": 4.0771219205511756e-06, "loss": 0.3003, "step": 687 }, { "epoch": 2.198083067092652, "grad_norm": 0.5307256800656307, "learning_rate": 4.04720713256302e-06, "loss": 0.29, "step": 688 }, { "epoch": 2.2012779552715656, "grad_norm": 0.5876558249492184, "learning_rate": 4.017374626908125e-06, "loss": 0.3012, "step": 689 }, { "epoch": 2.2044728434504792, "grad_norm": 0.5288296732250444, "learning_rate": 3.987624815945575e-06, "loss": 0.3387, "step": 690 }, { "epoch": 2.207667731629393, "grad_norm": 0.5334676159553654, "learning_rate": 3.957958110891395e-06, "loss": 0.3118, "step": 691 }, { "epoch": 2.2108626198083066, "grad_norm": 0.5018942342913997, "learning_rate": 3.9283749218128885e-06, "loss": 0.2442, "step": 692 }, { "epoch": 2.2140575079872207, "grad_norm": 0.5460501175933514, "learning_rate": 3.898875657622963e-06, "loss": 0.2751, "step": 693 }, { "epoch": 2.2172523961661343, "grad_norm": 0.5238545161103514, "learning_rate": 3.8694607260744745e-06, "loss": 0.3171, "step": 694 }, { "epoch": 2.220447284345048, "grad_norm": 0.5648156161744206, "learning_rate": 3.840130533754585e-06, "loss": 0.3184, "step": 695 }, { "epoch": 2.2236421725239617, "grad_norm": 0.5818824089460198, "learning_rate": 3.8108854860791657e-06, "loss": 0.2745, "step": 696 }, { "epoch": 2.2268370607028753, "grad_norm": 0.5940846743967164, "learning_rate": 3.781725987287166e-06, "loss": 0.2612, "step": 697 }, { "epoch": 2.230031948881789, "grad_norm": 0.48818409466248414, "learning_rate": 3.7526524404350407e-06, "loss": 0.2459, "step": 698 }, { "epoch": 2.2332268370607027, "grad_norm": 0.553611582334448, "learning_rate": 3.7236652473911817e-06, "loss": 0.2894, "step": 699 }, { "epoch": 2.236421725239617, "grad_norm": 0.639389907063906, "learning_rate": 3.6947648088303523e-06, "loss": 0.2926, "step": 700 }, { "epoch": 2.2396166134185305, "grad_norm": 0.5480348447624199, "learning_rate": 3.6659515242281496e-06, "loss": 0.2128, "step": 701 }, { "epoch": 2.242811501597444, "grad_norm": 0.5565653754577661, "learning_rate": 3.6372257918555e-06, "loss": 0.2949, "step": 702 }, { "epoch": 2.246006389776358, "grad_norm": 0.6313270010180382, "learning_rate": 3.6085880087731316e-06, "loss": 0.2719, "step": 703 }, { "epoch": 2.2492012779552715, "grad_norm": 0.5254468501533311, "learning_rate": 3.580038570826093e-06, "loss": 0.2646, "step": 704 }, { "epoch": 2.252396166134185, "grad_norm": 0.5449168135123015, "learning_rate": 3.5515778726382967e-06, "loss": 0.3125, "step": 705 }, { "epoch": 2.255591054313099, "grad_norm": 0.577769744560431, "learning_rate": 3.5232063076070377e-06, "loss": 0.2877, "step": 706 }, { "epoch": 2.258785942492013, "grad_norm": 0.5337396711928698, "learning_rate": 3.494924267897585e-06, "loss": 0.2254, "step": 707 }, { "epoch": 2.2619808306709266, "grad_norm": 0.5557081825995588, "learning_rate": 3.4667321444377334e-06, "loss": 0.2455, "step": 708 }, { "epoch": 2.2651757188498403, "grad_norm": 0.5763386812267897, "learning_rate": 3.4386303269124142e-06, "loss": 0.2458, "step": 709 }, { "epoch": 2.268370607028754, "grad_norm": 0.5750561585588471, "learning_rate": 3.4106192037583176e-06, "loss": 0.2817, "step": 710 }, { "epoch": 2.2715654952076676, "grad_norm": 0.5403613324177238, "learning_rate": 3.382699162158498e-06, "loss": 0.2796, "step": 711 }, { "epoch": 2.2747603833865817, "grad_norm": 0.49899972805059195, "learning_rate": 3.3548705880370538e-06, "loss": 0.2606, "step": 712 }, { "epoch": 2.2779552715654954, "grad_norm": 0.5211130243599169, "learning_rate": 3.327133866053758e-06, "loss": 0.2393, "step": 713 }, { "epoch": 2.281150159744409, "grad_norm": 0.5841971667091718, "learning_rate": 3.299489379598777e-06, "loss": 0.2831, "step": 714 }, { "epoch": 2.2843450479233227, "grad_norm": 0.5719620976677046, "learning_rate": 3.271937510787343e-06, "loss": 0.2411, "step": 715 }, { "epoch": 2.2875399361022364, "grad_norm": 0.540191979436821, "learning_rate": 3.2444786404544828e-06, "loss": 0.2698, "step": 716 }, { "epoch": 2.29073482428115, "grad_norm": 0.6112105115397988, "learning_rate": 3.217113148149765e-06, "loss": 0.3048, "step": 717 }, { "epoch": 2.2939297124600637, "grad_norm": 0.5771913514730167, "learning_rate": 3.1898414121320277e-06, "loss": 0.2752, "step": 718 }, { "epoch": 2.297124600638978, "grad_norm": 0.5501110706666438, "learning_rate": 3.162663809364178e-06, "loss": 0.1977, "step": 719 }, { "epoch": 2.3003194888178915, "grad_norm": 0.46927847463459355, "learning_rate": 3.135580715507961e-06, "loss": 0.2813, "step": 720 }, { "epoch": 2.303514376996805, "grad_norm": 0.5043271742582294, "learning_rate": 3.1085925049187738e-06, "loss": 0.2815, "step": 721 }, { "epoch": 2.306709265175719, "grad_norm": 0.47860976047812787, "learning_rate": 3.0816995506405e-06, "loss": 0.2394, "step": 722 }, { "epoch": 2.3099041533546325, "grad_norm": 0.5772481927495404, "learning_rate": 3.0549022244003368e-06, "loss": 0.2441, "step": 723 }, { "epoch": 2.313099041533546, "grad_norm": 0.5196025768297203, "learning_rate": 3.0282008966036647e-06, "loss": 0.2814, "step": 724 }, { "epoch": 2.31629392971246, "grad_norm": 0.600980799509102, "learning_rate": 3.001595936328939e-06, "loss": 0.3186, "step": 725 }, { "epoch": 2.319488817891374, "grad_norm": 0.5384170460565109, "learning_rate": 2.975087711322567e-06, "loss": 0.281, "step": 726 }, { "epoch": 2.3226837060702876, "grad_norm": 0.5532736843981153, "learning_rate": 2.948676587993834e-06, "loss": 0.2801, "step": 727 }, { "epoch": 2.3258785942492013, "grad_norm": 0.5448894944106978, "learning_rate": 2.9223629314098514e-06, "loss": 0.2722, "step": 728 }, { "epoch": 2.329073482428115, "grad_norm": 0.5493854347649725, "learning_rate": 2.8961471052904855e-06, "loss": 0.2564, "step": 729 }, { "epoch": 2.3322683706070286, "grad_norm": 0.5697309480878635, "learning_rate": 2.8700294720033486e-06, "loss": 0.2384, "step": 730 }, { "epoch": 2.3354632587859427, "grad_norm": 0.5481372632300319, "learning_rate": 2.8440103925587904e-06, "loss": 0.3108, "step": 731 }, { "epoch": 2.3386581469648564, "grad_norm": 0.5951546959209056, "learning_rate": 2.8180902266048947e-06, "loss": 0.288, "step": 732 }, { "epoch": 2.34185303514377, "grad_norm": 0.5593159640593378, "learning_rate": 2.792269332422517e-06, "loss": 0.2797, "step": 733 }, { "epoch": 2.3450479233226837, "grad_norm": 0.548130461510605, "learning_rate": 2.7665480669203383e-06, "loss": 0.291, "step": 734 }, { "epoch": 2.3482428115015974, "grad_norm": 0.50647579879316, "learning_rate": 2.740926785629915e-06, "loss": 0.2884, "step": 735 }, { "epoch": 2.351437699680511, "grad_norm": 0.5098520788004228, "learning_rate": 2.7154058427007822e-06, "loss": 0.2657, "step": 736 }, { "epoch": 2.3546325878594248, "grad_norm": 0.5226212845956969, "learning_rate": 2.6899855908955464e-06, "loss": 0.3284, "step": 737 }, { "epoch": 2.357827476038339, "grad_norm": 0.5311184509918885, "learning_rate": 2.6646663815850092e-06, "loss": 0.2598, "step": 738 }, { "epoch": 2.3610223642172525, "grad_norm": 0.5427537786360982, "learning_rate": 2.639448564743328e-06, "loss": 0.253, "step": 739 }, { "epoch": 2.364217252396166, "grad_norm": 0.5185119895770091, "learning_rate": 2.614332488943152e-06, "loss": 0.2433, "step": 740 }, { "epoch": 2.36741214057508, "grad_norm": 0.5551939240347452, "learning_rate": 2.5893185013508195e-06, "loss": 0.2509, "step": 741 }, { "epoch": 2.3706070287539935, "grad_norm": 0.5190495921348532, "learning_rate": 2.564406947721566e-06, "loss": 0.2895, "step": 742 }, { "epoch": 2.373801916932907, "grad_norm": 0.5382445162395503, "learning_rate": 2.539598172394727e-06, "loss": 0.2921, "step": 743 }, { "epoch": 2.376996805111821, "grad_norm": 0.5782994469073974, "learning_rate": 2.514892518288988e-06, "loss": 0.2628, "step": 744 }, { "epoch": 2.380191693290735, "grad_norm": 0.5004402464282985, "learning_rate": 2.490290326897653e-06, "loss": 0.2707, "step": 745 }, { "epoch": 2.3833865814696487, "grad_norm": 0.6037080436677754, "learning_rate": 2.4657919382839034e-06, "loss": 0.2648, "step": 746 }, { "epoch": 2.3865814696485623, "grad_norm": 0.538405984748164, "learning_rate": 2.4413976910761117e-06, "loss": 0.3182, "step": 747 }, { "epoch": 2.389776357827476, "grad_norm": 0.5439877759038688, "learning_rate": 2.417107922463169e-06, "loss": 0.3184, "step": 748 }, { "epoch": 2.3929712460063897, "grad_norm": 0.583035854090823, "learning_rate": 2.3929229681898005e-06, "loss": 0.2202, "step": 749 }, { "epoch": 2.3961661341853033, "grad_norm": 0.5884416492925523, "learning_rate": 2.3688431625519417e-06, "loss": 0.283, "step": 750 }, { "epoch": 2.3993610223642174, "grad_norm": 0.5355399468230221, "learning_rate": 2.3448688383921183e-06, "loss": 0.2694, "step": 751 }, { "epoch": 2.402555910543131, "grad_norm": 0.47662118367586076, "learning_rate": 2.3210003270948367e-06, "loss": 0.2694, "step": 752 }, { "epoch": 2.405750798722045, "grad_norm": 0.4805391075501156, "learning_rate": 2.297237958582005e-06, "loss": 0.2775, "step": 753 }, { "epoch": 2.4089456869009584, "grad_norm": 0.5202726458297047, "learning_rate": 2.2735820613083837e-06, "loss": 0.2687, "step": 754 }, { "epoch": 2.412140575079872, "grad_norm": 0.5392847644844359, "learning_rate": 2.2500329622570296e-06, "loss": 0.2961, "step": 755 }, { "epoch": 2.415335463258786, "grad_norm": 0.5240783463371834, "learning_rate": 2.2265909869347823e-06, "loss": 0.3119, "step": 756 }, { "epoch": 2.4185303514377, "grad_norm": 0.5569505587936721, "learning_rate": 2.2032564593677773e-06, "loss": 0.261, "step": 757 }, { "epoch": 2.4217252396166136, "grad_norm": 0.515386283950919, "learning_rate": 2.180029702096946e-06, "loss": 0.2909, "step": 758 }, { "epoch": 2.4249201277955272, "grad_norm": 0.5645060312581792, "learning_rate": 2.156911036173568e-06, "loss": 0.2765, "step": 759 }, { "epoch": 2.428115015974441, "grad_norm": 0.5536403998357785, "learning_rate": 2.1339007811548395e-06, "loss": 0.2152, "step": 760 }, { "epoch": 2.4313099041533546, "grad_norm": 0.5157472478536549, "learning_rate": 2.110999255099444e-06, "loss": 0.2869, "step": 761 }, { "epoch": 2.4345047923322682, "grad_norm": 0.5421226829224044, "learning_rate": 2.088206774563161e-06, "loss": 0.2442, "step": 762 }, { "epoch": 2.437699680511182, "grad_norm": 0.5221858315858158, "learning_rate": 2.065523654594497e-06, "loss": 0.2905, "step": 763 }, { "epoch": 2.440894568690096, "grad_norm": 0.5401675654904415, "learning_rate": 2.0429502087303164e-06, "loss": 0.2916, "step": 764 }, { "epoch": 2.4440894568690097, "grad_norm": 0.5219535133275498, "learning_rate": 2.020486748991526e-06, "loss": 0.268, "step": 765 }, { "epoch": 2.4472843450479234, "grad_norm": 0.5709710992260904, "learning_rate": 1.998133585878743e-06, "loss": 0.2779, "step": 766 }, { "epoch": 2.450479233226837, "grad_norm": 0.4824676976132383, "learning_rate": 1.9758910283680134e-06, "loss": 0.2553, "step": 767 }, { "epoch": 2.4536741214057507, "grad_norm": 0.5109201333375051, "learning_rate": 1.9537593839065484e-06, "loss": 0.3015, "step": 768 }, { "epoch": 2.4568690095846644, "grad_norm": 0.5863480432635425, "learning_rate": 1.931738958408457e-06, "loss": 0.2495, "step": 769 }, { "epoch": 2.460063897763578, "grad_norm": 0.5220281797772327, "learning_rate": 1.9098300562505266e-06, "loss": 0.2843, "step": 770 }, { "epoch": 2.463258785942492, "grad_norm": 0.5508369426630649, "learning_rate": 1.8880329802680253e-06, "loss": 0.2824, "step": 771 }, { "epoch": 2.466453674121406, "grad_norm": 0.5514083887348668, "learning_rate": 1.866348031750499e-06, "loss": 0.2532, "step": 772 }, { "epoch": 2.4696485623003195, "grad_norm": 0.5007229218626282, "learning_rate": 1.844775510437613e-06, "loss": 0.2775, "step": 773 }, { "epoch": 2.472843450479233, "grad_norm": 0.5654037998263154, "learning_rate": 1.8233157145150183e-06, "loss": 0.2169, "step": 774 }, { "epoch": 2.476038338658147, "grad_norm": 0.5079251173769247, "learning_rate": 1.8019689406102125e-06, "loss": 0.275, "step": 775 }, { "epoch": 2.479233226837061, "grad_norm": 0.7598992351846287, "learning_rate": 1.7807354837884583e-06, "loss": 0.2886, "step": 776 }, { "epoch": 2.4824281150159746, "grad_norm": 0.5402500169735379, "learning_rate": 1.759615637548686e-06, "loss": 0.2356, "step": 777 }, { "epoch": 2.4856230031948883, "grad_norm": 0.5153724379287912, "learning_rate": 1.7386096938194585e-06, "loss": 0.3273, "step": 778 }, { "epoch": 2.488817891373802, "grad_norm": 0.5656481880772636, "learning_rate": 1.717717942954914e-06, "loss": 0.2978, "step": 779 }, { "epoch": 2.4920127795527156, "grad_norm": 0.5600985228279312, "learning_rate": 1.6969406737307625e-06, "loss": 0.2914, "step": 780 }, { "epoch": 2.4952076677316293, "grad_norm": 0.5271143479597251, "learning_rate": 1.6762781733403034e-06, "loss": 0.256, "step": 781 }, { "epoch": 2.498402555910543, "grad_norm": 0.5290034655652396, "learning_rate": 1.6557307273904355e-06, "loss": 0.3164, "step": 782 }, { "epoch": 2.501597444089457, "grad_norm": 0.5282890946530933, "learning_rate": 1.6352986198977327e-06, "loss": 0.2548, "step": 783 }, { "epoch": 2.5047923322683707, "grad_norm": 0.5103345129350556, "learning_rate": 1.614982133284495e-06, "loss": 0.2616, "step": 784 }, { "epoch": 2.5079872204472844, "grad_norm": 0.49079582574254493, "learning_rate": 1.5947815483748573e-06, "loss": 0.2876, "step": 785 }, { "epoch": 2.511182108626198, "grad_norm": 0.5492629339702724, "learning_rate": 1.5746971443909143e-06, "loss": 0.2269, "step": 786 }, { "epoch": 2.5143769968051117, "grad_norm": 0.5084290858342608, "learning_rate": 1.5547291989488444e-06, "loss": 0.2656, "step": 787 }, { "epoch": 2.5175718849840254, "grad_norm": 0.4983976795381335, "learning_rate": 1.5348779880550812e-06, "loss": 0.2894, "step": 788 }, { "epoch": 2.520766773162939, "grad_norm": 0.5238056353329883, "learning_rate": 1.5151437861025032e-06, "loss": 0.2362, "step": 789 }, { "epoch": 2.523961661341853, "grad_norm": 0.518108096230708, "learning_rate": 1.49552686586663e-06, "loss": 0.2854, "step": 790 }, { "epoch": 2.527156549520767, "grad_norm": 0.5381430513007909, "learning_rate": 1.4760274985018619e-06, "loss": 0.2353, "step": 791 }, { "epoch": 2.5303514376996805, "grad_norm": 0.4929746328002691, "learning_rate": 1.4566459535377254e-06, "loss": 0.2667, "step": 792 }, { "epoch": 2.533546325878594, "grad_norm": 0.5046421029775986, "learning_rate": 1.4373824988751473e-06, "loss": 0.3179, "step": 793 }, { "epoch": 2.536741214057508, "grad_norm": 0.5157889946614715, "learning_rate": 1.4182374007827605e-06, "loss": 0.2662, "step": 794 }, { "epoch": 2.539936102236422, "grad_norm": 0.523857835711976, "learning_rate": 1.3992109238932106e-06, "loss": 0.2656, "step": 795 }, { "epoch": 2.543130990415335, "grad_norm": 0.5628959482047667, "learning_rate": 1.3803033311995072e-06, "loss": 0.3533, "step": 796 }, { "epoch": 2.5463258785942493, "grad_norm": 0.5176298006219728, "learning_rate": 1.361514884051388e-06, "loss": 0.2392, "step": 797 }, { "epoch": 2.549520766773163, "grad_norm": 0.5305179409745193, "learning_rate": 1.3428458421517031e-06, "loss": 0.2874, "step": 798 }, { "epoch": 2.5527156549520766, "grad_norm": 0.5133886968214696, "learning_rate": 1.324296463552821e-06, "loss": 0.2852, "step": 799 }, { "epoch": 2.5559105431309903, "grad_norm": 0.5299916695554814, "learning_rate": 1.3058670046530775e-06, "loss": 0.2755, "step": 800 }, { "epoch": 2.559105431309904, "grad_norm": 0.48847954753189066, "learning_rate": 1.2875577201932132e-06, "loss": 0.2565, "step": 801 }, { "epoch": 2.562300319488818, "grad_norm": 0.5247751090570331, "learning_rate": 1.2693688632528623e-06, "loss": 0.286, "step": 802 }, { "epoch": 2.5654952076677318, "grad_norm": 0.5452507249325443, "learning_rate": 1.2513006852470554e-06, "loss": 0.2564, "step": 803 }, { "epoch": 2.5686900958466454, "grad_norm": 0.5355228334038434, "learning_rate": 1.2333534359227383e-06, "loss": 0.248, "step": 804 }, { "epoch": 2.571884984025559, "grad_norm": 0.5077476188872533, "learning_rate": 1.2155273633553222e-06, "loss": 0.2174, "step": 805 }, { "epoch": 2.5750798722044728, "grad_norm": 0.5408303047151787, "learning_rate": 1.1978227139452624e-06, "loss": 0.2371, "step": 806 }, { "epoch": 2.5782747603833864, "grad_norm": 0.5799598867804451, "learning_rate": 1.1802397324146375e-06, "loss": 0.2508, "step": 807 }, { "epoch": 2.5814696485623, "grad_norm": 0.5223479303361819, "learning_rate": 1.1627786618037761e-06, "loss": 0.2861, "step": 808 }, { "epoch": 2.584664536741214, "grad_norm": 0.539996486264599, "learning_rate": 1.1454397434679022e-06, "loss": 0.2267, "step": 809 }, { "epoch": 2.587859424920128, "grad_norm": 0.5811749141190907, "learning_rate": 1.1282232170737862e-06, "loss": 0.2569, "step": 810 }, { "epoch": 2.5910543130990416, "grad_norm": 0.5179064965950383, "learning_rate": 1.1111293205964413e-06, "loss": 0.3336, "step": 811 }, { "epoch": 2.594249201277955, "grad_norm": 0.5171565494729409, "learning_rate": 1.0941582903158344e-06, "loss": 0.2548, "step": 812 }, { "epoch": 2.597444089456869, "grad_norm": 0.5396558702613092, "learning_rate": 1.0773103608136126e-06, "loss": 0.2928, "step": 813 }, { "epoch": 2.600638977635783, "grad_norm": 0.48371322074600365, "learning_rate": 1.060585764969867e-06, "loss": 0.2658, "step": 814 }, { "epoch": 2.6038338658146962, "grad_norm": 0.5173966192381088, "learning_rate": 1.0439847339599173e-06, "loss": 0.2858, "step": 815 }, { "epoch": 2.6070287539936103, "grad_norm": 0.6191072192150288, "learning_rate": 1.0275074972511034e-06, "loss": 0.3068, "step": 816 }, { "epoch": 2.610223642172524, "grad_norm": 0.49890486021543723, "learning_rate": 1.0111542825996245e-06, "loss": 0.2881, "step": 817 }, { "epoch": 2.6134185303514377, "grad_norm": 0.4826406747558694, "learning_rate": 9.949253160473914e-07, "loss": 0.3192, "step": 818 }, { "epoch": 2.6166134185303513, "grad_norm": 0.49492978407234944, "learning_rate": 9.788208219188932e-07, "loss": 0.2752, "step": 819 }, { "epoch": 2.619808306709265, "grad_norm": 0.5247273176934111, "learning_rate": 9.628410228181085e-07, "loss": 0.2852, "step": 820 }, { "epoch": 2.623003194888179, "grad_norm": 0.5305670666351503, "learning_rate": 9.469861396254154e-07, "loss": 0.2736, "step": 821 }, { "epoch": 2.626198083067093, "grad_norm": 0.5187931213108388, "learning_rate": 9.312563914945461e-07, "loss": 0.2896, "step": 822 }, { "epoch": 2.6293929712460065, "grad_norm": 0.49768159911322457, "learning_rate": 9.156519958495602e-07, "loss": 0.2846, "step": 823 }, { "epoch": 2.63258785942492, "grad_norm": 0.5290288470627704, "learning_rate": 9.001731683818338e-07, "loss": 0.277, "step": 824 }, { "epoch": 2.635782747603834, "grad_norm": 0.4902749919843502, "learning_rate": 8.848201230470777e-07, "loss": 0.3221, "step": 825 }, { "epoch": 2.6389776357827475, "grad_norm": 0.5040777631439952, "learning_rate": 8.695930720623857e-07, "loss": 0.3183, "step": 826 }, { "epoch": 2.642172523961661, "grad_norm": 0.5523788380454391, "learning_rate": 8.544922259032951e-07, "loss": 0.2696, "step": 827 }, { "epoch": 2.6453674121405752, "grad_norm": 0.5077339512969868, "learning_rate": 8.395177933008802e-07, "loss": 0.2786, "step": 828 }, { "epoch": 2.648562300319489, "grad_norm": 0.5148137411171129, "learning_rate": 8.246699812388714e-07, "loss": 0.2808, "step": 829 }, { "epoch": 2.6517571884984026, "grad_norm": 0.5326084811298537, "learning_rate": 8.099489949507843e-07, "loss": 0.3013, "step": 830 }, { "epoch": 2.6549520766773163, "grad_norm": 0.499399523060783, "learning_rate": 7.953550379170893e-07, "loss": 0.2509, "step": 831 }, { "epoch": 2.65814696485623, "grad_norm": 0.4895467308504159, "learning_rate": 7.808883118624012e-07, "loss": 0.2573, "step": 832 }, { "epoch": 2.661341853035144, "grad_norm": 0.46971413535806866, "learning_rate": 7.665490167526857e-07, "loss": 0.2546, "step": 833 }, { "epoch": 2.6645367412140573, "grad_norm": 0.5062579399060294, "learning_rate": 7.523373507924947e-07, "loss": 0.2895, "step": 834 }, { "epoch": 2.6677316293929714, "grad_norm": 0.49408089475640665, "learning_rate": 7.382535104222366e-07, "loss": 0.3381, "step": 835 }, { "epoch": 2.670926517571885, "grad_norm": 0.5079489630283403, "learning_rate": 7.242976903154442e-07, "loss": 0.2792, "step": 836 }, { "epoch": 2.6741214057507987, "grad_norm": 0.5367172556542773, "learning_rate": 7.104700833761014e-07, "loss": 0.2786, "step": 837 }, { "epoch": 2.6773162939297124, "grad_norm": 0.5173230966750194, "learning_rate": 6.967708807359664e-07, "loss": 0.2594, "step": 838 }, { "epoch": 2.680511182108626, "grad_norm": 0.5793340060415133, "learning_rate": 6.83200271751927e-07, "loss": 0.2284, "step": 839 }, { "epoch": 2.68370607028754, "grad_norm": 0.5594468901566473, "learning_rate": 6.697584440033989e-07, "loss": 0.2981, "step": 840 }, { "epoch": 2.686900958466454, "grad_norm": 0.49965854247621927, "learning_rate": 6.564455832897099e-07, "loss": 0.3002, "step": 841 }, { "epoch": 2.6900958466453675, "grad_norm": 0.626455409492764, "learning_rate": 6.432618736275553e-07, "loss": 0.2974, "step": 842 }, { "epoch": 2.693290734824281, "grad_norm": 0.5263238830296258, "learning_rate": 6.302074972484362e-07, "loss": 0.2628, "step": 843 }, { "epoch": 2.696485623003195, "grad_norm": 0.5454158557360471, "learning_rate": 6.17282634596148e-07, "loss": 0.2253, "step": 844 }, { "epoch": 2.6996805111821085, "grad_norm": 0.5073777391128576, "learning_rate": 6.044874643242904e-07, "loss": 0.2957, "step": 845 }, { "epoch": 2.702875399361022, "grad_norm": 0.4905589432687944, "learning_rate": 5.91822163293787e-07, "loss": 0.2735, "step": 846 }, { "epoch": 2.7060702875399363, "grad_norm": 0.5447677478507619, "learning_rate": 5.792869065704553e-07, "loss": 0.2489, "step": 847 }, { "epoch": 2.70926517571885, "grad_norm": 0.5367437184679952, "learning_rate": 5.668818674225684e-07, "loss": 0.2868, "step": 848 }, { "epoch": 2.7124600638977636, "grad_norm": 0.4634315253223173, "learning_rate": 5.546072173184791e-07, "loss": 0.2666, "step": 849 }, { "epoch": 2.7156549520766773, "grad_norm": 0.48495758547770107, "learning_rate": 5.424631259242352e-07, "loss": 0.2408, "step": 850 }, { "epoch": 2.718849840255591, "grad_norm": 0.5365013143998166, "learning_rate": 5.304497611012415e-07, "loss": 0.2742, "step": 851 }, { "epoch": 2.722044728434505, "grad_norm": 0.5619740516471143, "learning_rate": 5.185672889039395e-07, "loss": 0.2492, "step": 852 }, { "epoch": 2.7252396166134183, "grad_norm": 0.4959551051863538, "learning_rate": 5.068158735775098e-07, "loss": 0.2601, "step": 853 }, { "epoch": 2.7284345047923324, "grad_norm": 0.5614699103586969, "learning_rate": 4.951956775556e-07, "loss": 0.2893, "step": 854 }, { "epoch": 2.731629392971246, "grad_norm": 0.5478783753425022, "learning_rate": 4.837068614580875e-07, "loss": 0.2832, "step": 855 }, { "epoch": 2.7348242811501597, "grad_norm": 0.5084538253699109, "learning_rate": 4.7234958408884925e-07, "loss": 0.2785, "step": 856 }, { "epoch": 2.7380191693290734, "grad_norm": 0.4833191042095689, "learning_rate": 4.611240024335706e-07, "loss": 0.2525, "step": 857 }, { "epoch": 2.741214057507987, "grad_norm": 0.5557156275435591, "learning_rate": 4.5003027165758216e-07, "loss": 0.2763, "step": 858 }, { "epoch": 2.744408945686901, "grad_norm": 0.49396712066666637, "learning_rate": 4.3906854510370245e-07, "loss": 0.2961, "step": 859 }, { "epoch": 2.747603833865815, "grad_norm": 0.5218855996258477, "learning_rate": 4.282389742901283e-07, "loss": 0.3145, "step": 860 }, { "epoch": 2.7507987220447285, "grad_norm": 0.4909336132792226, "learning_rate": 4.1754170890833777e-07, "loss": 0.2901, "step": 861 }, { "epoch": 2.753993610223642, "grad_norm": 0.532978016110172, "learning_rate": 4.069768968210186e-07, "loss": 0.2767, "step": 862 }, { "epoch": 2.757188498402556, "grad_norm": 0.5284804740851335, "learning_rate": 3.96544684060024e-07, "loss": 0.3042, "step": 863 }, { "epoch": 2.7603833865814695, "grad_norm": 0.5625186952805096, "learning_rate": 3.862452148243623e-07, "loss": 0.3032, "step": 864 }, { "epoch": 2.763578274760383, "grad_norm": 0.5526705896948308, "learning_rate": 3.760786314781917e-07, "loss": 0.2162, "step": 865 }, { "epoch": 2.7667731629392973, "grad_norm": 0.49005060610573936, "learning_rate": 3.6604507454886083e-07, "loss": 0.3009, "step": 866 }, { "epoch": 2.769968051118211, "grad_norm": 0.5006073443686324, "learning_rate": 3.561446827249659e-07, "loss": 0.2863, "step": 867 }, { "epoch": 2.7731629392971247, "grad_norm": 0.5313378923234043, "learning_rate": 3.463775928544288e-07, "loss": 0.2681, "step": 868 }, { "epoch": 2.7763578274760383, "grad_norm": 0.45657247655517647, "learning_rate": 3.367439399426087e-07, "loss": 0.2833, "step": 869 }, { "epoch": 2.779552715654952, "grad_norm": 0.5438646304278568, "learning_rate": 3.2724385715043885e-07, "loss": 0.2821, "step": 870 }, { "epoch": 2.7827476038338657, "grad_norm": 0.4861986736846954, "learning_rate": 3.1787747579257623e-07, "loss": 0.302, "step": 871 }, { "epoch": 2.7859424920127793, "grad_norm": 0.5013004618464064, "learning_rate": 3.0864492533560167e-07, "loss": 0.3073, "step": 872 }, { "epoch": 2.7891373801916934, "grad_norm": 0.5139512438606135, "learning_rate": 2.9954633339621564e-07, "loss": 0.3061, "step": 873 }, { "epoch": 2.792332268370607, "grad_norm": 0.4967084848014566, "learning_rate": 2.905818257394799e-07, "loss": 0.2385, "step": 874 }, { "epoch": 2.7955271565495208, "grad_norm": 0.5583189331928233, "learning_rate": 2.8175152627708425e-07, "loss": 0.2576, "step": 875 }, { "epoch": 2.7987220447284344, "grad_norm": 0.5220482463238036, "learning_rate": 2.730555570656246e-07, "loss": 0.2427, "step": 876 }, { "epoch": 2.801916932907348, "grad_norm": 0.47836855934512534, "learning_rate": 2.6449403830492105e-07, "loss": 0.2559, "step": 877 }, { "epoch": 2.8051118210862622, "grad_norm": 0.49773019706098176, "learning_rate": 2.560670883363592e-07, "loss": 0.2364, "step": 878 }, { "epoch": 2.8083067092651754, "grad_norm": 0.49288457060529334, "learning_rate": 2.4777482364124695e-07, "loss": 0.2964, "step": 879 }, { "epoch": 2.8115015974440896, "grad_norm": 0.5198746225532588, "learning_rate": 2.39617358839207e-07, "loss": 0.2658, "step": 880 }, { "epoch": 2.8146964856230032, "grad_norm": 0.4799151728900049, "learning_rate": 2.315948066866003e-07, "loss": 0.2402, "step": 881 }, { "epoch": 2.817891373801917, "grad_norm": 0.465390503048871, "learning_rate": 2.2370727807495496e-07, "loss": 0.3056, "step": 882 }, { "epoch": 2.8210862619808306, "grad_norm": 0.5382332293961877, "learning_rate": 2.1595488202944104e-07, "loss": 0.2653, "step": 883 }, { "epoch": 2.8242811501597442, "grad_norm": 0.5068975952175465, "learning_rate": 2.0833772570736376e-07, "loss": 0.3053, "step": 884 }, { "epoch": 2.8274760383386583, "grad_norm": 0.4757131988181981, "learning_rate": 2.0085591439667928e-07, "loss": 0.346, "step": 885 }, { "epoch": 2.830670926517572, "grad_norm": 0.5434560209318039, "learning_rate": 1.935095515145391e-07, "loss": 0.2642, "step": 886 }, { "epoch": 2.8338658146964857, "grad_norm": 0.5099744081674852, "learning_rate": 1.8629873860586567e-07, "loss": 0.2416, "step": 887 }, { "epoch": 2.8370607028753994, "grad_norm": 0.46460652007791287, "learning_rate": 1.7922357534194356e-07, "loss": 0.2865, "step": 888 }, { "epoch": 2.840255591054313, "grad_norm": 0.6038585562899795, "learning_rate": 1.7228415951904165e-07, "loss": 0.317, "step": 889 }, { "epoch": 2.8434504792332267, "grad_norm": 0.4970433072203688, "learning_rate": 1.6548058705706528e-07, "loss": 0.2634, "step": 890 }, { "epoch": 2.8466453674121404, "grad_norm": 0.45087779573435216, "learning_rate": 1.5881295199822953e-07, "loss": 0.2869, "step": 891 }, { "epoch": 2.8498402555910545, "grad_norm": 0.4712297048952234, "learning_rate": 1.5228134650575265e-07, "loss": 0.3082, "step": 892 }, { "epoch": 2.853035143769968, "grad_norm": 0.5542853030279212, "learning_rate": 1.458858608625957e-07, "loss": 0.2609, "step": 893 }, { "epoch": 2.856230031948882, "grad_norm": 0.5299351884184478, "learning_rate": 1.3962658347019819e-07, "loss": 0.2796, "step": 894 }, { "epoch": 2.8594249201277955, "grad_norm": 1.7034789905383376, "learning_rate": 1.335036008472701e-07, "loss": 0.3549, "step": 895 }, { "epoch": 2.862619808306709, "grad_norm": 0.4941137486267854, "learning_rate": 1.2751699762858837e-07, "loss": 0.2473, "step": 896 }, { "epoch": 2.8658146964856233, "grad_norm": 0.4652931456111103, "learning_rate": 1.2166685656382905e-07, "loss": 0.3353, "step": 897 }, { "epoch": 2.8690095846645365, "grad_norm": 0.4942389320669494, "learning_rate": 1.1595325851642137e-07, "loss": 0.2944, "step": 898 }, { "epoch": 2.8722044728434506, "grad_norm": 0.48547745486914246, "learning_rate": 1.103762824624377e-07, "loss": 0.2688, "step": 899 }, { "epoch": 2.8753993610223643, "grad_norm": 0.5753452793142592, "learning_rate": 1.0493600548948879e-07, "loss": 0.2566, "step": 900 }, { "epoch": 2.878594249201278, "grad_norm": 0.4972571947816792, "learning_rate": 9.963250279567239e-08, "loss": 0.2943, "step": 901 }, { "epoch": 2.8817891373801916, "grad_norm": 0.5467418327312579, "learning_rate": 9.446584768852407e-08, "loss": 0.3408, "step": 902 }, { "epoch": 2.8849840255591053, "grad_norm": 0.540785739575046, "learning_rate": 8.943611158400479e-08, "loss": 0.2589, "step": 903 }, { "epoch": 2.8881789137380194, "grad_norm": 0.4998267472039897, "learning_rate": 8.454336400552154e-08, "loss": 0.302, "step": 904 }, { "epoch": 2.891373801916933, "grad_norm": 0.5000407109550236, "learning_rate": 7.978767258295494e-08, "loss": 0.2799, "step": 905 }, { "epoch": 2.8945686900958467, "grad_norm": 0.4826382651974826, "learning_rate": 7.51691030517343e-08, "loss": 0.3049, "step": 906 }, { "epoch": 2.8977635782747604, "grad_norm": 0.5533789623454506, "learning_rate": 7.068771925192286e-08, "loss": 0.2886, "step": 907 }, { "epoch": 2.900958466453674, "grad_norm": 0.4856062356070542, "learning_rate": 6.634358312733957e-08, "loss": 0.3344, "step": 908 }, { "epoch": 2.9041533546325877, "grad_norm": 0.5234411083127519, "learning_rate": 6.21367547246976e-08, "loss": 0.2975, "step": 909 }, { "epoch": 2.9073482428115014, "grad_norm": 0.5042769247684148, "learning_rate": 5.806729219278051e-08, "loss": 0.3385, "step": 910 }, { "epoch": 2.9105431309904155, "grad_norm": 0.4910677124873368, "learning_rate": 5.413525178163292e-08, "loss": 0.3113, "step": 911 }, { "epoch": 2.913738019169329, "grad_norm": 0.5181476682018931, "learning_rate": 5.034068784178892e-08, "loss": 0.2742, "step": 912 }, { "epoch": 2.916932907348243, "grad_norm": 0.5220994506844163, "learning_rate": 4.6683652823513725e-08, "loss": 0.2551, "step": 913 }, { "epoch": 2.9201277955271565, "grad_norm": 0.5176699460691047, "learning_rate": 4.316419727608434e-08, "loss": 0.2465, "step": 914 }, { "epoch": 2.92332268370607, "grad_norm": 0.5253749788385152, "learning_rate": 3.9782369847088944e-08, "loss": 0.2342, "step": 915 }, { "epoch": 2.9265175718849843, "grad_norm": 0.49606392975017294, "learning_rate": 3.653821728175522e-08, "loss": 0.2832, "step": 916 }, { "epoch": 2.9297124600638975, "grad_norm": 0.48632265213078496, "learning_rate": 3.3431784422300884e-08, "loss": 0.2869, "step": 917 }, { "epoch": 2.9329073482428116, "grad_norm": 0.5506996313847263, "learning_rate": 3.046311420731751e-08, "loss": 0.2121, "step": 918 }, { "epoch": 2.9361022364217253, "grad_norm": 0.49721273548429923, "learning_rate": 2.763224767117767e-08, "loss": 0.3057, "step": 919 }, { "epoch": 2.939297124600639, "grad_norm": 0.5242997933141604, "learning_rate": 2.4939223943463153e-08, "loss": 0.2932, "step": 920 }, { "epoch": 2.9424920127795526, "grad_norm": 0.48394661634360336, "learning_rate": 2.2384080248429863e-08, "loss": 0.3203, "step": 921 }, { "epoch": 2.9456869009584663, "grad_norm": 0.5321486189218536, "learning_rate": 1.9966851904487104e-08, "loss": 0.2875, "step": 922 }, { "epoch": 2.9488817891373804, "grad_norm": 0.54904053688612, "learning_rate": 1.768757232371576e-08, "loss": 0.2645, "step": 923 }, { "epoch": 2.952076677316294, "grad_norm": 0.48065835080140534, "learning_rate": 1.554627301140199e-08, "loss": 0.2675, "step": 924 }, { "epoch": 2.9552715654952078, "grad_norm": 0.5041424505746223, "learning_rate": 1.3542983565600909e-08, "loss": 0.2912, "step": 925 }, { "epoch": 2.9584664536741214, "grad_norm": 0.5222550703481469, "learning_rate": 1.1677731676733584e-08, "loss": 0.2605, "step": 926 }, { "epoch": 2.961661341853035, "grad_norm": 0.5298221402096054, "learning_rate": 9.950543127198454e-09, "loss": 0.2484, "step": 927 }, { "epoch": 2.9648562300319488, "grad_norm": 0.5060955481899465, "learning_rate": 8.361441791016056e-09, "loss": 0.2615, "step": 928 }, { "epoch": 2.9680511182108624, "grad_norm": 0.5013811699156504, "learning_rate": 6.910449633501515e-09, "loss": 0.2774, "step": 929 }, { "epoch": 2.9712460063897765, "grad_norm": 0.49244223044762647, "learning_rate": 5.597586710957004e-09, "loss": 0.323, "step": 930 }, { "epoch": 2.97444089456869, "grad_norm": 0.46269666655252023, "learning_rate": 4.422871170398635e-09, "loss": 0.2762, "step": 931 }, { "epoch": 2.977635782747604, "grad_norm": 0.4975777250067196, "learning_rate": 3.386319249303327e-09, "loss": 0.2915, "step": 932 }, { "epoch": 2.9808306709265175, "grad_norm": 0.4774284368895711, "learning_rate": 2.48794527538454e-09, "loss": 0.3059, "step": 933 }, { "epoch": 2.984025559105431, "grad_norm": 0.5051541199990559, "learning_rate": 1.7277616663946562e-09, "loss": 0.2846, "step": 934 }, { "epoch": 2.987220447284345, "grad_norm": 0.5314827456627474, "learning_rate": 1.1057789299517841e-09, "loss": 0.2911, "step": 935 }, { "epoch": 2.9904153354632586, "grad_norm": 0.5364041253421318, "learning_rate": 6.220056633987615e-10, "loss": 0.2783, "step": 936 }, { "epoch": 2.9936102236421727, "grad_norm": 0.5356243310916116, "learning_rate": 2.764485536776995e-10, "loss": 0.3151, "step": 937 }, { "epoch": 2.9968051118210863, "grad_norm": 0.49531175646741604, "learning_rate": 6.911237724560593e-11, "loss": 0.2669, "step": 938 }, { "epoch": 3.0, "grad_norm": 0.47710393066482504, "learning_rate": 0.0, "loss": 0.262, "step": 939 }, { "epoch": 3.0, "step": 939, "total_flos": 406391461183488.0, "train_loss": 0.4046504932661026, "train_runtime": 10117.0351, "train_samples_per_second": 2.964, "train_steps_per_second": 0.093 } ], "logging_steps": 1.0, "max_steps": 939, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 406391461183488.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }