{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002, "grad_norm": 226.99563598632812, "learning_rate": 6.666666666666668e-08, "loss": 21.9448, "step": 1 }, { "epoch": 0.004, "grad_norm": 236.27398681640625, "learning_rate": 1.3333333333333336e-07, "loss": 22.6687, "step": 2 }, { "epoch": 0.006, "grad_norm": 243.7906036376953, "learning_rate": 2.0000000000000002e-07, "loss": 23.7182, "step": 3 }, { "epoch": 0.008, "grad_norm": 233.99176025390625, "learning_rate": 2.666666666666667e-07, "loss": 22.4062, "step": 4 }, { "epoch": 0.01, "grad_norm": 247.73858642578125, "learning_rate": 3.3333333333333335e-07, "loss": 23.4446, "step": 5 }, { "epoch": 0.012, "grad_norm": 224.78480529785156, "learning_rate": 4.0000000000000003e-07, "loss": 22.0816, "step": 6 }, { "epoch": 0.014, "grad_norm": 239.08424377441406, "learning_rate": 4.666666666666667e-07, "loss": 22.2544, "step": 7 }, { "epoch": 0.016, "grad_norm": 243.3548126220703, "learning_rate": 5.333333333333335e-07, "loss": 22.3676, "step": 8 }, { "epoch": 0.018, "grad_norm": 228.88418579101562, "learning_rate": 6.000000000000001e-07, "loss": 21.8324, "step": 9 }, { "epoch": 0.02, "grad_norm": 255.19561767578125, "learning_rate": 6.666666666666667e-07, "loss": 21.7817, "step": 10 }, { "epoch": 0.022, "grad_norm": 256.03631591796875, "learning_rate": 7.333333333333334e-07, "loss": 22.0901, "step": 11 }, { "epoch": 0.024, "grad_norm": 256.8511047363281, "learning_rate": 8.000000000000001e-07, "loss": 22.1082, "step": 12 }, { "epoch": 0.026, "grad_norm": 276.2884521484375, "learning_rate": 8.666666666666668e-07, "loss": 20.2445, "step": 13 }, { "epoch": 0.028, "grad_norm": 274.6839599609375, "learning_rate": 9.333333333333334e-07, "loss": 18.7186, "step": 14 }, { "epoch": 0.03, "grad_norm": 282.9738464355469, "learning_rate": 1.0000000000000002e-06, "loss": 19.2151, "step": 15 }, { "epoch": 0.032, "grad_norm": 273.26934814453125, "learning_rate": 1.066666666666667e-06, "loss": 18.8715, "step": 16 }, { "epoch": 0.034, "grad_norm": 222.71524047851562, "learning_rate": 1.1333333333333334e-06, "loss": 14.6642, "step": 17 }, { "epoch": 0.036, "grad_norm": 218.20606994628906, "learning_rate": 1.2000000000000002e-06, "loss": 13.549, "step": 18 }, { "epoch": 0.038, "grad_norm": 229.51235961914062, "learning_rate": 1.2666666666666669e-06, "loss": 13.3927, "step": 19 }, { "epoch": 0.04, "grad_norm": 228.68344116210938, "learning_rate": 1.3333333333333334e-06, "loss": 12.2413, "step": 20 }, { "epoch": 0.042, "grad_norm": 222.99270629882812, "learning_rate": 1.4000000000000001e-06, "loss": 11.1357, "step": 21 }, { "epoch": 0.044, "grad_norm": 236.81912231445312, "learning_rate": 1.4666666666666669e-06, "loss": 10.6309, "step": 22 }, { "epoch": 0.046, "grad_norm": 253.3773193359375, "learning_rate": 1.5333333333333334e-06, "loss": 6.1736, "step": 23 }, { "epoch": 0.048, "grad_norm": 132.65660095214844, "learning_rate": 1.6000000000000001e-06, "loss": 3.7401, "step": 24 }, { "epoch": 0.05, "grad_norm": 171.21932983398438, "learning_rate": 1.6666666666666667e-06, "loss": 3.2601, "step": 25 }, { "epoch": 0.052, "grad_norm": 112.79869079589844, "learning_rate": 1.7333333333333336e-06, "loss": 2.6378, "step": 26 }, { "epoch": 0.054, "grad_norm": 72.08856201171875, "learning_rate": 1.8000000000000001e-06, "loss": 2.2122, "step": 27 }, { "epoch": 0.056, "grad_norm": 47.903202056884766, "learning_rate": 1.8666666666666669e-06, "loss": 1.9545, "step": 28 }, { "epoch": 0.058, "grad_norm": 42.68925094604492, "learning_rate": 1.9333333333333336e-06, "loss": 1.909, "step": 29 }, { "epoch": 0.06, "grad_norm": 37.55009460449219, "learning_rate": 2.0000000000000003e-06, "loss": 1.7777, "step": 30 }, { "epoch": 0.062, "grad_norm": 27.396568298339844, "learning_rate": 2.0666666666666666e-06, "loss": 1.8359, "step": 31 }, { "epoch": 0.064, "grad_norm": 30.148645401000977, "learning_rate": 2.133333333333334e-06, "loss": 1.7342, "step": 32 }, { "epoch": 0.066, "grad_norm": 28.60604476928711, "learning_rate": 2.2e-06, "loss": 1.8821, "step": 33 }, { "epoch": 0.068, "grad_norm": 22.30388641357422, "learning_rate": 2.266666666666667e-06, "loss": 1.6301, "step": 34 }, { "epoch": 0.07, "grad_norm": 28.496145248413086, "learning_rate": 2.3333333333333336e-06, "loss": 1.8361, "step": 35 }, { "epoch": 0.072, "grad_norm": 34.518287658691406, "learning_rate": 2.4000000000000003e-06, "loss": 1.6929, "step": 36 }, { "epoch": 0.074, "grad_norm": 23.480363845825195, "learning_rate": 2.466666666666667e-06, "loss": 1.5912, "step": 37 }, { "epoch": 0.076, "grad_norm": 26.371000289916992, "learning_rate": 2.5333333333333338e-06, "loss": 1.6452, "step": 38 }, { "epoch": 0.078, "grad_norm": 26.6358585357666, "learning_rate": 2.6e-06, "loss": 1.7357, "step": 39 }, { "epoch": 0.08, "grad_norm": 21.163368225097656, "learning_rate": 2.666666666666667e-06, "loss": 1.8135, "step": 40 }, { "epoch": 0.082, "grad_norm": 38.06758499145508, "learning_rate": 2.7333333333333336e-06, "loss": 1.7483, "step": 41 }, { "epoch": 0.084, "grad_norm": 20.108041763305664, "learning_rate": 2.8000000000000003e-06, "loss": 1.7272, "step": 42 }, { "epoch": 0.086, "grad_norm": 37.66496658325195, "learning_rate": 2.866666666666667e-06, "loss": 1.8263, "step": 43 }, { "epoch": 0.088, "grad_norm": 17.51531219482422, "learning_rate": 2.9333333333333338e-06, "loss": 1.7223, "step": 44 }, { "epoch": 0.09, "grad_norm": 22.1541748046875, "learning_rate": 3e-06, "loss": 1.6006, "step": 45 }, { "epoch": 0.092, "grad_norm": 19.2305908203125, "learning_rate": 3.066666666666667e-06, "loss": 1.6562, "step": 46 }, { "epoch": 0.094, "grad_norm": 16.655160903930664, "learning_rate": 3.133333333333334e-06, "loss": 1.7491, "step": 47 }, { "epoch": 0.096, "grad_norm": 20.247623443603516, "learning_rate": 3.2000000000000003e-06, "loss": 1.6449, "step": 48 }, { "epoch": 0.098, "grad_norm": 17.47651481628418, "learning_rate": 3.266666666666667e-06, "loss": 1.6761, "step": 49 }, { "epoch": 0.1, "grad_norm": 26.51736068725586, "learning_rate": 3.3333333333333333e-06, "loss": 1.7986, "step": 50 }, { "epoch": 0.102, "grad_norm": 21.225101470947266, "learning_rate": 3.4000000000000005e-06, "loss": 1.6329, "step": 51 }, { "epoch": 0.104, "grad_norm": 22.989187240600586, "learning_rate": 3.4666666666666672e-06, "loss": 1.7155, "step": 52 }, { "epoch": 0.106, "grad_norm": 18.516536712646484, "learning_rate": 3.5333333333333335e-06, "loss": 1.6831, "step": 53 }, { "epoch": 0.108, "grad_norm": 19.069442749023438, "learning_rate": 3.6000000000000003e-06, "loss": 1.6225, "step": 54 }, { "epoch": 0.11, "grad_norm": 20.60556983947754, "learning_rate": 3.6666666666666666e-06, "loss": 1.6726, "step": 55 }, { "epoch": 0.112, "grad_norm": 19.45859146118164, "learning_rate": 3.7333333333333337e-06, "loss": 1.6408, "step": 56 }, { "epoch": 0.114, "grad_norm": 26.011234283447266, "learning_rate": 3.8000000000000005e-06, "loss": 1.6057, "step": 57 }, { "epoch": 0.116, "grad_norm": 17.240083694458008, "learning_rate": 3.866666666666667e-06, "loss": 1.677, "step": 58 }, { "epoch": 0.118, "grad_norm": 22.089248657226562, "learning_rate": 3.9333333333333335e-06, "loss": 1.676, "step": 59 }, { "epoch": 0.12, "grad_norm": 33.735233306884766, "learning_rate": 4.000000000000001e-06, "loss": 1.6742, "step": 60 }, { "epoch": 0.122, "grad_norm": 27.743314743041992, "learning_rate": 4.066666666666667e-06, "loss": 1.6956, "step": 61 }, { "epoch": 0.124, "grad_norm": 25.890151977539062, "learning_rate": 4.133333333333333e-06, "loss": 1.56, "step": 62 }, { "epoch": 0.126, "grad_norm": 16.895370483398438, "learning_rate": 4.2000000000000004e-06, "loss": 1.5794, "step": 63 }, { "epoch": 0.128, "grad_norm": 19.82956314086914, "learning_rate": 4.266666666666668e-06, "loss": 1.4879, "step": 64 }, { "epoch": 0.13, "grad_norm": 15.571148872375488, "learning_rate": 4.333333333333334e-06, "loss": 1.7724, "step": 65 }, { "epoch": 0.132, "grad_norm": 14.684285163879395, "learning_rate": 4.4e-06, "loss": 1.6943, "step": 66 }, { "epoch": 0.134, "grad_norm": 13.911795616149902, "learning_rate": 4.4666666666666665e-06, "loss": 1.6627, "step": 67 }, { "epoch": 0.136, "grad_norm": 15.752901077270508, "learning_rate": 4.533333333333334e-06, "loss": 1.7536, "step": 68 }, { "epoch": 0.138, "grad_norm": 24.22060203552246, "learning_rate": 4.600000000000001e-06, "loss": 1.6225, "step": 69 }, { "epoch": 0.14, "grad_norm": 13.675880432128906, "learning_rate": 4.666666666666667e-06, "loss": 1.6696, "step": 70 }, { "epoch": 0.142, "grad_norm": 20.69804573059082, "learning_rate": 4.7333333333333335e-06, "loss": 1.6295, "step": 71 }, { "epoch": 0.144, "grad_norm": 28.758134841918945, "learning_rate": 4.800000000000001e-06, "loss": 1.6698, "step": 72 }, { "epoch": 0.146, "grad_norm": 14.652379035949707, "learning_rate": 4.866666666666667e-06, "loss": 1.606, "step": 73 }, { "epoch": 0.148, "grad_norm": 22.868322372436523, "learning_rate": 4.933333333333334e-06, "loss": 1.6168, "step": 74 }, { "epoch": 0.15, "grad_norm": 14.871725082397461, "learning_rate": 5e-06, "loss": 1.6506, "step": 75 }, { "epoch": 0.152, "grad_norm": 21.552318572998047, "learning_rate": 5.0666666666666676e-06, "loss": 1.6366, "step": 76 }, { "epoch": 0.154, "grad_norm": 15.857409477233887, "learning_rate": 5.133333333333334e-06, "loss": 1.6794, "step": 77 }, { "epoch": 0.156, "grad_norm": 17.261104583740234, "learning_rate": 5.2e-06, "loss": 1.6165, "step": 78 }, { "epoch": 0.158, "grad_norm": 14.66672420501709, "learning_rate": 5.2666666666666665e-06, "loss": 1.6562, "step": 79 }, { "epoch": 0.16, "grad_norm": 22.905376434326172, "learning_rate": 5.333333333333334e-06, "loss": 1.6468, "step": 80 }, { "epoch": 0.162, "grad_norm": 17.427703857421875, "learning_rate": 5.400000000000001e-06, "loss": 1.6618, "step": 81 }, { "epoch": 0.164, "grad_norm": 15.554149627685547, "learning_rate": 5.466666666666667e-06, "loss": 1.6368, "step": 82 }, { "epoch": 0.166, "grad_norm": 16.746028900146484, "learning_rate": 5.533333333333334e-06, "loss": 1.6583, "step": 83 }, { "epoch": 0.168, "grad_norm": 17.273025512695312, "learning_rate": 5.600000000000001e-06, "loss": 1.625, "step": 84 }, { "epoch": 0.17, "grad_norm": 20.172046661376953, "learning_rate": 5.666666666666667e-06, "loss": 1.6733, "step": 85 }, { "epoch": 0.172, "grad_norm": 13.854218482971191, "learning_rate": 5.733333333333334e-06, "loss": 1.649, "step": 86 }, { "epoch": 0.174, "grad_norm": 28.214906692504883, "learning_rate": 5.8e-06, "loss": 1.5396, "step": 87 }, { "epoch": 0.176, "grad_norm": 17.772258758544922, "learning_rate": 5.8666666666666675e-06, "loss": 1.5781, "step": 88 }, { "epoch": 0.178, "grad_norm": 14.36558723449707, "learning_rate": 5.933333333333335e-06, "loss": 1.6857, "step": 89 }, { "epoch": 0.18, "grad_norm": 15.606388092041016, "learning_rate": 6e-06, "loss": 1.6192, "step": 90 }, { "epoch": 0.182, "grad_norm": 16.106149673461914, "learning_rate": 6.066666666666667e-06, "loss": 1.568, "step": 91 }, { "epoch": 0.184, "grad_norm": 11.074296951293945, "learning_rate": 6.133333333333334e-06, "loss": 1.631, "step": 92 }, { "epoch": 0.186, "grad_norm": 15.625429153442383, "learning_rate": 6.200000000000001e-06, "loss": 1.6064, "step": 93 }, { "epoch": 0.188, "grad_norm": 15.9578218460083, "learning_rate": 6.266666666666668e-06, "loss": 1.6646, "step": 94 }, { "epoch": 0.19, "grad_norm": 19.56605339050293, "learning_rate": 6.333333333333333e-06, "loss": 1.5587, "step": 95 }, { "epoch": 0.192, "grad_norm": 14.848856925964355, "learning_rate": 6.4000000000000006e-06, "loss": 1.6036, "step": 96 }, { "epoch": 0.194, "grad_norm": 22.246797561645508, "learning_rate": 6.466666666666667e-06, "loss": 1.6709, "step": 97 }, { "epoch": 0.196, "grad_norm": 18.220890045166016, "learning_rate": 6.533333333333334e-06, "loss": 1.6603, "step": 98 }, { "epoch": 0.198, "grad_norm": 18.9663028717041, "learning_rate": 6.600000000000001e-06, "loss": 1.7576, "step": 99 }, { "epoch": 0.2, "grad_norm": 22.304872512817383, "learning_rate": 6.666666666666667e-06, "loss": 1.7278, "step": 100 }, { "epoch": 0.202, "grad_norm": 22.000619888305664, "learning_rate": 6.733333333333334e-06, "loss": 1.7162, "step": 101 }, { "epoch": 0.204, "grad_norm": 19.02189064025879, "learning_rate": 6.800000000000001e-06, "loss": 1.6294, "step": 102 }, { "epoch": 0.206, "grad_norm": 17.57056999206543, "learning_rate": 6.866666666666667e-06, "loss": 1.651, "step": 103 }, { "epoch": 0.208, "grad_norm": 20.44572639465332, "learning_rate": 6.9333333333333344e-06, "loss": 1.6798, "step": 104 }, { "epoch": 0.21, "grad_norm": 21.554231643676758, "learning_rate": 7e-06, "loss": 1.676, "step": 105 }, { "epoch": 0.212, "grad_norm": 18.977502822875977, "learning_rate": 7.066666666666667e-06, "loss": 1.6736, "step": 106 }, { "epoch": 0.214, "grad_norm": 18.957950592041016, "learning_rate": 7.133333333333334e-06, "loss": 1.7151, "step": 107 }, { "epoch": 0.216, "grad_norm": 19.2664852142334, "learning_rate": 7.2000000000000005e-06, "loss": 1.6875, "step": 108 }, { "epoch": 0.218, "grad_norm": 18.646236419677734, "learning_rate": 7.266666666666668e-06, "loss": 1.6509, "step": 109 }, { "epoch": 0.22, "grad_norm": 21.31477928161621, "learning_rate": 7.333333333333333e-06, "loss": 1.597, "step": 110 }, { "epoch": 0.222, "grad_norm": 17.566238403320312, "learning_rate": 7.4e-06, "loss": 1.7574, "step": 111 }, { "epoch": 0.224, "grad_norm": 13.890412330627441, "learning_rate": 7.4666666666666675e-06, "loss": 1.5845, "step": 112 }, { "epoch": 0.226, "grad_norm": 16.55223846435547, "learning_rate": 7.533333333333334e-06, "loss": 1.5888, "step": 113 }, { "epoch": 0.228, "grad_norm": 26.499820709228516, "learning_rate": 7.600000000000001e-06, "loss": 1.584, "step": 114 }, { "epoch": 0.23, "grad_norm": 31.17205810546875, "learning_rate": 7.666666666666667e-06, "loss": 1.6443, "step": 115 }, { "epoch": 0.232, "grad_norm": 31.60078239440918, "learning_rate": 7.733333333333334e-06, "loss": 1.7377, "step": 116 }, { "epoch": 0.234, "grad_norm": 27.893613815307617, "learning_rate": 7.800000000000002e-06, "loss": 1.779, "step": 117 }, { "epoch": 0.236, "grad_norm": 16.482492446899414, "learning_rate": 7.866666666666667e-06, "loss": 1.7002, "step": 118 }, { "epoch": 0.238, "grad_norm": 24.259340286254883, "learning_rate": 7.933333333333334e-06, "loss": 1.5989, "step": 119 }, { "epoch": 0.24, "grad_norm": 18.561723709106445, "learning_rate": 8.000000000000001e-06, "loss": 1.7336, "step": 120 }, { "epoch": 0.242, "grad_norm": 15.461645126342773, "learning_rate": 8.066666666666667e-06, "loss": 1.7121, "step": 121 }, { "epoch": 0.244, "grad_norm": 22.555328369140625, "learning_rate": 8.133333333333334e-06, "loss": 1.7554, "step": 122 }, { "epoch": 0.246, "grad_norm": 14.69766902923584, "learning_rate": 8.2e-06, "loss": 1.6397, "step": 123 }, { "epoch": 0.248, "grad_norm": 15.5711030960083, "learning_rate": 8.266666666666667e-06, "loss": 1.7016, "step": 124 }, { "epoch": 0.25, "grad_norm": 10.80448055267334, "learning_rate": 8.333333333333334e-06, "loss": 1.6544, "step": 125 }, { "epoch": 0.252, "grad_norm": 18.54620361328125, "learning_rate": 8.400000000000001e-06, "loss": 1.623, "step": 126 }, { "epoch": 0.254, "grad_norm": 34.518768310546875, "learning_rate": 8.466666666666668e-06, "loss": 1.6958, "step": 127 }, { "epoch": 0.256, "grad_norm": 21.02113151550293, "learning_rate": 8.533333333333335e-06, "loss": 1.7662, "step": 128 }, { "epoch": 0.258, "grad_norm": 14.756684303283691, "learning_rate": 8.6e-06, "loss": 1.5814, "step": 129 }, { "epoch": 0.26, "grad_norm": 16.498056411743164, "learning_rate": 8.666666666666668e-06, "loss": 1.7248, "step": 130 }, { "epoch": 0.262, "grad_norm": 19.80323028564453, "learning_rate": 8.733333333333333e-06, "loss": 1.7451, "step": 131 }, { "epoch": 0.264, "grad_norm": 19.826833724975586, "learning_rate": 8.8e-06, "loss": 1.6475, "step": 132 }, { "epoch": 0.266, "grad_norm": 15.338292121887207, "learning_rate": 8.866666666666668e-06, "loss": 1.623, "step": 133 }, { "epoch": 0.268, "grad_norm": 12.683643341064453, "learning_rate": 8.933333333333333e-06, "loss": 1.6932, "step": 134 }, { "epoch": 0.27, "grad_norm": 15.072694778442383, "learning_rate": 9e-06, "loss": 1.6978, "step": 135 }, { "epoch": 0.272, "grad_norm": 17.20955467224121, "learning_rate": 9.066666666666667e-06, "loss": 1.6464, "step": 136 }, { "epoch": 0.274, "grad_norm": 16.128379821777344, "learning_rate": 9.133333333333335e-06, "loss": 1.6678, "step": 137 }, { "epoch": 0.276, "grad_norm": 15.234736442565918, "learning_rate": 9.200000000000002e-06, "loss": 1.4748, "step": 138 }, { "epoch": 0.278, "grad_norm": 12.608988761901855, "learning_rate": 9.266666666666667e-06, "loss": 1.6285, "step": 139 }, { "epoch": 0.28, "grad_norm": 20.843536376953125, "learning_rate": 9.333333333333334e-06, "loss": 1.699, "step": 140 }, { "epoch": 0.282, "grad_norm": 18.524633407592773, "learning_rate": 9.4e-06, "loss": 1.6608, "step": 141 }, { "epoch": 0.284, "grad_norm": 13.68975830078125, "learning_rate": 9.466666666666667e-06, "loss": 1.6153, "step": 142 }, { "epoch": 0.286, "grad_norm": 14.197174072265625, "learning_rate": 9.533333333333334e-06, "loss": 1.7042, "step": 143 }, { "epoch": 0.288, "grad_norm": 18.283809661865234, "learning_rate": 9.600000000000001e-06, "loss": 1.7929, "step": 144 }, { "epoch": 0.29, "grad_norm": 14.084982872009277, "learning_rate": 9.666666666666667e-06, "loss": 1.7448, "step": 145 }, { "epoch": 0.292, "grad_norm": 15.236905097961426, "learning_rate": 9.733333333333334e-06, "loss": 1.7063, "step": 146 }, { "epoch": 0.294, "grad_norm": 14.99785327911377, "learning_rate": 9.800000000000001e-06, "loss": 1.6991, "step": 147 }, { "epoch": 0.296, "grad_norm": 18.394018173217773, "learning_rate": 9.866666666666668e-06, "loss": 1.5488, "step": 148 }, { "epoch": 0.298, "grad_norm": 11.244776725769043, "learning_rate": 9.933333333333334e-06, "loss": 1.6293, "step": 149 }, { "epoch": 0.3, "grad_norm": 20.61504364013672, "learning_rate": 1e-05, "loss": 1.703, "step": 150 }, { "epoch": 0.302, "grad_norm": 20.268144607543945, "learning_rate": 9.99998646145412e-06, "loss": 1.8211, "step": 151 }, { "epoch": 0.304, "grad_norm": 17.08184242248535, "learning_rate": 9.999945845889795e-06, "loss": 1.5821, "step": 152 }, { "epoch": 0.306, "grad_norm": 18.60841178894043, "learning_rate": 9.999878153526974e-06, "loss": 1.7683, "step": 153 }, { "epoch": 0.308, "grad_norm": 11.097747802734375, "learning_rate": 9.999783384732242e-06, "loss": 1.6726, "step": 154 }, { "epoch": 0.31, "grad_norm": 24.11589241027832, "learning_rate": 9.999661540018812e-06, "loss": 1.7708, "step": 155 }, { "epoch": 0.312, "grad_norm": 14.22172737121582, "learning_rate": 9.999512620046523e-06, "loss": 1.6218, "step": 156 }, { "epoch": 0.314, "grad_norm": 15.114810943603516, "learning_rate": 9.999336625621836e-06, "loss": 1.7024, "step": 157 }, { "epoch": 0.316, "grad_norm": 11.347399711608887, "learning_rate": 9.99913355769784e-06, "loss": 1.7575, "step": 158 }, { "epoch": 0.318, "grad_norm": 13.826147079467773, "learning_rate": 9.998903417374228e-06, "loss": 1.727, "step": 159 }, { "epoch": 0.32, "grad_norm": 14.352263450622559, "learning_rate": 9.99864620589731e-06, "loss": 1.5935, "step": 160 }, { "epoch": 0.322, "grad_norm": 12.488923072814941, "learning_rate": 9.998361924659989e-06, "loss": 1.5656, "step": 161 }, { "epoch": 0.324, "grad_norm": 18.001474380493164, "learning_rate": 9.998050575201772e-06, "loss": 1.6436, "step": 162 }, { "epoch": 0.326, "grad_norm": 16.588930130004883, "learning_rate": 9.997712159208745e-06, "loss": 1.6952, "step": 163 }, { "epoch": 0.328, "grad_norm": 15.841776847839355, "learning_rate": 9.99734667851357e-06, "loss": 1.6058, "step": 164 }, { "epoch": 0.33, "grad_norm": 17.352310180664062, "learning_rate": 9.99695413509548e-06, "loss": 1.6962, "step": 165 }, { "epoch": 0.332, "grad_norm": 12.201016426086426, "learning_rate": 9.99653453108026e-06, "loss": 1.5996, "step": 166 }, { "epoch": 0.334, "grad_norm": 20.2280216217041, "learning_rate": 9.996087868740244e-06, "loss": 1.7222, "step": 167 }, { "epoch": 0.336, "grad_norm": 15.186982154846191, "learning_rate": 9.995614150494293e-06, "loss": 1.7376, "step": 168 }, { "epoch": 0.338, "grad_norm": 17.792768478393555, "learning_rate": 9.995113378907791e-06, "loss": 1.6865, "step": 169 }, { "epoch": 0.34, "grad_norm": 12.01579761505127, "learning_rate": 9.994585556692624e-06, "loss": 1.6595, "step": 170 }, { "epoch": 0.342, "grad_norm": 16.191465377807617, "learning_rate": 9.994030686707171e-06, "loss": 1.5984, "step": 171 }, { "epoch": 0.344, "grad_norm": 15.65661334991455, "learning_rate": 9.993448771956285e-06, "loss": 1.6659, "step": 172 }, { "epoch": 0.346, "grad_norm": 673.6796264648438, "learning_rate": 9.99283981559128e-06, "loss": 2.1504, "step": 173 }, { "epoch": 0.348, "grad_norm": 11.247122764587402, "learning_rate": 9.992203820909906e-06, "loss": 1.7253, "step": 174 }, { "epoch": 0.35, "grad_norm": 17.28847312927246, "learning_rate": 9.991540791356342e-06, "loss": 1.7098, "step": 175 }, { "epoch": 0.352, "grad_norm": 13.881318092346191, "learning_rate": 9.99085073052117e-06, "loss": 1.5808, "step": 176 }, { "epoch": 0.354, "grad_norm": 14.093522071838379, "learning_rate": 9.990133642141359e-06, "loss": 1.613, "step": 177 }, { "epoch": 0.356, "grad_norm": 10.097461700439453, "learning_rate": 9.989389530100242e-06, "loss": 1.7024, "step": 178 }, { "epoch": 0.358, "grad_norm": 12.82798957824707, "learning_rate": 9.988618398427495e-06, "loss": 1.6558, "step": 179 }, { "epoch": 0.36, "grad_norm": 12.32880687713623, "learning_rate": 9.987820251299121e-06, "loss": 1.4778, "step": 180 }, { "epoch": 0.362, "grad_norm": 8.794655799865723, "learning_rate": 9.986995093037422e-06, "loss": 1.6181, "step": 181 }, { "epoch": 0.364, "grad_norm": 11.556446075439453, "learning_rate": 9.986142928110972e-06, "loss": 1.7423, "step": 182 }, { "epoch": 0.366, "grad_norm": 23.587114334106445, "learning_rate": 9.985263761134602e-06, "loss": 1.7441, "step": 183 }, { "epoch": 0.368, "grad_norm": 14.703253746032715, "learning_rate": 9.984357596869369e-06, "loss": 1.6989, "step": 184 }, { "epoch": 0.37, "grad_norm": 11.022418975830078, "learning_rate": 9.98342444022253e-06, "loss": 1.6443, "step": 185 }, { "epoch": 0.372, "grad_norm": 16.842967987060547, "learning_rate": 9.982464296247523e-06, "loss": 1.7194, "step": 186 }, { "epoch": 0.374, "grad_norm": 14.422867774963379, "learning_rate": 9.981477170143924e-06, "loss": 1.6715, "step": 187 }, { "epoch": 0.376, "grad_norm": 15.538463592529297, "learning_rate": 9.980463067257437e-06, "loss": 1.6339, "step": 188 }, { "epoch": 0.378, "grad_norm": 10.891705513000488, "learning_rate": 9.979421993079853e-06, "loss": 1.6721, "step": 189 }, { "epoch": 0.38, "grad_norm": 12.947959899902344, "learning_rate": 9.978353953249023e-06, "loss": 1.7966, "step": 190 }, { "epoch": 0.382, "grad_norm": 9.348876953125, "learning_rate": 9.977258953548831e-06, "loss": 1.6517, "step": 191 }, { "epoch": 0.384, "grad_norm": 8.993280410766602, "learning_rate": 9.976136999909156e-06, "loss": 1.6427, "step": 192 }, { "epoch": 0.386, "grad_norm": 8.9512300491333, "learning_rate": 9.97498809840585e-06, "loss": 1.6192, "step": 193 }, { "epoch": 0.388, "grad_norm": 9.874775886535645, "learning_rate": 9.973812255260692e-06, "loss": 1.6766, "step": 194 }, { "epoch": 0.39, "grad_norm": 11.946427345275879, "learning_rate": 9.972609476841368e-06, "loss": 1.6853, "step": 195 }, { "epoch": 0.392, "grad_norm": 11.508259773254395, "learning_rate": 9.971379769661422e-06, "loss": 1.6911, "step": 196 }, { "epoch": 0.394, "grad_norm": 10.715835571289062, "learning_rate": 9.970123140380237e-06, "loss": 1.6015, "step": 197 }, { "epoch": 0.396, "grad_norm": 11.727666854858398, "learning_rate": 9.968839595802982e-06, "loss": 1.6955, "step": 198 }, { "epoch": 0.398, "grad_norm": 10.18913745880127, "learning_rate": 9.967529142880592e-06, "loss": 1.5734, "step": 199 }, { "epoch": 0.4, "grad_norm": 10.491742134094238, "learning_rate": 9.966191788709716e-06, "loss": 1.659, "step": 200 }, { "epoch": 0.402, "grad_norm": 8.228020668029785, "learning_rate": 9.964827540532685e-06, "loss": 1.5536, "step": 201 }, { "epoch": 0.404, "grad_norm": 15.777739524841309, "learning_rate": 9.963436405737476e-06, "loss": 1.536, "step": 202 }, { "epoch": 0.406, "grad_norm": 11.057550430297852, "learning_rate": 9.962018391857665e-06, "loss": 1.583, "step": 203 }, { "epoch": 0.408, "grad_norm": 11.679120063781738, "learning_rate": 9.960573506572391e-06, "loss": 1.5931, "step": 204 }, { "epoch": 0.41, "grad_norm": 18.90032386779785, "learning_rate": 9.959101757706308e-06, "loss": 1.7411, "step": 205 }, { "epoch": 0.412, "grad_norm": 18.84972381591797, "learning_rate": 9.957603153229559e-06, "loss": 1.6154, "step": 206 }, { "epoch": 0.414, "grad_norm": 13.331605911254883, "learning_rate": 9.95607770125771e-06, "loss": 1.677, "step": 207 }, { "epoch": 0.416, "grad_norm": 13.559438705444336, "learning_rate": 9.95452541005172e-06, "loss": 1.6305, "step": 208 }, { "epoch": 0.418, "grad_norm": 14.361797332763672, "learning_rate": 9.952946288017899e-06, "loss": 1.6408, "step": 209 }, { "epoch": 0.42, "grad_norm": 16.02930450439453, "learning_rate": 9.951340343707852e-06, "loss": 1.7221, "step": 210 }, { "epoch": 0.422, "grad_norm": 15.836488723754883, "learning_rate": 9.94970758581844e-06, "loss": 1.5559, "step": 211 }, { "epoch": 0.424, "grad_norm": 13.252622604370117, "learning_rate": 9.948048023191728e-06, "loss": 1.6659, "step": 212 }, { "epoch": 0.426, "grad_norm": 10.738286972045898, "learning_rate": 9.946361664814942e-06, "loss": 1.5846, "step": 213 }, { "epoch": 0.428, "grad_norm": 12.482677459716797, "learning_rate": 9.94464851982042e-06, "loss": 1.7871, "step": 214 }, { "epoch": 0.43, "grad_norm": 10.500411987304688, "learning_rate": 9.942908597485558e-06, "loss": 1.7273, "step": 215 }, { "epoch": 0.432, "grad_norm": 10.43774700164795, "learning_rate": 9.941141907232766e-06, "loss": 1.7749, "step": 216 }, { "epoch": 0.434, "grad_norm": 10.52614688873291, "learning_rate": 9.939348458629406e-06, "loss": 1.6396, "step": 217 }, { "epoch": 0.436, "grad_norm": 9.133844375610352, "learning_rate": 9.937528261387753e-06, "loss": 1.6775, "step": 218 }, { "epoch": 0.438, "grad_norm": 10.179513931274414, "learning_rate": 9.93568132536494e-06, "loss": 1.6362, "step": 219 }, { "epoch": 0.44, "grad_norm": 10.565604209899902, "learning_rate": 9.933807660562898e-06, "loss": 1.6125, "step": 220 }, { "epoch": 0.442, "grad_norm": 13.952414512634277, "learning_rate": 9.9319072771283e-06, "loss": 1.6762, "step": 221 }, { "epoch": 0.444, "grad_norm": 13.159649848937988, "learning_rate": 9.929980185352525e-06, "loss": 1.7111, "step": 222 }, { "epoch": 0.446, "grad_norm": 15.363605499267578, "learning_rate": 9.928026395671577e-06, "loss": 1.687, "step": 223 }, { "epoch": 0.448, "grad_norm": 12.805254936218262, "learning_rate": 9.926045918666045e-06, "loss": 1.6426, "step": 224 }, { "epoch": 0.45, "grad_norm": 9.896926879882812, "learning_rate": 9.924038765061042e-06, "loss": 1.657, "step": 225 }, { "epoch": 0.452, "grad_norm": 8.098036766052246, "learning_rate": 9.92200494572614e-06, "loss": 1.5689, "step": 226 }, { "epoch": 0.454, "grad_norm": 9.297266960144043, "learning_rate": 9.919944471675328e-06, "loss": 1.5641, "step": 227 }, { "epoch": 0.456, "grad_norm": 11.86732292175293, "learning_rate": 9.91785735406693e-06, "loss": 1.5646, "step": 228 }, { "epoch": 0.458, "grad_norm": 15.542577743530273, "learning_rate": 9.915743604203563e-06, "loss": 1.7098, "step": 229 }, { "epoch": 0.46, "grad_norm": 15.652758598327637, "learning_rate": 9.913603233532067e-06, "loss": 1.6008, "step": 230 }, { "epoch": 0.462, "grad_norm": 14.435830116271973, "learning_rate": 9.911436253643445e-06, "loss": 1.6086, "step": 231 }, { "epoch": 0.464, "grad_norm": 12.681479454040527, "learning_rate": 9.909242676272797e-06, "loss": 1.7047, "step": 232 }, { "epoch": 0.466, "grad_norm": 13.065205574035645, "learning_rate": 9.907022513299264e-06, "loss": 1.6552, "step": 233 }, { "epoch": 0.468, "grad_norm": 10.368819236755371, "learning_rate": 9.904775776745959e-06, "loss": 1.6101, "step": 234 }, { "epoch": 0.47, "grad_norm": 17.001413345336914, "learning_rate": 9.902502478779897e-06, "loss": 1.7077, "step": 235 }, { "epoch": 0.472, "grad_norm": 9.429941177368164, "learning_rate": 9.90020263171194e-06, "loss": 1.5966, "step": 236 }, { "epoch": 0.474, "grad_norm": 9.692448616027832, "learning_rate": 9.89787624799672e-06, "loss": 1.6065, "step": 237 }, { "epoch": 0.476, "grad_norm": 12.412395477294922, "learning_rate": 9.89552334023258e-06, "loss": 1.8045, "step": 238 }, { "epoch": 0.478, "grad_norm": 10.290336608886719, "learning_rate": 9.893143921161501e-06, "loss": 1.5883, "step": 239 }, { "epoch": 0.48, "grad_norm": 16.823965072631836, "learning_rate": 9.890738003669029e-06, "loss": 1.6098, "step": 240 }, { "epoch": 0.482, "grad_norm": 10.46012020111084, "learning_rate": 9.888305600784217e-06, "loss": 1.6483, "step": 241 }, { "epoch": 0.484, "grad_norm": 9.59673023223877, "learning_rate": 9.88584672567954e-06, "loss": 1.5618, "step": 242 }, { "epoch": 0.486, "grad_norm": 11.56871223449707, "learning_rate": 9.883361391670841e-06, "loss": 1.698, "step": 243 }, { "epoch": 0.488, "grad_norm": 6.657703399658203, "learning_rate": 9.880849612217238e-06, "loss": 1.689, "step": 244 }, { "epoch": 0.49, "grad_norm": 11.670310020446777, "learning_rate": 9.878311400921072e-06, "loss": 1.694, "step": 245 }, { "epoch": 0.492, "grad_norm": 9.138250350952148, "learning_rate": 9.875746771527817e-06, "loss": 1.6979, "step": 246 }, { "epoch": 0.494, "grad_norm": 10.32927131652832, "learning_rate": 9.873155737926014e-06, "loss": 1.6198, "step": 247 }, { "epoch": 0.496, "grad_norm": 8.178282737731934, "learning_rate": 9.870538314147194e-06, "loss": 1.6388, "step": 248 }, { "epoch": 0.498, "grad_norm": 10.270970344543457, "learning_rate": 9.867894514365802e-06, "loss": 1.5972, "step": 249 }, { "epoch": 0.5, "grad_norm": 8.778549194335938, "learning_rate": 9.86522435289912e-06, "loss": 1.7337, "step": 250 }, { "epoch": 0.502, "grad_norm": 7.87909460067749, "learning_rate": 9.862527844207189e-06, "loss": 1.6102, "step": 251 }, { "epoch": 0.504, "grad_norm": 9.182847023010254, "learning_rate": 9.859805002892733e-06, "loss": 1.7548, "step": 252 }, { "epoch": 0.506, "grad_norm": 7.970770835876465, "learning_rate": 9.857055843701073e-06, "loss": 1.6482, "step": 253 }, { "epoch": 0.508, "grad_norm": 13.17197322845459, "learning_rate": 9.85428038152006e-06, "loss": 1.663, "step": 254 }, { "epoch": 0.51, "grad_norm": 9.727519035339355, "learning_rate": 9.851478631379982e-06, "loss": 1.613, "step": 255 }, { "epoch": 0.512, "grad_norm": 11.023371696472168, "learning_rate": 9.84865060845349e-06, "loss": 1.6383, "step": 256 }, { "epoch": 0.514, "grad_norm": 11.075227737426758, "learning_rate": 9.845796328055505e-06, "loss": 1.6506, "step": 257 }, { "epoch": 0.516, "grad_norm": 10.624147415161133, "learning_rate": 9.842915805643156e-06, "loss": 1.6063, "step": 258 }, { "epoch": 0.518, "grad_norm": 8.243653297424316, "learning_rate": 9.840009056815674e-06, "loss": 1.6393, "step": 259 }, { "epoch": 0.52, "grad_norm": 9.168231964111328, "learning_rate": 9.83707609731432e-06, "loss": 1.5481, "step": 260 }, { "epoch": 0.522, "grad_norm": 10.238672256469727, "learning_rate": 9.834116943022299e-06, "loss": 1.703, "step": 261 }, { "epoch": 0.524, "grad_norm": 6.614677906036377, "learning_rate": 9.831131609964664e-06, "loss": 1.5414, "step": 262 }, { "epoch": 0.526, "grad_norm": 7.963540554046631, "learning_rate": 9.828120114308248e-06, "loss": 1.603, "step": 263 }, { "epoch": 0.528, "grad_norm": 11.000412940979004, "learning_rate": 9.825082472361558e-06, "loss": 1.7098, "step": 264 }, { "epoch": 0.53, "grad_norm": 10.786978721618652, "learning_rate": 9.822018700574696e-06, "loss": 1.6766, "step": 265 }, { "epoch": 0.532, "grad_norm": 9.823018074035645, "learning_rate": 9.818928815539266e-06, "loss": 1.6362, "step": 266 }, { "epoch": 0.534, "grad_norm": 10.396644592285156, "learning_rate": 9.815812833988292e-06, "loss": 1.8225, "step": 267 }, { "epoch": 0.536, "grad_norm": 11.278525352478027, "learning_rate": 9.812670772796113e-06, "loss": 1.7868, "step": 268 }, { "epoch": 0.538, "grad_norm": 8.690960884094238, "learning_rate": 9.809502648978311e-06, "loss": 1.567, "step": 269 }, { "epoch": 0.54, "grad_norm": 11.479389190673828, "learning_rate": 9.806308479691595e-06, "loss": 1.6175, "step": 270 }, { "epoch": 0.542, "grad_norm": 8.428929328918457, "learning_rate": 9.803088282233733e-06, "loss": 1.6575, "step": 271 }, { "epoch": 0.544, "grad_norm": 8.638998985290527, "learning_rate": 9.799842074043438e-06, "loss": 1.6002, "step": 272 }, { "epoch": 0.546, "grad_norm": 9.341121673583984, "learning_rate": 9.796569872700287e-06, "loss": 1.672, "step": 273 }, { "epoch": 0.548, "grad_norm": 9.511893272399902, "learning_rate": 9.793271695924621e-06, "loss": 1.7315, "step": 274 }, { "epoch": 0.55, "grad_norm": 12.620855331420898, "learning_rate": 9.789947561577445e-06, "loss": 1.63, "step": 275 }, { "epoch": 0.552, "grad_norm": 14.644619941711426, "learning_rate": 9.786597487660336e-06, "loss": 1.649, "step": 276 }, { "epoch": 0.554, "grad_norm": 8.164085388183594, "learning_rate": 9.78322149231535e-06, "loss": 1.6249, "step": 277 }, { "epoch": 0.556, "grad_norm": 12.925362586975098, "learning_rate": 9.779819593824909e-06, "loss": 1.572, "step": 278 }, { "epoch": 0.558, "grad_norm": 10.382940292358398, "learning_rate": 9.776391810611719e-06, "loss": 1.6225, "step": 279 }, { "epoch": 0.56, "grad_norm": 14.114727973937988, "learning_rate": 9.77293816123866e-06, "loss": 1.4772, "step": 280 }, { "epoch": 0.562, "grad_norm": 9.955523490905762, "learning_rate": 9.769458664408689e-06, "loss": 1.6424, "step": 281 }, { "epoch": 0.564, "grad_norm": 10.805880546569824, "learning_rate": 9.765953338964736e-06, "loss": 1.5621, "step": 282 }, { "epoch": 0.566, "grad_norm": 9.07789134979248, "learning_rate": 9.762422203889604e-06, "loss": 1.6892, "step": 283 }, { "epoch": 0.568, "grad_norm": 13.484134674072266, "learning_rate": 9.75886527830587e-06, "loss": 1.7468, "step": 284 }, { "epoch": 0.57, "grad_norm": 10.758011817932129, "learning_rate": 9.755282581475769e-06, "loss": 1.699, "step": 285 }, { "epoch": 0.572, "grad_norm": 8.382722854614258, "learning_rate": 9.751674132801106e-06, "loss": 1.5586, "step": 286 }, { "epoch": 0.574, "grad_norm": 7.959819316864014, "learning_rate": 9.748039951823141e-06, "loss": 1.5625, "step": 287 }, { "epoch": 0.576, "grad_norm": 8.60888671875, "learning_rate": 9.744380058222483e-06, "loss": 1.5166, "step": 288 }, { "epoch": 0.578, "grad_norm": 10.173426628112793, "learning_rate": 9.740694471818988e-06, "loss": 1.7414, "step": 289 }, { "epoch": 0.58, "grad_norm": 8.694028854370117, "learning_rate": 9.736983212571646e-06, "loss": 1.5719, "step": 290 }, { "epoch": 0.582, "grad_norm": 12.50915241241455, "learning_rate": 9.733246300578482e-06, "loss": 1.5379, "step": 291 }, { "epoch": 0.584, "grad_norm": 7.476858615875244, "learning_rate": 9.729483756076436e-06, "loss": 1.5798, "step": 292 }, { "epoch": 0.586, "grad_norm": 8.575016975402832, "learning_rate": 9.72569559944126e-06, "loss": 1.6348, "step": 293 }, { "epoch": 0.588, "grad_norm": 10.078662872314453, "learning_rate": 9.721881851187406e-06, "loss": 1.6831, "step": 294 }, { "epoch": 0.59, "grad_norm": 12.232658386230469, "learning_rate": 9.718042531967918e-06, "loss": 1.687, "step": 295 }, { "epoch": 0.592, "grad_norm": 10.713579177856445, "learning_rate": 9.714177662574316e-06, "loss": 1.6168, "step": 296 }, { "epoch": 0.594, "grad_norm": 10.275739669799805, "learning_rate": 9.710287263936485e-06, "loss": 1.7538, "step": 297 }, { "epoch": 0.596, "grad_norm": 7.381926536560059, "learning_rate": 9.70637135712256e-06, "loss": 1.5313, "step": 298 }, { "epoch": 0.598, "grad_norm": 10.051726341247559, "learning_rate": 9.702429963338812e-06, "loss": 1.7786, "step": 299 }, { "epoch": 0.6, "grad_norm": 10.447134017944336, "learning_rate": 9.698463103929542e-06, "loss": 1.6418, "step": 300 }, { "epoch": 0.602, "grad_norm": 13.121293067932129, "learning_rate": 9.694470800376951e-06, "loss": 1.7257, "step": 301 }, { "epoch": 0.604, "grad_norm": 8.965837478637695, "learning_rate": 9.690453074301035e-06, "loss": 1.6921, "step": 302 }, { "epoch": 0.606, "grad_norm": 13.510847091674805, "learning_rate": 9.68640994745946e-06, "loss": 1.7176, "step": 303 }, { "epoch": 0.608, "grad_norm": 10.887725830078125, "learning_rate": 9.682341441747446e-06, "loss": 1.7061, "step": 304 }, { "epoch": 0.61, "grad_norm": 8.40693473815918, "learning_rate": 9.678247579197658e-06, "loss": 1.6644, "step": 305 }, { "epoch": 0.612, "grad_norm": 11.772154808044434, "learning_rate": 9.674128381980073e-06, "loss": 1.7006, "step": 306 }, { "epoch": 0.614, "grad_norm": 10.446290969848633, "learning_rate": 9.669983872401868e-06, "loss": 1.7206, "step": 307 }, { "epoch": 0.616, "grad_norm": 10.989177703857422, "learning_rate": 9.665814072907293e-06, "loss": 1.6525, "step": 308 }, { "epoch": 0.618, "grad_norm": 11.617728233337402, "learning_rate": 9.661619006077562e-06, "loss": 1.6366, "step": 309 }, { "epoch": 0.62, "grad_norm": 12.805769920349121, "learning_rate": 9.657398694630713e-06, "loss": 1.7126, "step": 310 }, { "epoch": 0.622, "grad_norm": 10.547633171081543, "learning_rate": 9.653153161421497e-06, "loss": 1.6447, "step": 311 }, { "epoch": 0.624, "grad_norm": 11.312098503112793, "learning_rate": 9.648882429441258e-06, "loss": 1.732, "step": 312 }, { "epoch": 0.626, "grad_norm": 11.163432121276855, "learning_rate": 9.644586521817792e-06, "loss": 1.5886, "step": 313 }, { "epoch": 0.628, "grad_norm": 14.074267387390137, "learning_rate": 9.640265461815235e-06, "loss": 1.6658, "step": 314 }, { "epoch": 0.63, "grad_norm": 8.643535614013672, "learning_rate": 9.635919272833938e-06, "loss": 1.5441, "step": 315 }, { "epoch": 0.632, "grad_norm": 17.20092010498047, "learning_rate": 9.63154797841033e-06, "loss": 1.6208, "step": 316 }, { "epoch": 0.634, "grad_norm": 9.415007591247559, "learning_rate": 9.627151602216801e-06, "loss": 1.5145, "step": 317 }, { "epoch": 0.636, "grad_norm": 14.709664344787598, "learning_rate": 9.622730168061568e-06, "loss": 1.7034, "step": 318 }, { "epoch": 0.638, "grad_norm": 13.328878402709961, "learning_rate": 9.618283699888543e-06, "loss": 1.7108, "step": 319 }, { "epoch": 0.64, "grad_norm": 9.980976104736328, "learning_rate": 9.613812221777212e-06, "loss": 1.6241, "step": 320 }, { "epoch": 0.642, "grad_norm": 11.059935569763184, "learning_rate": 9.609315757942504e-06, "loss": 1.7884, "step": 321 }, { "epoch": 0.644, "grad_norm": 8.424722671508789, "learning_rate": 9.604794332734647e-06, "loss": 1.5701, "step": 322 }, { "epoch": 0.646, "grad_norm": 13.05691146850586, "learning_rate": 9.600247970639053e-06, "loss": 1.6983, "step": 323 }, { "epoch": 0.648, "grad_norm": 10.937281608581543, "learning_rate": 9.595676696276173e-06, "loss": 1.7072, "step": 324 }, { "epoch": 0.65, "grad_norm": 10.963797569274902, "learning_rate": 9.591080534401371e-06, "loss": 1.7535, "step": 325 }, { "epoch": 0.652, "grad_norm": 11.561660766601562, "learning_rate": 9.586459509904786e-06, "loss": 1.564, "step": 326 }, { "epoch": 0.654, "grad_norm": 9.038650512695312, "learning_rate": 9.581813647811199e-06, "loss": 1.7599, "step": 327 }, { "epoch": 0.656, "grad_norm": 9.181880950927734, "learning_rate": 9.577142973279896e-06, "loss": 1.5822, "step": 328 }, { "epoch": 0.658, "grad_norm": 9.09460735321045, "learning_rate": 9.572447511604536e-06, "loss": 1.6547, "step": 329 }, { "epoch": 0.66, "grad_norm": 8.515859603881836, "learning_rate": 9.567727288213005e-06, "loss": 1.59, "step": 330 }, { "epoch": 0.662, "grad_norm": 11.761797904968262, "learning_rate": 9.56298232866729e-06, "loss": 1.7411, "step": 331 }, { "epoch": 0.664, "grad_norm": 10.45743179321289, "learning_rate": 9.55821265866333e-06, "loss": 1.6905, "step": 332 }, { "epoch": 0.666, "grad_norm": 8.767599105834961, "learning_rate": 9.553418304030886e-06, "loss": 1.624, "step": 333 }, { "epoch": 0.668, "grad_norm": 8.311175346374512, "learning_rate": 9.548599290733393e-06, "loss": 1.7281, "step": 334 }, { "epoch": 0.67, "grad_norm": 10.42666244506836, "learning_rate": 9.543755644867823e-06, "loss": 1.6878, "step": 335 }, { "epoch": 0.672, "grad_norm": 7.798083782196045, "learning_rate": 9.538887392664544e-06, "loss": 1.623, "step": 336 }, { "epoch": 0.674, "grad_norm": 10.574175834655762, "learning_rate": 9.53399456048718e-06, "loss": 1.6386, "step": 337 }, { "epoch": 0.676, "grad_norm": 8.909551620483398, "learning_rate": 9.529077174832466e-06, "loss": 1.7324, "step": 338 }, { "epoch": 0.678, "grad_norm": 7.710360527038574, "learning_rate": 9.524135262330098e-06, "loss": 1.6373, "step": 339 }, { "epoch": 0.68, "grad_norm": 13.496674537658691, "learning_rate": 9.519168849742603e-06, "loss": 1.6202, "step": 340 }, { "epoch": 0.682, "grad_norm": 7.529291152954102, "learning_rate": 9.514177963965181e-06, "loss": 1.7156, "step": 341 }, { "epoch": 0.684, "grad_norm": 7.229175567626953, "learning_rate": 9.50916263202557e-06, "loss": 1.663, "step": 342 }, { "epoch": 0.686, "grad_norm": 12.523305892944336, "learning_rate": 9.504122881083886e-06, "loss": 1.6902, "step": 343 }, { "epoch": 0.688, "grad_norm": 10.700648307800293, "learning_rate": 9.499058738432492e-06, "loss": 1.6298, "step": 344 }, { "epoch": 0.69, "grad_norm": 7.542520523071289, "learning_rate": 9.493970231495836e-06, "loss": 1.5829, "step": 345 }, { "epoch": 0.692, "grad_norm": 9.304771423339844, "learning_rate": 9.488857387830315e-06, "loss": 1.6467, "step": 346 }, { "epoch": 0.694, "grad_norm": 11.418816566467285, "learning_rate": 9.483720235124113e-06, "loss": 1.7727, "step": 347 }, { "epoch": 0.696, "grad_norm": 7.398611068725586, "learning_rate": 9.478558801197065e-06, "loss": 1.5846, "step": 348 }, { "epoch": 0.698, "grad_norm": 10.512649536132812, "learning_rate": 9.473373114000493e-06, "loss": 1.7182, "step": 349 }, { "epoch": 0.7, "grad_norm": 9.7409029006958, "learning_rate": 9.468163201617063e-06, "loss": 1.6869, "step": 350 }, { "epoch": 0.702, "grad_norm": 11.419679641723633, "learning_rate": 9.46292909226063e-06, "loss": 1.6018, "step": 351 }, { "epoch": 0.704, "grad_norm": 12.171987533569336, "learning_rate": 9.457670814276083e-06, "loss": 1.5906, "step": 352 }, { "epoch": 0.706, "grad_norm": 10.171733856201172, "learning_rate": 9.452388396139202e-06, "loss": 1.6307, "step": 353 }, { "epoch": 0.708, "grad_norm": 11.288997650146484, "learning_rate": 9.44708186645649e-06, "loss": 1.627, "step": 354 }, { "epoch": 0.71, "grad_norm": 10.578997611999512, "learning_rate": 9.441751253965022e-06, "loss": 1.7024, "step": 355 }, { "epoch": 0.712, "grad_norm": 8.29975414276123, "learning_rate": 9.436396587532297e-06, "loss": 1.7009, "step": 356 }, { "epoch": 0.714, "grad_norm": 9.676756858825684, "learning_rate": 9.431017896156074e-06, "loss": 1.6421, "step": 357 }, { "epoch": 0.716, "grad_norm": 8.530579566955566, "learning_rate": 9.425615208964217e-06, "loss": 1.619, "step": 358 }, { "epoch": 0.718, "grad_norm": 6.746725082397461, "learning_rate": 9.420188555214537e-06, "loss": 1.5287, "step": 359 }, { "epoch": 0.72, "grad_norm": 7.753213405609131, "learning_rate": 9.414737964294636e-06, "loss": 1.53, "step": 360 }, { "epoch": 0.722, "grad_norm": 7.940869331359863, "learning_rate": 9.40926346572174e-06, "loss": 1.5859, "step": 361 }, { "epoch": 0.724, "grad_norm": 8.257153511047363, "learning_rate": 9.403765089142554e-06, "loss": 1.559, "step": 362 }, { "epoch": 0.726, "grad_norm": 8.980607032775879, "learning_rate": 9.398242864333084e-06, "loss": 1.6408, "step": 363 }, { "epoch": 0.728, "grad_norm": 12.678767204284668, "learning_rate": 9.392696821198488e-06, "loss": 1.6613, "step": 364 }, { "epoch": 0.73, "grad_norm": 9.08625316619873, "learning_rate": 9.38712698977291e-06, "loss": 1.6557, "step": 365 }, { "epoch": 0.732, "grad_norm": 9.90092658996582, "learning_rate": 9.381533400219319e-06, "loss": 1.7143, "step": 366 }, { "epoch": 0.734, "grad_norm": 15.684215545654297, "learning_rate": 9.375916082829341e-06, "loss": 1.6149, "step": 367 }, { "epoch": 0.736, "grad_norm": 8.9537935256958, "learning_rate": 9.370275068023097e-06, "loss": 1.5268, "step": 368 }, { "epoch": 0.738, "grad_norm": 10.356677055358887, "learning_rate": 9.364610386349048e-06, "loss": 1.7228, "step": 369 }, { "epoch": 0.74, "grad_norm": 10.680842399597168, "learning_rate": 9.358922068483813e-06, "loss": 1.762, "step": 370 }, { "epoch": 0.742, "grad_norm": 9.59675121307373, "learning_rate": 9.35321014523201e-06, "loss": 1.6749, "step": 371 }, { "epoch": 0.744, "grad_norm": 11.908790588378906, "learning_rate": 9.347474647526095e-06, "loss": 1.67, "step": 372 }, { "epoch": 0.746, "grad_norm": 10.074137687683105, "learning_rate": 9.34171560642619e-06, "loss": 1.6049, "step": 373 }, { "epoch": 0.748, "grad_norm": 9.519111633300781, "learning_rate": 9.335933053119906e-06, "loss": 1.7383, "step": 374 }, { "epoch": 0.75, "grad_norm": 11.124157905578613, "learning_rate": 9.330127018922195e-06, "loss": 1.5772, "step": 375 }, { "epoch": 0.752, "grad_norm": 9.715572357177734, "learning_rate": 9.324297535275156e-06, "loss": 1.6024, "step": 376 }, { "epoch": 0.754, "grad_norm": 14.774503707885742, "learning_rate": 9.318444633747884e-06, "loss": 1.614, "step": 377 }, { "epoch": 0.756, "grad_norm": 15.559951782226562, "learning_rate": 9.312568346036288e-06, "loss": 1.6546, "step": 378 }, { "epoch": 0.758, "grad_norm": 9.326455116271973, "learning_rate": 9.306668703962927e-06, "loss": 1.6567, "step": 379 }, { "epoch": 0.76, "grad_norm": 10.113893508911133, "learning_rate": 9.30074573947683e-06, "loss": 1.6268, "step": 380 }, { "epoch": 0.762, "grad_norm": 10.376815795898438, "learning_rate": 9.294799484653323e-06, "loss": 1.649, "step": 381 }, { "epoch": 0.764, "grad_norm": 12.478555679321289, "learning_rate": 9.288829971693869e-06, "loss": 1.5325, "step": 382 }, { "epoch": 0.766, "grad_norm": 8.862922668457031, "learning_rate": 9.282837232925876e-06, "loss": 1.5611, "step": 383 }, { "epoch": 0.768, "grad_norm": 9.65459156036377, "learning_rate": 9.276821300802535e-06, "loss": 1.6238, "step": 384 }, { "epoch": 0.77, "grad_norm": 10.751141548156738, "learning_rate": 9.27078220790263e-06, "loss": 1.629, "step": 385 }, { "epoch": 0.772, "grad_norm": 11.15700912475586, "learning_rate": 9.264719986930376e-06, "loss": 1.757, "step": 386 }, { "epoch": 0.774, "grad_norm": 11.11645221710205, "learning_rate": 9.25863467071524e-06, "loss": 1.7262, "step": 387 }, { "epoch": 0.776, "grad_norm": 9.191835403442383, "learning_rate": 9.25252629221175e-06, "loss": 1.5161, "step": 388 }, { "epoch": 0.778, "grad_norm": 8.911935806274414, "learning_rate": 9.246394884499334e-06, "loss": 1.6185, "step": 389 }, { "epoch": 0.78, "grad_norm": 10.155871391296387, "learning_rate": 9.24024048078213e-06, "loss": 1.6151, "step": 390 }, { "epoch": 0.782, "grad_norm": 7.377295970916748, "learning_rate": 9.234063114388809e-06, "loss": 1.5232, "step": 391 }, { "epoch": 0.784, "grad_norm": 10.363668441772461, "learning_rate": 9.227862818772392e-06, "loss": 1.6332, "step": 392 }, { "epoch": 0.786, "grad_norm": 8.449118614196777, "learning_rate": 9.221639627510076e-06, "loss": 1.5769, "step": 393 }, { "epoch": 0.788, "grad_norm": 8.908105850219727, "learning_rate": 9.215393574303043e-06, "loss": 1.7008, "step": 394 }, { "epoch": 0.79, "grad_norm": 9.45577335357666, "learning_rate": 9.209124692976287e-06, "loss": 1.6707, "step": 395 }, { "epoch": 0.792, "grad_norm": 9.608206748962402, "learning_rate": 9.202833017478421e-06, "loss": 1.6652, "step": 396 }, { "epoch": 0.794, "grad_norm": 9.177596092224121, "learning_rate": 9.196518581881502e-06, "loss": 1.7067, "step": 397 }, { "epoch": 0.796, "grad_norm": 5.568012714385986, "learning_rate": 9.190181420380838e-06, "loss": 1.6233, "step": 398 }, { "epoch": 0.798, "grad_norm": 10.35659122467041, "learning_rate": 9.18382156729481e-06, "loss": 1.6694, "step": 399 }, { "epoch": 0.8, "grad_norm": 5.523914813995361, "learning_rate": 9.177439057064684e-06, "loss": 1.6281, "step": 400 }, { "epoch": 0.802, "grad_norm": 7.155726432800293, "learning_rate": 9.171033924254421e-06, "loss": 1.6411, "step": 401 }, { "epoch": 0.804, "grad_norm": 11.25847339630127, "learning_rate": 9.164606203550498e-06, "loss": 1.5891, "step": 402 }, { "epoch": 0.806, "grad_norm": 12.162601470947266, "learning_rate": 9.15815592976171e-06, "loss": 1.5304, "step": 403 }, { "epoch": 0.808, "grad_norm": 9.000397682189941, "learning_rate": 9.151683137818989e-06, "loss": 1.6407, "step": 404 }, { "epoch": 0.81, "grad_norm": 8.760917663574219, "learning_rate": 9.145187862775208e-06, "loss": 1.6462, "step": 405 }, { "epoch": 0.812, "grad_norm": 9.060429573059082, "learning_rate": 9.138670139805004e-06, "loss": 1.6845, "step": 406 }, { "epoch": 0.814, "grad_norm": 8.54736614227295, "learning_rate": 9.132130004204569e-06, "loss": 1.7051, "step": 407 }, { "epoch": 0.816, "grad_norm": 10.724129676818848, "learning_rate": 9.125567491391476e-06, "loss": 1.6058, "step": 408 }, { "epoch": 0.818, "grad_norm": 9.786903381347656, "learning_rate": 9.118982636904476e-06, "loss": 1.6181, "step": 409 }, { "epoch": 0.82, "grad_norm": 9.084239959716797, "learning_rate": 9.112375476403313e-06, "loss": 1.7077, "step": 410 }, { "epoch": 0.822, "grad_norm": 8.961618423461914, "learning_rate": 9.10574604566852e-06, "loss": 1.6431, "step": 411 }, { "epoch": 0.824, "grad_norm": 8.100645065307617, "learning_rate": 9.099094380601244e-06, "loss": 1.6346, "step": 412 }, { "epoch": 0.826, "grad_norm": 10.302638053894043, "learning_rate": 9.09242051722303e-06, "loss": 1.728, "step": 413 }, { "epoch": 0.828, "grad_norm": 8.645998001098633, "learning_rate": 9.085724491675642e-06, "loss": 1.6298, "step": 414 }, { "epoch": 0.83, "grad_norm": 6.636702060699463, "learning_rate": 9.079006340220862e-06, "loss": 1.7458, "step": 415 }, { "epoch": 0.832, "grad_norm": 9.560812950134277, "learning_rate": 9.072266099240286e-06, "loss": 1.6252, "step": 416 }, { "epoch": 0.834, "grad_norm": 8.170851707458496, "learning_rate": 9.065503805235139e-06, "loss": 1.5759, "step": 417 }, { "epoch": 0.836, "grad_norm": 10.425228118896484, "learning_rate": 9.058719494826076e-06, "loss": 1.7297, "step": 418 }, { "epoch": 0.838, "grad_norm": 8.824040412902832, "learning_rate": 9.051913204752972e-06, "loss": 1.7309, "step": 419 }, { "epoch": 0.84, "grad_norm": 9.032853126525879, "learning_rate": 9.045084971874738e-06, "loss": 1.5986, "step": 420 }, { "epoch": 0.842, "grad_norm": 8.494746208190918, "learning_rate": 9.03823483316911e-06, "loss": 1.6812, "step": 421 }, { "epoch": 0.844, "grad_norm": 9.491890907287598, "learning_rate": 9.031362825732456e-06, "loss": 1.6809, "step": 422 }, { "epoch": 0.846, "grad_norm": 10.895613670349121, "learning_rate": 9.02446898677957e-06, "loss": 1.6665, "step": 423 }, { "epoch": 0.848, "grad_norm": 6.491031169891357, "learning_rate": 9.017553353643479e-06, "loss": 1.6723, "step": 424 }, { "epoch": 0.85, "grad_norm": 8.402596473693848, "learning_rate": 9.01061596377522e-06, "loss": 1.5644, "step": 425 }, { "epoch": 0.852, "grad_norm": 8.312848091125488, "learning_rate": 9.003656854743667e-06, "loss": 1.5704, "step": 426 }, { "epoch": 0.854, "grad_norm": 10.550932884216309, "learning_rate": 8.996676064235308e-06, "loss": 1.6347, "step": 427 }, { "epoch": 0.856, "grad_norm": 7.069148540496826, "learning_rate": 8.989673630054044e-06, "loss": 1.5883, "step": 428 }, { "epoch": 0.858, "grad_norm": 8.47697639465332, "learning_rate": 8.982649590120982e-06, "loss": 1.6156, "step": 429 }, { "epoch": 0.86, "grad_norm": 10.076146125793457, "learning_rate": 8.97560398247424e-06, "loss": 1.6816, "step": 430 }, { "epoch": 0.862, "grad_norm": 9.539214134216309, "learning_rate": 8.96853684526873e-06, "loss": 1.6875, "step": 431 }, { "epoch": 0.864, "grad_norm": 10.413458824157715, "learning_rate": 8.961448216775955e-06, "loss": 1.6266, "step": 432 }, { "epoch": 0.866, "grad_norm": 9.216937065124512, "learning_rate": 8.954338135383804e-06, "loss": 1.6485, "step": 433 }, { "epoch": 0.868, "grad_norm": 8.924684524536133, "learning_rate": 8.947206639596346e-06, "loss": 1.5875, "step": 434 }, { "epoch": 0.87, "grad_norm": 9.03341007232666, "learning_rate": 8.94005376803361e-06, "loss": 1.562, "step": 435 }, { "epoch": 0.872, "grad_norm": 7.471339225769043, "learning_rate": 8.932879559431392e-06, "loss": 1.5646, "step": 436 }, { "epoch": 0.874, "grad_norm": 12.005831718444824, "learning_rate": 8.925684052641027e-06, "loss": 1.7528, "step": 437 }, { "epoch": 0.876, "grad_norm": 7.764755725860596, "learning_rate": 8.9184672866292e-06, "loss": 1.6857, "step": 438 }, { "epoch": 0.878, "grad_norm": 7.975648403167725, "learning_rate": 8.911229300477716e-06, "loss": 1.7224, "step": 439 }, { "epoch": 0.88, "grad_norm": 7.063853740692139, "learning_rate": 8.903970133383297e-06, "loss": 1.67, "step": 440 }, { "epoch": 0.882, "grad_norm": 9.282867431640625, "learning_rate": 8.896689824657371e-06, "loss": 1.5403, "step": 441 }, { "epoch": 0.884, "grad_norm": 7.045912265777588, "learning_rate": 8.889388413725857e-06, "loss": 1.5697, "step": 442 }, { "epoch": 0.886, "grad_norm": 7.684753894805908, "learning_rate": 8.882065940128946e-06, "loss": 1.6201, "step": 443 }, { "epoch": 0.888, "grad_norm": 11.638023376464844, "learning_rate": 8.874722443520898e-06, "loss": 1.6565, "step": 444 }, { "epoch": 0.89, "grad_norm": 8.81474781036377, "learning_rate": 8.867357963669821e-06, "loss": 1.6399, "step": 445 }, { "epoch": 0.892, "grad_norm": 9.439010620117188, "learning_rate": 8.859972540457451e-06, "loss": 1.7231, "step": 446 }, { "epoch": 0.894, "grad_norm": 8.762112617492676, "learning_rate": 8.852566213878947e-06, "loss": 1.5841, "step": 447 }, { "epoch": 0.896, "grad_norm": 8.807167053222656, "learning_rate": 8.845139024042664e-06, "loss": 1.6177, "step": 448 }, { "epoch": 0.898, "grad_norm": 7.984807014465332, "learning_rate": 8.837691011169944e-06, "loss": 1.6911, "step": 449 }, { "epoch": 0.9, "grad_norm": 9.620468139648438, "learning_rate": 8.83022221559489e-06, "loss": 1.6385, "step": 450 }, { "epoch": 0.902, "grad_norm": 8.629189491271973, "learning_rate": 8.822732677764158e-06, "loss": 1.6133, "step": 451 }, { "epoch": 0.904, "grad_norm": 10.831542015075684, "learning_rate": 8.815222438236726e-06, "loss": 1.655, "step": 452 }, { "epoch": 0.906, "grad_norm": 7.558828353881836, "learning_rate": 8.807691537683685e-06, "loss": 1.5384, "step": 453 }, { "epoch": 0.908, "grad_norm": 10.0075044631958, "learning_rate": 8.800140016888009e-06, "loss": 1.6663, "step": 454 }, { "epoch": 0.91, "grad_norm": 8.153804779052734, "learning_rate": 8.792567916744346e-06, "loss": 1.73, "step": 455 }, { "epoch": 0.912, "grad_norm": 7.489542007446289, "learning_rate": 8.784975278258783e-06, "loss": 1.6016, "step": 456 }, { "epoch": 0.914, "grad_norm": 6.917328357696533, "learning_rate": 8.777362142548636e-06, "loss": 1.6948, "step": 457 }, { "epoch": 0.916, "grad_norm": 6.830845832824707, "learning_rate": 8.769728550842217e-06, "loss": 1.6222, "step": 458 }, { "epoch": 0.918, "grad_norm": 7.163386821746826, "learning_rate": 8.762074544478622e-06, "loss": 1.6015, "step": 459 }, { "epoch": 0.92, "grad_norm": 8.106656074523926, "learning_rate": 8.754400164907496e-06, "loss": 1.6755, "step": 460 }, { "epoch": 0.922, "grad_norm": 8.358505249023438, "learning_rate": 8.746705453688815e-06, "loss": 1.5817, "step": 461 }, { "epoch": 0.924, "grad_norm": 10.56636905670166, "learning_rate": 8.73899045249266e-06, "loss": 1.6605, "step": 462 }, { "epoch": 0.926, "grad_norm": 8.793217658996582, "learning_rate": 8.73125520309899e-06, "loss": 1.6739, "step": 463 }, { "epoch": 0.928, "grad_norm": 7.576605319976807, "learning_rate": 8.723499747397415e-06, "loss": 1.5703, "step": 464 }, { "epoch": 0.93, "grad_norm": 9.332236289978027, "learning_rate": 8.715724127386971e-06, "loss": 1.5906, "step": 465 }, { "epoch": 0.932, "grad_norm": 8.388971328735352, "learning_rate": 8.707928385175898e-06, "loss": 1.6549, "step": 466 }, { "epoch": 0.934, "grad_norm": 7.761672019958496, "learning_rate": 8.700112562981398e-06, "loss": 1.5616, "step": 467 }, { "epoch": 0.936, "grad_norm": 9.178117752075195, "learning_rate": 8.692276703129421e-06, "loss": 1.7963, "step": 468 }, { "epoch": 0.938, "grad_norm": 6.7894287109375, "learning_rate": 8.68442084805442e-06, "loss": 1.579, "step": 469 }, { "epoch": 0.94, "grad_norm": 8.14854907989502, "learning_rate": 8.676545040299145e-06, "loss": 1.7024, "step": 470 }, { "epoch": 0.942, "grad_norm": 8.014143943786621, "learning_rate": 8.668649322514382e-06, "loss": 1.6969, "step": 471 }, { "epoch": 0.944, "grad_norm": 7.022289752960205, "learning_rate": 8.660733737458751e-06, "loss": 1.5763, "step": 472 }, { "epoch": 0.946, "grad_norm": 7.455463886260986, "learning_rate": 8.652798327998458e-06, "loss": 1.6751, "step": 473 }, { "epoch": 0.948, "grad_norm": 10.453346252441406, "learning_rate": 8.644843137107058e-06, "loss": 1.6393, "step": 474 }, { "epoch": 0.95, "grad_norm": 10.011568069458008, "learning_rate": 8.636868207865244e-06, "loss": 1.6452, "step": 475 }, { "epoch": 0.952, "grad_norm": 10.156243324279785, "learning_rate": 8.628873583460593e-06, "loss": 1.6624, "step": 476 }, { "epoch": 0.954, "grad_norm": 8.829718589782715, "learning_rate": 8.620859307187339e-06, "loss": 1.6908, "step": 477 }, { "epoch": 0.956, "grad_norm": 10.530731201171875, "learning_rate": 8.61282542244614e-06, "loss": 1.6371, "step": 478 }, { "epoch": 0.958, "grad_norm": 9.92332935333252, "learning_rate": 8.604771972743848e-06, "loss": 1.6542, "step": 479 }, { "epoch": 0.96, "grad_norm": 6.439817905426025, "learning_rate": 8.596699001693257e-06, "loss": 1.7083, "step": 480 }, { "epoch": 0.962, "grad_norm": 7.941191673278809, "learning_rate": 8.588606553012884e-06, "loss": 1.5819, "step": 481 }, { "epoch": 0.964, "grad_norm": 7.217868328094482, "learning_rate": 8.580494670526725e-06, "loss": 1.7362, "step": 482 }, { "epoch": 0.966, "grad_norm": 8.529379844665527, "learning_rate": 8.572363398164017e-06, "loss": 1.7196, "step": 483 }, { "epoch": 0.968, "grad_norm": 8.774826049804688, "learning_rate": 8.564212779959003e-06, "loss": 1.6536, "step": 484 }, { "epoch": 0.97, "grad_norm": 8.855049133300781, "learning_rate": 8.556042860050686e-06, "loss": 1.6759, "step": 485 }, { "epoch": 0.972, "grad_norm": 7.4477620124816895, "learning_rate": 8.547853682682605e-06, "loss": 1.6095, "step": 486 }, { "epoch": 0.974, "grad_norm": 8.155046463012695, "learning_rate": 8.539645292202579e-06, "loss": 1.6397, "step": 487 }, { "epoch": 0.976, "grad_norm": 9.956945419311523, "learning_rate": 8.531417733062476e-06, "loss": 1.5748, "step": 488 }, { "epoch": 0.978, "grad_norm": 9.320968627929688, "learning_rate": 8.523171049817974e-06, "loss": 1.6588, "step": 489 }, { "epoch": 0.98, "grad_norm": 7.276108264923096, "learning_rate": 8.51490528712831e-06, "loss": 1.6456, "step": 490 }, { "epoch": 0.982, "grad_norm": 11.987957954406738, "learning_rate": 8.506620489756045e-06, "loss": 1.6061, "step": 491 }, { "epoch": 0.984, "grad_norm": 9.098200798034668, "learning_rate": 8.498316702566828e-06, "loss": 1.6903, "step": 492 }, { "epoch": 0.986, "grad_norm": 7.814845561981201, "learning_rate": 8.489993970529137e-06, "loss": 1.4855, "step": 493 }, { "epoch": 0.988, "grad_norm": 7.096703052520752, "learning_rate": 8.481652338714048e-06, "loss": 1.6221, "step": 494 }, { "epoch": 0.99, "grad_norm": 6.307745456695557, "learning_rate": 8.473291852294986e-06, "loss": 1.5903, "step": 495 }, { "epoch": 0.992, "grad_norm": 7.393770694732666, "learning_rate": 8.464912556547486e-06, "loss": 1.5778, "step": 496 }, { "epoch": 0.994, "grad_norm": 21.961185455322266, "learning_rate": 8.456514496848938e-06, "loss": 1.7559, "step": 497 }, { "epoch": 0.996, "grad_norm": 12.971484184265137, "learning_rate": 8.44809771867835e-06, "loss": 1.6592, "step": 498 }, { "epoch": 0.998, "grad_norm": 12.674059867858887, "learning_rate": 8.439662267616093e-06, "loss": 1.6422, "step": 499 }, { "epoch": 1.0, "grad_norm": 10.149175643920898, "learning_rate": 8.43120818934367e-06, "loss": 1.6264, "step": 500 }, { "epoch": 1.002, "grad_norm": 14.11556625366211, "learning_rate": 8.422735529643445e-06, "loss": 1.5558, "step": 501 }, { "epoch": 1.004, "grad_norm": 10.902586936950684, "learning_rate": 8.414244334398418e-06, "loss": 1.6304, "step": 502 }, { "epoch": 1.006, "grad_norm": 9.201044082641602, "learning_rate": 8.405734649591964e-06, "loss": 1.7151, "step": 503 }, { "epoch": 1.008, "grad_norm": 8.613140106201172, "learning_rate": 8.397206521307584e-06, "loss": 1.6446, "step": 504 }, { "epoch": 1.01, "grad_norm": 11.33934497833252, "learning_rate": 8.388659995728662e-06, "loss": 1.5727, "step": 505 }, { "epoch": 1.012, "grad_norm": 9.204126358032227, "learning_rate": 8.380095119138209e-06, "loss": 1.6427, "step": 506 }, { "epoch": 1.014, "grad_norm": 13.722872734069824, "learning_rate": 8.371511937918616e-06, "loss": 1.5922, "step": 507 }, { "epoch": 1.016, "grad_norm": 9.983489036560059, "learning_rate": 8.362910498551402e-06, "loss": 1.5911, "step": 508 }, { "epoch": 1.018, "grad_norm": 6.286109447479248, "learning_rate": 8.354290847616954e-06, "loss": 1.581, "step": 509 }, { "epoch": 1.02, "grad_norm": 8.981342315673828, "learning_rate": 8.345653031794292e-06, "loss": 1.6099, "step": 510 }, { "epoch": 1.022, "grad_norm": 10.48516845703125, "learning_rate": 8.3369970978608e-06, "loss": 1.4918, "step": 511 }, { "epoch": 1.024, "grad_norm": 10.683873176574707, "learning_rate": 8.328323092691985e-06, "loss": 1.7634, "step": 512 }, { "epoch": 1.026, "grad_norm": 9.128839492797852, "learning_rate": 8.319631063261209e-06, "loss": 1.5248, "step": 513 }, { "epoch": 1.028, "grad_norm": 7.288445949554443, "learning_rate": 8.310921056639451e-06, "loss": 1.597, "step": 514 }, { "epoch": 1.03, "grad_norm": 7.43460750579834, "learning_rate": 8.302193119995038e-06, "loss": 1.469, "step": 515 }, { "epoch": 1.032, "grad_norm": 11.071965217590332, "learning_rate": 8.293447300593402e-06, "loss": 1.6112, "step": 516 }, { "epoch": 1.034, "grad_norm": 9.424286842346191, "learning_rate": 8.284683645796814e-06, "loss": 1.6145, "step": 517 }, { "epoch": 1.036, "grad_norm": 10.420470237731934, "learning_rate": 8.275902203064125e-06, "loss": 1.7171, "step": 518 }, { "epoch": 1.038, "grad_norm": 9.002021789550781, "learning_rate": 8.267103019950529e-06, "loss": 1.603, "step": 519 }, { "epoch": 1.04, "grad_norm": 12.653961181640625, "learning_rate": 8.258286144107277e-06, "loss": 1.6203, "step": 520 }, { "epoch": 1.042, "grad_norm": 7.975899696350098, "learning_rate": 8.249451623281444e-06, "loss": 1.5552, "step": 521 }, { "epoch": 1.044, "grad_norm": 7.326491832733154, "learning_rate": 8.240599505315656e-06, "loss": 1.6063, "step": 522 }, { "epoch": 1.046, "grad_norm": 7.846243381500244, "learning_rate": 8.231729838147833e-06, "loss": 1.5635, "step": 523 }, { "epoch": 1.048, "grad_norm": 9.442638397216797, "learning_rate": 8.222842669810936e-06, "loss": 1.5853, "step": 524 }, { "epoch": 1.05, "grad_norm": 8.490560531616211, "learning_rate": 8.213938048432697e-06, "loss": 1.5719, "step": 525 }, { "epoch": 1.052, "grad_norm": 9.718538284301758, "learning_rate": 8.205016022235368e-06, "loss": 1.6211, "step": 526 }, { "epoch": 1.054, "grad_norm": 8.703619956970215, "learning_rate": 8.196076639535453e-06, "loss": 1.6081, "step": 527 }, { "epoch": 1.056, "grad_norm": 11.14815902709961, "learning_rate": 8.18711994874345e-06, "loss": 1.6158, "step": 528 }, { "epoch": 1.058, "grad_norm": 9.741596221923828, "learning_rate": 8.178145998363585e-06, "loss": 1.5314, "step": 529 }, { "epoch": 1.06, "grad_norm": 9.530781745910645, "learning_rate": 8.16915483699355e-06, "loss": 1.5866, "step": 530 }, { "epoch": 1.062, "grad_norm": 8.731345176696777, "learning_rate": 8.160146513324256e-06, "loss": 1.5666, "step": 531 }, { "epoch": 1.064, "grad_norm": 11.338554382324219, "learning_rate": 8.151121076139534e-06, "loss": 1.5834, "step": 532 }, { "epoch": 1.066, "grad_norm": 20.54311752319336, "learning_rate": 8.142078574315907e-06, "loss": 1.5652, "step": 533 }, { "epoch": 1.068, "grad_norm": 11.977243423461914, "learning_rate": 8.133019056822303e-06, "loss": 1.4832, "step": 534 }, { "epoch": 1.07, "grad_norm": 10.514939308166504, "learning_rate": 8.123942572719801e-06, "loss": 1.6011, "step": 535 }, { "epoch": 1.072, "grad_norm": 9.276664733886719, "learning_rate": 8.11484917116136e-06, "loss": 1.5909, "step": 536 }, { "epoch": 1.074, "grad_norm": 11.073919296264648, "learning_rate": 8.105738901391553e-06, "loss": 1.5826, "step": 537 }, { "epoch": 1.076, "grad_norm": 9.078001022338867, "learning_rate": 8.096611812746302e-06, "loss": 1.641, "step": 538 }, { "epoch": 1.078, "grad_norm": 10.105594635009766, "learning_rate": 8.087467954652608e-06, "loss": 1.6196, "step": 539 }, { "epoch": 1.08, "grad_norm": 13.397587776184082, "learning_rate": 8.078307376628292e-06, "loss": 1.562, "step": 540 }, { "epoch": 1.082, "grad_norm": 13.389184951782227, "learning_rate": 8.069130128281714e-06, "loss": 1.5884, "step": 541 }, { "epoch": 1.084, "grad_norm": 10.954330444335938, "learning_rate": 8.059936259311514e-06, "loss": 1.6104, "step": 542 }, { "epoch": 1.086, "grad_norm": 8.864795684814453, "learning_rate": 8.05072581950634e-06, "loss": 1.6024, "step": 543 }, { "epoch": 1.088, "grad_norm": 9.769396781921387, "learning_rate": 8.041498858744572e-06, "loss": 1.553, "step": 544 }, { "epoch": 1.09, "grad_norm": 8.69478988647461, "learning_rate": 8.032255426994069e-06, "loss": 1.6915, "step": 545 }, { "epoch": 1.092, "grad_norm": 8.354500770568848, "learning_rate": 8.022995574311876e-06, "loss": 1.5773, "step": 546 }, { "epoch": 1.094, "grad_norm": 9.368426322937012, "learning_rate": 8.013719350843969e-06, "loss": 1.6233, "step": 547 }, { "epoch": 1.096, "grad_norm": 8.086129188537598, "learning_rate": 8.004426806824985e-06, "loss": 1.4905, "step": 548 }, { "epoch": 1.098, "grad_norm": 8.991674423217773, "learning_rate": 7.99511799257793e-06, "loss": 1.4947, "step": 549 }, { "epoch": 1.1, "grad_norm": 9.440938949584961, "learning_rate": 7.985792958513932e-06, "loss": 1.6062, "step": 550 }, { "epoch": 1.102, "grad_norm": 7.645297527313232, "learning_rate": 7.97645175513195e-06, "loss": 1.5178, "step": 551 }, { "epoch": 1.104, "grad_norm": 8.402973175048828, "learning_rate": 7.967094433018508e-06, "loss": 1.571, "step": 552 }, { "epoch": 1.106, "grad_norm": 9.366331100463867, "learning_rate": 7.95772104284742e-06, "loss": 1.4459, "step": 553 }, { "epoch": 1.108, "grad_norm": 9.664446830749512, "learning_rate": 7.948331635379517e-06, "loss": 1.5755, "step": 554 }, { "epoch": 1.11, "grad_norm": 10.556148529052734, "learning_rate": 7.938926261462366e-06, "loss": 1.5738, "step": 555 }, { "epoch": 1.112, "grad_norm": 10.46329116821289, "learning_rate": 7.929504972030003e-06, "loss": 1.7149, "step": 556 }, { "epoch": 1.114, "grad_norm": 9.829599380493164, "learning_rate": 7.920067818102652e-06, "loss": 1.5629, "step": 557 }, { "epoch": 1.116, "grad_norm": 8.370344161987305, "learning_rate": 7.910614850786448e-06, "loss": 1.6013, "step": 558 }, { "epoch": 1.1179999999999999, "grad_norm": 9.646671295166016, "learning_rate": 7.901146121273165e-06, "loss": 1.6073, "step": 559 }, { "epoch": 1.12, "grad_norm": 8.669224739074707, "learning_rate": 7.891661680839932e-06, "loss": 1.5163, "step": 560 }, { "epoch": 1.1219999999999999, "grad_norm": 11.22138500213623, "learning_rate": 7.882161580848966e-06, "loss": 1.6723, "step": 561 }, { "epoch": 1.124, "grad_norm": 9.086870193481445, "learning_rate": 7.872645872747281e-06, "loss": 1.4885, "step": 562 }, { "epoch": 1.126, "grad_norm": 11.861745834350586, "learning_rate": 7.863114608066417e-06, "loss": 1.6129, "step": 563 }, { "epoch": 1.1280000000000001, "grad_norm": 7.978770732879639, "learning_rate": 7.85356783842216e-06, "loss": 1.6875, "step": 564 }, { "epoch": 1.13, "grad_norm": 9.279024124145508, "learning_rate": 7.84400561551426e-06, "loss": 1.5874, "step": 565 }, { "epoch": 1.1320000000000001, "grad_norm": 10.196725845336914, "learning_rate": 7.834427991126155e-06, "loss": 1.5669, "step": 566 }, { "epoch": 1.134, "grad_norm": 12.994475364685059, "learning_rate": 7.82483501712469e-06, "loss": 1.6626, "step": 567 }, { "epoch": 1.1360000000000001, "grad_norm": 11.740025520324707, "learning_rate": 7.815226745459831e-06, "loss": 1.5378, "step": 568 }, { "epoch": 1.138, "grad_norm": 8.01264762878418, "learning_rate": 7.80560322816439e-06, "loss": 1.592, "step": 569 }, { "epoch": 1.1400000000000001, "grad_norm": 9.302233695983887, "learning_rate": 7.795964517353734e-06, "loss": 1.5832, "step": 570 }, { "epoch": 1.142, "grad_norm": 9.458878517150879, "learning_rate": 7.786310665225522e-06, "loss": 1.6201, "step": 571 }, { "epoch": 1.144, "grad_norm": 19.037525177001953, "learning_rate": 7.776641724059398e-06, "loss": 1.7447, "step": 572 }, { "epoch": 1.146, "grad_norm": 11.674372673034668, "learning_rate": 7.76695774621672e-06, "loss": 1.6556, "step": 573 }, { "epoch": 1.148, "grad_norm": 11.437299728393555, "learning_rate": 7.757258784140286e-06, "loss": 1.6109, "step": 574 }, { "epoch": 1.15, "grad_norm": 10.853487014770508, "learning_rate": 7.747544890354031e-06, "loss": 1.5528, "step": 575 }, { "epoch": 1.152, "grad_norm": 6.265218734741211, "learning_rate": 7.737816117462752e-06, "loss": 1.5691, "step": 576 }, { "epoch": 1.154, "grad_norm": 8.988601684570312, "learning_rate": 7.728072518151826e-06, "loss": 1.646, "step": 577 }, { "epoch": 1.156, "grad_norm": 10.684486389160156, "learning_rate": 7.718314145186918e-06, "loss": 1.6239, "step": 578 }, { "epoch": 1.158, "grad_norm": 8.397745132446289, "learning_rate": 7.7085410514137e-06, "loss": 1.5766, "step": 579 }, { "epoch": 1.16, "grad_norm": 8.635762214660645, "learning_rate": 7.698753289757565e-06, "loss": 1.5302, "step": 580 }, { "epoch": 1.162, "grad_norm": 11.5667085647583, "learning_rate": 7.688950913223336e-06, "loss": 1.5859, "step": 581 }, { "epoch": 1.164, "grad_norm": 6.988083839416504, "learning_rate": 7.679133974894984e-06, "loss": 1.6023, "step": 582 }, { "epoch": 1.166, "grad_norm": 8.85438346862793, "learning_rate": 7.669302527935334e-06, "loss": 1.7383, "step": 583 }, { "epoch": 1.168, "grad_norm": 8.871450424194336, "learning_rate": 7.65945662558579e-06, "loss": 1.5443, "step": 584 }, { "epoch": 1.17, "grad_norm": 6.740708827972412, "learning_rate": 7.649596321166024e-06, "loss": 1.581, "step": 585 }, { "epoch": 1.172, "grad_norm": 7.056929111480713, "learning_rate": 7.639721668073718e-06, "loss": 1.4831, "step": 586 }, { "epoch": 1.174, "grad_norm": 9.177947998046875, "learning_rate": 7.629832719784245e-06, "loss": 1.5719, "step": 587 }, { "epoch": 1.176, "grad_norm": 9.50424861907959, "learning_rate": 7.619929529850397e-06, "loss": 1.5951, "step": 588 }, { "epoch": 1.178, "grad_norm": 7.7955498695373535, "learning_rate": 7.610012151902091e-06, "loss": 1.6338, "step": 589 }, { "epoch": 1.18, "grad_norm": 9.096726417541504, "learning_rate": 7.600080639646077e-06, "loss": 1.6591, "step": 590 }, { "epoch": 1.182, "grad_norm": 8.232474327087402, "learning_rate": 7.590135046865652e-06, "loss": 1.4726, "step": 591 }, { "epoch": 1.184, "grad_norm": 8.038969039916992, "learning_rate": 7.580175427420358e-06, "loss": 1.7216, "step": 592 }, { "epoch": 1.186, "grad_norm": 10.49866771697998, "learning_rate": 7.570201835245703e-06, "loss": 1.55, "step": 593 }, { "epoch": 1.188, "grad_norm": 8.377806663513184, "learning_rate": 7.560214324352858e-06, "loss": 1.5407, "step": 594 }, { "epoch": 1.19, "grad_norm": 11.32930850982666, "learning_rate": 7.550212948828377e-06, "loss": 1.5192, "step": 595 }, { "epoch": 1.192, "grad_norm": 9.1795015335083, "learning_rate": 7.54019776283389e-06, "loss": 1.5734, "step": 596 }, { "epoch": 1.194, "grad_norm": 8.019397735595703, "learning_rate": 7.530168820605819e-06, "loss": 1.5216, "step": 597 }, { "epoch": 1.196, "grad_norm": 25.26226234436035, "learning_rate": 7.520126176455084e-06, "loss": 1.6976, "step": 598 }, { "epoch": 1.198, "grad_norm": 9.354122161865234, "learning_rate": 7.510069884766802e-06, "loss": 1.5146, "step": 599 }, { "epoch": 1.2, "grad_norm": 6.69480037689209, "learning_rate": 7.500000000000001e-06, "loss": 1.5678, "step": 600 }, { "epoch": 1.202, "grad_norm": 8.49975299835205, "learning_rate": 7.489916576687318e-06, "loss": 1.5507, "step": 601 }, { "epoch": 1.204, "grad_norm": 6.0479416847229, "learning_rate": 7.479819669434712e-06, "loss": 1.5298, "step": 602 }, { "epoch": 1.206, "grad_norm": 12.379949569702148, "learning_rate": 7.469709332921155e-06, "loss": 1.5703, "step": 603 }, { "epoch": 1.208, "grad_norm": 9.304234504699707, "learning_rate": 7.459585621898353e-06, "loss": 1.4965, "step": 604 }, { "epoch": 1.21, "grad_norm": 8.749626159667969, "learning_rate": 7.449448591190436e-06, "loss": 1.5667, "step": 605 }, { "epoch": 1.212, "grad_norm": 12.647852897644043, "learning_rate": 7.4392982956936644e-06, "loss": 1.518, "step": 606 }, { "epoch": 1.214, "grad_norm": 11.38586139678955, "learning_rate": 7.429134790376136e-06, "loss": 1.5827, "step": 607 }, { "epoch": 1.216, "grad_norm": 9.004302978515625, "learning_rate": 7.418958130277483e-06, "loss": 1.5788, "step": 608 }, { "epoch": 1.218, "grad_norm": 12.129622459411621, "learning_rate": 7.408768370508577e-06, "loss": 1.6698, "step": 609 }, { "epoch": 1.22, "grad_norm": 8.174181938171387, "learning_rate": 7.398565566251232e-06, "loss": 1.6104, "step": 610 }, { "epoch": 1.222, "grad_norm": 9.542710304260254, "learning_rate": 7.3883497727579e-06, "loss": 1.527, "step": 611 }, { "epoch": 1.224, "grad_norm": 8.519256591796875, "learning_rate": 7.378121045351378e-06, "loss": 1.6646, "step": 612 }, { "epoch": 1.226, "grad_norm": 10.969525337219238, "learning_rate": 7.3678794394245e-06, "loss": 1.5838, "step": 613 }, { "epoch": 1.228, "grad_norm": 12.213068008422852, "learning_rate": 7.357625010439853e-06, "loss": 1.5068, "step": 614 }, { "epoch": 1.23, "grad_norm": 14.603283882141113, "learning_rate": 7.347357813929455e-06, "loss": 1.7072, "step": 615 }, { "epoch": 1.232, "grad_norm": 9.047625541687012, "learning_rate": 7.337077905494472e-06, "loss": 1.5516, "step": 616 }, { "epoch": 1.234, "grad_norm": 8.961601257324219, "learning_rate": 7.326785340804908e-06, "loss": 1.6053, "step": 617 }, { "epoch": 1.236, "grad_norm": 10.050729751586914, "learning_rate": 7.31648017559931e-06, "loss": 1.5849, "step": 618 }, { "epoch": 1.238, "grad_norm": 6.469438076019287, "learning_rate": 7.3061624656844544e-06, "loss": 1.6075, "step": 619 }, { "epoch": 1.24, "grad_norm": 11.897672653198242, "learning_rate": 7.295832266935059e-06, "loss": 1.5038, "step": 620 }, { "epoch": 1.242, "grad_norm": 7.534898281097412, "learning_rate": 7.285489635293472e-06, "loss": 1.5344, "step": 621 }, { "epoch": 1.244, "grad_norm": 8.457494735717773, "learning_rate": 7.275134626769369e-06, "loss": 1.6063, "step": 622 }, { "epoch": 1.246, "grad_norm": 7.499537467956543, "learning_rate": 7.264767297439455e-06, "loss": 1.5974, "step": 623 }, { "epoch": 1.248, "grad_norm": 10.001435279846191, "learning_rate": 7.254387703447154e-06, "loss": 1.6185, "step": 624 }, { "epoch": 1.25, "grad_norm": 10.439517974853516, "learning_rate": 7.243995901002312e-06, "loss": 1.6371, "step": 625 }, { "epoch": 1.252, "grad_norm": 8.107089042663574, "learning_rate": 7.233591946380884e-06, "loss": 1.5729, "step": 626 }, { "epoch": 1.254, "grad_norm": 9.131059646606445, "learning_rate": 7.223175895924638e-06, "loss": 1.5677, "step": 627 }, { "epoch": 1.256, "grad_norm": 6.757850646972656, "learning_rate": 7.212747806040845e-06, "loss": 1.5922, "step": 628 }, { "epoch": 1.258, "grad_norm": 9.472972869873047, "learning_rate": 7.2023077332019755e-06, "loss": 1.6055, "step": 629 }, { "epoch": 1.26, "grad_norm": 7.784339427947998, "learning_rate": 7.191855733945388e-06, "loss": 1.461, "step": 630 }, { "epoch": 1.262, "grad_norm": 9.098430633544922, "learning_rate": 7.181391864873034e-06, "loss": 1.5214, "step": 631 }, { "epoch": 1.264, "grad_norm": 8.37144947052002, "learning_rate": 7.170916182651141e-06, "loss": 1.5766, "step": 632 }, { "epoch": 1.266, "grad_norm": 9.366813659667969, "learning_rate": 7.160428744009913e-06, "loss": 1.5955, "step": 633 }, { "epoch": 1.268, "grad_norm": 11.189432144165039, "learning_rate": 7.149929605743214e-06, "loss": 1.6196, "step": 634 }, { "epoch": 1.27, "grad_norm": 9.027175903320312, "learning_rate": 7.1394188247082715e-06, "loss": 1.5104, "step": 635 }, { "epoch": 1.272, "grad_norm": 7.792656898498535, "learning_rate": 7.128896457825364e-06, "loss": 1.6022, "step": 636 }, { "epoch": 1.274, "grad_norm": 7.127891540527344, "learning_rate": 7.118362562077508e-06, "loss": 1.624, "step": 637 }, { "epoch": 1.276, "grad_norm": 7.897225379943848, "learning_rate": 7.107817194510157e-06, "loss": 1.6961, "step": 638 }, { "epoch": 1.278, "grad_norm": 8.84518051147461, "learning_rate": 7.0972604122308865e-06, "loss": 1.4578, "step": 639 }, { "epoch": 1.28, "grad_norm": 7.849598407745361, "learning_rate": 7.08669227240909e-06, "loss": 1.6101, "step": 640 }, { "epoch": 1.282, "grad_norm": 9.6469144821167, "learning_rate": 7.076112832275667e-06, "loss": 1.6093, "step": 641 }, { "epoch": 1.284, "grad_norm": 5.705495834350586, "learning_rate": 7.06552214912271e-06, "loss": 1.5104, "step": 642 }, { "epoch": 1.286, "grad_norm": 7.658092975616455, "learning_rate": 7.054920280303199e-06, "loss": 1.666, "step": 643 }, { "epoch": 1.288, "grad_norm": 7.542886734008789, "learning_rate": 7.04430728323069e-06, "loss": 1.6318, "step": 644 }, { "epoch": 1.29, "grad_norm": 9.100532531738281, "learning_rate": 7.033683215379002e-06, "loss": 1.6983, "step": 645 }, { "epoch": 1.292, "grad_norm": 7.658532619476318, "learning_rate": 7.023048134281907e-06, "loss": 1.5868, "step": 646 }, { "epoch": 1.294, "grad_norm": 10.516380310058594, "learning_rate": 7.012402097532815e-06, "loss": 1.489, "step": 647 }, { "epoch": 1.296, "grad_norm": 7.941004276275635, "learning_rate": 7.0017451627844765e-06, "loss": 1.6114, "step": 648 }, { "epoch": 1.298, "grad_norm": 8.073655128479004, "learning_rate": 6.991077387748643e-06, "loss": 1.6023, "step": 649 }, { "epoch": 1.3, "grad_norm": 8.82116985321045, "learning_rate": 6.980398830195785e-06, "loss": 1.5371, "step": 650 }, { "epoch": 1.302, "grad_norm": 10.482192039489746, "learning_rate": 6.9697095479547564e-06, "loss": 1.6303, "step": 651 }, { "epoch": 1.304, "grad_norm": 12.892633438110352, "learning_rate": 6.959009598912493e-06, "loss": 1.4729, "step": 652 }, { "epoch": 1.306, "grad_norm": 7.209946632385254, "learning_rate": 6.948299041013695e-06, "loss": 1.493, "step": 653 }, { "epoch": 1.308, "grad_norm": 7.52318811416626, "learning_rate": 6.9375779322605154e-06, "loss": 1.5806, "step": 654 }, { "epoch": 1.31, "grad_norm": 10.265365600585938, "learning_rate": 6.9268463307122425e-06, "loss": 1.5159, "step": 655 }, { "epoch": 1.312, "grad_norm": 10.420729637145996, "learning_rate": 6.916104294484988e-06, "loss": 1.5097, "step": 656 }, { "epoch": 1.314, "grad_norm": 7.465059280395508, "learning_rate": 6.905351881751372e-06, "loss": 1.6499, "step": 657 }, { "epoch": 1.316, "grad_norm": 9.007606506347656, "learning_rate": 6.8945891507402075e-06, "loss": 1.5783, "step": 658 }, { "epoch": 1.318, "grad_norm": 11.8942289352417, "learning_rate": 6.883816159736187e-06, "loss": 1.6481, "step": 659 }, { "epoch": 1.32, "grad_norm": 10.41662311553955, "learning_rate": 6.873032967079562e-06, "loss": 1.6615, "step": 660 }, { "epoch": 1.322, "grad_norm": 11.450822830200195, "learning_rate": 6.862239631165831e-06, "loss": 1.6254, "step": 661 }, { "epoch": 1.324, "grad_norm": 8.604270935058594, "learning_rate": 6.851436210445427e-06, "loss": 1.6191, "step": 662 }, { "epoch": 1.326, "grad_norm": 7.488931655883789, "learning_rate": 6.840622763423391e-06, "loss": 1.6588, "step": 663 }, { "epoch": 1.328, "grad_norm": 9.091219902038574, "learning_rate": 6.829799348659061e-06, "loss": 1.6639, "step": 664 }, { "epoch": 1.33, "grad_norm": 8.698543548583984, "learning_rate": 6.818966024765758e-06, "loss": 1.5772, "step": 665 }, { "epoch": 1.332, "grad_norm": 13.658153533935547, "learning_rate": 6.808122850410461e-06, "loss": 1.4672, "step": 666 }, { "epoch": 1.334, "grad_norm": 26.972639083862305, "learning_rate": 6.7972698843135e-06, "loss": 1.6459, "step": 667 }, { "epoch": 1.336, "grad_norm": 9.072081565856934, "learning_rate": 6.7864071852482205e-06, "loss": 1.6711, "step": 668 }, { "epoch": 1.338, "grad_norm": 8.552204132080078, "learning_rate": 6.775534812040686e-06, "loss": 1.5927, "step": 669 }, { "epoch": 1.34, "grad_norm": 9.554261207580566, "learning_rate": 6.7646528235693445e-06, "loss": 1.6078, "step": 670 }, { "epoch": 1.342, "grad_norm": 9.100646018981934, "learning_rate": 6.753761278764719e-06, "loss": 1.6746, "step": 671 }, { "epoch": 1.3439999999999999, "grad_norm": 10.204363822937012, "learning_rate": 6.7428602366090764e-06, "loss": 1.626, "step": 672 }, { "epoch": 1.346, "grad_norm": 10.063047409057617, "learning_rate": 6.7319497561361245e-06, "loss": 1.6601, "step": 673 }, { "epoch": 1.3479999999999999, "grad_norm": 13.726625442504883, "learning_rate": 6.721029896430678e-06, "loss": 1.583, "step": 674 }, { "epoch": 1.35, "grad_norm": 11.784052848815918, "learning_rate": 6.710100716628345e-06, "loss": 1.6688, "step": 675 }, { "epoch": 1.3519999999999999, "grad_norm": 8.559849739074707, "learning_rate": 6.699162275915208e-06, "loss": 1.6578, "step": 676 }, { "epoch": 1.354, "grad_norm": 9.396211624145508, "learning_rate": 6.6882146335274955e-06, "loss": 1.6294, "step": 677 }, { "epoch": 1.3559999999999999, "grad_norm": 7.74012565612793, "learning_rate": 6.677257848751276e-06, "loss": 1.5466, "step": 678 }, { "epoch": 1.358, "grad_norm": 10.365443229675293, "learning_rate": 6.666291980922122e-06, "loss": 1.5296, "step": 679 }, { "epoch": 1.3599999999999999, "grad_norm": 7.977329254150391, "learning_rate": 6.655317089424791e-06, "loss": 1.5837, "step": 680 }, { "epoch": 1.362, "grad_norm": 8.795341491699219, "learning_rate": 6.644333233692917e-06, "loss": 1.5884, "step": 681 }, { "epoch": 1.3639999999999999, "grad_norm": 13.247668266296387, "learning_rate": 6.633340473208673e-06, "loss": 1.5977, "step": 682 }, { "epoch": 1.366, "grad_norm": 8.479852676391602, "learning_rate": 6.622338867502452e-06, "loss": 1.5983, "step": 683 }, { "epoch": 1.3679999999999999, "grad_norm": 8.384832382202148, "learning_rate": 6.611328476152557e-06, "loss": 1.5377, "step": 684 }, { "epoch": 1.37, "grad_norm": 7.661881923675537, "learning_rate": 6.600309358784858e-06, "loss": 1.4785, "step": 685 }, { "epoch": 1.3719999999999999, "grad_norm": 9.14194107055664, "learning_rate": 6.58928157507249e-06, "loss": 1.5439, "step": 686 }, { "epoch": 1.374, "grad_norm": 8.849069595336914, "learning_rate": 6.578245184735513e-06, "loss": 1.5466, "step": 687 }, { "epoch": 1.376, "grad_norm": 8.54023551940918, "learning_rate": 6.567200247540599e-06, "loss": 1.5738, "step": 688 }, { "epoch": 1.3780000000000001, "grad_norm": 13.907352447509766, "learning_rate": 6.556146823300701e-06, "loss": 1.6541, "step": 689 }, { "epoch": 1.38, "grad_norm": 10.435159683227539, "learning_rate": 6.545084971874738e-06, "loss": 1.6579, "step": 690 }, { "epoch": 1.3820000000000001, "grad_norm": 9.772438049316406, "learning_rate": 6.534014753167263e-06, "loss": 1.6995, "step": 691 }, { "epoch": 1.384, "grad_norm": 9.34780502319336, "learning_rate": 6.522936227128139e-06, "loss": 1.5157, "step": 692 }, { "epoch": 1.3860000000000001, "grad_norm": 8.92451000213623, "learning_rate": 6.5118494537522235e-06, "loss": 1.6148, "step": 693 }, { "epoch": 1.388, "grad_norm": 6.892347812652588, "learning_rate": 6.500754493079029e-06, "loss": 1.5511, "step": 694 }, { "epoch": 1.3900000000000001, "grad_norm": 10.89332389831543, "learning_rate": 6.48965140519241e-06, "loss": 1.5804, "step": 695 }, { "epoch": 1.392, "grad_norm": 7.6322221755981445, "learning_rate": 6.4785402502202345e-06, "loss": 1.6086, "step": 696 }, { "epoch": 1.3940000000000001, "grad_norm": 8.542652130126953, "learning_rate": 6.467421088334052e-06, "loss": 1.6496, "step": 697 }, { "epoch": 1.396, "grad_norm": 8.015932083129883, "learning_rate": 6.456293979748778e-06, "loss": 1.617, "step": 698 }, { "epoch": 1.3980000000000001, "grad_norm": 7.57454776763916, "learning_rate": 6.445158984722358e-06, "loss": 1.567, "step": 699 }, { "epoch": 1.4, "grad_norm": 8.950959205627441, "learning_rate": 6.434016163555452e-06, "loss": 1.5682, "step": 700 }, { "epoch": 1.4020000000000001, "grad_norm": 8.991364479064941, "learning_rate": 6.422865576591096e-06, "loss": 1.5012, "step": 701 }, { "epoch": 1.404, "grad_norm": 7.7304182052612305, "learning_rate": 6.411707284214384e-06, "loss": 1.5961, "step": 702 }, { "epoch": 1.4060000000000001, "grad_norm": 7.197605609893799, "learning_rate": 6.400541346852136e-06, "loss": 1.6654, "step": 703 }, { "epoch": 1.408, "grad_norm": 7.552187442779541, "learning_rate": 6.389367824972575e-06, "loss": 1.5137, "step": 704 }, { "epoch": 1.41, "grad_norm": 8.499285697937012, "learning_rate": 6.378186779084996e-06, "loss": 1.5841, "step": 705 }, { "epoch": 1.412, "grad_norm": 9.511728286743164, "learning_rate": 6.366998269739442e-06, "loss": 1.5504, "step": 706 }, { "epoch": 1.414, "grad_norm": 8.45433521270752, "learning_rate": 6.35580235752637e-06, "loss": 1.5983, "step": 707 }, { "epoch": 1.416, "grad_norm": 9.685602188110352, "learning_rate": 6.344599103076329e-06, "loss": 1.5785, "step": 708 }, { "epoch": 1.418, "grad_norm": 9.533397674560547, "learning_rate": 6.3333885670596285e-06, "loss": 1.5921, "step": 709 }, { "epoch": 1.42, "grad_norm": 10.1636381149292, "learning_rate": 6.322170810186013e-06, "loss": 1.6896, "step": 710 }, { "epoch": 1.422, "grad_norm": 9.969033241271973, "learning_rate": 6.310945893204324e-06, "loss": 1.6277, "step": 711 }, { "epoch": 1.424, "grad_norm": 13.378399848937988, "learning_rate": 6.299713876902188e-06, "loss": 1.6618, "step": 712 }, { "epoch": 1.426, "grad_norm": 7.862060546875, "learning_rate": 6.28847482210567e-06, "loss": 1.5287, "step": 713 }, { "epoch": 1.428, "grad_norm": 8.11763858795166, "learning_rate": 6.277228789678953e-06, "loss": 1.6417, "step": 714 }, { "epoch": 1.43, "grad_norm": 8.063044548034668, "learning_rate": 6.26597584052401e-06, "loss": 1.6309, "step": 715 }, { "epoch": 1.432, "grad_norm": 6.405232906341553, "learning_rate": 6.254716035580264e-06, "loss": 1.5913, "step": 716 }, { "epoch": 1.434, "grad_norm": 6.900762557983398, "learning_rate": 6.243449435824276e-06, "loss": 1.5563, "step": 717 }, { "epoch": 1.436, "grad_norm": 10.574906349182129, "learning_rate": 6.23217610226939e-06, "loss": 1.601, "step": 718 }, { "epoch": 1.438, "grad_norm": 7.271671295166016, "learning_rate": 6.220896095965428e-06, "loss": 1.5809, "step": 719 }, { "epoch": 1.44, "grad_norm": 8.701942443847656, "learning_rate": 6.209609477998339e-06, "loss": 1.588, "step": 720 }, { "epoch": 1.442, "grad_norm": 8.735617637634277, "learning_rate": 6.198316309489886e-06, "loss": 1.5435, "step": 721 }, { "epoch": 1.444, "grad_norm": 7.150242328643799, "learning_rate": 6.187016651597299e-06, "loss": 1.6309, "step": 722 }, { "epoch": 1.446, "grad_norm": 9.113722801208496, "learning_rate": 6.17571056551295e-06, "loss": 1.6128, "step": 723 }, { "epoch": 1.448, "grad_norm": 9.70335865020752, "learning_rate": 6.16439811246403e-06, "loss": 1.5812, "step": 724 }, { "epoch": 1.45, "grad_norm": 7.478845596313477, "learning_rate": 6.153079353712201e-06, "loss": 1.653, "step": 725 }, { "epoch": 1.452, "grad_norm": 7.009204864501953, "learning_rate": 6.141754350553279e-06, "loss": 1.432, "step": 726 }, { "epoch": 1.454, "grad_norm": 8.384693145751953, "learning_rate": 6.130423164316893e-06, "loss": 1.6201, "step": 727 }, { "epoch": 1.456, "grad_norm": 12.53249454498291, "learning_rate": 6.119085856366158e-06, "loss": 1.6526, "step": 728 }, { "epoch": 1.458, "grad_norm": 7.682470321655273, "learning_rate": 6.107742488097338e-06, "loss": 1.5562, "step": 729 }, { "epoch": 1.46, "grad_norm": 10.752216339111328, "learning_rate": 6.0963931209395165e-06, "loss": 1.6478, "step": 730 }, { "epoch": 1.462, "grad_norm": 9.087852478027344, "learning_rate": 6.085037816354269e-06, "loss": 1.5929, "step": 731 }, { "epoch": 1.464, "grad_norm": 7.526221752166748, "learning_rate": 6.073676635835317e-06, "loss": 1.588, "step": 732 }, { "epoch": 1.466, "grad_norm": 7.148344993591309, "learning_rate": 6.062309640908206e-06, "loss": 1.575, "step": 733 }, { "epoch": 1.468, "grad_norm": 10.243959426879883, "learning_rate": 6.05093689312997e-06, "loss": 1.5805, "step": 734 }, { "epoch": 1.47, "grad_norm": 8.489969253540039, "learning_rate": 6.039558454088796e-06, "loss": 1.4469, "step": 735 }, { "epoch": 1.472, "grad_norm": 7.80888557434082, "learning_rate": 6.028174385403693e-06, "loss": 1.6703, "step": 736 }, { "epoch": 1.474, "grad_norm": 8.337100982666016, "learning_rate": 6.016784748724153e-06, "loss": 1.5589, "step": 737 }, { "epoch": 1.476, "grad_norm": 7.308524131774902, "learning_rate": 6.005389605729824e-06, "loss": 1.5458, "step": 738 }, { "epoch": 1.478, "grad_norm": 10.208418846130371, "learning_rate": 5.993989018130173e-06, "loss": 1.5389, "step": 739 }, { "epoch": 1.48, "grad_norm": 6.792656421661377, "learning_rate": 5.982583047664151e-06, "loss": 1.574, "step": 740 }, { "epoch": 1.482, "grad_norm": 7.216954708099365, "learning_rate": 5.97117175609986e-06, "loss": 1.6285, "step": 741 }, { "epoch": 1.484, "grad_norm": 7.5151238441467285, "learning_rate": 5.9597552052342174e-06, "loss": 1.536, "step": 742 }, { "epoch": 1.486, "grad_norm": 8.329843521118164, "learning_rate": 5.948333456892624e-06, "loss": 1.564, "step": 743 }, { "epoch": 1.488, "grad_norm": 7.076045036315918, "learning_rate": 5.936906572928625e-06, "loss": 1.5722, "step": 744 }, { "epoch": 1.49, "grad_norm": 8.269755363464355, "learning_rate": 5.925474615223573e-06, "loss": 1.6673, "step": 745 }, { "epoch": 1.492, "grad_norm": 9.766873359680176, "learning_rate": 5.914037645686308e-06, "loss": 1.4451, "step": 746 }, { "epoch": 1.494, "grad_norm": 7.705737590789795, "learning_rate": 5.902595726252801e-06, "loss": 1.7146, "step": 747 }, { "epoch": 1.496, "grad_norm": 6.665014266967773, "learning_rate": 5.891148918885834e-06, "loss": 1.498, "step": 748 }, { "epoch": 1.498, "grad_norm": 8.403691291809082, "learning_rate": 5.879697285574655e-06, "loss": 1.6839, "step": 749 }, { "epoch": 1.5, "grad_norm": 6.893798351287842, "learning_rate": 5.8682408883346535e-06, "loss": 1.535, "step": 750 }, { "epoch": 1.502, "grad_norm": 7.8562421798706055, "learning_rate": 5.85677978920701e-06, "loss": 1.5439, "step": 751 }, { "epoch": 1.504, "grad_norm": 9.73377513885498, "learning_rate": 5.84531405025837e-06, "loss": 1.576, "step": 752 }, { "epoch": 1.506, "grad_norm": 9.49201774597168, "learning_rate": 5.8338437335805124e-06, "loss": 1.7001, "step": 753 }, { "epoch": 1.508, "grad_norm": 7.50467586517334, "learning_rate": 5.8223689012899945e-06, "loss": 1.6096, "step": 754 }, { "epoch": 1.51, "grad_norm": 7.861015319824219, "learning_rate": 5.810889615527839e-06, "loss": 1.6206, "step": 755 }, { "epoch": 1.512, "grad_norm": 9.48017692565918, "learning_rate": 5.799405938459175e-06, "loss": 1.5395, "step": 756 }, { "epoch": 1.514, "grad_norm": 8.29472827911377, "learning_rate": 5.787917932272922e-06, "loss": 1.567, "step": 757 }, { "epoch": 1.516, "grad_norm": 9.242623329162598, "learning_rate": 5.776425659181438e-06, "loss": 1.4826, "step": 758 }, { "epoch": 1.518, "grad_norm": 8.55409049987793, "learning_rate": 5.764929181420191e-06, "loss": 1.462, "step": 759 }, { "epoch": 1.52, "grad_norm": 8.088449478149414, "learning_rate": 5.753428561247416e-06, "loss": 1.6413, "step": 760 }, { "epoch": 1.522, "grad_norm": 7.811099529266357, "learning_rate": 5.741923860943783e-06, "loss": 1.5257, "step": 761 }, { "epoch": 1.524, "grad_norm": 7.643144130706787, "learning_rate": 5.730415142812059e-06, "loss": 1.638, "step": 762 }, { "epoch": 1.526, "grad_norm": 8.378085136413574, "learning_rate": 5.718902469176765e-06, "loss": 1.6547, "step": 763 }, { "epoch": 1.528, "grad_norm": 8.745218276977539, "learning_rate": 5.707385902383845e-06, "loss": 1.5213, "step": 764 }, { "epoch": 1.53, "grad_norm": 11.519359588623047, "learning_rate": 5.695865504800328e-06, "loss": 1.6252, "step": 765 }, { "epoch": 1.532, "grad_norm": 9.585226058959961, "learning_rate": 5.684341338813986e-06, "loss": 1.5071, "step": 766 }, { "epoch": 1.534, "grad_norm": 7.059016704559326, "learning_rate": 5.672813466832998e-06, "loss": 1.6355, "step": 767 }, { "epoch": 1.536, "grad_norm": 7.046347618103027, "learning_rate": 5.661281951285613e-06, "loss": 1.7801, "step": 768 }, { "epoch": 1.538, "grad_norm": 6.897932052612305, "learning_rate": 5.649746854619814e-06, "loss": 1.5833, "step": 769 }, { "epoch": 1.54, "grad_norm": 8.469090461730957, "learning_rate": 5.638208239302975e-06, "loss": 1.6203, "step": 770 }, { "epoch": 1.542, "grad_norm": 10.167533874511719, "learning_rate": 5.626666167821522e-06, "loss": 1.5861, "step": 771 }, { "epoch": 1.544, "grad_norm": 7.769595146179199, "learning_rate": 5.615120702680604e-06, "loss": 1.5514, "step": 772 }, { "epoch": 1.546, "grad_norm": 7.826175689697266, "learning_rate": 5.6035719064037446e-06, "loss": 1.6232, "step": 773 }, { "epoch": 1.548, "grad_norm": 7.963818550109863, "learning_rate": 5.592019841532507e-06, "loss": 1.5703, "step": 774 }, { "epoch": 1.55, "grad_norm": 8.253701210021973, "learning_rate": 5.5804645706261515e-06, "loss": 1.5996, "step": 775 }, { "epoch": 1.552, "grad_norm": 10.267999649047852, "learning_rate": 5.568906156261309e-06, "loss": 1.5394, "step": 776 }, { "epoch": 1.554, "grad_norm": 9.601211547851562, "learning_rate": 5.557344661031628e-06, "loss": 1.511, "step": 777 }, { "epoch": 1.556, "grad_norm": 9.681934356689453, "learning_rate": 5.54578014754744e-06, "loss": 1.5503, "step": 778 }, { "epoch": 1.558, "grad_norm": 8.93424129486084, "learning_rate": 5.5342126784354265e-06, "loss": 1.561, "step": 779 }, { "epoch": 1.56, "grad_norm": 9.094952583312988, "learning_rate": 5.522642316338268e-06, "loss": 1.5146, "step": 780 }, { "epoch": 1.562, "grad_norm": 8.412158012390137, "learning_rate": 5.511069123914319e-06, "loss": 1.6634, "step": 781 }, { "epoch": 1.564, "grad_norm": 9.692502975463867, "learning_rate": 5.499493163837258e-06, "loss": 1.5547, "step": 782 }, { "epoch": 1.5659999999999998, "grad_norm": 6.936313152313232, "learning_rate": 5.487914498795748e-06, "loss": 1.5515, "step": 783 }, { "epoch": 1.568, "grad_norm": 7.49924373626709, "learning_rate": 5.476333191493108e-06, "loss": 1.5958, "step": 784 }, { "epoch": 1.5699999999999998, "grad_norm": 7.130555152893066, "learning_rate": 5.464749304646963e-06, "loss": 1.607, "step": 785 }, { "epoch": 1.572, "grad_norm": 8.909390449523926, "learning_rate": 5.453162900988902e-06, "loss": 1.5895, "step": 786 }, { "epoch": 1.5739999999999998, "grad_norm": 9.79128360748291, "learning_rate": 5.44157404326415e-06, "loss": 1.6367, "step": 787 }, { "epoch": 1.576, "grad_norm": 8.873651504516602, "learning_rate": 5.429982794231221e-06, "loss": 1.5858, "step": 788 }, { "epoch": 1.5779999999999998, "grad_norm": 9.033736228942871, "learning_rate": 5.41838921666158e-06, "loss": 1.6132, "step": 789 }, { "epoch": 1.58, "grad_norm": 8.515612602233887, "learning_rate": 5.406793373339292e-06, "loss": 1.5178, "step": 790 }, { "epoch": 1.5819999999999999, "grad_norm": 8.608535766601562, "learning_rate": 5.395195327060707e-06, "loss": 1.5546, "step": 791 }, { "epoch": 1.584, "grad_norm": 9.189397811889648, "learning_rate": 5.383595140634093e-06, "loss": 1.64, "step": 792 }, { "epoch": 1.5859999999999999, "grad_norm": 9.675806045532227, "learning_rate": 5.371992876879318e-06, "loss": 1.6044, "step": 793 }, { "epoch": 1.588, "grad_norm": 10.594379425048828, "learning_rate": 5.360388598627487e-06, "loss": 1.6525, "step": 794 }, { "epoch": 1.5899999999999999, "grad_norm": 7.640488624572754, "learning_rate": 5.348782368720627e-06, "loss": 1.5487, "step": 795 }, { "epoch": 1.592, "grad_norm": 8.902362823486328, "learning_rate": 5.337174250011326e-06, "loss": 1.5257, "step": 796 }, { "epoch": 1.5939999999999999, "grad_norm": 9.681800842285156, "learning_rate": 5.325564305362404e-06, "loss": 1.5424, "step": 797 }, { "epoch": 1.596, "grad_norm": 7.919555187225342, "learning_rate": 5.3139525976465675e-06, "loss": 1.4885, "step": 798 }, { "epoch": 1.5979999999999999, "grad_norm": 11.47796630859375, "learning_rate": 5.3023391897460715e-06, "loss": 1.6287, "step": 799 }, { "epoch": 1.6, "grad_norm": 8.65668773651123, "learning_rate": 5.290724144552379e-06, "loss": 1.6287, "step": 800 }, { "epoch": 1.6019999999999999, "grad_norm": 8.334498405456543, "learning_rate": 5.27910752496582e-06, "loss": 1.5546, "step": 801 }, { "epoch": 1.604, "grad_norm": 9.96323013305664, "learning_rate": 5.267489393895247e-06, "loss": 1.658, "step": 802 }, { "epoch": 1.6059999999999999, "grad_norm": 11.58132266998291, "learning_rate": 5.255869814257701e-06, "loss": 1.5074, "step": 803 }, { "epoch": 1.608, "grad_norm": 10.45251178741455, "learning_rate": 5.244248848978067e-06, "loss": 1.6304, "step": 804 }, { "epoch": 1.6099999999999999, "grad_norm": 9.885043144226074, "learning_rate": 5.232626560988735e-06, "loss": 1.6131, "step": 805 }, { "epoch": 1.612, "grad_norm": 6.945827960968018, "learning_rate": 5.221003013229253e-06, "loss": 1.5726, "step": 806 }, { "epoch": 1.6139999999999999, "grad_norm": 10.760123252868652, "learning_rate": 5.209378268645998e-06, "loss": 1.6255, "step": 807 }, { "epoch": 1.616, "grad_norm": 8.588194847106934, "learning_rate": 5.197752390191827e-06, "loss": 1.6538, "step": 808 }, { "epoch": 1.6179999999999999, "grad_norm": 7.07314920425415, "learning_rate": 5.18612544082573e-06, "loss": 1.5529, "step": 809 }, { "epoch": 1.62, "grad_norm": 9.155885696411133, "learning_rate": 5.174497483512506e-06, "loss": 1.7452, "step": 810 }, { "epoch": 1.6219999999999999, "grad_norm": 5.848752498626709, "learning_rate": 5.162868581222407e-06, "loss": 1.6038, "step": 811 }, { "epoch": 1.624, "grad_norm": 6.948676109313965, "learning_rate": 5.151238796930804e-06, "loss": 1.5384, "step": 812 }, { "epoch": 1.626, "grad_norm": 7.449524402618408, "learning_rate": 5.139608193617846e-06, "loss": 1.7695, "step": 813 }, { "epoch": 1.6280000000000001, "grad_norm": 9.431687355041504, "learning_rate": 5.127976834268112e-06, "loss": 1.6164, "step": 814 }, { "epoch": 1.63, "grad_norm": 11.5729341506958, "learning_rate": 5.116344781870282e-06, "loss": 1.5993, "step": 815 }, { "epoch": 1.6320000000000001, "grad_norm": 10.78165054321289, "learning_rate": 5.1047120994167855e-06, "loss": 1.483, "step": 816 }, { "epoch": 1.634, "grad_norm": 7.827059745788574, "learning_rate": 5.093078849903464e-06, "loss": 1.675, "step": 817 }, { "epoch": 1.6360000000000001, "grad_norm": 8.432703971862793, "learning_rate": 5.081445096329229e-06, "loss": 1.5087, "step": 818 }, { "epoch": 1.638, "grad_norm": 8.069623947143555, "learning_rate": 5.069810901695727e-06, "loss": 1.6049, "step": 819 }, { "epoch": 1.6400000000000001, "grad_norm": 8.646689414978027, "learning_rate": 5.0581763290069865e-06, "loss": 1.5509, "step": 820 }, { "epoch": 1.642, "grad_norm": 7.84726619720459, "learning_rate": 5.046541441269085e-06, "loss": 1.5971, "step": 821 }, { "epoch": 1.6440000000000001, "grad_norm": 9.143599510192871, "learning_rate": 5.034906301489808e-06, "loss": 1.5909, "step": 822 }, { "epoch": 1.646, "grad_norm": 7.073722839355469, "learning_rate": 5.0232709726783065e-06, "loss": 1.4608, "step": 823 }, { "epoch": 1.6480000000000001, "grad_norm": 7.9401397705078125, "learning_rate": 5.011635517844753e-06, "loss": 1.6247, "step": 824 }, { "epoch": 1.65, "grad_norm": 7.164168834686279, "learning_rate": 5e-06, "loss": 1.5878, "step": 825 }, { "epoch": 1.6520000000000001, "grad_norm": 8.36746883392334, "learning_rate": 4.988364482155249e-06, "loss": 1.5431, "step": 826 }, { "epoch": 1.654, "grad_norm": 7.337077617645264, "learning_rate": 4.976729027321694e-06, "loss": 1.574, "step": 827 }, { "epoch": 1.6560000000000001, "grad_norm": 8.89696979522705, "learning_rate": 4.965093698510192e-06, "loss": 1.6283, "step": 828 }, { "epoch": 1.658, "grad_norm": 10.577298164367676, "learning_rate": 4.953458558730917e-06, "loss": 1.6708, "step": 829 }, { "epoch": 1.6600000000000001, "grad_norm": 9.475666046142578, "learning_rate": 4.941823670993016e-06, "loss": 1.5338, "step": 830 }, { "epoch": 1.662, "grad_norm": 9.556079864501953, "learning_rate": 4.9301890983042744e-06, "loss": 1.5048, "step": 831 }, { "epoch": 1.6640000000000001, "grad_norm": 6.210651397705078, "learning_rate": 4.9185549036707715e-06, "loss": 1.5286, "step": 832 }, { "epoch": 1.666, "grad_norm": 8.869709968566895, "learning_rate": 4.906921150096538e-06, "loss": 1.6674, "step": 833 }, { "epoch": 1.6680000000000001, "grad_norm": 7.895263671875, "learning_rate": 4.895287900583216e-06, "loss": 1.5605, "step": 834 }, { "epoch": 1.67, "grad_norm": 12.988919258117676, "learning_rate": 4.883655218129719e-06, "loss": 1.7044, "step": 835 }, { "epoch": 1.6720000000000002, "grad_norm": 7.571807861328125, "learning_rate": 4.87202316573189e-06, "loss": 1.5871, "step": 836 }, { "epoch": 1.674, "grad_norm": 10.251978874206543, "learning_rate": 4.860391806382157e-06, "loss": 1.4593, "step": 837 }, { "epoch": 1.6760000000000002, "grad_norm": 7.005547046661377, "learning_rate": 4.8487612030691975e-06, "loss": 1.5433, "step": 838 }, { "epoch": 1.678, "grad_norm": 10.263413429260254, "learning_rate": 4.837131418777595e-06, "loss": 1.617, "step": 839 }, { "epoch": 1.6800000000000002, "grad_norm": 6.690262794494629, "learning_rate": 4.825502516487497e-06, "loss": 1.6738, "step": 840 }, { "epoch": 1.682, "grad_norm": 9.945344924926758, "learning_rate": 4.813874559174271e-06, "loss": 1.5897, "step": 841 }, { "epoch": 1.6840000000000002, "grad_norm": 6.344468116760254, "learning_rate": 4.802247609808175e-06, "loss": 1.5544, "step": 842 }, { "epoch": 1.686, "grad_norm": 6.898435592651367, "learning_rate": 4.7906217313540035e-06, "loss": 1.667, "step": 843 }, { "epoch": 1.688, "grad_norm": 7.620174407958984, "learning_rate": 4.778996986770747e-06, "loss": 1.6153, "step": 844 }, { "epoch": 1.69, "grad_norm": 8.348420143127441, "learning_rate": 4.767373439011267e-06, "loss": 1.5119, "step": 845 }, { "epoch": 1.692, "grad_norm": 6.85077428817749, "learning_rate": 4.755751151021934e-06, "loss": 1.6303, "step": 846 }, { "epoch": 1.694, "grad_norm": 6.931562900543213, "learning_rate": 4.744130185742301e-06, "loss": 1.535, "step": 847 }, { "epoch": 1.696, "grad_norm": 7.309017658233643, "learning_rate": 4.732510606104754e-06, "loss": 1.6259, "step": 848 }, { "epoch": 1.698, "grad_norm": 8.722718238830566, "learning_rate": 4.720892475034181e-06, "loss": 1.5336, "step": 849 }, { "epoch": 1.7, "grad_norm": 9.0813627243042, "learning_rate": 4.7092758554476215e-06, "loss": 1.5323, "step": 850 }, { "epoch": 1.702, "grad_norm": 8.41715145111084, "learning_rate": 4.6976608102539285e-06, "loss": 1.5765, "step": 851 }, { "epoch": 1.704, "grad_norm": 9.829058647155762, "learning_rate": 4.686047402353433e-06, "loss": 1.4992, "step": 852 }, { "epoch": 1.706, "grad_norm": 8.533679962158203, "learning_rate": 4.674435694637597e-06, "loss": 1.6102, "step": 853 }, { "epoch": 1.708, "grad_norm": 8.357962608337402, "learning_rate": 4.662825749988675e-06, "loss": 1.5547, "step": 854 }, { "epoch": 1.71, "grad_norm": 6.73824405670166, "learning_rate": 4.651217631279374e-06, "loss": 1.5051, "step": 855 }, { "epoch": 1.712, "grad_norm": 13.358663558959961, "learning_rate": 4.639611401372514e-06, "loss": 1.6093, "step": 856 }, { "epoch": 1.714, "grad_norm": 9.384882926940918, "learning_rate": 4.6280071231206845e-06, "loss": 1.5455, "step": 857 }, { "epoch": 1.716, "grad_norm": 10.82388687133789, "learning_rate": 4.6164048593659076e-06, "loss": 1.5374, "step": 858 }, { "epoch": 1.718, "grad_norm": 6.8409743309021, "learning_rate": 4.604804672939295e-06, "loss": 1.6312, "step": 859 }, { "epoch": 1.72, "grad_norm": 7.881139755249023, "learning_rate": 4.59320662666071e-06, "loss": 1.5693, "step": 860 }, { "epoch": 1.722, "grad_norm": 9.830620765686035, "learning_rate": 4.581610783338424e-06, "loss": 1.7035, "step": 861 }, { "epoch": 1.724, "grad_norm": 7.074033737182617, "learning_rate": 4.570017205768779e-06, "loss": 1.5734, "step": 862 }, { "epoch": 1.726, "grad_norm": 10.345975875854492, "learning_rate": 4.5584259567358505e-06, "loss": 1.474, "step": 863 }, { "epoch": 1.728, "grad_norm": 8.930947303771973, "learning_rate": 4.546837099011101e-06, "loss": 1.5575, "step": 864 }, { "epoch": 1.73, "grad_norm": 8.436446189880371, "learning_rate": 4.53525069535304e-06, "loss": 1.6057, "step": 865 }, { "epoch": 1.732, "grad_norm": 8.274133682250977, "learning_rate": 4.523666808506893e-06, "loss": 1.6484, "step": 866 }, { "epoch": 1.734, "grad_norm": 8.548066139221191, "learning_rate": 4.512085501204254e-06, "loss": 1.5368, "step": 867 }, { "epoch": 1.736, "grad_norm": 8.803610801696777, "learning_rate": 4.500506836162746e-06, "loss": 1.6525, "step": 868 }, { "epoch": 1.738, "grad_norm": 8.110183715820312, "learning_rate": 4.4889308760856826e-06, "loss": 1.5937, "step": 869 }, { "epoch": 1.74, "grad_norm": 11.329172134399414, "learning_rate": 4.477357683661734e-06, "loss": 1.6, "step": 870 }, { "epoch": 1.742, "grad_norm": 7.598846435546875, "learning_rate": 4.465787321564576e-06, "loss": 1.546, "step": 871 }, { "epoch": 1.744, "grad_norm": 6.989120960235596, "learning_rate": 4.45421985245256e-06, "loss": 1.6186, "step": 872 }, { "epoch": 1.746, "grad_norm": 9.312042236328125, "learning_rate": 4.442655338968373e-06, "loss": 1.6592, "step": 873 }, { "epoch": 1.748, "grad_norm": 5.902655124664307, "learning_rate": 4.431093843738693e-06, "loss": 1.5185, "step": 874 }, { "epoch": 1.75, "grad_norm": 6.853744029998779, "learning_rate": 4.4195354293738484e-06, "loss": 1.5383, "step": 875 }, { "epoch": 1.752, "grad_norm": 8.076836585998535, "learning_rate": 4.4079801584674955e-06, "loss": 1.4652, "step": 876 }, { "epoch": 1.754, "grad_norm": 6.988975524902344, "learning_rate": 4.396428093596258e-06, "loss": 1.5636, "step": 877 }, { "epoch": 1.756, "grad_norm": 7.838133335113525, "learning_rate": 4.384879297319398e-06, "loss": 1.5891, "step": 878 }, { "epoch": 1.758, "grad_norm": 6.956199645996094, "learning_rate": 4.373333832178478e-06, "loss": 1.5573, "step": 879 }, { "epoch": 1.76, "grad_norm": 7.702134132385254, "learning_rate": 4.361791760697027e-06, "loss": 1.6047, "step": 880 }, { "epoch": 1.762, "grad_norm": 8.466774940490723, "learning_rate": 4.3502531453801885e-06, "loss": 1.6113, "step": 881 }, { "epoch": 1.764, "grad_norm": 7.680281639099121, "learning_rate": 4.3387180487143875e-06, "loss": 1.5202, "step": 882 }, { "epoch": 1.766, "grad_norm": 8.012104988098145, "learning_rate": 4.3271865331670036e-06, "loss": 1.6169, "step": 883 }, { "epoch": 1.768, "grad_norm": 9.084745407104492, "learning_rate": 4.315658661186016e-06, "loss": 1.4818, "step": 884 }, { "epoch": 1.77, "grad_norm": 8.11100959777832, "learning_rate": 4.304134495199675e-06, "loss": 1.6068, "step": 885 }, { "epoch": 1.772, "grad_norm": 7.566526412963867, "learning_rate": 4.2926140976161555e-06, "loss": 1.5804, "step": 886 }, { "epoch": 1.774, "grad_norm": 6.527312755584717, "learning_rate": 4.281097530823237e-06, "loss": 1.5348, "step": 887 }, { "epoch": 1.776, "grad_norm": 8.60621166229248, "learning_rate": 4.269584857187942e-06, "loss": 1.6018, "step": 888 }, { "epoch": 1.778, "grad_norm": 6.590750217437744, "learning_rate": 4.258076139056217e-06, "loss": 1.5499, "step": 889 }, { "epoch": 1.78, "grad_norm": 8.905508995056152, "learning_rate": 4.246571438752585e-06, "loss": 1.7064, "step": 890 }, { "epoch": 1.782, "grad_norm": 8.529908180236816, "learning_rate": 4.23507081857981e-06, "loss": 1.534, "step": 891 }, { "epoch": 1.784, "grad_norm": 9.130575180053711, "learning_rate": 4.2235743408185635e-06, "loss": 1.6427, "step": 892 }, { "epoch": 1.786, "grad_norm": 8.667200088500977, "learning_rate": 4.212082067727079e-06, "loss": 1.4933, "step": 893 }, { "epoch": 1.788, "grad_norm": 8.957341194152832, "learning_rate": 4.200594061540827e-06, "loss": 1.591, "step": 894 }, { "epoch": 1.79, "grad_norm": 7.048657417297363, "learning_rate": 4.189110384472164e-06, "loss": 1.5583, "step": 895 }, { "epoch": 1.792, "grad_norm": 8.028057098388672, "learning_rate": 4.1776310987100054e-06, "loss": 1.4866, "step": 896 }, { "epoch": 1.794, "grad_norm": 8.738648414611816, "learning_rate": 4.166156266419489e-06, "loss": 1.6125, "step": 897 }, { "epoch": 1.796, "grad_norm": 8.001562118530273, "learning_rate": 4.154685949741631e-06, "loss": 1.4699, "step": 898 }, { "epoch": 1.798, "grad_norm": 8.324260711669922, "learning_rate": 4.143220210792993e-06, "loss": 1.5949, "step": 899 }, { "epoch": 1.8, "grad_norm": 11.599272727966309, "learning_rate": 4.131759111665349e-06, "loss": 1.6686, "step": 900 }, { "epoch": 1.802, "grad_norm": 7.560389518737793, "learning_rate": 4.1203027144253466e-06, "loss": 1.5041, "step": 901 }, { "epoch": 1.804, "grad_norm": 7.603427886962891, "learning_rate": 4.108851081114169e-06, "loss": 1.4909, "step": 902 }, { "epoch": 1.806, "grad_norm": 6.763309955596924, "learning_rate": 4.0974042737472005e-06, "loss": 1.5467, "step": 903 }, { "epoch": 1.808, "grad_norm": 9.465215682983398, "learning_rate": 4.0859623543136935e-06, "loss": 1.5266, "step": 904 }, { "epoch": 1.81, "grad_norm": 7.232966423034668, "learning_rate": 4.074525384776428e-06, "loss": 1.5645, "step": 905 }, { "epoch": 1.812, "grad_norm": 7.509275436401367, "learning_rate": 4.063093427071376e-06, "loss": 1.5804, "step": 906 }, { "epoch": 1.814, "grad_norm": 8.851505279541016, "learning_rate": 4.051666543107377e-06, "loss": 1.5903, "step": 907 }, { "epoch": 1.8159999999999998, "grad_norm": 7.193111419677734, "learning_rate": 4.040244794765783e-06, "loss": 1.6361, "step": 908 }, { "epoch": 1.818, "grad_norm": 13.53274154663086, "learning_rate": 4.028828243900141e-06, "loss": 1.5629, "step": 909 }, { "epoch": 1.8199999999999998, "grad_norm": 8.000975608825684, "learning_rate": 4.017416952335849e-06, "loss": 1.55, "step": 910 }, { "epoch": 1.822, "grad_norm": 9.145268440246582, "learning_rate": 4.006010981869829e-06, "loss": 1.5949, "step": 911 }, { "epoch": 1.8239999999999998, "grad_norm": 9.474929809570312, "learning_rate": 3.994610394270178e-06, "loss": 1.5537, "step": 912 }, { "epoch": 1.826, "grad_norm": 8.653277397155762, "learning_rate": 3.983215251275847e-06, "loss": 1.6259, "step": 913 }, { "epoch": 1.8279999999999998, "grad_norm": 7.7228617668151855, "learning_rate": 3.971825614596308e-06, "loss": 1.5946, "step": 914 }, { "epoch": 1.83, "grad_norm": 9.129724502563477, "learning_rate": 3.960441545911205e-06, "loss": 1.5925, "step": 915 }, { "epoch": 1.8319999999999999, "grad_norm": 9.415990829467773, "learning_rate": 3.949063106870031e-06, "loss": 1.5227, "step": 916 }, { "epoch": 1.834, "grad_norm": 9.796669006347656, "learning_rate": 3.9376903590917945e-06, "loss": 1.5435, "step": 917 }, { "epoch": 1.8359999999999999, "grad_norm": 9.218023300170898, "learning_rate": 3.926323364164684e-06, "loss": 1.4847, "step": 918 }, { "epoch": 1.838, "grad_norm": 6.569913387298584, "learning_rate": 3.914962183645733e-06, "loss": 1.598, "step": 919 }, { "epoch": 1.8399999999999999, "grad_norm": 8.879981994628906, "learning_rate": 3.903606879060483e-06, "loss": 1.625, "step": 920 }, { "epoch": 1.842, "grad_norm": 6.769296169281006, "learning_rate": 3.892257511902664e-06, "loss": 1.5214, "step": 921 }, { "epoch": 1.8439999999999999, "grad_norm": 7.948380470275879, "learning_rate": 3.880914143633844e-06, "loss": 1.5543, "step": 922 }, { "epoch": 1.846, "grad_norm": 6.776271820068359, "learning_rate": 3.869576835683109e-06, "loss": 1.5748, "step": 923 }, { "epoch": 1.8479999999999999, "grad_norm": 8.43204402923584, "learning_rate": 3.8582456494467214e-06, "loss": 1.5795, "step": 924 }, { "epoch": 1.85, "grad_norm": 9.641144752502441, "learning_rate": 3.8469206462878e-06, "loss": 1.5457, "step": 925 }, { "epoch": 1.8519999999999999, "grad_norm": 9.725312232971191, "learning_rate": 3.835601887535971e-06, "loss": 1.5125, "step": 926 }, { "epoch": 1.854, "grad_norm": 10.626245498657227, "learning_rate": 3.82428943448705e-06, "loss": 1.6646, "step": 927 }, { "epoch": 1.8559999999999999, "grad_norm": 8.524333000183105, "learning_rate": 3.812983348402703e-06, "loss": 1.6704, "step": 928 }, { "epoch": 1.858, "grad_norm": 8.207454681396484, "learning_rate": 3.8016836905101157e-06, "loss": 1.6347, "step": 929 }, { "epoch": 1.8599999999999999, "grad_norm": 9.849774360656738, "learning_rate": 3.790390522001662e-06, "loss": 1.5782, "step": 930 }, { "epoch": 1.862, "grad_norm": 8.06809139251709, "learning_rate": 3.7791039040345743e-06, "loss": 1.6321, "step": 931 }, { "epoch": 1.8639999999999999, "grad_norm": 8.71696949005127, "learning_rate": 3.767823897730612e-06, "loss": 1.66, "step": 932 }, { "epoch": 1.866, "grad_norm": 7.418612480163574, "learning_rate": 3.756550564175727e-06, "loss": 1.5837, "step": 933 }, { "epoch": 1.8679999999999999, "grad_norm": 7.238306999206543, "learning_rate": 3.745283964419736e-06, "loss": 1.5742, "step": 934 }, { "epoch": 1.87, "grad_norm": 7.4198899269104, "learning_rate": 3.7340241594759917e-06, "loss": 1.7463, "step": 935 }, { "epoch": 1.8719999999999999, "grad_norm": 6.619482517242432, "learning_rate": 3.7227712103210485e-06, "loss": 1.738, "step": 936 }, { "epoch": 1.874, "grad_norm": 7.6474223136901855, "learning_rate": 3.7115251778943314e-06, "loss": 1.5902, "step": 937 }, { "epoch": 1.876, "grad_norm": 9.21036434173584, "learning_rate": 3.700286123097814e-06, "loss": 1.5674, "step": 938 }, { "epoch": 1.8780000000000001, "grad_norm": 7.860700607299805, "learning_rate": 3.6890541067956775e-06, "loss": 1.6321, "step": 939 }, { "epoch": 1.88, "grad_norm": 7.944121837615967, "learning_rate": 3.6778291898139907e-06, "loss": 1.6441, "step": 940 }, { "epoch": 1.8820000000000001, "grad_norm": 9.204889297485352, "learning_rate": 3.6666114329403723e-06, "loss": 1.5388, "step": 941 }, { "epoch": 1.884, "grad_norm": 6.569108009338379, "learning_rate": 3.655400896923672e-06, "loss": 1.5072, "step": 942 }, { "epoch": 1.8860000000000001, "grad_norm": 6.905144691467285, "learning_rate": 3.6441976424736315e-06, "loss": 1.591, "step": 943 }, { "epoch": 1.888, "grad_norm": 9.775949478149414, "learning_rate": 3.633001730260558e-06, "loss": 1.5154, "step": 944 }, { "epoch": 1.8900000000000001, "grad_norm": 8.357000350952148, "learning_rate": 3.6218132209150047e-06, "loss": 1.7292, "step": 945 }, { "epoch": 1.892, "grad_norm": 10.853804588317871, "learning_rate": 3.6106321750274275e-06, "loss": 1.6071, "step": 946 }, { "epoch": 1.8940000000000001, "grad_norm": 7.702592849731445, "learning_rate": 3.5994586531478672e-06, "loss": 1.5247, "step": 947 }, { "epoch": 1.896, "grad_norm": 8.18514633178711, "learning_rate": 3.5882927157856175e-06, "loss": 1.5674, "step": 948 }, { "epoch": 1.8980000000000001, "grad_norm": 6.387351989746094, "learning_rate": 3.577134423408906e-06, "loss": 1.5946, "step": 949 }, { "epoch": 1.9, "grad_norm": 7.652561187744141, "learning_rate": 3.5659838364445505e-06, "loss": 1.5211, "step": 950 }, { "epoch": 1.9020000000000001, "grad_norm": 6.697483062744141, "learning_rate": 3.5548410152776414e-06, "loss": 1.6415, "step": 951 }, { "epoch": 1.904, "grad_norm": 7.114959716796875, "learning_rate": 3.543706020251223e-06, "loss": 1.5515, "step": 952 }, { "epoch": 1.9060000000000001, "grad_norm": 7.068447113037109, "learning_rate": 3.5325789116659493e-06, "loss": 1.6781, "step": 953 }, { "epoch": 1.908, "grad_norm": 6.634418487548828, "learning_rate": 3.521459749779769e-06, "loss": 1.5848, "step": 954 }, { "epoch": 1.9100000000000001, "grad_norm": 7.493570327758789, "learning_rate": 3.51034859480759e-06, "loss": 1.6289, "step": 955 }, { "epoch": 1.912, "grad_norm": 9.925775527954102, "learning_rate": 3.4992455069209717e-06, "loss": 1.5641, "step": 956 }, { "epoch": 1.9140000000000001, "grad_norm": 10.061575889587402, "learning_rate": 3.488150546247778e-06, "loss": 1.5989, "step": 957 }, { "epoch": 1.916, "grad_norm": 8.778186798095703, "learning_rate": 3.4770637728718608e-06, "loss": 1.7018, "step": 958 }, { "epoch": 1.9180000000000001, "grad_norm": 8.304412841796875, "learning_rate": 3.465985246832739e-06, "loss": 1.6029, "step": 959 }, { "epoch": 1.92, "grad_norm": 7.178003311157227, "learning_rate": 3.4549150281252635e-06, "loss": 1.5976, "step": 960 }, { "epoch": 1.9220000000000002, "grad_norm": 9.42418384552002, "learning_rate": 3.4438531766993012e-06, "loss": 1.5237, "step": 961 }, { "epoch": 1.924, "grad_norm": 8.352529525756836, "learning_rate": 3.4327997524594026e-06, "loss": 1.6348, "step": 962 }, { "epoch": 1.9260000000000002, "grad_norm": 10.59170913696289, "learning_rate": 3.4217548152644887e-06, "loss": 1.5547, "step": 963 }, { "epoch": 1.928, "grad_norm": 7.5832343101501465, "learning_rate": 3.4107184249275114e-06, "loss": 1.5294, "step": 964 }, { "epoch": 1.9300000000000002, "grad_norm": 9.453595161437988, "learning_rate": 3.399690641215142e-06, "loss": 1.6929, "step": 965 }, { "epoch": 1.932, "grad_norm": 8.6815185546875, "learning_rate": 3.3886715238474454e-06, "loss": 1.6084, "step": 966 }, { "epoch": 1.9340000000000002, "grad_norm": 6.6434149742126465, "learning_rate": 3.3776611324975496e-06, "loss": 1.4945, "step": 967 }, { "epoch": 1.936, "grad_norm": 9.94503402709961, "learning_rate": 3.3666595267913293e-06, "loss": 1.587, "step": 968 }, { "epoch": 1.938, "grad_norm": 7.814234256744385, "learning_rate": 3.355666766307084e-06, "loss": 1.6181, "step": 969 }, { "epoch": 1.94, "grad_norm": 8.182234764099121, "learning_rate": 3.3446829105752103e-06, "loss": 1.5493, "step": 970 }, { "epoch": 1.942, "grad_norm": 7.502097129821777, "learning_rate": 3.3337080190778816e-06, "loss": 1.4921, "step": 971 }, { "epoch": 1.944, "grad_norm": 6.4139604568481445, "learning_rate": 3.322742151248726e-06, "loss": 1.5533, "step": 972 }, { "epoch": 1.946, "grad_norm": 7.537562370300293, "learning_rate": 3.311785366472506e-06, "loss": 1.5851, "step": 973 }, { "epoch": 1.948, "grad_norm": 8.741056442260742, "learning_rate": 3.3008377240847955e-06, "loss": 1.5118, "step": 974 }, { "epoch": 1.95, "grad_norm": 7.846621513366699, "learning_rate": 3.289899283371657e-06, "loss": 1.5772, "step": 975 }, { "epoch": 1.952, "grad_norm": 9.379910469055176, "learning_rate": 3.2789701035693242e-06, "loss": 1.5339, "step": 976 }, { "epoch": 1.954, "grad_norm": 9.359630584716797, "learning_rate": 3.268050243863877e-06, "loss": 1.6382, "step": 977 }, { "epoch": 1.956, "grad_norm": 11.636152267456055, "learning_rate": 3.2571397633909252e-06, "loss": 1.5506, "step": 978 }, { "epoch": 1.958, "grad_norm": 8.888190269470215, "learning_rate": 3.246238721235283e-06, "loss": 1.4885, "step": 979 }, { "epoch": 1.96, "grad_norm": 8.042271614074707, "learning_rate": 3.2353471764306567e-06, "loss": 1.5741, "step": 980 }, { "epoch": 1.962, "grad_norm": 7.534533500671387, "learning_rate": 3.224465187959316e-06, "loss": 1.5664, "step": 981 }, { "epoch": 1.964, "grad_norm": 6.7709479331970215, "learning_rate": 3.2135928147517803e-06, "loss": 1.5943, "step": 982 }, { "epoch": 1.966, "grad_norm": 8.381340026855469, "learning_rate": 3.2027301156865015e-06, "loss": 1.6603, "step": 983 }, { "epoch": 1.968, "grad_norm": 9.999662399291992, "learning_rate": 3.1918771495895395e-06, "loss": 1.7147, "step": 984 }, { "epoch": 1.97, "grad_norm": 11.495597839355469, "learning_rate": 3.1810339752342446e-06, "loss": 1.6706, "step": 985 }, { "epoch": 1.972, "grad_norm": 9.23818588256836, "learning_rate": 3.1702006513409393e-06, "loss": 1.6012, "step": 986 }, { "epoch": 1.974, "grad_norm": 10.354619026184082, "learning_rate": 3.1593772365766107e-06, "loss": 1.6126, "step": 987 }, { "epoch": 1.976, "grad_norm": 9.37294864654541, "learning_rate": 3.148563789554575e-06, "loss": 1.6156, "step": 988 }, { "epoch": 1.978, "grad_norm": 10.598193168640137, "learning_rate": 3.137760368834169e-06, "loss": 1.7388, "step": 989 }, { "epoch": 1.98, "grad_norm": 9.898024559020996, "learning_rate": 3.12696703292044e-06, "loss": 1.6421, "step": 990 }, { "epoch": 1.982, "grad_norm": 9.265511512756348, "learning_rate": 3.1161838402638158e-06, "loss": 1.5518, "step": 991 }, { "epoch": 1.984, "grad_norm": 7.578424453735352, "learning_rate": 3.105410849259796e-06, "loss": 1.5866, "step": 992 }, { "epoch": 1.986, "grad_norm": 10.20102310180664, "learning_rate": 3.09464811824863e-06, "loss": 1.6485, "step": 993 }, { "epoch": 1.988, "grad_norm": 7.367624759674072, "learning_rate": 3.0838957055150136e-06, "loss": 1.6409, "step": 994 }, { "epoch": 1.99, "grad_norm": 7.860062599182129, "learning_rate": 3.0731536692877596e-06, "loss": 1.535, "step": 995 }, { "epoch": 1.992, "grad_norm": 10.079544067382812, "learning_rate": 3.0624220677394854e-06, "loss": 1.5819, "step": 996 }, { "epoch": 1.994, "grad_norm": 6.722997665405273, "learning_rate": 3.0517009589863057e-06, "loss": 1.6481, "step": 997 }, { "epoch": 1.996, "grad_norm": 11.341468811035156, "learning_rate": 3.040990401087508e-06, "loss": 1.695, "step": 998 }, { "epoch": 1.998, "grad_norm": 6.343021392822266, "learning_rate": 3.030290452045245e-06, "loss": 1.6707, "step": 999 }, { "epoch": 2.0, "grad_norm": 11.605053901672363, "learning_rate": 3.019601169804216e-06, "loss": 1.5376, "step": 1000 }, { "epoch": 2.002, "grad_norm": 9.129085540771484, "learning_rate": 3.0089226122513583e-06, "loss": 1.3932, "step": 1001 }, { "epoch": 2.004, "grad_norm": 7.879745006561279, "learning_rate": 2.9982548372155264e-06, "loss": 1.4429, "step": 1002 }, { "epoch": 2.006, "grad_norm": 7.70796537399292, "learning_rate": 2.9875979024671846e-06, "loss": 1.4238, "step": 1003 }, { "epoch": 2.008, "grad_norm": 8.425646781921387, "learning_rate": 2.9769518657180953e-06, "loss": 1.3642, "step": 1004 }, { "epoch": 2.01, "grad_norm": 7.608778953552246, "learning_rate": 2.966316784621e-06, "loss": 1.3483, "step": 1005 }, { "epoch": 2.012, "grad_norm": 9.824869155883789, "learning_rate": 2.9556927167693107e-06, "loss": 1.4165, "step": 1006 }, { "epoch": 2.014, "grad_norm": 7.270105838775635, "learning_rate": 2.945079719696802e-06, "loss": 1.3235, "step": 1007 }, { "epoch": 2.016, "grad_norm": 10.12110424041748, "learning_rate": 2.934477850877292e-06, "loss": 1.3102, "step": 1008 }, { "epoch": 2.018, "grad_norm": 8.959726333618164, "learning_rate": 2.9238871677243354e-06, "loss": 1.2703, "step": 1009 }, { "epoch": 2.02, "grad_norm": 8.507242202758789, "learning_rate": 2.9133077275909112e-06, "loss": 1.1562, "step": 1010 }, { "epoch": 2.022, "grad_norm": 11.486827850341797, "learning_rate": 2.9027395877691143e-06, "loss": 1.2312, "step": 1011 }, { "epoch": 2.024, "grad_norm": 9.935345649719238, "learning_rate": 2.892182805489846e-06, "loss": 1.1142, "step": 1012 }, { "epoch": 2.026, "grad_norm": 9.927567481994629, "learning_rate": 2.8816374379224932e-06, "loss": 1.0935, "step": 1013 }, { "epoch": 2.028, "grad_norm": 11.30349349975586, "learning_rate": 2.871103542174637e-06, "loss": 1.0115, "step": 1014 }, { "epoch": 2.03, "grad_norm": 11.31701946258545, "learning_rate": 2.86058117529173e-06, "loss": 1.0013, "step": 1015 }, { "epoch": 2.032, "grad_norm": 11.658658027648926, "learning_rate": 2.8500703942567874e-06, "loss": 1.0056, "step": 1016 }, { "epoch": 2.034, "grad_norm": 15.923133850097656, "learning_rate": 2.839571255990088e-06, "loss": 1.1577, "step": 1017 }, { "epoch": 2.036, "grad_norm": 16.112945556640625, "learning_rate": 2.82908381734886e-06, "loss": 1.0646, "step": 1018 }, { "epoch": 2.038, "grad_norm": 16.52297019958496, "learning_rate": 2.818608135126967e-06, "loss": 1.1384, "step": 1019 }, { "epoch": 2.04, "grad_norm": 17.75345802307129, "learning_rate": 2.8081442660546126e-06, "loss": 1.0859, "step": 1020 }, { "epoch": 2.042, "grad_norm": 18.585851669311523, "learning_rate": 2.797692266798027e-06, "loss": 1.0863, "step": 1021 }, { "epoch": 2.044, "grad_norm": 17.183208465576172, "learning_rate": 2.7872521939591556e-06, "loss": 1.0416, "step": 1022 }, { "epoch": 2.046, "grad_norm": 18.7218074798584, "learning_rate": 2.776824104075364e-06, "loss": 0.8856, "step": 1023 }, { "epoch": 2.048, "grad_norm": 20.450927734375, "learning_rate": 2.7664080536191178e-06, "loss": 1.1884, "step": 1024 }, { "epoch": 2.05, "grad_norm": 17.043212890625, "learning_rate": 2.7560040989976894e-06, "loss": 1.0495, "step": 1025 }, { "epoch": 2.052, "grad_norm": 15.958205223083496, "learning_rate": 2.7456122965528475e-06, "loss": 0.9814, "step": 1026 }, { "epoch": 2.054, "grad_norm": 15.604181289672852, "learning_rate": 2.7352327025605464e-06, "loss": 0.9974, "step": 1027 }, { "epoch": 2.056, "grad_norm": 14.632473945617676, "learning_rate": 2.724865373230632e-06, "loss": 1.002, "step": 1028 }, { "epoch": 2.058, "grad_norm": 14.43281364440918, "learning_rate": 2.714510364706531e-06, "loss": 0.9437, "step": 1029 }, { "epoch": 2.06, "grad_norm": 17.825098037719727, "learning_rate": 2.7041677330649408e-06, "loss": 1.0608, "step": 1030 }, { "epoch": 2.062, "grad_norm": 15.265949249267578, "learning_rate": 2.6938375343155464e-06, "loss": 1.1083, "step": 1031 }, { "epoch": 2.064, "grad_norm": 13.251029014587402, "learning_rate": 2.683519824400693e-06, "loss": 1.0742, "step": 1032 }, { "epoch": 2.066, "grad_norm": 13.636655807495117, "learning_rate": 2.6732146591950924e-06, "loss": 1.0812, "step": 1033 }, { "epoch": 2.068, "grad_norm": 14.354068756103516, "learning_rate": 2.662922094505529e-06, "loss": 1.1726, "step": 1034 }, { "epoch": 2.07, "grad_norm": 13.713906288146973, "learning_rate": 2.6526421860705474e-06, "loss": 1.1794, "step": 1035 }, { "epoch": 2.072, "grad_norm": 15.267590522766113, "learning_rate": 2.6423749895601494e-06, "loss": 1.1732, "step": 1036 }, { "epoch": 2.074, "grad_norm": 13.415190696716309, "learning_rate": 2.6321205605755002e-06, "loss": 1.1165, "step": 1037 }, { "epoch": 2.076, "grad_norm": 12.924043655395508, "learning_rate": 2.6218789546486235e-06, "loss": 1.1036, "step": 1038 }, { "epoch": 2.078, "grad_norm": 11.99133014678955, "learning_rate": 2.611650227242102e-06, "loss": 1.0163, "step": 1039 }, { "epoch": 2.08, "grad_norm": 11.737472534179688, "learning_rate": 2.601434433748771e-06, "loss": 0.9887, "step": 1040 }, { "epoch": 2.082, "grad_norm": 14.03266429901123, "learning_rate": 2.5912316294914232e-06, "loss": 1.0226, "step": 1041 }, { "epoch": 2.084, "grad_norm": 13.012822151184082, "learning_rate": 2.581041869722519e-06, "loss": 1.1165, "step": 1042 }, { "epoch": 2.086, "grad_norm": 13.519615173339844, "learning_rate": 2.5708652096238674e-06, "loss": 1.1066, "step": 1043 }, { "epoch": 2.088, "grad_norm": 11.919333457946777, "learning_rate": 2.560701704306336e-06, "loss": 0.9055, "step": 1044 }, { "epoch": 2.09, "grad_norm": 11.374495506286621, "learning_rate": 2.550551408809566e-06, "loss": 1.024, "step": 1045 }, { "epoch": 2.092, "grad_norm": 12.566267013549805, "learning_rate": 2.540414378101647e-06, "loss": 1.1746, "step": 1046 }, { "epoch": 2.094, "grad_norm": 13.454764366149902, "learning_rate": 2.5302906670788463e-06, "loss": 0.9756, "step": 1047 }, { "epoch": 2.096, "grad_norm": 12.35405445098877, "learning_rate": 2.52018033056529e-06, "loss": 1.0644, "step": 1048 }, { "epoch": 2.098, "grad_norm": 12.588571548461914, "learning_rate": 2.5100834233126827e-06, "loss": 1.1582, "step": 1049 }, { "epoch": 2.1, "grad_norm": 13.59836196899414, "learning_rate": 2.5000000000000015e-06, "loss": 1.0943, "step": 1050 }, { "epoch": 2.102, "grad_norm": 12.1016206741333, "learning_rate": 2.489930115233199e-06, "loss": 1.0855, "step": 1051 }, { "epoch": 2.104, "grad_norm": 13.028971672058105, "learning_rate": 2.4798738235449164e-06, "loss": 1.1361, "step": 1052 }, { "epoch": 2.106, "grad_norm": 13.81318187713623, "learning_rate": 2.469831179394182e-06, "loss": 1.0715, "step": 1053 }, { "epoch": 2.108, "grad_norm": 12.744100570678711, "learning_rate": 2.4598022371661113e-06, "loss": 1.1322, "step": 1054 }, { "epoch": 2.11, "grad_norm": 13.312298774719238, "learning_rate": 2.4497870511716237e-06, "loss": 1.0601, "step": 1055 }, { "epoch": 2.112, "grad_norm": 12.346517562866211, "learning_rate": 2.4397856756471435e-06, "loss": 1.0386, "step": 1056 }, { "epoch": 2.114, "grad_norm": 11.398303985595703, "learning_rate": 2.429798164754299e-06, "loss": 1.1074, "step": 1057 }, { "epoch": 2.116, "grad_norm": 14.021239280700684, "learning_rate": 2.4198245725796427e-06, "loss": 1.1536, "step": 1058 }, { "epoch": 2.118, "grad_norm": 13.355626106262207, "learning_rate": 2.40986495313435e-06, "loss": 1.1139, "step": 1059 }, { "epoch": 2.12, "grad_norm": 11.496376037597656, "learning_rate": 2.3999193603539234e-06, "loss": 1.1207, "step": 1060 }, { "epoch": 2.122, "grad_norm": 12.229351043701172, "learning_rate": 2.3899878480979098e-06, "loss": 1.0847, "step": 1061 }, { "epoch": 2.124, "grad_norm": 12.624899864196777, "learning_rate": 2.380070470149605e-06, "loss": 1.1526, "step": 1062 }, { "epoch": 2.126, "grad_norm": 12.152226448059082, "learning_rate": 2.3701672802157567e-06, "loss": 1.0407, "step": 1063 }, { "epoch": 2.128, "grad_norm": 12.016361236572266, "learning_rate": 2.3602783319262847e-06, "loss": 1.0879, "step": 1064 }, { "epoch": 2.13, "grad_norm": 11.808967590332031, "learning_rate": 2.3504036788339763e-06, "loss": 1.1118, "step": 1065 }, { "epoch": 2.132, "grad_norm": 11.515680313110352, "learning_rate": 2.340543374414212e-06, "loss": 1.0536, "step": 1066 }, { "epoch": 2.134, "grad_norm": 10.819012641906738, "learning_rate": 2.330697472064667e-06, "loss": 1.0542, "step": 1067 }, { "epoch": 2.136, "grad_norm": 12.405117988586426, "learning_rate": 2.320866025105016e-06, "loss": 1.0506, "step": 1068 }, { "epoch": 2.138, "grad_norm": 11.67018985748291, "learning_rate": 2.3110490867766644e-06, "loss": 1.101, "step": 1069 }, { "epoch": 2.14, "grad_norm": 12.757566452026367, "learning_rate": 2.3012467102424373e-06, "loss": 1.093, "step": 1070 }, { "epoch": 2.142, "grad_norm": 12.47193717956543, "learning_rate": 2.2914589485863015e-06, "loss": 1.1508, "step": 1071 }, { "epoch": 2.144, "grad_norm": 11.597710609436035, "learning_rate": 2.2816858548130837e-06, "loss": 1.1538, "step": 1072 }, { "epoch": 2.146, "grad_norm": 12.810328483581543, "learning_rate": 2.2719274818481767e-06, "loss": 1.0238, "step": 1073 }, { "epoch": 2.148, "grad_norm": 13.841683387756348, "learning_rate": 2.2621838825372496e-06, "loss": 1.202, "step": 1074 }, { "epoch": 2.15, "grad_norm": 12.471292495727539, "learning_rate": 2.2524551096459703e-06, "loss": 1.0969, "step": 1075 }, { "epoch": 2.152, "grad_norm": 11.695208549499512, "learning_rate": 2.2427412158597133e-06, "loss": 1.0346, "step": 1076 }, { "epoch": 2.154, "grad_norm": 12.01142406463623, "learning_rate": 2.23304225378328e-06, "loss": 1.091, "step": 1077 }, { "epoch": 2.156, "grad_norm": 14.317254066467285, "learning_rate": 2.2233582759406065e-06, "loss": 1.1849, "step": 1078 }, { "epoch": 2.158, "grad_norm": 11.322208404541016, "learning_rate": 2.213689334774479e-06, "loss": 1.0879, "step": 1079 }, { "epoch": 2.16, "grad_norm": 14.340149879455566, "learning_rate": 2.204035482646267e-06, "loss": 1.1532, "step": 1080 }, { "epoch": 2.162, "grad_norm": 12.387587547302246, "learning_rate": 2.1943967718356123e-06, "loss": 1.0726, "step": 1081 }, { "epoch": 2.164, "grad_norm": 12.853300094604492, "learning_rate": 2.184773254540169e-06, "loss": 1.1011, "step": 1082 }, { "epoch": 2.166, "grad_norm": 12.014087677001953, "learning_rate": 2.175164982875311e-06, "loss": 1.1168, "step": 1083 }, { "epoch": 2.168, "grad_norm": 13.224556922912598, "learning_rate": 2.165572008873845e-06, "loss": 1.1332, "step": 1084 }, { "epoch": 2.17, "grad_norm": 10.886467933654785, "learning_rate": 2.155994384485742e-06, "loss": 0.9585, "step": 1085 }, { "epoch": 2.172, "grad_norm": 13.911365509033203, "learning_rate": 2.146432161577842e-06, "loss": 1.1193, "step": 1086 }, { "epoch": 2.174, "grad_norm": 11.698256492614746, "learning_rate": 2.1368853919335835e-06, "loss": 1.1248, "step": 1087 }, { "epoch": 2.176, "grad_norm": 11.102900505065918, "learning_rate": 2.12735412725272e-06, "loss": 1.1625, "step": 1088 }, { "epoch": 2.178, "grad_norm": 11.995986938476562, "learning_rate": 2.1178384191510344e-06, "loss": 1.1379, "step": 1089 }, { "epoch": 2.18, "grad_norm": 10.211990356445312, "learning_rate": 2.1083383191600676e-06, "loss": 1.0004, "step": 1090 }, { "epoch": 2.182, "grad_norm": 11.571313858032227, "learning_rate": 2.0988538787268374e-06, "loss": 1.1064, "step": 1091 }, { "epoch": 2.184, "grad_norm": 12.789441108703613, "learning_rate": 2.0893851492135536e-06, "loss": 1.1492, "step": 1092 }, { "epoch": 2.186, "grad_norm": 11.842870712280273, "learning_rate": 2.0799321818973488e-06, "loss": 1.1277, "step": 1093 }, { "epoch": 2.188, "grad_norm": 12.860237121582031, "learning_rate": 2.0704950279699986e-06, "loss": 1.112, "step": 1094 }, { "epoch": 2.19, "grad_norm": 12.648483276367188, "learning_rate": 2.061073738537635e-06, "loss": 0.9933, "step": 1095 }, { "epoch": 2.192, "grad_norm": 12.022543907165527, "learning_rate": 2.0516683646204836e-06, "loss": 1.0655, "step": 1096 }, { "epoch": 2.194, "grad_norm": 12.181710243225098, "learning_rate": 2.0422789571525813e-06, "loss": 1.1167, "step": 1097 }, { "epoch": 2.196, "grad_norm": 12.39856243133545, "learning_rate": 2.0329055669814936e-06, "loss": 1.1543, "step": 1098 }, { "epoch": 2.198, "grad_norm": 10.813960075378418, "learning_rate": 2.023548244868051e-06, "loss": 1.031, "step": 1099 }, { "epoch": 2.2, "grad_norm": 9.889363288879395, "learning_rate": 2.0142070414860704e-06, "loss": 1.0438, "step": 1100 }, { "epoch": 2.202, "grad_norm": 11.947040557861328, "learning_rate": 2.0048820074220716e-06, "loss": 1.0579, "step": 1101 }, { "epoch": 2.204, "grad_norm": 10.942092895507812, "learning_rate": 1.9955731931750182e-06, "loss": 1.1079, "step": 1102 }, { "epoch": 2.206, "grad_norm": 10.924947738647461, "learning_rate": 1.9862806491560315e-06, "loss": 1.1075, "step": 1103 }, { "epoch": 2.208, "grad_norm": 11.159443855285645, "learning_rate": 1.977004425688126e-06, "loss": 1.0548, "step": 1104 }, { "epoch": 2.21, "grad_norm": 11.316449165344238, "learning_rate": 1.9677445730059348e-06, "loss": 1.1579, "step": 1105 }, { "epoch": 2.212, "grad_norm": 12.146157264709473, "learning_rate": 1.958501141255427e-06, "loss": 1.136, "step": 1106 }, { "epoch": 2.214, "grad_norm": 10.939238548278809, "learning_rate": 1.9492741804936623e-06, "loss": 1.0623, "step": 1107 }, { "epoch": 2.216, "grad_norm": 12.354981422424316, "learning_rate": 1.9400637406884875e-06, "loss": 1.1298, "step": 1108 }, { "epoch": 2.218, "grad_norm": 10.884641647338867, "learning_rate": 1.9308698717182874e-06, "loss": 0.9976, "step": 1109 }, { "epoch": 2.22, "grad_norm": 12.507721900939941, "learning_rate": 1.9216926233717087e-06, "loss": 1.1589, "step": 1110 }, { "epoch": 2.222, "grad_norm": 12.828377723693848, "learning_rate": 1.9125320453473923e-06, "loss": 1.0951, "step": 1111 }, { "epoch": 2.224, "grad_norm": 12.240373611450195, "learning_rate": 1.9033881872537009e-06, "loss": 1.1288, "step": 1112 }, { "epoch": 2.226, "grad_norm": 11.162793159484863, "learning_rate": 1.8942610986084487e-06, "loss": 1.1229, "step": 1113 }, { "epoch": 2.228, "grad_norm": 12.57970905303955, "learning_rate": 1.88515082883864e-06, "loss": 1.0853, "step": 1114 }, { "epoch": 2.23, "grad_norm": 12.0857515335083, "learning_rate": 1.8760574272802002e-06, "loss": 1.1373, "step": 1115 }, { "epoch": 2.232, "grad_norm": 10.519421577453613, "learning_rate": 1.8669809431776991e-06, "loss": 1.1059, "step": 1116 }, { "epoch": 2.234, "grad_norm": 11.660649299621582, "learning_rate": 1.8579214256840938e-06, "loss": 1.2105, "step": 1117 }, { "epoch": 2.2359999999999998, "grad_norm": 11.070806503295898, "learning_rate": 1.8488789238604676e-06, "loss": 1.0679, "step": 1118 }, { "epoch": 2.238, "grad_norm": 11.053348541259766, "learning_rate": 1.8398534866757455e-06, "loss": 1.0581, "step": 1119 }, { "epoch": 2.24, "grad_norm": 12.008465766906738, "learning_rate": 1.8308451630064484e-06, "loss": 1.1217, "step": 1120 }, { "epoch": 2.242, "grad_norm": 12.64566707611084, "learning_rate": 1.8218540016364178e-06, "loss": 1.0691, "step": 1121 }, { "epoch": 2.2439999999999998, "grad_norm": 12.815990447998047, "learning_rate": 1.8128800512565514e-06, "loss": 1.0881, "step": 1122 }, { "epoch": 2.246, "grad_norm": 12.682955741882324, "learning_rate": 1.8039233604645468e-06, "loss": 1.0893, "step": 1123 }, { "epoch": 2.248, "grad_norm": 11.445165634155273, "learning_rate": 1.7949839777646327e-06, "loss": 1.1113, "step": 1124 }, { "epoch": 2.25, "grad_norm": 11.555493354797363, "learning_rate": 1.7860619515673034e-06, "loss": 1.1268, "step": 1125 }, { "epoch": 2.252, "grad_norm": 12.906476974487305, "learning_rate": 1.7771573301890666e-06, "loss": 1.1886, "step": 1126 }, { "epoch": 2.254, "grad_norm": 11.322837829589844, "learning_rate": 1.7682701618521687e-06, "loss": 1.0748, "step": 1127 }, { "epoch": 2.2560000000000002, "grad_norm": 14.397950172424316, "learning_rate": 1.7594004946843458e-06, "loss": 1.174, "step": 1128 }, { "epoch": 2.258, "grad_norm": 12.00009536743164, "learning_rate": 1.7505483767185583e-06, "loss": 1.0512, "step": 1129 }, { "epoch": 2.26, "grad_norm": 13.125904083251953, "learning_rate": 1.7417138558927244e-06, "loss": 1.1748, "step": 1130 }, { "epoch": 2.262, "grad_norm": 11.499984741210938, "learning_rate": 1.7328969800494727e-06, "loss": 0.9966, "step": 1131 }, { "epoch": 2.2640000000000002, "grad_norm": 11.893850326538086, "learning_rate": 1.7240977969358757e-06, "loss": 1.1099, "step": 1132 }, { "epoch": 2.266, "grad_norm": 13.56734848022461, "learning_rate": 1.7153163542031881e-06, "loss": 1.1765, "step": 1133 }, { "epoch": 2.268, "grad_norm": 11.519514083862305, "learning_rate": 1.7065526994065973e-06, "loss": 1.0598, "step": 1134 }, { "epoch": 2.27, "grad_norm": 13.245732307434082, "learning_rate": 1.6978068800049624e-06, "loss": 0.9934, "step": 1135 }, { "epoch": 2.2720000000000002, "grad_norm": 12.127825736999512, "learning_rate": 1.6890789433605508e-06, "loss": 1.1238, "step": 1136 }, { "epoch": 2.274, "grad_norm": 12.772904396057129, "learning_rate": 1.680368936738792e-06, "loss": 1.1137, "step": 1137 }, { "epoch": 2.276, "grad_norm": 13.146944046020508, "learning_rate": 1.671676907308018e-06, "loss": 1.1266, "step": 1138 }, { "epoch": 2.278, "grad_norm": 11.267017364501953, "learning_rate": 1.6630029021392007e-06, "loss": 1.1258, "step": 1139 }, { "epoch": 2.2800000000000002, "grad_norm": 11.609925270080566, "learning_rate": 1.6543469682057105e-06, "loss": 1.097, "step": 1140 }, { "epoch": 2.282, "grad_norm": 12.532341957092285, "learning_rate": 1.645709152383046e-06, "loss": 1.193, "step": 1141 }, { "epoch": 2.284, "grad_norm": 11.958260536193848, "learning_rate": 1.6370895014486e-06, "loss": 1.0575, "step": 1142 }, { "epoch": 2.286, "grad_norm": 12.34417724609375, "learning_rate": 1.6284880620813847e-06, "loss": 1.0509, "step": 1143 }, { "epoch": 2.288, "grad_norm": 12.214522361755371, "learning_rate": 1.6199048808617896e-06, "loss": 1.0467, "step": 1144 }, { "epoch": 2.29, "grad_norm": 12.764805793762207, "learning_rate": 1.611340004271339e-06, "loss": 1.0441, "step": 1145 }, { "epoch": 2.292, "grad_norm": 11.342555046081543, "learning_rate": 1.6027934786924187e-06, "loss": 1.1076, "step": 1146 }, { "epoch": 2.294, "grad_norm": 13.207151412963867, "learning_rate": 1.594265350408039e-06, "loss": 1.2069, "step": 1147 }, { "epoch": 2.296, "grad_norm": 11.775276184082031, "learning_rate": 1.5857556656015837e-06, "loss": 1.0898, "step": 1148 }, { "epoch": 2.298, "grad_norm": 11.530139923095703, "learning_rate": 1.5772644703565564e-06, "loss": 1.098, "step": 1149 }, { "epoch": 2.3, "grad_norm": 11.4733247756958, "learning_rate": 1.5687918106563326e-06, "loss": 1.0295, "step": 1150 }, { "epoch": 2.302, "grad_norm": 11.146039962768555, "learning_rate": 1.5603377323839069e-06, "loss": 1.0014, "step": 1151 }, { "epoch": 2.304, "grad_norm": 13.669384002685547, "learning_rate": 1.551902281321651e-06, "loss": 1.2271, "step": 1152 }, { "epoch": 2.306, "grad_norm": 12.079588890075684, "learning_rate": 1.5434855031510626e-06, "loss": 1.0904, "step": 1153 }, { "epoch": 2.308, "grad_norm": 11.25638198852539, "learning_rate": 1.5350874434525142e-06, "loss": 1.0801, "step": 1154 }, { "epoch": 2.31, "grad_norm": 11.719844818115234, "learning_rate": 1.5267081477050132e-06, "loss": 1.0488, "step": 1155 }, { "epoch": 2.312, "grad_norm": 11.61817455291748, "learning_rate": 1.5183476612859538e-06, "loss": 1.0197, "step": 1156 }, { "epoch": 2.314, "grad_norm": 12.426447868347168, "learning_rate": 1.5100060294708647e-06, "loss": 1.155, "step": 1157 }, { "epoch": 2.316, "grad_norm": 11.935876846313477, "learning_rate": 1.5016832974331725e-06, "loss": 1.0953, "step": 1158 }, { "epoch": 2.318, "grad_norm": 15.38991641998291, "learning_rate": 1.4933795102439558e-06, "loss": 1.212, "step": 1159 }, { "epoch": 2.32, "grad_norm": 11.428476333618164, "learning_rate": 1.4850947128716914e-06, "loss": 1.1117, "step": 1160 }, { "epoch": 2.322, "grad_norm": 11.7312593460083, "learning_rate": 1.4768289501820265e-06, "loss": 1.0994, "step": 1161 }, { "epoch": 2.324, "grad_norm": 11.2421236038208, "learning_rate": 1.4685822669375239e-06, "loss": 1.1205, "step": 1162 }, { "epoch": 2.326, "grad_norm": 10.058606147766113, "learning_rate": 1.4603547077974217e-06, "loss": 1.0785, "step": 1163 }, { "epoch": 2.328, "grad_norm": 11.7360258102417, "learning_rate": 1.4521463173173966e-06, "loss": 1.097, "step": 1164 }, { "epoch": 2.33, "grad_norm": 13.815256118774414, "learning_rate": 1.4439571399493146e-06, "loss": 1.0887, "step": 1165 }, { "epoch": 2.332, "grad_norm": 12.337494850158691, "learning_rate": 1.4357872200409988e-06, "loss": 1.1101, "step": 1166 }, { "epoch": 2.334, "grad_norm": 11.706501007080078, "learning_rate": 1.4276366018359845e-06, "loss": 1.0746, "step": 1167 }, { "epoch": 2.336, "grad_norm": 12.499887466430664, "learning_rate": 1.4195053294732757e-06, "loss": 1.1403, "step": 1168 }, { "epoch": 2.338, "grad_norm": 12.3418550491333, "learning_rate": 1.4113934469871166e-06, "loss": 1.1632, "step": 1169 }, { "epoch": 2.34, "grad_norm": 12.219500541687012, "learning_rate": 1.4033009983067454e-06, "loss": 1.1484, "step": 1170 }, { "epoch": 2.342, "grad_norm": 11.66283893585205, "learning_rate": 1.3952280272561541e-06, "loss": 1.0012, "step": 1171 }, { "epoch": 2.344, "grad_norm": 13.044709205627441, "learning_rate": 1.3871745775538598e-06, "loss": 1.2462, "step": 1172 }, { "epoch": 2.346, "grad_norm": 11.258277893066406, "learning_rate": 1.3791406928126638e-06, "loss": 1.1165, "step": 1173 }, { "epoch": 2.348, "grad_norm": 11.9083833694458, "learning_rate": 1.371126416539409e-06, "loss": 1.0926, "step": 1174 }, { "epoch": 2.35, "grad_norm": 12.25865364074707, "learning_rate": 1.3631317921347564e-06, "loss": 1.1481, "step": 1175 }, { "epoch": 2.352, "grad_norm": 11.566573143005371, "learning_rate": 1.3551568628929434e-06, "loss": 1.0938, "step": 1176 }, { "epoch": 2.354, "grad_norm": 12.630902290344238, "learning_rate": 1.3472016720015447e-06, "loss": 1.1382, "step": 1177 }, { "epoch": 2.356, "grad_norm": 12.472233772277832, "learning_rate": 1.339266262541249e-06, "loss": 1.149, "step": 1178 }, { "epoch": 2.358, "grad_norm": 11.453776359558105, "learning_rate": 1.3313506774856177e-06, "loss": 1.0327, "step": 1179 }, { "epoch": 2.36, "grad_norm": 10.934307098388672, "learning_rate": 1.3234549597008572e-06, "loss": 1.1564, "step": 1180 }, { "epoch": 2.362, "grad_norm": 13.961017608642578, "learning_rate": 1.3155791519455812e-06, "loss": 1.0754, "step": 1181 }, { "epoch": 2.364, "grad_norm": 12.650055885314941, "learning_rate": 1.3077232968705805e-06, "loss": 1.1337, "step": 1182 }, { "epoch": 2.366, "grad_norm": 12.112220764160156, "learning_rate": 1.2998874370186026e-06, "loss": 1.0891, "step": 1183 }, { "epoch": 2.368, "grad_norm": 11.407106399536133, "learning_rate": 1.2920716148241036e-06, "loss": 1.1488, "step": 1184 }, { "epoch": 2.37, "grad_norm": 11.474730491638184, "learning_rate": 1.2842758726130283e-06, "loss": 1.1326, "step": 1185 }, { "epoch": 2.372, "grad_norm": 12.02614974975586, "learning_rate": 1.2765002526025871e-06, "loss": 1.0903, "step": 1186 }, { "epoch": 2.374, "grad_norm": 13.217632293701172, "learning_rate": 1.2687447969010113e-06, "loss": 1.1562, "step": 1187 }, { "epoch": 2.376, "grad_norm": 13.802258491516113, "learning_rate": 1.2610095475073415e-06, "loss": 1.1038, "step": 1188 }, { "epoch": 2.378, "grad_norm": 11.107643127441406, "learning_rate": 1.2532945463111856e-06, "loss": 1.0914, "step": 1189 }, { "epoch": 2.38, "grad_norm": 12.443587303161621, "learning_rate": 1.2455998350925042e-06, "loss": 1.1518, "step": 1190 }, { "epoch": 2.382, "grad_norm": 12.572113990783691, "learning_rate": 1.2379254555213788e-06, "loss": 1.1528, "step": 1191 }, { "epoch": 2.384, "grad_norm": 11.02535629272461, "learning_rate": 1.2302714491577834e-06, "loss": 1.1726, "step": 1192 }, { "epoch": 2.386, "grad_norm": 12.016326904296875, "learning_rate": 1.2226378574513654e-06, "loss": 1.1441, "step": 1193 }, { "epoch": 2.388, "grad_norm": 11.50503921508789, "learning_rate": 1.2150247217412186e-06, "loss": 1.2099, "step": 1194 }, { "epoch": 2.39, "grad_norm": 11.974237442016602, "learning_rate": 1.2074320832556558e-06, "loss": 1.1982, "step": 1195 }, { "epoch": 2.392, "grad_norm": 11.554012298583984, "learning_rate": 1.1998599831119912e-06, "loss": 1.0467, "step": 1196 }, { "epoch": 2.394, "grad_norm": 13.131747245788574, "learning_rate": 1.1923084623163172e-06, "loss": 1.143, "step": 1197 }, { "epoch": 2.396, "grad_norm": 10.756204605102539, "learning_rate": 1.1847775617632746e-06, "loss": 1.1161, "step": 1198 }, { "epoch": 2.398, "grad_norm": 12.911304473876953, "learning_rate": 1.1772673222358421e-06, "loss": 1.1532, "step": 1199 }, { "epoch": 2.4, "grad_norm": 10.806382179260254, "learning_rate": 1.1697777844051105e-06, "loss": 1.0573, "step": 1200 }, { "epoch": 2.402, "grad_norm": 11.249881744384766, "learning_rate": 1.162308988830057e-06, "loss": 1.1033, "step": 1201 }, { "epoch": 2.404, "grad_norm": 11.274486541748047, "learning_rate": 1.1548609759573375e-06, "loss": 1.2207, "step": 1202 }, { "epoch": 2.406, "grad_norm": 10.787302017211914, "learning_rate": 1.1474337861210543e-06, "loss": 1.1259, "step": 1203 }, { "epoch": 2.408, "grad_norm": 10.975671768188477, "learning_rate": 1.1400274595425499e-06, "loss": 1.0418, "step": 1204 }, { "epoch": 2.41, "grad_norm": 10.228716850280762, "learning_rate": 1.132642036330181e-06, "loss": 1.087, "step": 1205 }, { "epoch": 2.412, "grad_norm": 13.028627395629883, "learning_rate": 1.1252775564791023e-06, "loss": 1.1651, "step": 1206 }, { "epoch": 2.414, "grad_norm": 10.35562801361084, "learning_rate": 1.1179340598710547e-06, "loss": 1.1121, "step": 1207 }, { "epoch": 2.416, "grad_norm": 11.16613483428955, "learning_rate": 1.1106115862741457e-06, "loss": 1.1242, "step": 1208 }, { "epoch": 2.418, "grad_norm": 11.81697940826416, "learning_rate": 1.1033101753426285e-06, "loss": 1.167, "step": 1209 }, { "epoch": 2.42, "grad_norm": 13.364669799804688, "learning_rate": 1.096029866616704e-06, "loss": 1.0961, "step": 1210 }, { "epoch": 2.422, "grad_norm": 11.796897888183594, "learning_rate": 1.0887706995222864e-06, "loss": 1.137, "step": 1211 }, { "epoch": 2.424, "grad_norm": 11.452189445495605, "learning_rate": 1.0815327133708015e-06, "loss": 1.1626, "step": 1212 }, { "epoch": 2.426, "grad_norm": 11.605690956115723, "learning_rate": 1.0743159473589738e-06, "loss": 1.1818, "step": 1213 }, { "epoch": 2.428, "grad_norm": 12.917741775512695, "learning_rate": 1.0671204405686108e-06, "loss": 1.0901, "step": 1214 }, { "epoch": 2.43, "grad_norm": 11.108345985412598, "learning_rate": 1.0599462319663906e-06, "loss": 1.0953, "step": 1215 }, { "epoch": 2.432, "grad_norm": 11.12553882598877, "learning_rate": 1.052793360403655e-06, "loss": 1.1915, "step": 1216 }, { "epoch": 2.434, "grad_norm": 12.073833465576172, "learning_rate": 1.0456618646161954e-06, "loss": 1.2316, "step": 1217 }, { "epoch": 2.436, "grad_norm": 11.272497177124023, "learning_rate": 1.0385517832240472e-06, "loss": 1.1353, "step": 1218 }, { "epoch": 2.438, "grad_norm": 11.970922470092773, "learning_rate": 1.0314631547312738e-06, "loss": 1.1071, "step": 1219 }, { "epoch": 2.44, "grad_norm": 11.06673812866211, "learning_rate": 1.0243960175257605e-06, "loss": 1.1197, "step": 1220 }, { "epoch": 2.442, "grad_norm": 11.756420135498047, "learning_rate": 1.0173504098790188e-06, "loss": 1.0385, "step": 1221 }, { "epoch": 2.444, "grad_norm": 11.263325691223145, "learning_rate": 1.010326369945957e-06, "loss": 1.0892, "step": 1222 }, { "epoch": 2.446, "grad_norm": 11.817276954650879, "learning_rate": 1.0033239357646913e-06, "loss": 1.138, "step": 1223 }, { "epoch": 2.448, "grad_norm": 11.443320274353027, "learning_rate": 9.963431452563331e-07, "loss": 1.0683, "step": 1224 }, { "epoch": 2.45, "grad_norm": 11.445216178894043, "learning_rate": 9.893840362247809e-07, "loss": 1.1286, "step": 1225 }, { "epoch": 2.452, "grad_norm": 12.109580039978027, "learning_rate": 9.824466463565246e-07, "loss": 1.2696, "step": 1226 }, { "epoch": 2.454, "grad_norm": 12.882740020751953, "learning_rate": 9.7553101322043e-07, "loss": 1.1698, "step": 1227 }, { "epoch": 2.456, "grad_norm": 12.377523422241211, "learning_rate": 9.686371742675443e-07, "loss": 1.2098, "step": 1228 }, { "epoch": 2.458, "grad_norm": 11.193984985351562, "learning_rate": 9.617651668308914e-07, "loss": 1.1167, "step": 1229 }, { "epoch": 2.46, "grad_norm": 11.221639633178711, "learning_rate": 9.549150281252633e-07, "loss": 1.1706, "step": 1230 }, { "epoch": 2.462, "grad_norm": 11.720982551574707, "learning_rate": 9.480867952470285e-07, "loss": 1.1991, "step": 1231 }, { "epoch": 2.464, "grad_norm": 9.958009719848633, "learning_rate": 9.412805051739266e-07, "loss": 1.0561, "step": 1232 }, { "epoch": 2.466, "grad_norm": 12.225953102111816, "learning_rate": 9.344961947648624e-07, "loss": 1.0812, "step": 1233 }, { "epoch": 2.468, "grad_norm": 11.826346397399902, "learning_rate": 9.277339007597158e-07, "loss": 1.1611, "step": 1234 }, { "epoch": 2.4699999999999998, "grad_norm": 11.533407211303711, "learning_rate": 9.209936597791407e-07, "loss": 1.1837, "step": 1235 }, { "epoch": 2.472, "grad_norm": 10.118829727172852, "learning_rate": 9.142755083243577e-07, "loss": 1.0882, "step": 1236 }, { "epoch": 2.474, "grad_norm": 13.343694686889648, "learning_rate": 9.075794827769696e-07, "loss": 1.1914, "step": 1237 }, { "epoch": 2.476, "grad_norm": 11.293721199035645, "learning_rate": 9.009056193987569e-07, "loss": 1.0637, "step": 1238 }, { "epoch": 2.4779999999999998, "grad_norm": 10.591233253479004, "learning_rate": 8.942539543314799e-07, "loss": 1.1345, "step": 1239 }, { "epoch": 2.48, "grad_norm": 13.138537406921387, "learning_rate": 8.876245235966884e-07, "loss": 1.2053, "step": 1240 }, { "epoch": 2.482, "grad_norm": 10.92805290222168, "learning_rate": 8.810173630955249e-07, "loss": 1.1372, "step": 1241 }, { "epoch": 2.484, "grad_norm": 11.212373733520508, "learning_rate": 8.744325086085248e-07, "loss": 1.0706, "step": 1242 }, { "epoch": 2.4859999999999998, "grad_norm": 11.673620223999023, "learning_rate": 8.678699957954323e-07, "loss": 1.1367, "step": 1243 }, { "epoch": 2.488, "grad_norm": 11.204456329345703, "learning_rate": 8.613298601949971e-07, "loss": 1.0237, "step": 1244 }, { "epoch": 2.49, "grad_norm": 12.066752433776855, "learning_rate": 8.54812137224792e-07, "loss": 1.2087, "step": 1245 }, { "epoch": 2.492, "grad_norm": 10.22867202758789, "learning_rate": 8.483168621810133e-07, "loss": 1.1078, "step": 1246 }, { "epoch": 2.4939999999999998, "grad_norm": 12.135293960571289, "learning_rate": 8.418440702382897e-07, "loss": 1.0668, "step": 1247 }, { "epoch": 2.496, "grad_norm": 9.851129531860352, "learning_rate": 8.353937964495029e-07, "loss": 1.0521, "step": 1248 }, { "epoch": 2.498, "grad_norm": 10.662277221679688, "learning_rate": 8.289660757455803e-07, "loss": 1.0326, "step": 1249 }, { "epoch": 2.5, "grad_norm": 13.312153816223145, "learning_rate": 8.225609429353187e-07, "loss": 1.0838, "step": 1250 }, { "epoch": 2.502, "grad_norm": 10.633809089660645, "learning_rate": 8.161784327051919e-07, "loss": 1.0238, "step": 1251 }, { "epoch": 2.504, "grad_norm": 11.557962417602539, "learning_rate": 8.098185796191632e-07, "loss": 1.1297, "step": 1252 }, { "epoch": 2.5060000000000002, "grad_norm": 10.614686965942383, "learning_rate": 8.034814181184996e-07, "loss": 1.1224, "step": 1253 }, { "epoch": 2.508, "grad_norm": 11.647783279418945, "learning_rate": 7.971669825215789e-07, "loss": 1.191, "step": 1254 }, { "epoch": 2.51, "grad_norm": 12.704024314880371, "learning_rate": 7.908753070237124e-07, "loss": 1.0431, "step": 1255 }, { "epoch": 2.512, "grad_norm": 11.132079124450684, "learning_rate": 7.846064256969571e-07, "loss": 1.1417, "step": 1256 }, { "epoch": 2.5140000000000002, "grad_norm": 12.66719913482666, "learning_rate": 7.783603724899258e-07, "loss": 1.1726, "step": 1257 }, { "epoch": 2.516, "grad_norm": 13.010668754577637, "learning_rate": 7.72137181227608e-07, "loss": 1.2599, "step": 1258 }, { "epoch": 2.518, "grad_norm": 13.327101707458496, "learning_rate": 7.659368856111926e-07, "loss": 1.0874, "step": 1259 }, { "epoch": 2.52, "grad_norm": 12.53348445892334, "learning_rate": 7.597595192178702e-07, "loss": 1.0867, "step": 1260 }, { "epoch": 2.5220000000000002, "grad_norm": 11.78579044342041, "learning_rate": 7.536051155006657e-07, "loss": 1.1224, "step": 1261 }, { "epoch": 2.524, "grad_norm": 12.989802360534668, "learning_rate": 7.47473707788251e-07, "loss": 1.0848, "step": 1262 }, { "epoch": 2.526, "grad_norm": 10.429177284240723, "learning_rate": 7.413653292847617e-07, "loss": 1.0456, "step": 1263 }, { "epoch": 2.528, "grad_norm": 12.047698020935059, "learning_rate": 7.352800130696253e-07, "loss": 1.1725, "step": 1264 }, { "epoch": 2.5300000000000002, "grad_norm": 13.212532043457031, "learning_rate": 7.292177920973726e-07, "loss": 1.0755, "step": 1265 }, { "epoch": 2.532, "grad_norm": 11.261322021484375, "learning_rate": 7.23178699197467e-07, "loss": 1.0894, "step": 1266 }, { "epoch": 2.534, "grad_norm": 12.5214262008667, "learning_rate": 7.171627670741243e-07, "loss": 1.0554, "step": 1267 }, { "epoch": 2.536, "grad_norm": 11.23154067993164, "learning_rate": 7.111700283061318e-07, "loss": 1.1117, "step": 1268 }, { "epoch": 2.5380000000000003, "grad_norm": 14.136534690856934, "learning_rate": 7.052005153466779e-07, "loss": 1.1256, "step": 1269 }, { "epoch": 2.54, "grad_norm": 11.402527809143066, "learning_rate": 6.992542605231739e-07, "loss": 1.1365, "step": 1270 }, { "epoch": 2.542, "grad_norm": 10.729368209838867, "learning_rate": 6.933312960370748e-07, "loss": 1.0958, "step": 1271 }, { "epoch": 2.544, "grad_norm": 12.732880592346191, "learning_rate": 6.874316539637127e-07, "loss": 1.1066, "step": 1272 }, { "epoch": 2.5460000000000003, "grad_norm": 11.761198043823242, "learning_rate": 6.815553662521185e-07, "loss": 1.0512, "step": 1273 }, { "epoch": 2.548, "grad_norm": 12.321219444274902, "learning_rate": 6.757024647248456e-07, "loss": 1.1105, "step": 1274 }, { "epoch": 2.55, "grad_norm": 10.499734878540039, "learning_rate": 6.698729810778065e-07, "loss": 1.1287, "step": 1275 }, { "epoch": 2.552, "grad_norm": 11.47294807434082, "learning_rate": 6.640669468800947e-07, "loss": 1.1906, "step": 1276 }, { "epoch": 2.5540000000000003, "grad_norm": 10.670211791992188, "learning_rate": 6.58284393573812e-07, "loss": 1.0593, "step": 1277 }, { "epoch": 2.556, "grad_norm": 11.801868438720703, "learning_rate": 6.52525352473905e-07, "loss": 1.0933, "step": 1278 }, { "epoch": 2.558, "grad_norm": 12.327805519104004, "learning_rate": 6.467898547679913e-07, "loss": 1.0864, "step": 1279 }, { "epoch": 2.56, "grad_norm": 12.2786226272583, "learning_rate": 6.410779315161885e-07, "loss": 1.1561, "step": 1280 }, { "epoch": 2.5620000000000003, "grad_norm": 11.15182876586914, "learning_rate": 6.353896136509524e-07, "loss": 1.1301, "step": 1281 }, { "epoch": 2.564, "grad_norm": 11.805049896240234, "learning_rate": 6.297249319769016e-07, "loss": 1.0785, "step": 1282 }, { "epoch": 2.566, "grad_norm": 11.21564769744873, "learning_rate": 6.240839171706608e-07, "loss": 1.2017, "step": 1283 }, { "epoch": 2.568, "grad_norm": 12.291101455688477, "learning_rate": 6.184665997806832e-07, "loss": 1.0942, "step": 1284 }, { "epoch": 2.57, "grad_norm": 11.54765510559082, "learning_rate": 6.128730102270897e-07, "loss": 1.1542, "step": 1285 }, { "epoch": 2.572, "grad_norm": 10.28836441040039, "learning_rate": 6.073031788015133e-07, "loss": 1.0565, "step": 1286 }, { "epoch": 2.574, "grad_norm": 12.515631675720215, "learning_rate": 6.017571356669183e-07, "loss": 1.0679, "step": 1287 }, { "epoch": 2.576, "grad_norm": 12.502131462097168, "learning_rate": 5.962349108574478e-07, "loss": 1.0424, "step": 1288 }, { "epoch": 2.578, "grad_norm": 11.3397855758667, "learning_rate": 5.9073653427826e-07, "loss": 1.0948, "step": 1289 }, { "epoch": 2.58, "grad_norm": 10.07397174835205, "learning_rate": 5.852620357053651e-07, "loss": 1.0421, "step": 1290 }, { "epoch": 2.582, "grad_norm": 11.410602569580078, "learning_rate": 5.798114447854636e-07, "loss": 1.1638, "step": 1291 }, { "epoch": 2.584, "grad_norm": 11.01577091217041, "learning_rate": 5.743847910357836e-07, "loss": 1.0889, "step": 1292 }, { "epoch": 2.586, "grad_norm": 12.289595603942871, "learning_rate": 5.689821038439264e-07, "loss": 1.1365, "step": 1293 }, { "epoch": 2.588, "grad_norm": 12.653968811035156, "learning_rate": 5.636034124677043e-07, "loss": 1.208, "step": 1294 }, { "epoch": 2.59, "grad_norm": 11.91626262664795, "learning_rate": 5.582487460349806e-07, "loss": 1.2131, "step": 1295 }, { "epoch": 2.592, "grad_norm": 12.287956237792969, "learning_rate": 5.529181335435124e-07, "loss": 1.096, "step": 1296 }, { "epoch": 2.594, "grad_norm": 14.091611862182617, "learning_rate": 5.476116038607993e-07, "loss": 1.0923, "step": 1297 }, { "epoch": 2.596, "grad_norm": 11.291535377502441, "learning_rate": 5.423291857239177e-07, "loss": 1.0751, "step": 1298 }, { "epoch": 2.598, "grad_norm": 10.707423210144043, "learning_rate": 5.370709077393721e-07, "loss": 1.1536, "step": 1299 }, { "epoch": 2.6, "grad_norm": 10.066755294799805, "learning_rate": 5.318367983829393e-07, "loss": 1.1073, "step": 1300 }, { "epoch": 2.602, "grad_norm": 12.727775573730469, "learning_rate": 5.266268859995083e-07, "loss": 1.0421, "step": 1301 }, { "epoch": 2.604, "grad_norm": 10.819839477539062, "learning_rate": 5.214411988029355e-07, "loss": 1.0953, "step": 1302 }, { "epoch": 2.606, "grad_norm": 11.471372604370117, "learning_rate": 5.162797648758877e-07, "loss": 1.1238, "step": 1303 }, { "epoch": 2.608, "grad_norm": 11.147175788879395, "learning_rate": 5.111426121696866e-07, "loss": 1.1459, "step": 1304 }, { "epoch": 2.61, "grad_norm": 10.736647605895996, "learning_rate": 5.06029768504166e-07, "loss": 1.0977, "step": 1305 }, { "epoch": 2.612, "grad_norm": 10.199406623840332, "learning_rate": 5.009412615675102e-07, "loss": 1.0806, "step": 1306 }, { "epoch": 2.614, "grad_norm": 10.652769088745117, "learning_rate": 4.958771189161149e-07, "loss": 1.1096, "step": 1307 }, { "epoch": 2.616, "grad_norm": 10.275524139404297, "learning_rate": 4.908373679744316e-07, "loss": 1.1095, "step": 1308 }, { "epoch": 2.618, "grad_norm": 11.423064231872559, "learning_rate": 4.858220360348187e-07, "loss": 1.1176, "step": 1309 }, { "epoch": 2.62, "grad_norm": 12.112739562988281, "learning_rate": 4.808311502573976e-07, "loss": 1.1444, "step": 1310 }, { "epoch": 2.622, "grad_norm": 10.57923412322998, "learning_rate": 4.758647376699033e-07, "loss": 1.1199, "step": 1311 }, { "epoch": 2.624, "grad_norm": 11.439313888549805, "learning_rate": 4.709228251675357e-07, "loss": 1.0995, "step": 1312 }, { "epoch": 2.626, "grad_norm": 11.128246307373047, "learning_rate": 4.6600543951281995e-07, "loss": 1.1438, "step": 1313 }, { "epoch": 2.628, "grad_norm": 12.107537269592285, "learning_rate": 4.6111260733545714e-07, "loss": 1.0989, "step": 1314 }, { "epoch": 2.63, "grad_norm": 11.53816032409668, "learning_rate": 4.562443551321788e-07, "loss": 1.1305, "step": 1315 }, { "epoch": 2.632, "grad_norm": 10.835124015808105, "learning_rate": 4.514007092666084e-07, "loss": 1.1451, "step": 1316 }, { "epoch": 2.634, "grad_norm": 11.95799446105957, "learning_rate": 4.4658169596911493e-07, "loss": 1.1823, "step": 1317 }, { "epoch": 2.636, "grad_norm": 10.443716049194336, "learning_rate": 4.417873413366702e-07, "loss": 1.0862, "step": 1318 }, { "epoch": 2.638, "grad_norm": 12.407042503356934, "learning_rate": 4.370176713327118e-07, "loss": 1.0715, "step": 1319 }, { "epoch": 2.64, "grad_norm": 12.52112102508545, "learning_rate": 4.322727117869951e-07, "loss": 1.1408, "step": 1320 }, { "epoch": 2.642, "grad_norm": 12.11574935913086, "learning_rate": 4.275524883954657e-07, "loss": 1.1602, "step": 1321 }, { "epoch": 2.644, "grad_norm": 11.21030330657959, "learning_rate": 4.228570267201049e-07, "loss": 1.0917, "step": 1322 }, { "epoch": 2.646, "grad_norm": 10.748331069946289, "learning_rate": 4.1818635218880186e-07, "loss": 1.0612, "step": 1323 }, { "epoch": 2.648, "grad_norm": 12.968511581420898, "learning_rate": 4.1354049009521504e-07, "loss": 1.1723, "step": 1324 }, { "epoch": 2.65, "grad_norm": 13.352898597717285, "learning_rate": 4.089194655986306e-07, "loss": 1.127, "step": 1325 }, { "epoch": 2.652, "grad_norm": 10.655840873718262, "learning_rate": 4.043233037238281e-07, "loss": 1.0859, "step": 1326 }, { "epoch": 2.654, "grad_norm": 10.203536033630371, "learning_rate": 3.99752029360948e-07, "loss": 1.1438, "step": 1327 }, { "epoch": 2.656, "grad_norm": 12.646231651306152, "learning_rate": 3.9520566726535367e-07, "loss": 1.2127, "step": 1328 }, { "epoch": 2.658, "grad_norm": 12.770830154418945, "learning_rate": 3.90684242057498e-07, "loss": 1.0962, "step": 1329 }, { "epoch": 2.66, "grad_norm": 11.451815605163574, "learning_rate": 3.8618777822278854e-07, "loss": 1.0883, "step": 1330 }, { "epoch": 2.662, "grad_norm": 11.632091522216797, "learning_rate": 3.8171630011145877e-07, "loss": 1.1129, "step": 1331 }, { "epoch": 2.664, "grad_norm": 11.543761253356934, "learning_rate": 3.772698319384349e-07, "loss": 1.0953, "step": 1332 }, { "epoch": 2.666, "grad_norm": 12.997357368469238, "learning_rate": 3.728483977831998e-07, "loss": 1.1444, "step": 1333 }, { "epoch": 2.668, "grad_norm": 11.470213890075684, "learning_rate": 3.684520215896703e-07, "loss": 1.1423, "step": 1334 }, { "epoch": 2.67, "grad_norm": 10.749590873718262, "learning_rate": 3.6408072716606346e-07, "loss": 1.1258, "step": 1335 }, { "epoch": 2.672, "grad_norm": 10.158390045166016, "learning_rate": 3.597345381847656e-07, "loss": 1.032, "step": 1336 }, { "epoch": 2.674, "grad_norm": 10.492109298706055, "learning_rate": 3.554134781822094e-07, "loss": 1.1678, "step": 1337 }, { "epoch": 2.676, "grad_norm": 12.122913360595703, "learning_rate": 3.511175705587433e-07, "loss": 1.1777, "step": 1338 }, { "epoch": 2.678, "grad_norm": 10.95935344696045, "learning_rate": 3.468468385785023e-07, "loss": 1.1732, "step": 1339 }, { "epoch": 2.68, "grad_norm": 12.425519943237305, "learning_rate": 3.426013053692878e-07, "loss": 1.1649, "step": 1340 }, { "epoch": 2.682, "grad_norm": 11.766524314880371, "learning_rate": 3.3838099392243915e-07, "loss": 1.0, "step": 1341 }, { "epoch": 2.684, "grad_norm": 12.33736515045166, "learning_rate": 3.341859270927067e-07, "loss": 1.148, "step": 1342 }, { "epoch": 2.686, "grad_norm": 10.662920951843262, "learning_rate": 3.30016127598134e-07, "loss": 1.0533, "step": 1343 }, { "epoch": 2.6879999999999997, "grad_norm": 11.27861213684082, "learning_rate": 3.258716180199278e-07, "loss": 1.114, "step": 1344 }, { "epoch": 2.69, "grad_norm": 11.482233047485352, "learning_rate": 3.2175242080234314e-07, "loss": 1.1744, "step": 1345 }, { "epoch": 2.692, "grad_norm": 11.544189453125, "learning_rate": 3.1765855825255543e-07, "loss": 1.0746, "step": 1346 }, { "epoch": 2.694, "grad_norm": 13.263728141784668, "learning_rate": 3.135900525405428e-07, "loss": 1.1174, "step": 1347 }, { "epoch": 2.6959999999999997, "grad_norm": 11.041204452514648, "learning_rate": 3.0954692569896585e-07, "loss": 1.1725, "step": 1348 }, { "epoch": 2.698, "grad_norm": 11.216727256774902, "learning_rate": 3.055291996230492e-07, "loss": 1.1185, "step": 1349 }, { "epoch": 2.7, "grad_norm": 11.795129776000977, "learning_rate": 3.015368960704584e-07, "loss": 1.1468, "step": 1350 }, { "epoch": 2.702, "grad_norm": 10.980323791503906, "learning_rate": 2.975700366611883e-07, "loss": 1.0928, "step": 1351 }, { "epoch": 2.7039999999999997, "grad_norm": 12.678966522216797, "learning_rate": 2.9362864287744266e-07, "loss": 1.1624, "step": 1352 }, { "epoch": 2.706, "grad_norm": 12.718985557556152, "learning_rate": 2.8971273606351656e-07, "loss": 1.141, "step": 1353 }, { "epoch": 2.708, "grad_norm": 11.222257614135742, "learning_rate": 2.858223374256841e-07, "loss": 1.1179, "step": 1354 }, { "epoch": 2.71, "grad_norm": 10.49862003326416, "learning_rate": 2.819574680320825e-07, "loss": 1.1192, "step": 1355 }, { "epoch": 2.7119999999999997, "grad_norm": 13.092155456542969, "learning_rate": 2.7811814881259503e-07, "loss": 1.1709, "step": 1356 }, { "epoch": 2.714, "grad_norm": 10.232337951660156, "learning_rate": 2.743044005587425e-07, "loss": 1.0475, "step": 1357 }, { "epoch": 2.716, "grad_norm": 12.870683670043945, "learning_rate": 2.705162439235648e-07, "loss": 1.0407, "step": 1358 }, { "epoch": 2.718, "grad_norm": 11.157443046569824, "learning_rate": 2.6675369942151864e-07, "loss": 1.1319, "step": 1359 }, { "epoch": 2.7199999999999998, "grad_norm": 10.935739517211914, "learning_rate": 2.63016787428354e-07, "loss": 1.1432, "step": 1360 }, { "epoch": 2.722, "grad_norm": 12.044547080993652, "learning_rate": 2.593055281810125e-07, "loss": 1.1435, "step": 1361 }, { "epoch": 2.724, "grad_norm": 10.721463203430176, "learning_rate": 2.556199417775174e-07, "loss": 1.1649, "step": 1362 }, { "epoch": 2.726, "grad_norm": 10.794017791748047, "learning_rate": 2.519600481768597e-07, "loss": 1.1491, "step": 1363 }, { "epoch": 2.7279999999999998, "grad_norm": 11.180771827697754, "learning_rate": 2.483258671988942e-07, "loss": 1.0898, "step": 1364 }, { "epoch": 2.73, "grad_norm": 11.11221694946289, "learning_rate": 2.447174185242324e-07, "loss": 1.1323, "step": 1365 }, { "epoch": 2.732, "grad_norm": 11.723511695861816, "learning_rate": 2.4113472169413176e-07, "loss": 1.2018, "step": 1366 }, { "epoch": 2.734, "grad_norm": 12.620880126953125, "learning_rate": 2.37577796110397e-07, "loss": 1.1411, "step": 1367 }, { "epoch": 2.7359999999999998, "grad_norm": 11.710165023803711, "learning_rate": 2.3404666103526542e-07, "loss": 1.0482, "step": 1368 }, { "epoch": 2.738, "grad_norm": 11.673789024353027, "learning_rate": 2.3054133559131163e-07, "loss": 1.2405, "step": 1369 }, { "epoch": 2.74, "grad_norm": 11.153477668762207, "learning_rate": 2.2706183876134047e-07, "loss": 1.1506, "step": 1370 }, { "epoch": 2.742, "grad_norm": 11.211380958557129, "learning_rate": 2.2360818938828189e-07, "loss": 1.046, "step": 1371 }, { "epoch": 2.7439999999999998, "grad_norm": 10.854473114013672, "learning_rate": 2.2018040617509174e-07, "loss": 1.0861, "step": 1372 }, { "epoch": 2.746, "grad_norm": 11.484323501586914, "learning_rate": 2.167785076846518e-07, "loss": 1.0931, "step": 1373 }, { "epoch": 2.748, "grad_norm": 12.862896919250488, "learning_rate": 2.134025123396638e-07, "loss": 1.215, "step": 1374 }, { "epoch": 2.75, "grad_norm": 12.15230941772461, "learning_rate": 2.1005243842255552e-07, "loss": 1.167, "step": 1375 }, { "epoch": 2.752, "grad_norm": 10.922401428222656, "learning_rate": 2.0672830407537925e-07, "loss": 1.1432, "step": 1376 }, { "epoch": 2.754, "grad_norm": 12.105512619018555, "learning_rate": 2.0343012729971244e-07, "loss": 1.1478, "step": 1377 }, { "epoch": 2.7560000000000002, "grad_norm": 10.566096305847168, "learning_rate": 2.0015792595656225e-07, "loss": 1.1038, "step": 1378 }, { "epoch": 2.758, "grad_norm": 12.602991104125977, "learning_rate": 1.9691171776626882e-07, "loss": 1.2035, "step": 1379 }, { "epoch": 2.76, "grad_norm": 10.561416625976562, "learning_rate": 1.9369152030840553e-07, "loss": 1.1399, "step": 1380 }, { "epoch": 2.762, "grad_norm": 12.064254760742188, "learning_rate": 1.904973510216912e-07, "loss": 1.2584, "step": 1381 }, { "epoch": 2.7640000000000002, "grad_norm": 10.637829780578613, "learning_rate": 1.873292272038868e-07, "loss": 1.0919, "step": 1382 }, { "epoch": 2.766, "grad_norm": 11.943845748901367, "learning_rate": 1.841871660117095e-07, "loss": 1.1898, "step": 1383 }, { "epoch": 2.768, "grad_norm": 11.168000221252441, "learning_rate": 1.8107118446073492e-07, "loss": 1.1557, "step": 1384 }, { "epoch": 2.77, "grad_norm": 10.773613929748535, "learning_rate": 1.779812994253055e-07, "loss": 1.1016, "step": 1385 }, { "epoch": 2.7720000000000002, "grad_norm": 10.496236801147461, "learning_rate": 1.7491752763844294e-07, "loss": 1.0975, "step": 1386 }, { "epoch": 2.774, "grad_norm": 11.72801399230957, "learning_rate": 1.7187988569175307e-07, "loss": 1.0915, "step": 1387 }, { "epoch": 2.776, "grad_norm": 11.977339744567871, "learning_rate": 1.688683900353366e-07, "loss": 1.1441, "step": 1388 }, { "epoch": 2.778, "grad_norm": 10.9048490524292, "learning_rate": 1.6588305697770313e-07, "loss": 1.0953, "step": 1389 }, { "epoch": 2.7800000000000002, "grad_norm": 10.485379219055176, "learning_rate": 1.6292390268568103e-07, "loss": 1.0981, "step": 1390 }, { "epoch": 2.782, "grad_norm": 10.757002830505371, "learning_rate": 1.5999094318432662e-07, "loss": 1.1695, "step": 1391 }, { "epoch": 2.784, "grad_norm": 11.401092529296875, "learning_rate": 1.5708419435684463e-07, "loss": 1.0509, "step": 1392 }, { "epoch": 2.786, "grad_norm": 11.38297176361084, "learning_rate": 1.5420367194449448e-07, "loss": 1.0742, "step": 1393 }, { "epoch": 2.7880000000000003, "grad_norm": 10.606424331665039, "learning_rate": 1.5134939154651196e-07, "loss": 1.1533, "step": 1394 }, { "epoch": 2.79, "grad_norm": 11.047826766967773, "learning_rate": 1.4852136862001766e-07, "loss": 1.15, "step": 1395 }, { "epoch": 2.792, "grad_norm": 11.345439910888672, "learning_rate": 1.4571961847993977e-07, "loss": 1.2146, "step": 1396 }, { "epoch": 2.794, "grad_norm": 11.464273452758789, "learning_rate": 1.4294415629892756e-07, "loss": 1.0851, "step": 1397 }, { "epoch": 2.7960000000000003, "grad_norm": 11.990900039672852, "learning_rate": 1.4019499710726913e-07, "loss": 1.1197, "step": 1398 }, { "epoch": 2.798, "grad_norm": 11.901836395263672, "learning_rate": 1.374721557928116e-07, "loss": 1.2454, "step": 1399 }, { "epoch": 2.8, "grad_norm": 10.661705017089844, "learning_rate": 1.3477564710088097e-07, "loss": 1.0545, "step": 1400 }, { "epoch": 2.802, "grad_norm": 11.410598754882812, "learning_rate": 1.3210548563419857e-07, "loss": 1.0326, "step": 1401 }, { "epoch": 2.8040000000000003, "grad_norm": 11.102482795715332, "learning_rate": 1.294616858528064e-07, "loss": 1.1085, "step": 1402 }, { "epoch": 2.806, "grad_norm": 10.520735740661621, "learning_rate": 1.268442620739868e-07, "loss": 1.1905, "step": 1403 }, { "epoch": 2.808, "grad_norm": 11.018781661987305, "learning_rate": 1.2425322847218368e-07, "loss": 1.1735, "step": 1404 }, { "epoch": 2.81, "grad_norm": 12.908723831176758, "learning_rate": 1.2168859907892904e-07, "loss": 1.1788, "step": 1405 }, { "epoch": 2.8120000000000003, "grad_norm": 11.897324562072754, "learning_rate": 1.1915038778276212e-07, "loss": 1.093, "step": 1406 }, { "epoch": 2.814, "grad_norm": 11.710148811340332, "learning_rate": 1.166386083291604e-07, "loss": 1.1192, "step": 1407 }, { "epoch": 2.816, "grad_norm": 11.686257362365723, "learning_rate": 1.1415327432046041e-07, "loss": 1.0824, "step": 1408 }, { "epoch": 2.818, "grad_norm": 11.252178192138672, "learning_rate": 1.1169439921578485e-07, "loss": 1.0743, "step": 1409 }, { "epoch": 2.82, "grad_norm": 10.60265064239502, "learning_rate": 1.0926199633097156e-07, "loss": 1.0421, "step": 1410 }, { "epoch": 2.822, "grad_norm": 12.824352264404297, "learning_rate": 1.0685607883850035e-07, "loss": 1.1632, "step": 1411 }, { "epoch": 2.824, "grad_norm": 11.57745361328125, "learning_rate": 1.044766597674196e-07, "loss": 1.1655, "step": 1412 }, { "epoch": 2.826, "grad_norm": 11.447077751159668, "learning_rate": 1.0212375200327973e-07, "loss": 1.0834, "step": 1413 }, { "epoch": 2.828, "grad_norm": 12.608105659484863, "learning_rate": 9.979736828806096e-08, "loss": 1.1674, "step": 1414 }, { "epoch": 2.83, "grad_norm": 11.511768341064453, "learning_rate": 9.749752122010347e-08, "loss": 1.2345, "step": 1415 }, { "epoch": 2.832, "grad_norm": 10.86443042755127, "learning_rate": 9.522422325404234e-08, "loss": 1.0663, "step": 1416 }, { "epoch": 2.834, "grad_norm": 11.792540550231934, "learning_rate": 9.297748670073658e-08, "loss": 1.0739, "step": 1417 }, { "epoch": 2.836, "grad_norm": 12.60987377166748, "learning_rate": 9.075732372720414e-08, "loss": 1.121, "step": 1418 }, { "epoch": 2.838, "grad_norm": 11.70394515991211, "learning_rate": 8.856374635655696e-08, "loss": 1.1565, "step": 1419 }, { "epoch": 2.84, "grad_norm": 11.06529426574707, "learning_rate": 8.639676646793382e-08, "loss": 1.0428, "step": 1420 }, { "epoch": 2.842, "grad_norm": 11.53507137298584, "learning_rate": 8.425639579643763e-08, "loss": 1.094, "step": 1421 }, { "epoch": 2.844, "grad_norm": 10.740865707397461, "learning_rate": 8.214264593307097e-08, "loss": 1.0513, "step": 1422 }, { "epoch": 2.846, "grad_norm": 11.995918273925781, "learning_rate": 8.00555283246729e-08, "loss": 1.1279, "step": 1423 }, { "epoch": 2.848, "grad_norm": 12.916094779968262, "learning_rate": 7.799505427386001e-08, "loss": 1.1227, "step": 1424 }, { "epoch": 2.85, "grad_norm": 10.555110931396484, "learning_rate": 7.59612349389599e-08, "loss": 1.0907, "step": 1425 }, { "epoch": 2.852, "grad_norm": 11.170307159423828, "learning_rate": 7.395408133395509e-08, "loss": 1.1291, "step": 1426 }, { "epoch": 2.854, "grad_norm": 10.48733901977539, "learning_rate": 7.197360432842359e-08, "loss": 1.1168, "step": 1427 }, { "epoch": 2.856, "grad_norm": 11.542524337768555, "learning_rate": 7.001981464747565e-08, "loss": 1.1142, "step": 1428 }, { "epoch": 2.858, "grad_norm": 10.778676986694336, "learning_rate": 6.809272287169988e-08, "loss": 1.1537, "step": 1429 }, { "epoch": 2.86, "grad_norm": 10.768854141235352, "learning_rate": 6.61923394371039e-08, "loss": 1.0235, "step": 1430 }, { "epoch": 2.862, "grad_norm": 10.812145233154297, "learning_rate": 6.431867463506047e-08, "loss": 1.1195, "step": 1431 }, { "epoch": 2.864, "grad_norm": 10.739622116088867, "learning_rate": 6.247173861224753e-08, "loss": 1.1552, "step": 1432 }, { "epoch": 2.866, "grad_norm": 10.252985954284668, "learning_rate": 6.065154137059603e-08, "loss": 1.1009, "step": 1433 }, { "epoch": 2.868, "grad_norm": 11.594807624816895, "learning_rate": 5.8858092767236084e-08, "loss": 1.1601, "step": 1434 }, { "epoch": 2.87, "grad_norm": 10.98749828338623, "learning_rate": 5.709140251444201e-08, "loss": 1.1708, "step": 1435 }, { "epoch": 2.872, "grad_norm": 11.422292709350586, "learning_rate": 5.535148017958014e-08, "loss": 1.0827, "step": 1436 }, { "epoch": 2.874, "grad_norm": 10.309944152832031, "learning_rate": 5.363833518505834e-08, "loss": 1.1094, "step": 1437 }, { "epoch": 2.876, "grad_norm": 11.423945426940918, "learning_rate": 5.19519768082738e-08, "loss": 1.2409, "step": 1438 }, { "epoch": 2.878, "grad_norm": 10.608428001403809, "learning_rate": 5.029241418156139e-08, "loss": 1.1391, "step": 1439 }, { "epoch": 2.88, "grad_norm": 13.38857650756836, "learning_rate": 4.865965629214819e-08, "loss": 1.1201, "step": 1440 }, { "epoch": 2.882, "grad_norm": 11.526823997497559, "learning_rate": 4.7053711982101294e-08, "loss": 1.145, "step": 1441 }, { "epoch": 2.884, "grad_norm": 11.45752239227295, "learning_rate": 4.5474589948280026e-08, "loss": 1.2218, "step": 1442 }, { "epoch": 2.886, "grad_norm": 10.46944808959961, "learning_rate": 4.392229874229159e-08, "loss": 1.2481, "step": 1443 }, { "epoch": 2.888, "grad_norm": 12.171918869018555, "learning_rate": 4.2396846770441644e-08, "loss": 1.1381, "step": 1444 }, { "epoch": 2.89, "grad_norm": 11.57470703125, "learning_rate": 4.0898242293691546e-08, "loss": 1.1412, "step": 1445 }, { "epoch": 2.892, "grad_norm": 11.620773315429688, "learning_rate": 3.9426493427611177e-08, "loss": 1.1664, "step": 1446 }, { "epoch": 2.894, "grad_norm": 12.99191951751709, "learning_rate": 3.7981608142335644e-08, "loss": 1.2922, "step": 1447 }, { "epoch": 2.896, "grad_norm": 11.910186767578125, "learning_rate": 3.65635942625242e-08, "loss": 1.285, "step": 1448 }, { "epoch": 2.898, "grad_norm": 11.08059310913086, "learning_rate": 3.517245946731529e-08, "loss": 1.1577, "step": 1449 }, { "epoch": 2.9, "grad_norm": 11.138588905334473, "learning_rate": 3.3808211290284886e-08, "loss": 1.1573, "step": 1450 }, { "epoch": 2.902, "grad_norm": 10.384156227111816, "learning_rate": 3.247085711940878e-08, "loss": 1.124, "step": 1451 }, { "epoch": 2.904, "grad_norm": 12.75297737121582, "learning_rate": 3.1160404197018155e-08, "loss": 1.1535, "step": 1452 }, { "epoch": 2.906, "grad_norm": 12.249570846557617, "learning_rate": 2.9876859619764606e-08, "loss": 1.1892, "step": 1453 }, { "epoch": 2.908, "grad_norm": 12.755828857421875, "learning_rate": 2.8620230338578526e-08, "loss": 1.0924, "step": 1454 }, { "epoch": 2.91, "grad_norm": 13.399191856384277, "learning_rate": 2.7390523158633552e-08, "loss": 1.0451, "step": 1455 }, { "epoch": 2.912, "grad_norm": 11.16951847076416, "learning_rate": 2.6187744739308297e-08, "loss": 1.0882, "step": 1456 }, { "epoch": 2.914, "grad_norm": 10.711827278137207, "learning_rate": 2.501190159415079e-08, "loss": 1.164, "step": 1457 }, { "epoch": 2.916, "grad_norm": 10.825516700744629, "learning_rate": 2.386300009084408e-08, "loss": 1.0708, "step": 1458 }, { "epoch": 2.918, "grad_norm": 11.371330261230469, "learning_rate": 2.27410464511707e-08, "loss": 1.1395, "step": 1459 }, { "epoch": 2.92, "grad_norm": 11.252603530883789, "learning_rate": 2.1646046750978255e-08, "loss": 1.1699, "step": 1460 }, { "epoch": 2.922, "grad_norm": 11.264942169189453, "learning_rate": 2.057800692014833e-08, "loss": 1.0813, "step": 1461 }, { "epoch": 2.924, "grad_norm": 11.598838806152344, "learning_rate": 1.953693274256374e-08, "loss": 1.124, "step": 1462 }, { "epoch": 2.926, "grad_norm": 10.306319236755371, "learning_rate": 1.8522829856076895e-08, "loss": 1.0401, "step": 1463 }, { "epoch": 2.928, "grad_norm": 11.464588165283203, "learning_rate": 1.753570375247815e-08, "loss": 1.154, "step": 1464 }, { "epoch": 2.93, "grad_norm": 11.513689994812012, "learning_rate": 1.657555977746972e-08, "loss": 1.0962, "step": 1465 }, { "epoch": 2.932, "grad_norm": 11.152294158935547, "learning_rate": 1.5642403130632367e-08, "loss": 1.119, "step": 1466 }, { "epoch": 2.934, "grad_norm": 11.795382499694824, "learning_rate": 1.4736238865398766e-08, "loss": 1.1594, "step": 1467 }, { "epoch": 2.936, "grad_norm": 13.539538383483887, "learning_rate": 1.3857071889029073e-08, "loss": 1.1061, "step": 1468 }, { "epoch": 2.9379999999999997, "grad_norm": 10.830578804016113, "learning_rate": 1.3004906962578723e-08, "loss": 1.114, "step": 1469 }, { "epoch": 2.94, "grad_norm": 12.558431625366211, "learning_rate": 1.2179748700879013e-08, "loss": 1.2487, "step": 1470 }, { "epoch": 2.942, "grad_norm": 11.865883827209473, "learning_rate": 1.1381601572505452e-08, "loss": 1.0191, "step": 1471 }, { "epoch": 2.944, "grad_norm": 10.461407661437988, "learning_rate": 1.0610469899760001e-08, "loss": 1.1221, "step": 1472 }, { "epoch": 2.9459999999999997, "grad_norm": 11.201630592346191, "learning_rate": 9.866357858642206e-09, "loss": 0.9862, "step": 1473 }, { "epoch": 2.948, "grad_norm": 11.596713066101074, "learning_rate": 9.14926947883088e-09, "loss": 1.2098, "step": 1474 }, { "epoch": 2.95, "grad_norm": 10.908271789550781, "learning_rate": 8.459208643659122e-09, "loss": 1.1182, "step": 1475 }, { "epoch": 2.952, "grad_norm": 11.701807975769043, "learning_rate": 7.796179090094891e-09, "loss": 1.2094, "step": 1476 }, { "epoch": 2.9539999999999997, "grad_norm": 13.19516658782959, "learning_rate": 7.160184408721571e-09, "loss": 1.1626, "step": 1477 }, { "epoch": 2.956, "grad_norm": 12.156759262084961, "learning_rate": 6.551228043715218e-09, "loss": 1.1171, "step": 1478 }, { "epoch": 2.958, "grad_norm": 10.075406074523926, "learning_rate": 5.969313292830126e-09, "loss": 0.9876, "step": 1479 }, { "epoch": 2.96, "grad_norm": 11.928035736083984, "learning_rate": 5.414443307377171e-09, "loss": 1.1166, "step": 1480 }, { "epoch": 2.9619999999999997, "grad_norm": 10.68159294128418, "learning_rate": 4.8866210922110525e-09, "loss": 1.058, "step": 1481 }, { "epoch": 2.964, "grad_norm": 11.216858863830566, "learning_rate": 4.385849505708084e-09, "loss": 1.1111, "step": 1482 }, { "epoch": 2.966, "grad_norm": 12.130516052246094, "learning_rate": 3.912131259757313e-09, "loss": 1.1494, "step": 1483 }, { "epoch": 2.968, "grad_norm": 10.646476745605469, "learning_rate": 3.4654689197405335e-09, "loss": 1.0957, "step": 1484 }, { "epoch": 2.9699999999999998, "grad_norm": 11.538399696350098, "learning_rate": 3.0458649045211897e-09, "loss": 1.1852, "step": 1485 }, { "epoch": 2.972, "grad_norm": 12.342727661132812, "learning_rate": 2.6533214864310485e-09, "loss": 1.1064, "step": 1486 }, { "epoch": 2.974, "grad_norm": 11.315055847167969, "learning_rate": 2.287840791256324e-09, "loss": 1.0906, "step": 1487 }, { "epoch": 2.976, "grad_norm": 12.045461654663086, "learning_rate": 1.9494247982282386e-09, "loss": 1.163, "step": 1488 }, { "epoch": 2.9779999999999998, "grad_norm": 10.301689147949219, "learning_rate": 1.638075340010814e-09, "loss": 1.1269, "step": 1489 }, { "epoch": 2.98, "grad_norm": 11.46061897277832, "learning_rate": 1.3537941026914302e-09, "loss": 1.0705, "step": 1490 }, { "epoch": 2.982, "grad_norm": 10.900529861450195, "learning_rate": 1.096582625772502e-09, "loss": 1.099, "step": 1491 }, { "epoch": 2.984, "grad_norm": 10.490682601928711, "learning_rate": 8.664423021614854e-10, "loss": 1.0517, "step": 1492 }, { "epoch": 2.9859999999999998, "grad_norm": 11.499110221862793, "learning_rate": 6.633743781642166e-10, "loss": 1.177, "step": 1493 }, { "epoch": 2.988, "grad_norm": 11.795324325561523, "learning_rate": 4.87379953478806e-10, "loss": 1.1048, "step": 1494 }, { "epoch": 2.99, "grad_norm": 11.121499061584473, "learning_rate": 3.384599811889766e-10, "loss": 1.1005, "step": 1495 }, { "epoch": 2.992, "grad_norm": 11.869853973388672, "learning_rate": 2.1661526775795804e-10, "loss": 1.1684, "step": 1496 }, { "epoch": 2.9939999999999998, "grad_norm": 13.046727180480957, "learning_rate": 1.2184647302626585e-10, "loss": 1.0746, "step": 1497 }, { "epoch": 2.996, "grad_norm": 12.80275821685791, "learning_rate": 5.4154110206150465e-11, "loss": 1.1424, "step": 1498 }, { "epoch": 2.998, "grad_norm": 13.787501335144043, "learning_rate": 1.3538545881042198e-11, "loss": 1.172, "step": 1499 }, { "epoch": 3.0, "grad_norm": 11.100687026977539, "learning_rate": 0.0, "loss": 1.1813, "step": 1500 }, { "epoch": 3.0, "step": 1500, "total_flos": 3.656864303572582e+16, "train_loss": 0.0, "train_runtime": 15.8388, "train_samples_per_second": 9091.618, "train_steps_per_second": 94.704 } ], "logging_steps": 1, "max_steps": 1500, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.656864303572582e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }