{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1359, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002207505518763797, "grad_norm": 2.7993483543395996, "learning_rate": 7.352941176470589e-08, "loss": 0.7615, "step": 1 }, { "epoch": 0.004415011037527594, "grad_norm": 2.814159393310547, "learning_rate": 1.4705882352941178e-07, "loss": 0.7759, "step": 2 }, { "epoch": 0.006622516556291391, "grad_norm": 2.932206392288208, "learning_rate": 2.2058823529411768e-07, "loss": 0.7897, "step": 3 }, { "epoch": 0.008830022075055188, "grad_norm": 2.909721851348877, "learning_rate": 2.9411764705882356e-07, "loss": 0.7698, "step": 4 }, { "epoch": 0.011037527593818985, "grad_norm": 2.855281114578247, "learning_rate": 3.6764705882352943e-07, "loss": 0.7673, "step": 5 }, { "epoch": 0.013245033112582781, "grad_norm": 2.930614709854126, "learning_rate": 4.4117647058823536e-07, "loss": 0.7914, "step": 6 }, { "epoch": 0.01545253863134658, "grad_norm": 2.8653368949890137, "learning_rate": 5.147058823529412e-07, "loss": 0.7713, "step": 7 }, { "epoch": 0.017660044150110375, "grad_norm": 2.7474284172058105, "learning_rate": 5.882352941176471e-07, "loss": 0.7713, "step": 8 }, { "epoch": 0.019867549668874173, "grad_norm": 2.6992416381835938, "learning_rate": 6.61764705882353e-07, "loss": 0.7549, "step": 9 }, { "epoch": 0.02207505518763797, "grad_norm": 2.718104362487793, "learning_rate": 7.352941176470589e-07, "loss": 0.757, "step": 10 }, { "epoch": 0.024282560706401765, "grad_norm": 2.670048952102661, "learning_rate": 8.088235294117648e-07, "loss": 0.7367, "step": 11 }, { "epoch": 0.026490066225165563, "grad_norm": 2.205970287322998, "learning_rate": 8.823529411764707e-07, "loss": 0.7397, "step": 12 }, { "epoch": 0.02869757174392936, "grad_norm": 2.2243752479553223, "learning_rate": 9.558823529411764e-07, "loss": 0.7565, "step": 13 }, { "epoch": 0.03090507726269316, "grad_norm": 2.2006163597106934, "learning_rate": 1.0294117647058825e-06, "loss": 0.744, "step": 14 }, { "epoch": 0.033112582781456956, "grad_norm": 2.0692009925842285, "learning_rate": 1.1029411764705884e-06, "loss": 0.7282, "step": 15 }, { "epoch": 0.03532008830022075, "grad_norm": 1.755692958831787, "learning_rate": 1.1764705882352942e-06, "loss": 0.7027, "step": 16 }, { "epoch": 0.037527593818984545, "grad_norm": 1.429651141166687, "learning_rate": 1.25e-06, "loss": 0.7028, "step": 17 }, { "epoch": 0.039735099337748346, "grad_norm": 1.4149061441421509, "learning_rate": 1.323529411764706e-06, "loss": 0.7027, "step": 18 }, { "epoch": 0.04194260485651214, "grad_norm": 1.4027491807937622, "learning_rate": 1.3970588235294119e-06, "loss": 0.7034, "step": 19 }, { "epoch": 0.04415011037527594, "grad_norm": 1.3382784128189087, "learning_rate": 1.4705882352941177e-06, "loss": 0.6838, "step": 20 }, { "epoch": 0.046357615894039736, "grad_norm": 1.2848469018936157, "learning_rate": 1.5441176470588238e-06, "loss": 0.6904, "step": 21 }, { "epoch": 0.04856512141280353, "grad_norm": 1.0581680536270142, "learning_rate": 1.6176470588235297e-06, "loss": 0.6658, "step": 22 }, { "epoch": 0.05077262693156733, "grad_norm": 1.0082190036773682, "learning_rate": 1.6911764705882356e-06, "loss": 0.653, "step": 23 }, { "epoch": 0.052980132450331126, "grad_norm": 0.979928195476532, "learning_rate": 1.7647058823529414e-06, "loss": 0.6501, "step": 24 }, { "epoch": 0.05518763796909492, "grad_norm": 0.9646239876747131, "learning_rate": 1.8382352941176473e-06, "loss": 0.6415, "step": 25 }, { "epoch": 0.05739514348785872, "grad_norm": 0.8932761549949646, "learning_rate": 1.9117647058823528e-06, "loss": 0.6523, "step": 26 }, { "epoch": 0.059602649006622516, "grad_norm": 0.8707468509674072, "learning_rate": 1.985294117647059e-06, "loss": 0.6434, "step": 27 }, { "epoch": 0.06181015452538632, "grad_norm": 0.8291523456573486, "learning_rate": 2.058823529411765e-06, "loss": 0.6333, "step": 28 }, { "epoch": 0.0640176600441501, "grad_norm": 0.7280705571174622, "learning_rate": 2.132352941176471e-06, "loss": 0.6198, "step": 29 }, { "epoch": 0.06622516556291391, "grad_norm": 0.6107202172279358, "learning_rate": 2.2058823529411767e-06, "loss": 0.6057, "step": 30 }, { "epoch": 0.0684326710816777, "grad_norm": 0.6679992079734802, "learning_rate": 2.2794117647058826e-06, "loss": 0.5983, "step": 31 }, { "epoch": 0.0706401766004415, "grad_norm": 0.6695526242256165, "learning_rate": 2.3529411764705885e-06, "loss": 0.5946, "step": 32 }, { "epoch": 0.0728476821192053, "grad_norm": 0.6435050964355469, "learning_rate": 2.4264705882352943e-06, "loss": 0.5975, "step": 33 }, { "epoch": 0.07505518763796909, "grad_norm": 0.6036580204963684, "learning_rate": 2.5e-06, "loss": 0.592, "step": 34 }, { "epoch": 0.0772626931567329, "grad_norm": 0.5135894417762756, "learning_rate": 2.5735294117647057e-06, "loss": 0.596, "step": 35 }, { "epoch": 0.07947019867549669, "grad_norm": 0.4571033716201782, "learning_rate": 2.647058823529412e-06, "loss": 0.5764, "step": 36 }, { "epoch": 0.08167770419426049, "grad_norm": 0.48609447479248047, "learning_rate": 2.720588235294118e-06, "loss": 0.5593, "step": 37 }, { "epoch": 0.08388520971302428, "grad_norm": 0.49421966075897217, "learning_rate": 2.7941176470588237e-06, "loss": 0.5634, "step": 38 }, { "epoch": 0.08609271523178808, "grad_norm": 0.48713281750679016, "learning_rate": 2.8676470588235296e-06, "loss": 0.5577, "step": 39 }, { "epoch": 0.08830022075055188, "grad_norm": 0.42998674511909485, "learning_rate": 2.9411764705882355e-06, "loss": 0.5602, "step": 40 }, { "epoch": 0.09050772626931568, "grad_norm": 0.39199528098106384, "learning_rate": 3.0147058823529413e-06, "loss": 0.5509, "step": 41 }, { "epoch": 0.09271523178807947, "grad_norm": 0.3977169692516327, "learning_rate": 3.0882352941176476e-06, "loss": 0.5408, "step": 42 }, { "epoch": 0.09492273730684327, "grad_norm": 0.3659592866897583, "learning_rate": 3.161764705882353e-06, "loss": 0.5468, "step": 43 }, { "epoch": 0.09713024282560706, "grad_norm": 0.41232988238334656, "learning_rate": 3.2352941176470594e-06, "loss": 0.5312, "step": 44 }, { "epoch": 0.09933774834437085, "grad_norm": 0.41818928718566895, "learning_rate": 3.308823529411765e-06, "loss": 0.542, "step": 45 }, { "epoch": 0.10154525386313466, "grad_norm": 0.38174012303352356, "learning_rate": 3.382352941176471e-06, "loss": 0.5341, "step": 46 }, { "epoch": 0.10375275938189846, "grad_norm": 0.3903500735759735, "learning_rate": 3.4558823529411766e-06, "loss": 0.5215, "step": 47 }, { "epoch": 0.10596026490066225, "grad_norm": 0.3658589720726013, "learning_rate": 3.529411764705883e-06, "loss": 0.5277, "step": 48 }, { "epoch": 0.10816777041942605, "grad_norm": 0.3160182237625122, "learning_rate": 3.6029411764705883e-06, "loss": 0.5193, "step": 49 }, { "epoch": 0.11037527593818984, "grad_norm": 0.29906004667282104, "learning_rate": 3.6764705882352946e-06, "loss": 0.525, "step": 50 }, { "epoch": 0.11258278145695365, "grad_norm": 0.28197285532951355, "learning_rate": 3.7500000000000005e-06, "loss": 0.5136, "step": 51 }, { "epoch": 0.11479028697571744, "grad_norm": 0.2733807861804962, "learning_rate": 3.8235294117647055e-06, "loss": 0.5126, "step": 52 }, { "epoch": 0.11699779249448124, "grad_norm": 0.2581369876861572, "learning_rate": 3.897058823529412e-06, "loss": 0.5102, "step": 53 }, { "epoch": 0.11920529801324503, "grad_norm": 0.24509315192699432, "learning_rate": 3.970588235294118e-06, "loss": 0.4959, "step": 54 }, { "epoch": 0.12141280353200883, "grad_norm": 0.23736661672592163, "learning_rate": 4.044117647058824e-06, "loss": 0.5024, "step": 55 }, { "epoch": 0.12362030905077263, "grad_norm": 0.2297072857618332, "learning_rate": 4.11764705882353e-06, "loss": 0.516, "step": 56 }, { "epoch": 0.12582781456953643, "grad_norm": 0.21641261875629425, "learning_rate": 4.191176470588236e-06, "loss": 0.5045, "step": 57 }, { "epoch": 0.1280353200883002, "grad_norm": 0.2174406796693802, "learning_rate": 4.264705882352942e-06, "loss": 0.4954, "step": 58 }, { "epoch": 0.13024282560706402, "grad_norm": 0.2182844579219818, "learning_rate": 4.3382352941176475e-06, "loss": 0.5073, "step": 59 }, { "epoch": 0.13245033112582782, "grad_norm": 0.18668220937252045, "learning_rate": 4.411764705882353e-06, "loss": 0.4927, "step": 60 }, { "epoch": 0.1346578366445916, "grad_norm": 0.20196162164211273, "learning_rate": 4.485294117647059e-06, "loss": 0.4962, "step": 61 }, { "epoch": 0.1368653421633554, "grad_norm": 0.20648355782032013, "learning_rate": 4.558823529411765e-06, "loss": 0.4965, "step": 62 }, { "epoch": 0.1390728476821192, "grad_norm": 0.1998893916606903, "learning_rate": 4.632352941176471e-06, "loss": 0.4857, "step": 63 }, { "epoch": 0.141280353200883, "grad_norm": 0.20052312314510345, "learning_rate": 4.705882352941177e-06, "loss": 0.4919, "step": 64 }, { "epoch": 0.1434878587196468, "grad_norm": 0.18653374910354614, "learning_rate": 4.779411764705883e-06, "loss": 0.4895, "step": 65 }, { "epoch": 0.1456953642384106, "grad_norm": 0.17638395726680756, "learning_rate": 4.852941176470589e-06, "loss": 0.4831, "step": 66 }, { "epoch": 0.1479028697571744, "grad_norm": 0.17102564871311188, "learning_rate": 4.9264705882352945e-06, "loss": 0.4817, "step": 67 }, { "epoch": 0.15011037527593818, "grad_norm": 0.15633539855480194, "learning_rate": 5e-06, "loss": 0.4882, "step": 68 }, { "epoch": 0.152317880794702, "grad_norm": 0.17420779168605804, "learning_rate": 5.073529411764706e-06, "loss": 0.4853, "step": 69 }, { "epoch": 0.1545253863134658, "grad_norm": 0.1659373641014099, "learning_rate": 5.147058823529411e-06, "loss": 0.492, "step": 70 }, { "epoch": 0.15673289183222958, "grad_norm": 0.17147059738636017, "learning_rate": 5.220588235294118e-06, "loss": 0.4744, "step": 71 }, { "epoch": 0.15894039735099338, "grad_norm": 0.15868496894836426, "learning_rate": 5.294117647058824e-06, "loss": 0.4845, "step": 72 }, { "epoch": 0.16114790286975716, "grad_norm": 0.16405610740184784, "learning_rate": 5.36764705882353e-06, "loss": 0.4806, "step": 73 }, { "epoch": 0.16335540838852097, "grad_norm": 0.16762660443782806, "learning_rate": 5.441176470588236e-06, "loss": 0.465, "step": 74 }, { "epoch": 0.16556291390728478, "grad_norm": 0.209846630692482, "learning_rate": 5.514705882352942e-06, "loss": 0.486, "step": 75 }, { "epoch": 0.16777041942604856, "grad_norm": 0.17751334607601166, "learning_rate": 5.588235294117647e-06, "loss": 0.4761, "step": 76 }, { "epoch": 0.16997792494481237, "grad_norm": 0.16121278703212738, "learning_rate": 5.661764705882353e-06, "loss": 0.4774, "step": 77 }, { "epoch": 0.17218543046357615, "grad_norm": 0.2625029385089874, "learning_rate": 5.735294117647059e-06, "loss": 0.4663, "step": 78 }, { "epoch": 0.17439293598233996, "grad_norm": 0.15058760344982147, "learning_rate": 5.808823529411766e-06, "loss": 0.4644, "step": 79 }, { "epoch": 0.17660044150110377, "grad_norm": 0.16306105256080627, "learning_rate": 5.882352941176471e-06, "loss": 0.4745, "step": 80 }, { "epoch": 0.17880794701986755, "grad_norm": 0.1497834473848343, "learning_rate": 5.955882352941177e-06, "loss": 0.4614, "step": 81 }, { "epoch": 0.18101545253863136, "grad_norm": 0.15127182006835938, "learning_rate": 6.029411764705883e-06, "loss": 0.468, "step": 82 }, { "epoch": 0.18322295805739514, "grad_norm": 0.14110144972801208, "learning_rate": 6.102941176470589e-06, "loss": 0.4709, "step": 83 }, { "epoch": 0.18543046357615894, "grad_norm": 0.150424063205719, "learning_rate": 6.176470588235295e-06, "loss": 0.4614, "step": 84 }, { "epoch": 0.18763796909492272, "grad_norm": 0.14802858233451843, "learning_rate": 6.25e-06, "loss": 0.4549, "step": 85 }, { "epoch": 0.18984547461368653, "grad_norm": 0.1567091941833496, "learning_rate": 6.323529411764706e-06, "loss": 0.4487, "step": 86 }, { "epoch": 0.19205298013245034, "grad_norm": 0.14341039955615997, "learning_rate": 6.397058823529412e-06, "loss": 0.4453, "step": 87 }, { "epoch": 0.19426048565121412, "grad_norm": 0.20370961725711823, "learning_rate": 6.470588235294119e-06, "loss": 0.4626, "step": 88 }, { "epoch": 0.19646799116997793, "grad_norm": 0.14623787999153137, "learning_rate": 6.544117647058824e-06, "loss": 0.4559, "step": 89 }, { "epoch": 0.1986754966887417, "grad_norm": 0.1428503692150116, "learning_rate": 6.61764705882353e-06, "loss": 0.4464, "step": 90 }, { "epoch": 0.20088300220750552, "grad_norm": 0.14742067456245422, "learning_rate": 6.6911764705882356e-06, "loss": 0.4663, "step": 91 }, { "epoch": 0.20309050772626933, "grad_norm": 0.16871479153633118, "learning_rate": 6.764705882352942e-06, "loss": 0.4549, "step": 92 }, { "epoch": 0.2052980132450331, "grad_norm": 0.14448532462120056, "learning_rate": 6.838235294117648e-06, "loss": 0.4544, "step": 93 }, { "epoch": 0.20750551876379691, "grad_norm": 0.17373333871364594, "learning_rate": 6.911764705882353e-06, "loss": 0.4557, "step": 94 }, { "epoch": 0.2097130242825607, "grad_norm": 0.15801453590393066, "learning_rate": 6.985294117647059e-06, "loss": 0.4507, "step": 95 }, { "epoch": 0.2119205298013245, "grad_norm": 0.15145470201969147, "learning_rate": 7.058823529411766e-06, "loss": 0.4499, "step": 96 }, { "epoch": 0.2141280353200883, "grad_norm": 0.14600904285907745, "learning_rate": 7.132352941176472e-06, "loss": 0.4548, "step": 97 }, { "epoch": 0.2163355408388521, "grad_norm": 0.15834525227546692, "learning_rate": 7.205882352941177e-06, "loss": 0.4505, "step": 98 }, { "epoch": 0.2185430463576159, "grad_norm": 0.15612734854221344, "learning_rate": 7.2794117647058826e-06, "loss": 0.4598, "step": 99 }, { "epoch": 0.22075055187637968, "grad_norm": 0.18165510892868042, "learning_rate": 7.352941176470589e-06, "loss": 0.4479, "step": 100 }, { "epoch": 0.2229580573951435, "grad_norm": 0.16820134222507477, "learning_rate": 7.426470588235295e-06, "loss": 0.4497, "step": 101 }, { "epoch": 0.2251655629139073, "grad_norm": 0.16453172266483307, "learning_rate": 7.500000000000001e-06, "loss": 0.4532, "step": 102 }, { "epoch": 0.22737306843267108, "grad_norm": 0.15813149511814117, "learning_rate": 7.573529411764706e-06, "loss": 0.4513, "step": 103 }, { "epoch": 0.22958057395143489, "grad_norm": 0.1652165949344635, "learning_rate": 7.647058823529411e-06, "loss": 0.445, "step": 104 }, { "epoch": 0.23178807947019867, "grad_norm": 0.14849768579006195, "learning_rate": 7.720588235294119e-06, "loss": 0.435, "step": 105 }, { "epoch": 0.23399558498896247, "grad_norm": 0.17478714883327484, "learning_rate": 7.794117647058825e-06, "loss": 0.4343, "step": 106 }, { "epoch": 0.23620309050772628, "grad_norm": 0.16727301478385925, "learning_rate": 7.86764705882353e-06, "loss": 0.4483, "step": 107 }, { "epoch": 0.23841059602649006, "grad_norm": 0.1674540638923645, "learning_rate": 7.941176470588236e-06, "loss": 0.4452, "step": 108 }, { "epoch": 0.24061810154525387, "grad_norm": 0.16588866710662842, "learning_rate": 8.014705882352942e-06, "loss": 0.4517, "step": 109 }, { "epoch": 0.24282560706401765, "grad_norm": 0.176283061504364, "learning_rate": 8.088235294117648e-06, "loss": 0.4434, "step": 110 }, { "epoch": 0.24503311258278146, "grad_norm": 0.17021676898002625, "learning_rate": 8.161764705882354e-06, "loss": 0.442, "step": 111 }, { "epoch": 0.24724061810154527, "grad_norm": 0.15938061475753784, "learning_rate": 8.23529411764706e-06, "loss": 0.4436, "step": 112 }, { "epoch": 0.24944812362030905, "grad_norm": 0.18990886211395264, "learning_rate": 8.308823529411766e-06, "loss": 0.439, "step": 113 }, { "epoch": 0.25165562913907286, "grad_norm": 0.16168883442878723, "learning_rate": 8.382352941176472e-06, "loss": 0.4392, "step": 114 }, { "epoch": 0.25386313465783666, "grad_norm": 0.2176610231399536, "learning_rate": 8.455882352941177e-06, "loss": 0.4488, "step": 115 }, { "epoch": 0.2560706401766004, "grad_norm": 0.19106176495552063, "learning_rate": 8.529411764705883e-06, "loss": 0.436, "step": 116 }, { "epoch": 0.2582781456953642, "grad_norm": 0.1480225920677185, "learning_rate": 8.60294117647059e-06, "loss": 0.4386, "step": 117 }, { "epoch": 0.26048565121412803, "grad_norm": 0.20528610050678253, "learning_rate": 8.676470588235295e-06, "loss": 0.4441, "step": 118 }, { "epoch": 0.26269315673289184, "grad_norm": 0.18629467487335205, "learning_rate": 8.750000000000001e-06, "loss": 0.4322, "step": 119 }, { "epoch": 0.26490066225165565, "grad_norm": 0.1764117330312729, "learning_rate": 8.823529411764707e-06, "loss": 0.4261, "step": 120 }, { "epoch": 0.2671081677704194, "grad_norm": 0.1975659281015396, "learning_rate": 8.897058823529413e-06, "loss": 0.4401, "step": 121 }, { "epoch": 0.2693156732891832, "grad_norm": 0.2136935442686081, "learning_rate": 8.970588235294119e-06, "loss": 0.4416, "step": 122 }, { "epoch": 0.271523178807947, "grad_norm": 0.18505676090717316, "learning_rate": 9.044117647058824e-06, "loss": 0.4423, "step": 123 }, { "epoch": 0.2737306843267108, "grad_norm": 0.176743283867836, "learning_rate": 9.11764705882353e-06, "loss": 0.4435, "step": 124 }, { "epoch": 0.27593818984547464, "grad_norm": 0.17542196810245514, "learning_rate": 9.191176470588236e-06, "loss": 0.4255, "step": 125 }, { "epoch": 0.2781456953642384, "grad_norm": 0.15908770263195038, "learning_rate": 9.264705882352942e-06, "loss": 0.438, "step": 126 }, { "epoch": 0.2803532008830022, "grad_norm": 0.16695120930671692, "learning_rate": 9.338235294117648e-06, "loss": 0.4418, "step": 127 }, { "epoch": 0.282560706401766, "grad_norm": 0.1714697927236557, "learning_rate": 9.411764705882354e-06, "loss": 0.4293, "step": 128 }, { "epoch": 0.2847682119205298, "grad_norm": 0.1601938009262085, "learning_rate": 9.48529411764706e-06, "loss": 0.4402, "step": 129 }, { "epoch": 0.2869757174392936, "grad_norm": 0.16425947844982147, "learning_rate": 9.558823529411766e-06, "loss": 0.4328, "step": 130 }, { "epoch": 0.2891832229580574, "grad_norm": 0.1931643933057785, "learning_rate": 9.632352941176471e-06, "loss": 0.4307, "step": 131 }, { "epoch": 0.2913907284768212, "grad_norm": 0.18675902485847473, "learning_rate": 9.705882352941177e-06, "loss": 0.4554, "step": 132 }, { "epoch": 0.293598233995585, "grad_norm": 0.16391406953334808, "learning_rate": 9.779411764705883e-06, "loss": 0.4272, "step": 133 }, { "epoch": 0.2958057395143488, "grad_norm": 0.20189572870731354, "learning_rate": 9.852941176470589e-06, "loss": 0.4258, "step": 134 }, { "epoch": 0.2980132450331126, "grad_norm": 0.16828037798404694, "learning_rate": 9.926470588235295e-06, "loss": 0.4287, "step": 135 }, { "epoch": 0.30022075055187636, "grad_norm": 0.1638776957988739, "learning_rate": 1e-05, "loss": 0.4326, "step": 136 }, { "epoch": 0.30242825607064017, "grad_norm": 0.20775483548641205, "learning_rate": 9.999983503697906e-06, "loss": 0.4274, "step": 137 }, { "epoch": 0.304635761589404, "grad_norm": 0.17668417096138, "learning_rate": 9.999934014900475e-06, "loss": 0.4277, "step": 138 }, { "epoch": 0.3068432671081678, "grad_norm": 0.20949822664260864, "learning_rate": 9.999851533934259e-06, "loss": 0.4277, "step": 139 }, { "epoch": 0.3090507726269316, "grad_norm": 0.13797180354595184, "learning_rate": 9.999736061343512e-06, "loss": 0.4072, "step": 140 }, { "epoch": 0.31125827814569534, "grad_norm": 0.20374132692813873, "learning_rate": 9.99958759789018e-06, "loss": 0.4302, "step": 141 }, { "epoch": 0.31346578366445915, "grad_norm": 0.1946762204170227, "learning_rate": 9.999406144553905e-06, "loss": 0.4213, "step": 142 }, { "epoch": 0.31567328918322296, "grad_norm": 0.21811267733573914, "learning_rate": 9.999191702532008e-06, "loss": 0.4285, "step": 143 }, { "epoch": 0.31788079470198677, "grad_norm": 0.20533326268196106, "learning_rate": 9.99894427323949e-06, "loss": 0.4251, "step": 144 }, { "epoch": 0.3200883002207506, "grad_norm": 0.20235486328601837, "learning_rate": 9.99866385830902e-06, "loss": 0.4237, "step": 145 }, { "epoch": 0.32229580573951433, "grad_norm": 0.20962080359458923, "learning_rate": 9.99835045959092e-06, "loss": 0.4266, "step": 146 }, { "epoch": 0.32450331125827814, "grad_norm": 0.22475510835647583, "learning_rate": 9.998004079153156e-06, "loss": 0.4263, "step": 147 }, { "epoch": 0.32671081677704195, "grad_norm": 0.20724737644195557, "learning_rate": 9.997624719281332e-06, "loss": 0.416, "step": 148 }, { "epoch": 0.32891832229580575, "grad_norm": 0.1801760494709015, "learning_rate": 9.997212382478658e-06, "loss": 0.4233, "step": 149 }, { "epoch": 0.33112582781456956, "grad_norm": 0.23205707967281342, "learning_rate": 9.996767071465947e-06, "loss": 0.4277, "step": 150 }, { "epoch": 0.3333333333333333, "grad_norm": 0.18537338078022003, "learning_rate": 9.996288789181595e-06, "loss": 0.4317, "step": 151 }, { "epoch": 0.3355408388520971, "grad_norm": 0.17498917877674103, "learning_rate": 9.995777538781556e-06, "loss": 0.4288, "step": 152 }, { "epoch": 0.33774834437086093, "grad_norm": 0.18248897790908813, "learning_rate": 9.995233323639326e-06, "loss": 0.4261, "step": 153 }, { "epoch": 0.33995584988962474, "grad_norm": 0.1849048137664795, "learning_rate": 9.994656147345922e-06, "loss": 0.4216, "step": 154 }, { "epoch": 0.34216335540838855, "grad_norm": 0.19169744849205017, "learning_rate": 9.994046013709852e-06, "loss": 0.423, "step": 155 }, { "epoch": 0.3443708609271523, "grad_norm": 0.18524128198623657, "learning_rate": 9.993402926757098e-06, "loss": 0.4213, "step": 156 }, { "epoch": 0.3465783664459161, "grad_norm": 0.16218866407871246, "learning_rate": 9.99272689073108e-06, "loss": 0.4252, "step": 157 }, { "epoch": 0.3487858719646799, "grad_norm": 0.17451681196689606, "learning_rate": 9.992017910092636e-06, "loss": 0.4251, "step": 158 }, { "epoch": 0.3509933774834437, "grad_norm": 0.15944437682628632, "learning_rate": 9.991275989519991e-06, "loss": 0.4123, "step": 159 }, { "epoch": 0.35320088300220753, "grad_norm": 0.17371642589569092, "learning_rate": 9.990501133908722e-06, "loss": 0.4234, "step": 160 }, { "epoch": 0.3554083885209713, "grad_norm": 0.1783660650253296, "learning_rate": 9.98969334837173e-06, "loss": 0.4234, "step": 161 }, { "epoch": 0.3576158940397351, "grad_norm": 0.1762082278728485, "learning_rate": 9.988852638239206e-06, "loss": 0.418, "step": 162 }, { "epoch": 0.3598233995584989, "grad_norm": 0.16491912305355072, "learning_rate": 9.987979009058593e-06, "loss": 0.4248, "step": 163 }, { "epoch": 0.3620309050772627, "grad_norm": 0.2407284379005432, "learning_rate": 9.98707246659455e-06, "loss": 0.4307, "step": 164 }, { "epoch": 0.36423841059602646, "grad_norm": 0.19042451679706573, "learning_rate": 9.986133016828916e-06, "loss": 0.4231, "step": 165 }, { "epoch": 0.36644591611479027, "grad_norm": 0.18169504404067993, "learning_rate": 9.985160665960672e-06, "loss": 0.4266, "step": 166 }, { "epoch": 0.3686534216335541, "grad_norm": 0.18646620213985443, "learning_rate": 9.984155420405895e-06, "loss": 0.4231, "step": 167 }, { "epoch": 0.3708609271523179, "grad_norm": 0.19154079258441925, "learning_rate": 9.983117286797718e-06, "loss": 0.4308, "step": 168 }, { "epoch": 0.3730684326710817, "grad_norm": 0.17594484984874725, "learning_rate": 9.982046271986287e-06, "loss": 0.4115, "step": 169 }, { "epoch": 0.37527593818984545, "grad_norm": 0.18167531490325928, "learning_rate": 9.980942383038717e-06, "loss": 0.424, "step": 170 }, { "epoch": 0.37748344370860926, "grad_norm": 0.1535561978816986, "learning_rate": 9.97980562723904e-06, "loss": 0.4296, "step": 171 }, { "epoch": 0.37969094922737306, "grad_norm": 0.15756377577781677, "learning_rate": 9.978636012088165e-06, "loss": 0.4169, "step": 172 }, { "epoch": 0.3818984547461369, "grad_norm": 0.15670788288116455, "learning_rate": 9.97743354530382e-06, "loss": 0.4394, "step": 173 }, { "epoch": 0.3841059602649007, "grad_norm": 0.16224409639835358, "learning_rate": 9.976198234820509e-06, "loss": 0.4228, "step": 174 }, { "epoch": 0.38631346578366443, "grad_norm": 0.14743737876415253, "learning_rate": 9.974930088789452e-06, "loss": 0.4144, "step": 175 }, { "epoch": 0.38852097130242824, "grad_norm": 0.1594422310590744, "learning_rate": 9.97362911557854e-06, "loss": 0.4168, "step": 176 }, { "epoch": 0.39072847682119205, "grad_norm": 0.16173714399337769, "learning_rate": 9.972295323772268e-06, "loss": 0.4166, "step": 177 }, { "epoch": 0.39293598233995586, "grad_norm": 0.1668204814195633, "learning_rate": 9.970928722171691e-06, "loss": 0.4252, "step": 178 }, { "epoch": 0.39514348785871967, "grad_norm": 0.15836164355278015, "learning_rate": 9.96952931979436e-06, "loss": 0.4209, "step": 179 }, { "epoch": 0.3973509933774834, "grad_norm": 0.1634080708026886, "learning_rate": 9.968097125874258e-06, "loss": 0.4076, "step": 180 }, { "epoch": 0.3995584988962472, "grad_norm": 0.16134855151176453, "learning_rate": 9.966632149861748e-06, "loss": 0.4276, "step": 181 }, { "epoch": 0.40176600441501104, "grad_norm": 0.15993578732013702, "learning_rate": 9.965134401423503e-06, "loss": 0.4308, "step": 182 }, { "epoch": 0.40397350993377484, "grad_norm": 0.19045297801494598, "learning_rate": 9.963603890442448e-06, "loss": 0.4185, "step": 183 }, { "epoch": 0.40618101545253865, "grad_norm": 0.16455209255218506, "learning_rate": 9.962040627017693e-06, "loss": 0.4232, "step": 184 }, { "epoch": 0.4083885209713024, "grad_norm": 0.16577620804309845, "learning_rate": 9.960444621464462e-06, "loss": 0.4149, "step": 185 }, { "epoch": 0.4105960264900662, "grad_norm": 0.25177431106567383, "learning_rate": 9.958815884314033e-06, "loss": 0.4172, "step": 186 }, { "epoch": 0.41280353200883, "grad_norm": 0.18712477385997772, "learning_rate": 9.957154426313662e-06, "loss": 0.4205, "step": 187 }, { "epoch": 0.41501103752759383, "grad_norm": 0.1667563021183014, "learning_rate": 9.955460258426512e-06, "loss": 0.4207, "step": 188 }, { "epoch": 0.41721854304635764, "grad_norm": 0.22865413129329681, "learning_rate": 9.953733391831586e-06, "loss": 0.4109, "step": 189 }, { "epoch": 0.4194260485651214, "grad_norm": 0.18536990880966187, "learning_rate": 9.951973837923652e-06, "loss": 0.4187, "step": 190 }, { "epoch": 0.4216335540838852, "grad_norm": 0.19504587352275848, "learning_rate": 9.950181608313158e-06, "loss": 0.4142, "step": 191 }, { "epoch": 0.423841059602649, "grad_norm": 0.1892482340335846, "learning_rate": 9.948356714826172e-06, "loss": 0.4142, "step": 192 }, { "epoch": 0.4260485651214128, "grad_norm": 0.1839127093553543, "learning_rate": 9.946499169504294e-06, "loss": 0.4161, "step": 193 }, { "epoch": 0.4282560706401766, "grad_norm": 0.20385828614234924, "learning_rate": 9.944608984604569e-06, "loss": 0.4124, "step": 194 }, { "epoch": 0.4304635761589404, "grad_norm": 0.1948205530643463, "learning_rate": 9.942686172599425e-06, "loss": 0.4251, "step": 195 }, { "epoch": 0.4326710816777042, "grad_norm": 0.19438982009887695, "learning_rate": 9.940730746176578e-06, "loss": 0.4158, "step": 196 }, { "epoch": 0.434878587196468, "grad_norm": 0.17213338613510132, "learning_rate": 9.93874271823895e-06, "loss": 0.4175, "step": 197 }, { "epoch": 0.4370860927152318, "grad_norm": 0.22870118916034698, "learning_rate": 9.936722101904582e-06, "loss": 0.4267, "step": 198 }, { "epoch": 0.4392935982339956, "grad_norm": 0.20383016765117645, "learning_rate": 9.934668910506555e-06, "loss": 0.422, "step": 199 }, { "epoch": 0.44150110375275936, "grad_norm": 0.16936808824539185, "learning_rate": 9.932583157592896e-06, "loss": 0.4144, "step": 200 }, { "epoch": 0.44370860927152317, "grad_norm": 0.19149592518806458, "learning_rate": 9.930464856926488e-06, "loss": 0.4077, "step": 201 }, { "epoch": 0.445916114790287, "grad_norm": 0.15204112231731415, "learning_rate": 9.928314022484982e-06, "loss": 0.4207, "step": 202 }, { "epoch": 0.4481236203090508, "grad_norm": 0.19697798788547516, "learning_rate": 9.926130668460702e-06, "loss": 0.4159, "step": 203 }, { "epoch": 0.4503311258278146, "grad_norm": 0.1751161813735962, "learning_rate": 9.92391480926056e-06, "loss": 0.4179, "step": 204 }, { "epoch": 0.45253863134657835, "grad_norm": 0.17122775316238403, "learning_rate": 9.921666459505944e-06, "loss": 0.4104, "step": 205 }, { "epoch": 0.45474613686534215, "grad_norm": 0.18898116052150726, "learning_rate": 9.91938563403264e-06, "loss": 0.4183, "step": 206 }, { "epoch": 0.45695364238410596, "grad_norm": 0.17628274857997894, "learning_rate": 9.917072347890721e-06, "loss": 0.411, "step": 207 }, { "epoch": 0.45916114790286977, "grad_norm": 0.19427300989627838, "learning_rate": 9.914726616344454e-06, "loss": 0.4144, "step": 208 }, { "epoch": 0.4613686534216336, "grad_norm": 0.19164274632930756, "learning_rate": 9.912348454872196e-06, "loss": 0.4067, "step": 209 }, { "epoch": 0.46357615894039733, "grad_norm": 0.17385149002075195, "learning_rate": 9.909937879166298e-06, "loss": 0.408, "step": 210 }, { "epoch": 0.46578366445916114, "grad_norm": 0.18648236989974976, "learning_rate": 9.907494905132994e-06, "loss": 0.4273, "step": 211 }, { "epoch": 0.46799116997792495, "grad_norm": 0.1911754459142685, "learning_rate": 9.905019548892296e-06, "loss": 0.4127, "step": 212 }, { "epoch": 0.47019867549668876, "grad_norm": 0.15749873220920563, "learning_rate": 9.902511826777895e-06, "loss": 0.4174, "step": 213 }, { "epoch": 0.47240618101545256, "grad_norm": 0.19831134378910065, "learning_rate": 9.899971755337049e-06, "loss": 0.4045, "step": 214 }, { "epoch": 0.4746136865342163, "grad_norm": 0.18536338210105896, "learning_rate": 9.897399351330471e-06, "loss": 0.4134, "step": 215 }, { "epoch": 0.4768211920529801, "grad_norm": 0.17010532319545746, "learning_rate": 9.894794631732223e-06, "loss": 0.4076, "step": 216 }, { "epoch": 0.47902869757174393, "grad_norm": 0.17602422833442688, "learning_rate": 9.8921576137296e-06, "loss": 0.4079, "step": 217 }, { "epoch": 0.48123620309050774, "grad_norm": 0.1983213573694229, "learning_rate": 9.889488314723024e-06, "loss": 0.413, "step": 218 }, { "epoch": 0.48344370860927155, "grad_norm": 0.17187613248825073, "learning_rate": 9.886786752325917e-06, "loss": 0.4229, "step": 219 }, { "epoch": 0.4856512141280353, "grad_norm": 0.2029353827238083, "learning_rate": 9.884052944364595e-06, "loss": 0.4242, "step": 220 }, { "epoch": 0.4878587196467991, "grad_norm": 0.18107262253761292, "learning_rate": 9.881286908878148e-06, "loss": 0.4157, "step": 221 }, { "epoch": 0.4900662251655629, "grad_norm": 0.18853497505187988, "learning_rate": 9.878488664118316e-06, "loss": 0.4153, "step": 222 }, { "epoch": 0.4922737306843267, "grad_norm": 0.1775682419538498, "learning_rate": 9.875658228549379e-06, "loss": 0.4109, "step": 223 }, { "epoch": 0.49448123620309054, "grad_norm": 0.18237102031707764, "learning_rate": 9.872795620848024e-06, "loss": 0.407, "step": 224 }, { "epoch": 0.4966887417218543, "grad_norm": 0.20489446818828583, "learning_rate": 9.869900859903225e-06, "loss": 0.4198, "step": 225 }, { "epoch": 0.4988962472406181, "grad_norm": 0.1481681913137436, "learning_rate": 9.866973964816126e-06, "loss": 0.4085, "step": 226 }, { "epoch": 0.5011037527593819, "grad_norm": 0.18315470218658447, "learning_rate": 9.864014954899905e-06, "loss": 0.4079, "step": 227 }, { "epoch": 0.5033112582781457, "grad_norm": 0.18866921961307526, "learning_rate": 9.861023849679648e-06, "loss": 0.4249, "step": 228 }, { "epoch": 0.5055187637969095, "grad_norm": 0.15121376514434814, "learning_rate": 9.858000668892226e-06, "loss": 0.4179, "step": 229 }, { "epoch": 0.5077262693156733, "grad_norm": 0.19818085432052612, "learning_rate": 9.85494543248616e-06, "loss": 0.4218, "step": 230 }, { "epoch": 0.5099337748344371, "grad_norm": 0.15964657068252563, "learning_rate": 9.851858160621496e-06, "loss": 0.4074, "step": 231 }, { "epoch": 0.5121412803532008, "grad_norm": 0.18371723592281342, "learning_rate": 9.848738873669653e-06, "loss": 0.4103, "step": 232 }, { "epoch": 0.5143487858719646, "grad_norm": 0.1854197084903717, "learning_rate": 9.845587592213318e-06, "loss": 0.4092, "step": 233 }, { "epoch": 0.5165562913907285, "grad_norm": 0.15952259302139282, "learning_rate": 9.842404337046284e-06, "loss": 0.4139, "step": 234 }, { "epoch": 0.5187637969094923, "grad_norm": 0.1830589473247528, "learning_rate": 9.839189129173328e-06, "loss": 0.4143, "step": 235 }, { "epoch": 0.5209713024282561, "grad_norm": 0.1684870719909668, "learning_rate": 9.835941989810065e-06, "loss": 0.4088, "step": 236 }, { "epoch": 0.5231788079470199, "grad_norm": 0.17676231265068054, "learning_rate": 9.832662940382813e-06, "loss": 0.4123, "step": 237 }, { "epoch": 0.5253863134657837, "grad_norm": 0.17995555698871613, "learning_rate": 9.829352002528449e-06, "loss": 0.4086, "step": 238 }, { "epoch": 0.5275938189845475, "grad_norm": 0.1694101095199585, "learning_rate": 9.826009198094262e-06, "loss": 0.4137, "step": 239 }, { "epoch": 0.5298013245033113, "grad_norm": 0.16819199919700623, "learning_rate": 9.822634549137819e-06, "loss": 0.4072, "step": 240 }, { "epoch": 0.5320088300220751, "grad_norm": 0.17408284544944763, "learning_rate": 9.81922807792681e-06, "loss": 0.4052, "step": 241 }, { "epoch": 0.5342163355408388, "grad_norm": 0.1843319982290268, "learning_rate": 9.815789806938909e-06, "loss": 0.3966, "step": 242 }, { "epoch": 0.5364238410596026, "grad_norm": 0.18125468492507935, "learning_rate": 9.812319758861616e-06, "loss": 0.4044, "step": 243 }, { "epoch": 0.5386313465783664, "grad_norm": 0.16861899197101593, "learning_rate": 9.808817956592115e-06, "loss": 0.4092, "step": 244 }, { "epoch": 0.5408388520971302, "grad_norm": 0.22598163783550262, "learning_rate": 9.805284423237126e-06, "loss": 0.4137, "step": 245 }, { "epoch": 0.543046357615894, "grad_norm": 0.22511421144008636, "learning_rate": 9.801719182112738e-06, "loss": 0.4073, "step": 246 }, { "epoch": 0.5452538631346578, "grad_norm": 0.1690152883529663, "learning_rate": 9.798122256744269e-06, "loss": 0.4097, "step": 247 }, { "epoch": 0.5474613686534217, "grad_norm": 0.22893239557743073, "learning_rate": 9.794493670866108e-06, "loss": 0.4054, "step": 248 }, { "epoch": 0.5496688741721855, "grad_norm": 0.17257164418697357, "learning_rate": 9.790833448421554e-06, "loss": 0.4109, "step": 249 }, { "epoch": 0.5518763796909493, "grad_norm": 0.1636303812265396, "learning_rate": 9.787141613562661e-06, "loss": 0.3995, "step": 250 }, { "epoch": 0.5540838852097131, "grad_norm": 0.17622527480125427, "learning_rate": 9.783418190650079e-06, "loss": 0.409, "step": 251 }, { "epoch": 0.5562913907284768, "grad_norm": 0.20090465247631073, "learning_rate": 9.779663204252887e-06, "loss": 0.4015, "step": 252 }, { "epoch": 0.5584988962472406, "grad_norm": 0.15368807315826416, "learning_rate": 9.775876679148449e-06, "loss": 0.4009, "step": 253 }, { "epoch": 0.5607064017660044, "grad_norm": 0.2135976403951645, "learning_rate": 9.772058640322221e-06, "loss": 0.407, "step": 254 }, { "epoch": 0.5629139072847682, "grad_norm": 0.17839573323726654, "learning_rate": 9.768209112967619e-06, "loss": 0.404, "step": 255 }, { "epoch": 0.565121412803532, "grad_norm": 0.1953095942735672, "learning_rate": 9.764328122485827e-06, "loss": 0.402, "step": 256 }, { "epoch": 0.5673289183222958, "grad_norm": 0.16529282927513123, "learning_rate": 9.76041569448564e-06, "loss": 0.4117, "step": 257 }, { "epoch": 0.5695364238410596, "grad_norm": 0.17458973824977875, "learning_rate": 9.756471854783297e-06, "loss": 0.4144, "step": 258 }, { "epoch": 0.5717439293598234, "grad_norm": 0.1776653230190277, "learning_rate": 9.752496629402307e-06, "loss": 0.4123, "step": 259 }, { "epoch": 0.5739514348785872, "grad_norm": 0.17428044974803925, "learning_rate": 9.748490044573275e-06, "loss": 0.3975, "step": 260 }, { "epoch": 0.5761589403973509, "grad_norm": 0.17150405049324036, "learning_rate": 9.744452126733739e-06, "loss": 0.4023, "step": 261 }, { "epoch": 0.5783664459161147, "grad_norm": 0.1668493002653122, "learning_rate": 9.740382902527981e-06, "loss": 0.409, "step": 262 }, { "epoch": 0.5805739514348786, "grad_norm": 0.17727597057819366, "learning_rate": 9.736282398806862e-06, "loss": 0.4061, "step": 263 }, { "epoch": 0.5827814569536424, "grad_norm": 0.15944141149520874, "learning_rate": 9.73215064262764e-06, "loss": 0.402, "step": 264 }, { "epoch": 0.5849889624724062, "grad_norm": 0.17232947051525116, "learning_rate": 9.727987661253796e-06, "loss": 0.4102, "step": 265 }, { "epoch": 0.58719646799117, "grad_norm": 0.2054961770772934, "learning_rate": 9.72379348215485e-06, "loss": 0.3994, "step": 266 }, { "epoch": 0.5894039735099338, "grad_norm": 0.1817290335893631, "learning_rate": 9.719568133006177e-06, "loss": 0.4033, "step": 267 }, { "epoch": 0.5916114790286976, "grad_norm": 0.16904965043067932, "learning_rate": 9.715311641688835e-06, "loss": 0.4082, "step": 268 }, { "epoch": 0.5938189845474614, "grad_norm": 0.17628441751003265, "learning_rate": 9.71102403628937e-06, "loss": 0.3969, "step": 269 }, { "epoch": 0.5960264900662252, "grad_norm": 0.20128114521503448, "learning_rate": 9.706705345099632e-06, "loss": 0.4081, "step": 270 }, { "epoch": 0.5982339955849889, "grad_norm": 0.17412817478179932, "learning_rate": 9.7023555966166e-06, "loss": 0.4076, "step": 271 }, { "epoch": 0.6004415011037527, "grad_norm": 0.18813695013523102, "learning_rate": 9.697974819542178e-06, "loss": 0.4042, "step": 272 }, { "epoch": 0.6026490066225165, "grad_norm": 0.17176660895347595, "learning_rate": 9.693563042783011e-06, "loss": 0.4099, "step": 273 }, { "epoch": 0.6048565121412803, "grad_norm": 0.17056933045387268, "learning_rate": 9.689120295450308e-06, "loss": 0.4045, "step": 274 }, { "epoch": 0.6070640176600441, "grad_norm": 0.20970946550369263, "learning_rate": 9.684646606859621e-06, "loss": 0.3944, "step": 275 }, { "epoch": 0.609271523178808, "grad_norm": 0.19319604337215424, "learning_rate": 9.680142006530684e-06, "loss": 0.4062, "step": 276 }, { "epoch": 0.6114790286975718, "grad_norm": 0.17683015763759613, "learning_rate": 9.675606524187192e-06, "loss": 0.4077, "step": 277 }, { "epoch": 0.6136865342163356, "grad_norm": 0.16389036178588867, "learning_rate": 9.671040189756623e-06, "loss": 0.412, "step": 278 }, { "epoch": 0.6158940397350994, "grad_norm": 0.15970537066459656, "learning_rate": 9.666443033370026e-06, "loss": 0.4068, "step": 279 }, { "epoch": 0.6181015452538632, "grad_norm": 0.17420579493045807, "learning_rate": 9.661815085361836e-06, "loss": 0.4046, "step": 280 }, { "epoch": 0.6203090507726269, "grad_norm": 0.16548947989940643, "learning_rate": 9.657156376269665e-06, "loss": 0.4122, "step": 281 }, { "epoch": 0.6225165562913907, "grad_norm": 0.147056445479393, "learning_rate": 9.652466936834101e-06, "loss": 0.4085, "step": 282 }, { "epoch": 0.6247240618101545, "grad_norm": 0.17280390858650208, "learning_rate": 9.647746797998508e-06, "loss": 0.4142, "step": 283 }, { "epoch": 0.6269315673289183, "grad_norm": 0.16386403143405914, "learning_rate": 9.642995990908817e-06, "loss": 0.4103, "step": 284 }, { "epoch": 0.6291390728476821, "grad_norm": 0.16252438724040985, "learning_rate": 9.638214546913333e-06, "loss": 0.4023, "step": 285 }, { "epoch": 0.6313465783664459, "grad_norm": 0.16047868132591248, "learning_rate": 9.633402497562512e-06, "loss": 0.4032, "step": 286 }, { "epoch": 0.6335540838852097, "grad_norm": 0.18797720968723297, "learning_rate": 9.628559874608761e-06, "loss": 0.4014, "step": 287 }, { "epoch": 0.6357615894039735, "grad_norm": 0.15865157544612885, "learning_rate": 9.62368671000623e-06, "loss": 0.4022, "step": 288 }, { "epoch": 0.6379690949227373, "grad_norm": 0.15757879614830017, "learning_rate": 9.618783035910596e-06, "loss": 0.4019, "step": 289 }, { "epoch": 0.6401766004415012, "grad_norm": 0.15986581146717072, "learning_rate": 9.613848884678851e-06, "loss": 0.3981, "step": 290 }, { "epoch": 0.6423841059602649, "grad_norm": 0.1824173629283905, "learning_rate": 9.608884288869103e-06, "loss": 0.407, "step": 291 }, { "epoch": 0.6445916114790287, "grad_norm": 0.185842826962471, "learning_rate": 9.603889281240334e-06, "loss": 0.4016, "step": 292 }, { "epoch": 0.6467991169977925, "grad_norm": 0.17280805110931396, "learning_rate": 9.59886389475221e-06, "loss": 0.4081, "step": 293 }, { "epoch": 0.6490066225165563, "grad_norm": 0.1815565824508667, "learning_rate": 9.593808162564845e-06, "loss": 0.4141, "step": 294 }, { "epoch": 0.6512141280353201, "grad_norm": 0.19602055847644806, "learning_rate": 9.588722118038595e-06, "loss": 0.4048, "step": 295 }, { "epoch": 0.6534216335540839, "grad_norm": 0.1916995495557785, "learning_rate": 9.583605794733833e-06, "loss": 0.3953, "step": 296 }, { "epoch": 0.6556291390728477, "grad_norm": 0.17578016221523285, "learning_rate": 9.578459226410722e-06, "loss": 0.4133, "step": 297 }, { "epoch": 0.6578366445916115, "grad_norm": 0.18694248795509338, "learning_rate": 9.573282447029e-06, "loss": 0.4023, "step": 298 }, { "epoch": 0.6600441501103753, "grad_norm": 0.20029006898403168, "learning_rate": 9.568075490747756e-06, "loss": 0.3967, "step": 299 }, { "epoch": 0.6622516556291391, "grad_norm": 0.15852802991867065, "learning_rate": 9.562838391925197e-06, "loss": 0.4053, "step": 300 }, { "epoch": 0.6644591611479028, "grad_norm": 0.18394002318382263, "learning_rate": 9.557571185118431e-06, "loss": 0.4001, "step": 301 }, { "epoch": 0.6666666666666666, "grad_norm": 0.1864984780550003, "learning_rate": 9.55227390508323e-06, "loss": 0.3964, "step": 302 }, { "epoch": 0.6688741721854304, "grad_norm": 0.15841950476169586, "learning_rate": 9.546946586773808e-06, "loss": 0.4045, "step": 303 }, { "epoch": 0.6710816777041942, "grad_norm": 0.19293592870235443, "learning_rate": 9.541589265342585e-06, "loss": 0.405, "step": 304 }, { "epoch": 0.673289183222958, "grad_norm": 0.16064338386058807, "learning_rate": 9.536201976139958e-06, "loss": 0.4098, "step": 305 }, { "epoch": 0.6754966887417219, "grad_norm": 0.16002054512500763, "learning_rate": 9.530784754714069e-06, "loss": 0.3837, "step": 306 }, { "epoch": 0.6777041942604857, "grad_norm": 0.16032235324382782, "learning_rate": 9.525337636810564e-06, "loss": 0.4095, "step": 307 }, { "epoch": 0.6799116997792495, "grad_norm": 0.19734519720077515, "learning_rate": 9.519860658372364e-06, "loss": 0.4049, "step": 308 }, { "epoch": 0.6821192052980133, "grad_norm": 0.1894775629043579, "learning_rate": 9.514353855539428e-06, "loss": 0.3926, "step": 309 }, { "epoch": 0.6843267108167771, "grad_norm": 0.2021895796060562, "learning_rate": 9.508817264648506e-06, "loss": 0.3983, "step": 310 }, { "epoch": 0.6865342163355408, "grad_norm": 0.20270881056785583, "learning_rate": 9.503250922232911e-06, "loss": 0.4105, "step": 311 }, { "epoch": 0.6887417218543046, "grad_norm": 0.17310801148414612, "learning_rate": 9.497654865022268e-06, "loss": 0.4026, "step": 312 }, { "epoch": 0.6909492273730684, "grad_norm": 0.22883597016334534, "learning_rate": 9.492029129942277e-06, "loss": 0.4062, "step": 313 }, { "epoch": 0.6931567328918322, "grad_norm": 0.1955188363790512, "learning_rate": 9.48637375411447e-06, "loss": 0.4044, "step": 314 }, { "epoch": 0.695364238410596, "grad_norm": 0.18373191356658936, "learning_rate": 9.48068877485596e-06, "loss": 0.4032, "step": 315 }, { "epoch": 0.6975717439293598, "grad_norm": 0.21621759235858917, "learning_rate": 9.474974229679201e-06, "loss": 0.3904, "step": 316 }, { "epoch": 0.6997792494481236, "grad_norm": 0.17569488286972046, "learning_rate": 9.469230156291742e-06, "loss": 0.4087, "step": 317 }, { "epoch": 0.7019867549668874, "grad_norm": 0.2064937800168991, "learning_rate": 9.463456592595966e-06, "loss": 0.396, "step": 318 }, { "epoch": 0.7041942604856513, "grad_norm": 0.24051177501678467, "learning_rate": 9.457653576688857e-06, "loss": 0.4062, "step": 319 }, { "epoch": 0.7064017660044151, "grad_norm": 0.16446927189826965, "learning_rate": 9.451821146861734e-06, "loss": 0.3987, "step": 320 }, { "epoch": 0.7086092715231788, "grad_norm": 0.20139139890670776, "learning_rate": 9.445959341600009e-06, "loss": 0.4036, "step": 321 }, { "epoch": 0.7108167770419426, "grad_norm": 0.18157215416431427, "learning_rate": 9.440068199582923e-06, "loss": 0.4124, "step": 322 }, { "epoch": 0.7130242825607064, "grad_norm": 0.15641391277313232, "learning_rate": 9.434147759683303e-06, "loss": 0.405, "step": 323 }, { "epoch": 0.7152317880794702, "grad_norm": 0.16391552984714508, "learning_rate": 9.428198060967294e-06, "loss": 0.4005, "step": 324 }, { "epoch": 0.717439293598234, "grad_norm": 0.179102823138237, "learning_rate": 9.422219142694104e-06, "loss": 0.3968, "step": 325 }, { "epoch": 0.7196467991169978, "grad_norm": 0.17109854519367218, "learning_rate": 9.416211044315754e-06, "loss": 0.4049, "step": 326 }, { "epoch": 0.7218543046357616, "grad_norm": 0.20250354707241058, "learning_rate": 9.410173805476804e-06, "loss": 0.4186, "step": 327 }, { "epoch": 0.7240618101545254, "grad_norm": 0.18576852977275848, "learning_rate": 9.404107466014101e-06, "loss": 0.4022, "step": 328 }, { "epoch": 0.7262693156732892, "grad_norm": 0.16930828988552094, "learning_rate": 9.398012065956512e-06, "loss": 0.3949, "step": 329 }, { "epoch": 0.7284768211920529, "grad_norm": 0.1966543048620224, "learning_rate": 9.39188764552466e-06, "loss": 0.4103, "step": 330 }, { "epoch": 0.7306843267108167, "grad_norm": 0.18573778867721558, "learning_rate": 9.385734245130664e-06, "loss": 0.4069, "step": 331 }, { "epoch": 0.7328918322295805, "grad_norm": 0.16225308179855347, "learning_rate": 9.379551905377863e-06, "loss": 0.4049, "step": 332 }, { "epoch": 0.7350993377483444, "grad_norm": 0.1815934032201767, "learning_rate": 9.373340667060553e-06, "loss": 0.3927, "step": 333 }, { "epoch": 0.7373068432671082, "grad_norm": 0.1688205897808075, "learning_rate": 9.367100571163722e-06, "loss": 0.4019, "step": 334 }, { "epoch": 0.739514348785872, "grad_norm": 0.18315370380878448, "learning_rate": 9.360831658862774e-06, "loss": 0.3989, "step": 335 }, { "epoch": 0.7417218543046358, "grad_norm": 0.17102378606796265, "learning_rate": 9.354533971523253e-06, "loss": 0.4122, "step": 336 }, { "epoch": 0.7439293598233996, "grad_norm": 0.17149822413921356, "learning_rate": 9.348207550700584e-06, "loss": 0.4087, "step": 337 }, { "epoch": 0.7461368653421634, "grad_norm": 0.19260001182556152, "learning_rate": 9.341852438139784e-06, "loss": 0.4064, "step": 338 }, { "epoch": 0.7483443708609272, "grad_norm": 0.18973691761493683, "learning_rate": 9.335468675775196e-06, "loss": 0.399, "step": 339 }, { "epoch": 0.7505518763796909, "grad_norm": 0.19365577399730682, "learning_rate": 9.329056305730211e-06, "loss": 0.3959, "step": 340 }, { "epoch": 0.7527593818984547, "grad_norm": 0.21033021807670593, "learning_rate": 9.322615370316986e-06, "loss": 0.4027, "step": 341 }, { "epoch": 0.7549668874172185, "grad_norm": 0.1825421005487442, "learning_rate": 9.316145912036165e-06, "loss": 0.3971, "step": 342 }, { "epoch": 0.7571743929359823, "grad_norm": 0.1825910061597824, "learning_rate": 9.309647973576605e-06, "loss": 0.3959, "step": 343 }, { "epoch": 0.7593818984547461, "grad_norm": 0.1806352436542511, "learning_rate": 9.30312159781509e-06, "loss": 0.4047, "step": 344 }, { "epoch": 0.7615894039735099, "grad_norm": 0.15455348789691925, "learning_rate": 9.296566827816044e-06, "loss": 0.3958, "step": 345 }, { "epoch": 0.7637969094922737, "grad_norm": 0.19140276312828064, "learning_rate": 9.289983706831254e-06, "loss": 0.3955, "step": 346 }, { "epoch": 0.7660044150110376, "grad_norm": 0.1777879148721695, "learning_rate": 9.28337227829958e-06, "loss": 0.3971, "step": 347 }, { "epoch": 0.7682119205298014, "grad_norm": 0.16399559378623962, "learning_rate": 9.276732585846673e-06, "loss": 0.3998, "step": 348 }, { "epoch": 0.7704194260485652, "grad_norm": 0.1594182848930359, "learning_rate": 9.270064673284681e-06, "loss": 0.3898, "step": 349 }, { "epoch": 0.7726269315673289, "grad_norm": 0.16792874038219452, "learning_rate": 9.263368584611965e-06, "loss": 0.4021, "step": 350 }, { "epoch": 0.7748344370860927, "grad_norm": 0.18502795696258545, "learning_rate": 9.256644364012803e-06, "loss": 0.3987, "step": 351 }, { "epoch": 0.7770419426048565, "grad_norm": 0.18240559101104736, "learning_rate": 9.249892055857107e-06, "loss": 0.4074, "step": 352 }, { "epoch": 0.7792494481236203, "grad_norm": 0.15263855457305908, "learning_rate": 9.243111704700126e-06, "loss": 0.3928, "step": 353 }, { "epoch": 0.7814569536423841, "grad_norm": 0.18149858713150024, "learning_rate": 9.236303355282142e-06, "loss": 0.4015, "step": 354 }, { "epoch": 0.7836644591611479, "grad_norm": 0.1617615818977356, "learning_rate": 9.229467052528191e-06, "loss": 0.4024, "step": 355 }, { "epoch": 0.7858719646799117, "grad_norm": 0.19588352739810944, "learning_rate": 9.222602841547766e-06, "loss": 0.3946, "step": 356 }, { "epoch": 0.7880794701986755, "grad_norm": 0.1633402556180954, "learning_rate": 9.2157107676345e-06, "loss": 0.4027, "step": 357 }, { "epoch": 0.7902869757174393, "grad_norm": 0.16059032082557678, "learning_rate": 9.208790876265887e-06, "loss": 0.3981, "step": 358 }, { "epoch": 0.7924944812362031, "grad_norm": 0.1558162271976471, "learning_rate": 9.201843213102976e-06, "loss": 0.3969, "step": 359 }, { "epoch": 0.7947019867549668, "grad_norm": 0.1607562005519867, "learning_rate": 9.194867823990069e-06, "loss": 0.3972, "step": 360 }, { "epoch": 0.7969094922737306, "grad_norm": 0.1428382843732834, "learning_rate": 9.187864754954412e-06, "loss": 0.3928, "step": 361 }, { "epoch": 0.7991169977924945, "grad_norm": 0.15147483348846436, "learning_rate": 9.180834052205903e-06, "loss": 0.392, "step": 362 }, { "epoch": 0.8013245033112583, "grad_norm": 0.14845693111419678, "learning_rate": 9.173775762136783e-06, "loss": 0.3989, "step": 363 }, { "epoch": 0.8035320088300221, "grad_norm": 0.1665990948677063, "learning_rate": 9.166689931321326e-06, "loss": 0.3928, "step": 364 }, { "epoch": 0.8057395143487859, "grad_norm": 0.14729240536689758, "learning_rate": 9.159576606515532e-06, "loss": 0.3953, "step": 365 }, { "epoch": 0.8079470198675497, "grad_norm": 0.14690490067005157, "learning_rate": 9.152435834656823e-06, "loss": 0.4023, "step": 366 }, { "epoch": 0.8101545253863135, "grad_norm": 0.15939104557037354, "learning_rate": 9.145267662863732e-06, "loss": 0.4013, "step": 367 }, { "epoch": 0.8123620309050773, "grad_norm": 0.1510556936264038, "learning_rate": 9.13807213843559e-06, "loss": 0.3992, "step": 368 }, { "epoch": 0.8145695364238411, "grad_norm": 0.1646500676870346, "learning_rate": 9.130849308852217e-06, "loss": 0.3925, "step": 369 }, { "epoch": 0.8167770419426048, "grad_norm": 0.17254091799259186, "learning_rate": 9.123599221773601e-06, "loss": 0.4071, "step": 370 }, { "epoch": 0.8189845474613686, "grad_norm": 0.1497507095336914, "learning_rate": 9.116321925039591e-06, "loss": 0.3883, "step": 371 }, { "epoch": 0.8211920529801324, "grad_norm": 0.16002339124679565, "learning_rate": 9.109017466669587e-06, "loss": 0.3953, "step": 372 }, { "epoch": 0.8233995584988962, "grad_norm": 0.14812716841697693, "learning_rate": 9.101685894862206e-06, "loss": 0.4021, "step": 373 }, { "epoch": 0.82560706401766, "grad_norm": 0.15898163616657257, "learning_rate": 9.094327257994978e-06, "loss": 0.4102, "step": 374 }, { "epoch": 0.8278145695364238, "grad_norm": 0.15488837659358978, "learning_rate": 9.086941604624022e-06, "loss": 0.3912, "step": 375 }, { "epoch": 0.8300220750551877, "grad_norm": 0.1613994836807251, "learning_rate": 9.079528983483726e-06, "loss": 0.4029, "step": 376 }, { "epoch": 0.8322295805739515, "grad_norm": 0.1536788046360016, "learning_rate": 9.072089443486425e-06, "loss": 0.3956, "step": 377 }, { "epoch": 0.8344370860927153, "grad_norm": 0.1693611443042755, "learning_rate": 9.064623033722077e-06, "loss": 0.3984, "step": 378 }, { "epoch": 0.8366445916114791, "grad_norm": 0.18005254864692688, "learning_rate": 9.057129803457943e-06, "loss": 0.4022, "step": 379 }, { "epoch": 0.8388520971302428, "grad_norm": 0.15852075815200806, "learning_rate": 9.049609802138262e-06, "loss": 0.3816, "step": 380 }, { "epoch": 0.8410596026490066, "grad_norm": 0.1933150738477707, "learning_rate": 9.042063079383916e-06, "loss": 0.4028, "step": 381 }, { "epoch": 0.8432671081677704, "grad_norm": 0.18579082190990448, "learning_rate": 9.034489684992112e-06, "loss": 0.4057, "step": 382 }, { "epoch": 0.8454746136865342, "grad_norm": 0.1803402453660965, "learning_rate": 9.026889668936054e-06, "loss": 0.3976, "step": 383 }, { "epoch": 0.847682119205298, "grad_norm": 0.17872066795825958, "learning_rate": 9.019263081364605e-06, "loss": 0.3908, "step": 384 }, { "epoch": 0.8498896247240618, "grad_norm": 0.18206505477428436, "learning_rate": 9.01160997260196e-06, "loss": 0.4072, "step": 385 }, { "epoch": 0.8520971302428256, "grad_norm": 0.1718129813671112, "learning_rate": 9.00393039314732e-06, "loss": 0.4036, "step": 386 }, { "epoch": 0.8543046357615894, "grad_norm": 0.21254006028175354, "learning_rate": 8.996224393674545e-06, "loss": 0.4097, "step": 387 }, { "epoch": 0.8565121412803532, "grad_norm": 0.15457090735435486, "learning_rate": 8.988492025031838e-06, "loss": 0.4001, "step": 388 }, { "epoch": 0.8587196467991169, "grad_norm": 0.18473166227340698, "learning_rate": 8.980733338241395e-06, "loss": 0.3965, "step": 389 }, { "epoch": 0.8609271523178808, "grad_norm": 0.19067919254302979, "learning_rate": 8.972948384499068e-06, "loss": 0.3996, "step": 390 }, { "epoch": 0.8631346578366446, "grad_norm": 0.199691504240036, "learning_rate": 8.965137215174037e-06, "loss": 0.4003, "step": 391 }, { "epoch": 0.8653421633554084, "grad_norm": 0.22586044669151306, "learning_rate": 8.957299881808471e-06, "loss": 0.393, "step": 392 }, { "epoch": 0.8675496688741722, "grad_norm": 0.1807902604341507, "learning_rate": 8.949436436117172e-06, "loss": 0.4039, "step": 393 }, { "epoch": 0.869757174392936, "grad_norm": 0.22098712623119354, "learning_rate": 8.941546929987253e-06, "loss": 0.39, "step": 394 }, { "epoch": 0.8719646799116998, "grad_norm": 0.19912059605121613, "learning_rate": 8.933631415477785e-06, "loss": 0.396, "step": 395 }, { "epoch": 0.8741721854304636, "grad_norm": 0.16147476434707642, "learning_rate": 8.925689944819452e-06, "loss": 0.4011, "step": 396 }, { "epoch": 0.8763796909492274, "grad_norm": 0.19136802852153778, "learning_rate": 8.917722570414217e-06, "loss": 0.3896, "step": 397 }, { "epoch": 0.8785871964679912, "grad_norm": 0.20598876476287842, "learning_rate": 8.909729344834965e-06, "loss": 0.3995, "step": 398 }, { "epoch": 0.8807947019867549, "grad_norm": 0.1776047945022583, "learning_rate": 8.901710320825161e-06, "loss": 0.4001, "step": 399 }, { "epoch": 0.8830022075055187, "grad_norm": 0.18616576492786407, "learning_rate": 8.893665551298502e-06, "loss": 0.3995, "step": 400 }, { "epoch": 0.8852097130242825, "grad_norm": 0.18072794377803802, "learning_rate": 8.885595089338567e-06, "loss": 0.3978, "step": 401 }, { "epoch": 0.8874172185430463, "grad_norm": 0.181128591299057, "learning_rate": 8.877498988198471e-06, "loss": 0.4112, "step": 402 }, { "epoch": 0.8896247240618101, "grad_norm": 0.17399437725543976, "learning_rate": 8.869377301300501e-06, "loss": 0.3937, "step": 403 }, { "epoch": 0.891832229580574, "grad_norm": 0.17823876440525055, "learning_rate": 8.86123008223578e-06, "loss": 0.4055, "step": 404 }, { "epoch": 0.8940397350993378, "grad_norm": 0.176737442612648, "learning_rate": 8.853057384763904e-06, "loss": 0.3991, "step": 405 }, { "epoch": 0.8962472406181016, "grad_norm": 0.1630028337240219, "learning_rate": 8.844859262812584e-06, "loss": 0.3903, "step": 406 }, { "epoch": 0.8984547461368654, "grad_norm": 0.18285702168941498, "learning_rate": 8.8366357704773e-06, "loss": 0.3904, "step": 407 }, { "epoch": 0.9006622516556292, "grad_norm": 0.1716790795326233, "learning_rate": 8.82838696202094e-06, "loss": 0.4033, "step": 408 }, { "epoch": 0.9028697571743929, "grad_norm": 0.15402917563915253, "learning_rate": 8.820112891873433e-06, "loss": 0.39, "step": 409 }, { "epoch": 0.9050772626931567, "grad_norm": 0.16390980780124664, "learning_rate": 8.811813614631411e-06, "loss": 0.3993, "step": 410 }, { "epoch": 0.9072847682119205, "grad_norm": 0.15821807086467743, "learning_rate": 8.803489185057822e-06, "loss": 0.3953, "step": 411 }, { "epoch": 0.9094922737306843, "grad_norm": 0.17032112181186676, "learning_rate": 8.795139658081586e-06, "loss": 0.4035, "step": 412 }, { "epoch": 0.9116997792494481, "grad_norm": 0.15845684707164764, "learning_rate": 8.786765088797238e-06, "loss": 0.4013, "step": 413 }, { "epoch": 0.9139072847682119, "grad_norm": 0.17293021082878113, "learning_rate": 8.778365532464543e-06, "loss": 0.3965, "step": 414 }, { "epoch": 0.9161147902869757, "grad_norm": 0.14891566336154938, "learning_rate": 8.76994104450815e-06, "loss": 0.392, "step": 415 }, { "epoch": 0.9183222958057395, "grad_norm": 0.15123558044433594, "learning_rate": 8.761491680517218e-06, "loss": 0.396, "step": 416 }, { "epoch": 0.9205298013245033, "grad_norm": 0.1682642549276352, "learning_rate": 8.75301749624505e-06, "loss": 0.3838, "step": 417 }, { "epoch": 0.9227373068432672, "grad_norm": 0.1535079926252365, "learning_rate": 8.744518547608732e-06, "loss": 0.3921, "step": 418 }, { "epoch": 0.9249448123620309, "grad_norm": 0.16250640153884888, "learning_rate": 8.735994890688749e-06, "loss": 0.3896, "step": 419 }, { "epoch": 0.9271523178807947, "grad_norm": 0.16567584872245789, "learning_rate": 8.72744658172863e-06, "loss": 0.3928, "step": 420 }, { "epoch": 0.9293598233995585, "grad_norm": 0.16492141783237457, "learning_rate": 8.718873677134569e-06, "loss": 0.405, "step": 421 }, { "epoch": 0.9315673289183223, "grad_norm": 0.14573417603969574, "learning_rate": 8.710276233475058e-06, "loss": 0.3955, "step": 422 }, { "epoch": 0.9337748344370861, "grad_norm": 0.15780314803123474, "learning_rate": 8.701654307480508e-06, "loss": 0.3868, "step": 423 }, { "epoch": 0.9359823399558499, "grad_norm": 0.14735905826091766, "learning_rate": 8.693007956042874e-06, "loss": 0.3924, "step": 424 }, { "epoch": 0.9381898454746137, "grad_norm": 0.16305825114250183, "learning_rate": 8.684337236215289e-06, "loss": 0.3866, "step": 425 }, { "epoch": 0.9403973509933775, "grad_norm": 0.1656455248594284, "learning_rate": 8.675642205211679e-06, "loss": 0.3965, "step": 426 }, { "epoch": 0.9426048565121413, "grad_norm": 0.15034431219100952, "learning_rate": 8.666922920406384e-06, "loss": 0.3982, "step": 427 }, { "epoch": 0.9448123620309051, "grad_norm": 0.161673903465271, "learning_rate": 8.65817943933379e-06, "loss": 0.3933, "step": 428 }, { "epoch": 0.9470198675496688, "grad_norm": 0.1589784324169159, "learning_rate": 8.649411819687936e-06, "loss": 0.3976, "step": 429 }, { "epoch": 0.9492273730684326, "grad_norm": 0.166986882686615, "learning_rate": 8.640620119322146e-06, "loss": 0.4003, "step": 430 }, { "epoch": 0.9514348785871964, "grad_norm": 0.15880529582500458, "learning_rate": 8.631804396248637e-06, "loss": 0.3926, "step": 431 }, { "epoch": 0.9536423841059603, "grad_norm": 0.1741640269756317, "learning_rate": 8.62296470863814e-06, "loss": 0.3978, "step": 432 }, { "epoch": 0.9558498896247241, "grad_norm": 0.15339982509613037, "learning_rate": 8.61410111481952e-06, "loss": 0.4059, "step": 433 }, { "epoch": 0.9580573951434879, "grad_norm": 0.1605585813522339, "learning_rate": 8.605213673279382e-06, "loss": 0.3897, "step": 434 }, { "epoch": 0.9602649006622517, "grad_norm": 0.154588520526886, "learning_rate": 8.5963024426617e-06, "loss": 0.3853, "step": 435 }, { "epoch": 0.9624724061810155, "grad_norm": 0.16428066790103912, "learning_rate": 8.587367481767409e-06, "loss": 0.3944, "step": 436 }, { "epoch": 0.9646799116997793, "grad_norm": 0.15117953717708588, "learning_rate": 8.578408849554037e-06, "loss": 0.3924, "step": 437 }, { "epoch": 0.9668874172185431, "grad_norm": 0.15928852558135986, "learning_rate": 8.569426605135307e-06, "loss": 0.399, "step": 438 }, { "epoch": 0.9690949227373068, "grad_norm": 0.16322654485702515, "learning_rate": 8.560420807780742e-06, "loss": 0.3863, "step": 439 }, { "epoch": 0.9713024282560706, "grad_norm": 0.1749914139509201, "learning_rate": 8.551391516915288e-06, "loss": 0.4018, "step": 440 }, { "epoch": 0.9735099337748344, "grad_norm": 0.15128456056118011, "learning_rate": 8.542338792118907e-06, "loss": 0.3784, "step": 441 }, { "epoch": 0.9757174392935982, "grad_norm": 0.19126002490520477, "learning_rate": 8.533262693126191e-06, "loss": 0.3923, "step": 442 }, { "epoch": 0.977924944812362, "grad_norm": 0.16031384468078613, "learning_rate": 8.52416327982597e-06, "loss": 0.4036, "step": 443 }, { "epoch": 0.9801324503311258, "grad_norm": 0.16994954645633698, "learning_rate": 8.515040612260912e-06, "loss": 0.3865, "step": 444 }, { "epoch": 0.9823399558498896, "grad_norm": 0.17667335271835327, "learning_rate": 8.505894750627128e-06, "loss": 0.3884, "step": 445 }, { "epoch": 0.9845474613686535, "grad_norm": 0.1745615154504776, "learning_rate": 8.496725755273778e-06, "loss": 0.3926, "step": 446 }, { "epoch": 0.9867549668874173, "grad_norm": 0.16568712890148163, "learning_rate": 8.487533686702668e-06, "loss": 0.393, "step": 447 }, { "epoch": 0.9889624724061811, "grad_norm": 0.1762251853942871, "learning_rate": 8.478318605567853e-06, "loss": 0.3868, "step": 448 }, { "epoch": 0.9911699779249448, "grad_norm": 0.18370142579078674, "learning_rate": 8.46908057267524e-06, "loss": 0.3939, "step": 449 }, { "epoch": 0.9933774834437086, "grad_norm": 0.1815333366394043, "learning_rate": 8.459819648982182e-06, "loss": 0.3849, "step": 450 }, { "epoch": 0.9955849889624724, "grad_norm": 0.19721707701683044, "learning_rate": 8.450535895597074e-06, "loss": 0.3953, "step": 451 }, { "epoch": 0.9977924944812362, "grad_norm": 0.16212257742881775, "learning_rate": 8.441229373778957e-06, "loss": 0.3933, "step": 452 }, { "epoch": 1.0, "grad_norm": 0.2004193216562271, "learning_rate": 8.43190014493711e-06, "loss": 0.3914, "step": 453 }, { "epoch": 1.0022075055187638, "grad_norm": 0.15856173634529114, "learning_rate": 8.422548270630646e-06, "loss": 0.3806, "step": 454 }, { "epoch": 1.0044150110375276, "grad_norm": 0.18225938081741333, "learning_rate": 8.413173812568099e-06, "loss": 0.3714, "step": 455 }, { "epoch": 1.0066225165562914, "grad_norm": 0.17915207147598267, "learning_rate": 8.403776832607028e-06, "loss": 0.3834, "step": 456 }, { "epoch": 1.0088300220750552, "grad_norm": 0.1916593611240387, "learning_rate": 8.394357392753599e-06, "loss": 0.3666, "step": 457 }, { "epoch": 1.011037527593819, "grad_norm": 0.166362464427948, "learning_rate": 8.384915555162183e-06, "loss": 0.3839, "step": 458 }, { "epoch": 1.0132450331125828, "grad_norm": 0.20503659546375275, "learning_rate": 8.375451382134942e-06, "loss": 0.381, "step": 459 }, { "epoch": 1.0154525386313467, "grad_norm": 0.18665875494480133, "learning_rate": 8.365964936121422e-06, "loss": 0.3893, "step": 460 }, { "epoch": 1.0176600441501105, "grad_norm": 0.15370848774909973, "learning_rate": 8.35645627971813e-06, "loss": 0.3796, "step": 461 }, { "epoch": 1.0198675496688743, "grad_norm": 0.17442859709262848, "learning_rate": 8.346925475668138e-06, "loss": 0.3633, "step": 462 }, { "epoch": 1.022075055187638, "grad_norm": 0.18989215791225433, "learning_rate": 8.337372586860651e-06, "loss": 0.3743, "step": 463 }, { "epoch": 1.0242825607064017, "grad_norm": 0.16941452026367188, "learning_rate": 8.327797676330604e-06, "loss": 0.369, "step": 464 }, { "epoch": 1.0264900662251655, "grad_norm": 0.14053016901016235, "learning_rate": 8.31820080725825e-06, "loss": 0.3724, "step": 465 }, { "epoch": 1.0286975717439293, "grad_norm": 0.17336276173591614, "learning_rate": 8.308582042968726e-06, "loss": 0.3903, "step": 466 }, { "epoch": 1.030905077262693, "grad_norm": 0.15840588510036469, "learning_rate": 8.298941446931646e-06, "loss": 0.3847, "step": 467 }, { "epoch": 1.033112582781457, "grad_norm": 0.16392916440963745, "learning_rate": 8.289279082760685e-06, "loss": 0.3893, "step": 468 }, { "epoch": 1.0353200883002207, "grad_norm": 0.17757326364517212, "learning_rate": 8.279595014213158e-06, "loss": 0.3876, "step": 469 }, { "epoch": 1.0375275938189845, "grad_norm": 0.1557520627975464, "learning_rate": 8.26988930518959e-06, "loss": 0.3782, "step": 470 }, { "epoch": 1.0397350993377483, "grad_norm": 0.16154958307743073, "learning_rate": 8.260162019733305e-06, "loss": 0.3933, "step": 471 }, { "epoch": 1.0419426048565121, "grad_norm": 0.18913333117961884, "learning_rate": 8.250413222029997e-06, "loss": 0.3888, "step": 472 }, { "epoch": 1.044150110375276, "grad_norm": 0.1763067990541458, "learning_rate": 8.240642976407313e-06, "loss": 0.3875, "step": 473 }, { "epoch": 1.0463576158940397, "grad_norm": 0.16598297655582428, "learning_rate": 8.230851347334424e-06, "loss": 0.3894, "step": 474 }, { "epoch": 1.0485651214128036, "grad_norm": 0.17947614192962646, "learning_rate": 8.221038399421592e-06, "loss": 0.3853, "step": 475 }, { "epoch": 1.0507726269315674, "grad_norm": 0.16084261238574982, "learning_rate": 8.211204197419766e-06, "loss": 0.3762, "step": 476 }, { "epoch": 1.0529801324503312, "grad_norm": 0.21188747882843018, "learning_rate": 8.201348806220127e-06, "loss": 0.3791, "step": 477 }, { "epoch": 1.055187637969095, "grad_norm": 0.15903352200984955, "learning_rate": 8.191472290853683e-06, "loss": 0.3799, "step": 478 }, { "epoch": 1.0573951434878588, "grad_norm": 0.2036534547805786, "learning_rate": 8.181574716490823e-06, "loss": 0.3909, "step": 479 }, { "epoch": 1.0596026490066226, "grad_norm": 0.16643071174621582, "learning_rate": 8.171656148440902e-06, "loss": 0.3842, "step": 480 }, { "epoch": 1.0618101545253864, "grad_norm": 0.16434621810913086, "learning_rate": 8.161716652151795e-06, "loss": 0.3748, "step": 481 }, { "epoch": 1.0640176600441502, "grad_norm": 0.18316681683063507, "learning_rate": 8.151756293209476e-06, "loss": 0.3795, "step": 482 }, { "epoch": 1.0662251655629138, "grad_norm": 0.14608311653137207, "learning_rate": 8.14177513733758e-06, "loss": 0.3749, "step": 483 }, { "epoch": 1.0684326710816776, "grad_norm": 0.1656610518693924, "learning_rate": 8.131773250396973e-06, "loss": 0.3794, "step": 484 }, { "epoch": 1.0706401766004414, "grad_norm": 0.14669524133205414, "learning_rate": 8.121750698385315e-06, "loss": 0.3825, "step": 485 }, { "epoch": 1.0728476821192052, "grad_norm": 0.17891530692577362, "learning_rate": 8.111707547436623e-06, "loss": 0.373, "step": 486 }, { "epoch": 1.075055187637969, "grad_norm": 0.15400467813014984, "learning_rate": 8.10164386382084e-06, "loss": 0.3792, "step": 487 }, { "epoch": 1.0772626931567328, "grad_norm": 0.17551693320274353, "learning_rate": 8.091559713943388e-06, "loss": 0.3908, "step": 488 }, { "epoch": 1.0794701986754967, "grad_norm": 0.15460826456546783, "learning_rate": 8.081455164344745e-06, "loss": 0.3838, "step": 489 }, { "epoch": 1.0816777041942605, "grad_norm": 0.15716975927352905, "learning_rate": 8.071330281699989e-06, "loss": 0.3934, "step": 490 }, { "epoch": 1.0838852097130243, "grad_norm": 0.15411889553070068, "learning_rate": 8.06118513281837e-06, "loss": 0.3792, "step": 491 }, { "epoch": 1.086092715231788, "grad_norm": 0.18962320685386658, "learning_rate": 8.051019784642864e-06, "loss": 0.3888, "step": 492 }, { "epoch": 1.0883002207505519, "grad_norm": 0.15896014869213104, "learning_rate": 8.040834304249733e-06, "loss": 0.3861, "step": 493 }, { "epoch": 1.0905077262693157, "grad_norm": 0.1523059904575348, "learning_rate": 8.03062875884808e-06, "loss": 0.3727, "step": 494 }, { "epoch": 1.0927152317880795, "grad_norm": 0.13958971202373505, "learning_rate": 8.02040321577941e-06, "loss": 0.3718, "step": 495 }, { "epoch": 1.0949227373068433, "grad_norm": 0.1423843950033188, "learning_rate": 8.010157742517185e-06, "loss": 0.3746, "step": 496 }, { "epoch": 1.0971302428256071, "grad_norm": 0.13685709238052368, "learning_rate": 7.99989240666637e-06, "loss": 0.3727, "step": 497 }, { "epoch": 1.099337748344371, "grad_norm": 0.15423151850700378, "learning_rate": 7.989607275963e-06, "loss": 0.3778, "step": 498 }, { "epoch": 1.1015452538631347, "grad_norm": 0.15014959871768951, "learning_rate": 7.979302418273723e-06, "loss": 0.3747, "step": 499 }, { "epoch": 1.1037527593818985, "grad_norm": 0.16919955611228943, "learning_rate": 7.968977901595355e-06, "loss": 0.3782, "step": 500 }, { "epoch": 1.1059602649006623, "grad_norm": 0.1481235921382904, "learning_rate": 7.958633794054439e-06, "loss": 0.3808, "step": 501 }, { "epoch": 1.108167770419426, "grad_norm": 0.17575468122959137, "learning_rate": 7.94827016390678e-06, "loss": 0.3708, "step": 502 }, { "epoch": 1.1103752759381897, "grad_norm": 0.139452263712883, "learning_rate": 7.93788707953701e-06, "loss": 0.376, "step": 503 }, { "epoch": 1.1125827814569536, "grad_norm": 0.14127641916275024, "learning_rate": 7.927484609458128e-06, "loss": 0.3847, "step": 504 }, { "epoch": 1.1147902869757174, "grad_norm": 0.15063825249671936, "learning_rate": 7.917062822311047e-06, "loss": 0.3842, "step": 505 }, { "epoch": 1.1169977924944812, "grad_norm": 0.15780918300151825, "learning_rate": 7.90662178686415e-06, "loss": 0.3737, "step": 506 }, { "epoch": 1.119205298013245, "grad_norm": 0.14620929956436157, "learning_rate": 7.896161572012824e-06, "loss": 0.3862, "step": 507 }, { "epoch": 1.1214128035320088, "grad_norm": 0.1493159383535385, "learning_rate": 7.885682246779016e-06, "loss": 0.3809, "step": 508 }, { "epoch": 1.1236203090507726, "grad_norm": 0.1471249759197235, "learning_rate": 7.875183880310772e-06, "loss": 0.3885, "step": 509 }, { "epoch": 1.1258278145695364, "grad_norm": 0.1509741246700287, "learning_rate": 7.86466654188178e-06, "loss": 0.3778, "step": 510 }, { "epoch": 1.1280353200883002, "grad_norm": 0.14812204241752625, "learning_rate": 7.854130300890921e-06, "loss": 0.3731, "step": 511 }, { "epoch": 1.130242825607064, "grad_norm": 0.1415746510028839, "learning_rate": 7.843575226861795e-06, "loss": 0.3798, "step": 512 }, { "epoch": 1.1324503311258278, "grad_norm": 0.14002346992492676, "learning_rate": 7.833001389442283e-06, "loss": 0.3833, "step": 513 }, { "epoch": 1.1346578366445916, "grad_norm": 0.15462878346443176, "learning_rate": 7.82240885840407e-06, "loss": 0.3785, "step": 514 }, { "epoch": 1.1368653421633554, "grad_norm": 0.15146006643772125, "learning_rate": 7.811797703642193e-06, "loss": 0.3767, "step": 515 }, { "epoch": 1.1390728476821192, "grad_norm": 0.1545468419790268, "learning_rate": 7.801167995174575e-06, "loss": 0.3739, "step": 516 }, { "epoch": 1.141280353200883, "grad_norm": 0.1515691876411438, "learning_rate": 7.790519803141572e-06, "loss": 0.394, "step": 517 }, { "epoch": 1.1434878587196469, "grad_norm": 0.1403704136610031, "learning_rate": 7.7798531978055e-06, "loss": 0.3844, "step": 518 }, { "epoch": 1.1456953642384107, "grad_norm": 0.12919290363788605, "learning_rate": 7.769168249550176e-06, "loss": 0.3746, "step": 519 }, { "epoch": 1.1479028697571745, "grad_norm": 0.137840136885643, "learning_rate": 7.758465028880455e-06, "loss": 0.3748, "step": 520 }, { "epoch": 1.150110375275938, "grad_norm": 0.13867942988872528, "learning_rate": 7.747743606421761e-06, "loss": 0.3813, "step": 521 }, { "epoch": 1.152317880794702, "grad_norm": 0.1563754677772522, "learning_rate": 7.737004052919623e-06, "loss": 0.3673, "step": 522 }, { "epoch": 1.1545253863134657, "grad_norm": 0.15125791728496552, "learning_rate": 7.726246439239209e-06, "loss": 0.3814, "step": 523 }, { "epoch": 1.1567328918322295, "grad_norm": 0.15860167145729065, "learning_rate": 7.715470836364857e-06, "loss": 0.3962, "step": 524 }, { "epoch": 1.1589403973509933, "grad_norm": 0.14636121690273285, "learning_rate": 7.704677315399607e-06, "loss": 0.3786, "step": 525 }, { "epoch": 1.161147902869757, "grad_norm": 0.1441173404455185, "learning_rate": 7.693865947564733e-06, "loss": 0.3737, "step": 526 }, { "epoch": 1.163355408388521, "grad_norm": 0.1591232568025589, "learning_rate": 7.68303680419927e-06, "loss": 0.3721, "step": 527 }, { "epoch": 1.1655629139072847, "grad_norm": 0.1598597913980484, "learning_rate": 7.672189956759546e-06, "loss": 0.3793, "step": 528 }, { "epoch": 1.1677704194260485, "grad_norm": 0.16569367051124573, "learning_rate": 7.661325476818708e-06, "loss": 0.3799, "step": 529 }, { "epoch": 1.1699779249448123, "grad_norm": 0.1879713088274002, "learning_rate": 7.65044343606626e-06, "loss": 0.3831, "step": 530 }, { "epoch": 1.1721854304635762, "grad_norm": 0.14313288033008575, "learning_rate": 7.639543906307565e-06, "loss": 0.3774, "step": 531 }, { "epoch": 1.17439293598234, "grad_norm": 0.20557354390621185, "learning_rate": 7.628626959463405e-06, "loss": 0.3766, "step": 532 }, { "epoch": 1.1766004415011038, "grad_norm": 0.14320634305477142, "learning_rate": 7.6176926675694786e-06, "loss": 0.3754, "step": 533 }, { "epoch": 1.1788079470198676, "grad_norm": 0.18957629799842834, "learning_rate": 7.606741102775936e-06, "loss": 0.3799, "step": 534 }, { "epoch": 1.1810154525386314, "grad_norm": 0.15119552612304688, "learning_rate": 7.595772337346912e-06, "loss": 0.3863, "step": 535 }, { "epoch": 1.1832229580573952, "grad_norm": 0.16690769791603088, "learning_rate": 7.584786443660028e-06, "loss": 0.371, "step": 536 }, { "epoch": 1.185430463576159, "grad_norm": 0.1685992032289505, "learning_rate": 7.573783494205936e-06, "loss": 0.3741, "step": 537 }, { "epoch": 1.1876379690949228, "grad_norm": 0.14501997828483582, "learning_rate": 7.562763561587824e-06, "loss": 0.3701, "step": 538 }, { "epoch": 1.1898454746136866, "grad_norm": 0.16112159192562103, "learning_rate": 7.55172671852095e-06, "loss": 0.3781, "step": 539 }, { "epoch": 1.1920529801324504, "grad_norm": 0.13081905245780945, "learning_rate": 7.5406730378321506e-06, "loss": 0.39, "step": 540 }, { "epoch": 1.1942604856512142, "grad_norm": 0.15677790343761444, "learning_rate": 7.5296025924593705e-06, "loss": 0.3814, "step": 541 }, { "epoch": 1.1964679911699778, "grad_norm": 0.1494779884815216, "learning_rate": 7.518515455451172e-06, "loss": 0.3757, "step": 542 }, { "epoch": 1.1986754966887416, "grad_norm": 0.1597052365541458, "learning_rate": 7.50741169996626e-06, "loss": 0.3772, "step": 543 }, { "epoch": 1.2008830022075054, "grad_norm": 0.1445734202861786, "learning_rate": 7.496291399273e-06, "loss": 0.3779, "step": 544 }, { "epoch": 1.2030905077262692, "grad_norm": 0.14188243448734283, "learning_rate": 7.485154626748924e-06, "loss": 0.3801, "step": 545 }, { "epoch": 1.205298013245033, "grad_norm": 0.15295171737670898, "learning_rate": 7.474001455880258e-06, "loss": 0.3786, "step": 546 }, { "epoch": 1.2075055187637969, "grad_norm": 0.12851738929748535, "learning_rate": 7.4628319602614315e-06, "loss": 0.3759, "step": 547 }, { "epoch": 1.2097130242825607, "grad_norm": 0.12803953886032104, "learning_rate": 7.451646213594597e-06, "loss": 0.3807, "step": 548 }, { "epoch": 1.2119205298013245, "grad_norm": 0.15474575757980347, "learning_rate": 7.440444289689135e-06, "loss": 0.3761, "step": 549 }, { "epoch": 1.2141280353200883, "grad_norm": 0.14079326391220093, "learning_rate": 7.429226262461175e-06, "loss": 0.3744, "step": 550 }, { "epoch": 1.216335540838852, "grad_norm": 0.13553716242313385, "learning_rate": 7.417992205933104e-06, "loss": 0.3813, "step": 551 }, { "epoch": 1.218543046357616, "grad_norm": 0.14058908820152283, "learning_rate": 7.406742194233074e-06, "loss": 0.3722, "step": 552 }, { "epoch": 1.2207505518763797, "grad_norm": 0.15452724695205688, "learning_rate": 7.3954763015945266e-06, "loss": 0.3628, "step": 553 }, { "epoch": 1.2229580573951435, "grad_norm": 0.14715242385864258, "learning_rate": 7.384194602355685e-06, "loss": 0.3779, "step": 554 }, { "epoch": 1.2251655629139073, "grad_norm": 0.15119391679763794, "learning_rate": 7.37289717095908e-06, "loss": 0.3787, "step": 555 }, { "epoch": 1.2273730684326711, "grad_norm": 0.1800910234451294, "learning_rate": 7.361584081951046e-06, "loss": 0.3781, "step": 556 }, { "epoch": 1.229580573951435, "grad_norm": 0.1489747166633606, "learning_rate": 7.350255409981237e-06, "loss": 0.3843, "step": 557 }, { "epoch": 1.2317880794701987, "grad_norm": 0.14379991590976715, "learning_rate": 7.338911229802133e-06, "loss": 0.3832, "step": 558 }, { "epoch": 1.2339955849889626, "grad_norm": 0.13160887360572815, "learning_rate": 7.327551616268541e-06, "loss": 0.3858, "step": 559 }, { "epoch": 1.2362030905077264, "grad_norm": 0.1353141814470291, "learning_rate": 7.316176644337107e-06, "loss": 0.3688, "step": 560 }, { "epoch": 1.23841059602649, "grad_norm": 0.16255126893520355, "learning_rate": 7.304786389065823e-06, "loss": 0.377, "step": 561 }, { "epoch": 1.240618101545254, "grad_norm": 0.13924263417720795, "learning_rate": 7.293380925613524e-06, "loss": 0.376, "step": 562 }, { "epoch": 1.2428256070640176, "grad_norm": 0.1347806751728058, "learning_rate": 7.281960329239398e-06, "loss": 0.3881, "step": 563 }, { "epoch": 1.2450331125827814, "grad_norm": 0.1494341343641281, "learning_rate": 7.270524675302491e-06, "loss": 0.3843, "step": 564 }, { "epoch": 1.2472406181015452, "grad_norm": 0.16562990844249725, "learning_rate": 7.259074039261199e-06, "loss": 0.3843, "step": 565 }, { "epoch": 1.249448123620309, "grad_norm": 0.1605864018201828, "learning_rate": 7.247608496672786e-06, "loss": 0.3741, "step": 566 }, { "epoch": 1.2516556291390728, "grad_norm": 0.1619981825351715, "learning_rate": 7.2361281231928725e-06, "loss": 0.392, "step": 567 }, { "epoch": 1.2538631346578366, "grad_norm": 0.13447479903697968, "learning_rate": 7.2246329945749425e-06, "loss": 0.3755, "step": 568 }, { "epoch": 1.2560706401766004, "grad_norm": 0.1463773399591446, "learning_rate": 7.213123186669842e-06, "loss": 0.3913, "step": 569 }, { "epoch": 1.2582781456953642, "grad_norm": 0.14680610597133636, "learning_rate": 7.201598775425278e-06, "loss": 0.3776, "step": 570 }, { "epoch": 1.260485651214128, "grad_norm": 0.14192719757556915, "learning_rate": 7.190059836885318e-06, "loss": 0.3829, "step": 571 }, { "epoch": 1.2626931567328918, "grad_norm": 0.1412820965051651, "learning_rate": 7.178506447189887e-06, "loss": 0.3782, "step": 572 }, { "epoch": 1.2649006622516556, "grad_norm": 0.1515144258737564, "learning_rate": 7.166938682574272e-06, "loss": 0.3779, "step": 573 }, { "epoch": 1.2671081677704195, "grad_norm": 0.148550346493721, "learning_rate": 7.155356619368604e-06, "loss": 0.376, "step": 574 }, { "epoch": 1.2693156732891833, "grad_norm": 0.14767727255821228, "learning_rate": 7.14376033399737e-06, "loss": 0.3768, "step": 575 }, { "epoch": 1.271523178807947, "grad_norm": 0.1577431708574295, "learning_rate": 7.132149902978902e-06, "loss": 0.373, "step": 576 }, { "epoch": 1.2737306843267109, "grad_norm": 0.15225905179977417, "learning_rate": 7.120525402924871e-06, "loss": 0.365, "step": 577 }, { "epoch": 1.2759381898454747, "grad_norm": 0.16508492827415466, "learning_rate": 7.108886910539781e-06, "loss": 0.3866, "step": 578 }, { "epoch": 1.2781456953642385, "grad_norm": 0.16345573961734772, "learning_rate": 7.097234502620468e-06, "loss": 0.3819, "step": 579 }, { "epoch": 1.280353200883002, "grad_norm": 0.15904562175273895, "learning_rate": 7.085568256055589e-06, "loss": 0.3877, "step": 580 }, { "epoch": 1.2825607064017661, "grad_norm": 0.17447535693645477, "learning_rate": 7.073888247825111e-06, "loss": 0.3902, "step": 581 }, { "epoch": 1.2847682119205297, "grad_norm": 0.1613752692937851, "learning_rate": 7.062194554999817e-06, "loss": 0.3765, "step": 582 }, { "epoch": 1.2869757174392937, "grad_norm": 0.1705675572156906, "learning_rate": 7.05048725474078e-06, "loss": 0.3699, "step": 583 }, { "epoch": 1.2891832229580573, "grad_norm": 0.20963701605796814, "learning_rate": 7.038766424298865e-06, "loss": 0.3909, "step": 584 }, { "epoch": 1.2913907284768211, "grad_norm": 0.16403751075267792, "learning_rate": 7.027032141014216e-06, "loss": 0.3753, "step": 585 }, { "epoch": 1.293598233995585, "grad_norm": 0.18310756981372833, "learning_rate": 7.0152844823157474e-06, "loss": 0.3807, "step": 586 }, { "epoch": 1.2958057395143487, "grad_norm": 0.16576433181762695, "learning_rate": 7.003523525720626e-06, "loss": 0.3659, "step": 587 }, { "epoch": 1.2980132450331126, "grad_norm": 0.1746741682291031, "learning_rate": 6.991749348833773e-06, "loss": 0.3783, "step": 588 }, { "epoch": 1.3002207505518764, "grad_norm": 0.16022977232933044, "learning_rate": 6.979962029347338e-06, "loss": 0.3745, "step": 589 }, { "epoch": 1.3024282560706402, "grad_norm": 0.1607898473739624, "learning_rate": 6.9681616450401936e-06, "loss": 0.3732, "step": 590 }, { "epoch": 1.304635761589404, "grad_norm": 0.1624252200126648, "learning_rate": 6.956348273777424e-06, "loss": 0.3815, "step": 591 }, { "epoch": 1.3068432671081678, "grad_norm": 0.14729204773902893, "learning_rate": 6.944521993509803e-06, "loss": 0.3734, "step": 592 }, { "epoch": 1.3090507726269316, "grad_norm": 0.18559885025024414, "learning_rate": 6.9326828822732885e-06, "loss": 0.386, "step": 593 }, { "epoch": 1.3112582781456954, "grad_norm": 0.15120132267475128, "learning_rate": 6.920831018188502e-06, "loss": 0.3702, "step": 594 }, { "epoch": 1.3134657836644592, "grad_norm": 0.18332970142364502, "learning_rate": 6.908966479460219e-06, "loss": 0.3807, "step": 595 }, { "epoch": 1.315673289183223, "grad_norm": 0.17234960198402405, "learning_rate": 6.89708934437684e-06, "loss": 0.3644, "step": 596 }, { "epoch": 1.3178807947019868, "grad_norm": 0.18331332504749298, "learning_rate": 6.885199691309892e-06, "loss": 0.3841, "step": 597 }, { "epoch": 1.3200883002207506, "grad_norm": 0.15819165110588074, "learning_rate": 6.873297598713497e-06, "loss": 0.3725, "step": 598 }, { "epoch": 1.3222958057395142, "grad_norm": 0.1651681512594223, "learning_rate": 6.8613831451238636e-06, "loss": 0.3762, "step": 599 }, { "epoch": 1.3245033112582782, "grad_norm": 0.1659359186887741, "learning_rate": 6.84945640915876e-06, "loss": 0.3757, "step": 600 }, { "epoch": 1.3267108167770418, "grad_norm": 0.15379931032657623, "learning_rate": 6.837517469517001e-06, "loss": 0.3735, "step": 601 }, { "epoch": 1.3289183222958059, "grad_norm": 0.1771930605173111, "learning_rate": 6.82556640497793e-06, "loss": 0.3798, "step": 602 }, { "epoch": 1.3311258278145695, "grad_norm": 0.16866551339626312, "learning_rate": 6.813603294400895e-06, "loss": 0.3724, "step": 603 }, { "epoch": 1.3333333333333333, "grad_norm": 0.16047048568725586, "learning_rate": 6.8016282167247325e-06, "loss": 0.3744, "step": 604 }, { "epoch": 1.335540838852097, "grad_norm": 0.13133041560649872, "learning_rate": 6.7896412509672385e-06, "loss": 0.3775, "step": 605 }, { "epoch": 1.3377483443708609, "grad_norm": 0.16929350793361664, "learning_rate": 6.777642476224658e-06, "loss": 0.3724, "step": 606 }, { "epoch": 1.3399558498896247, "grad_norm": 0.14370004832744598, "learning_rate": 6.765631971671156e-06, "loss": 0.3795, "step": 607 }, { "epoch": 1.3421633554083885, "grad_norm": 0.18519333004951477, "learning_rate": 6.753609816558297e-06, "loss": 0.3789, "step": 608 }, { "epoch": 1.3443708609271523, "grad_norm": 0.15061011910438538, "learning_rate": 6.741576090214526e-06, "loss": 0.3779, "step": 609 }, { "epoch": 1.346578366445916, "grad_norm": 0.16702063381671906, "learning_rate": 6.729530872044629e-06, "loss": 0.3791, "step": 610 }, { "epoch": 1.34878587196468, "grad_norm": 0.1492423117160797, "learning_rate": 6.717474241529235e-06, "loss": 0.3821, "step": 611 }, { "epoch": 1.3509933774834437, "grad_norm": 0.15660499036312103, "learning_rate": 6.705406278224269e-06, "loss": 0.3764, "step": 612 }, { "epoch": 1.3532008830022075, "grad_norm": 0.13613760471343994, "learning_rate": 6.69332706176044e-06, "loss": 0.3842, "step": 613 }, { "epoch": 1.3554083885209713, "grad_norm": 0.1655375063419342, "learning_rate": 6.681236671842709e-06, "loss": 0.3861, "step": 614 }, { "epoch": 1.3576158940397351, "grad_norm": 0.15369166433811188, "learning_rate": 6.669135188249767e-06, "loss": 0.3753, "step": 615 }, { "epoch": 1.359823399558499, "grad_norm": 0.1611892580986023, "learning_rate": 6.657022690833503e-06, "loss": 0.3732, "step": 616 }, { "epoch": 1.3620309050772628, "grad_norm": 0.15144729614257812, "learning_rate": 6.644899259518485e-06, "loss": 0.3756, "step": 617 }, { "epoch": 1.3642384105960264, "grad_norm": 0.1591017246246338, "learning_rate": 6.632764974301429e-06, "loss": 0.3643, "step": 618 }, { "epoch": 1.3664459161147904, "grad_norm": 0.15775880217552185, "learning_rate": 6.620619915250666e-06, "loss": 0.3773, "step": 619 }, { "epoch": 1.368653421633554, "grad_norm": 0.14554527401924133, "learning_rate": 6.608464162505621e-06, "loss": 0.3788, "step": 620 }, { "epoch": 1.370860927152318, "grad_norm": 0.15284626185894012, "learning_rate": 6.596297796276284e-06, "loss": 0.3743, "step": 621 }, { "epoch": 1.3730684326710816, "grad_norm": 0.14736317098140717, "learning_rate": 6.584120896842675e-06, "loss": 0.3763, "step": 622 }, { "epoch": 1.3752759381898454, "grad_norm": 0.15955647826194763, "learning_rate": 6.571933544554319e-06, "loss": 0.3664, "step": 623 }, { "epoch": 1.3774834437086092, "grad_norm": 0.16919533908367157, "learning_rate": 6.559735819829713e-06, "loss": 0.3755, "step": 624 }, { "epoch": 1.379690949227373, "grad_norm": 0.15306392312049866, "learning_rate": 6.5475278031558e-06, "loss": 0.3772, "step": 625 }, { "epoch": 1.3818984547461368, "grad_norm": 0.1504804790019989, "learning_rate": 6.5353095750874295e-06, "loss": 0.3709, "step": 626 }, { "epoch": 1.3841059602649006, "grad_norm": 0.14429201185703278, "learning_rate": 6.523081216246838e-06, "loss": 0.3709, "step": 627 }, { "epoch": 1.3863134657836644, "grad_norm": 0.14882300794124603, "learning_rate": 6.5108428073231e-06, "loss": 0.3768, "step": 628 }, { "epoch": 1.3885209713024282, "grad_norm": 0.1360517293214798, "learning_rate": 6.498594429071618e-06, "loss": 0.3768, "step": 629 }, { "epoch": 1.390728476821192, "grad_norm": 0.13942170143127441, "learning_rate": 6.486336162313568e-06, "loss": 0.3796, "step": 630 }, { "epoch": 1.3929359823399559, "grad_norm": 0.14323705434799194, "learning_rate": 6.474068087935379e-06, "loss": 0.379, "step": 631 }, { "epoch": 1.3951434878587197, "grad_norm": 0.15375126898288727, "learning_rate": 6.461790286888196e-06, "loss": 0.3721, "step": 632 }, { "epoch": 1.3973509933774835, "grad_norm": 0.15478521585464478, "learning_rate": 6.449502840187344e-06, "loss": 0.3726, "step": 633 }, { "epoch": 1.3995584988962473, "grad_norm": 0.16403432190418243, "learning_rate": 6.437205828911797e-06, "loss": 0.3765, "step": 634 }, { "epoch": 1.401766004415011, "grad_norm": 0.1496008038520813, "learning_rate": 6.424899334203641e-06, "loss": 0.3785, "step": 635 }, { "epoch": 1.403973509933775, "grad_norm": 0.152191162109375, "learning_rate": 6.4125834372675355e-06, "loss": 0.3752, "step": 636 }, { "epoch": 1.4061810154525387, "grad_norm": 0.15051637589931488, "learning_rate": 6.400258219370183e-06, "loss": 0.3743, "step": 637 }, { "epoch": 1.4083885209713025, "grad_norm": 0.15709619224071503, "learning_rate": 6.387923761839794e-06, "loss": 0.3697, "step": 638 }, { "epoch": 1.410596026490066, "grad_norm": 0.14248456060886383, "learning_rate": 6.375580146065539e-06, "loss": 0.373, "step": 639 }, { "epoch": 1.4128035320088301, "grad_norm": 0.1371423304080963, "learning_rate": 6.363227453497029e-06, "loss": 0.3819, "step": 640 }, { "epoch": 1.4150110375275937, "grad_norm": 0.16379839181900024, "learning_rate": 6.350865765643759e-06, "loss": 0.3793, "step": 641 }, { "epoch": 1.4172185430463577, "grad_norm": 0.14373619854450226, "learning_rate": 6.3384951640745865e-06, "loss": 0.3779, "step": 642 }, { "epoch": 1.4194260485651213, "grad_norm": 0.14477743208408356, "learning_rate": 6.32611573041718e-06, "loss": 0.3786, "step": 643 }, { "epoch": 1.4216335540838851, "grad_norm": 0.17920951545238495, "learning_rate": 6.313727546357498e-06, "loss": 0.3901, "step": 644 }, { "epoch": 1.423841059602649, "grad_norm": 0.14486798644065857, "learning_rate": 6.301330693639224e-06, "loss": 0.3786, "step": 645 }, { "epoch": 1.4260485651214128, "grad_norm": 0.13817182183265686, "learning_rate": 6.288925254063249e-06, "loss": 0.3725, "step": 646 }, { "epoch": 1.4282560706401766, "grad_norm": 0.14681169390678406, "learning_rate": 6.27651130948713e-06, "loss": 0.3851, "step": 647 }, { "epoch": 1.4304635761589404, "grad_norm": 0.14171209931373596, "learning_rate": 6.264088941824534e-06, "loss": 0.3802, "step": 648 }, { "epoch": 1.4326710816777042, "grad_norm": 0.14188838005065918, "learning_rate": 6.251658233044715e-06, "loss": 0.3764, "step": 649 }, { "epoch": 1.434878587196468, "grad_norm": 0.15146797895431519, "learning_rate": 6.239219265171964e-06, "loss": 0.3791, "step": 650 }, { "epoch": 1.4370860927152318, "grad_norm": 0.1602582484483719, "learning_rate": 6.22677212028507e-06, "loss": 0.3773, "step": 651 }, { "epoch": 1.4392935982339956, "grad_norm": 0.14558373391628265, "learning_rate": 6.214316880516775e-06, "loss": 0.3631, "step": 652 }, { "epoch": 1.4415011037527594, "grad_norm": 0.17893047630786896, "learning_rate": 6.201853628053243e-06, "loss": 0.374, "step": 653 }, { "epoch": 1.4437086092715232, "grad_norm": 0.14639593660831451, "learning_rate": 6.189382445133502e-06, "loss": 0.3753, "step": 654 }, { "epoch": 1.445916114790287, "grad_norm": 0.14472749829292297, "learning_rate": 6.17690341404891e-06, "loss": 0.3796, "step": 655 }, { "epoch": 1.4481236203090508, "grad_norm": 0.1566278040409088, "learning_rate": 6.16441661714262e-06, "loss": 0.3741, "step": 656 }, { "epoch": 1.4503311258278146, "grad_norm": 0.14471283555030823, "learning_rate": 6.151922136809017e-06, "loss": 0.3808, "step": 657 }, { "epoch": 1.4525386313465782, "grad_norm": 0.15734338760375977, "learning_rate": 6.13942005549319e-06, "loss": 0.3781, "step": 658 }, { "epoch": 1.4547461368653423, "grad_norm": 0.13064594566822052, "learning_rate": 6.126910455690384e-06, "loss": 0.377, "step": 659 }, { "epoch": 1.4569536423841059, "grad_norm": 0.15152592957019806, "learning_rate": 6.114393419945455e-06, "loss": 0.3784, "step": 660 }, { "epoch": 1.4591611479028699, "grad_norm": 0.14837685227394104, "learning_rate": 6.101869030852321e-06, "loss": 0.3889, "step": 661 }, { "epoch": 1.4613686534216335, "grad_norm": 0.14178897440433502, "learning_rate": 6.089337371053429e-06, "loss": 0.3714, "step": 662 }, { "epoch": 1.4635761589403973, "grad_norm": 0.15140818059444427, "learning_rate": 6.076798523239194e-06, "loss": 0.3816, "step": 663 }, { "epoch": 1.465783664459161, "grad_norm": 0.14195586740970612, "learning_rate": 6.064252570147464e-06, "loss": 0.3738, "step": 664 }, { "epoch": 1.467991169977925, "grad_norm": 0.13242661952972412, "learning_rate": 6.051699594562976e-06, "loss": 0.3669, "step": 665 }, { "epoch": 1.4701986754966887, "grad_norm": 0.1379902958869934, "learning_rate": 6.039139679316797e-06, "loss": 0.362, "step": 666 }, { "epoch": 1.4724061810154525, "grad_norm": 0.1443612426519394, "learning_rate": 6.026572907285791e-06, "loss": 0.3711, "step": 667 }, { "epoch": 1.4746136865342163, "grad_norm": 0.14281077682971954, "learning_rate": 6.013999361392064e-06, "loss": 0.3753, "step": 668 }, { "epoch": 1.4768211920529801, "grad_norm": 0.14133213460445404, "learning_rate": 6.001419124602422e-06, "loss": 0.377, "step": 669 }, { "epoch": 1.479028697571744, "grad_norm": 0.14672201871871948, "learning_rate": 5.988832279927818e-06, "loss": 0.3763, "step": 670 }, { "epoch": 1.4812362030905077, "grad_norm": 0.13457442820072174, "learning_rate": 5.976238910422811e-06, "loss": 0.381, "step": 671 }, { "epoch": 1.4834437086092715, "grad_norm": 0.15843527019023895, "learning_rate": 5.963639099185011e-06, "loss": 0.3773, "step": 672 }, { "epoch": 1.4856512141280354, "grad_norm": 0.15779507160186768, "learning_rate": 5.9510329293545356e-06, "loss": 0.3789, "step": 673 }, { "epoch": 1.4878587196467992, "grad_norm": 0.14860793948173523, "learning_rate": 5.938420484113458e-06, "loss": 0.3765, "step": 674 }, { "epoch": 1.490066225165563, "grad_norm": 0.13150885701179504, "learning_rate": 5.925801846685262e-06, "loss": 0.37, "step": 675 }, { "epoch": 1.4922737306843268, "grad_norm": 0.13564042747020721, "learning_rate": 5.913177100334287e-06, "loss": 0.3788, "step": 676 }, { "epoch": 1.4944812362030906, "grad_norm": 0.14382179081439972, "learning_rate": 5.900546328365189e-06, "loss": 0.381, "step": 677 }, { "epoch": 1.4966887417218544, "grad_norm": 0.14150989055633545, "learning_rate": 5.887909614122378e-06, "loss": 0.3683, "step": 678 }, { "epoch": 1.498896247240618, "grad_norm": 0.13479964435100555, "learning_rate": 5.875267040989475e-06, "loss": 0.3651, "step": 679 }, { "epoch": 1.501103752759382, "grad_norm": 0.14743578433990479, "learning_rate": 5.862618692388765e-06, "loss": 0.3676, "step": 680 }, { "epoch": 1.5033112582781456, "grad_norm": 0.1419716477394104, "learning_rate": 5.849964651780637e-06, "loss": 0.367, "step": 681 }, { "epoch": 1.5055187637969096, "grad_norm": 0.15260837972164154, "learning_rate": 5.837305002663043e-06, "loss": 0.3814, "step": 682 }, { "epoch": 1.5077262693156732, "grad_norm": 0.12932439148426056, "learning_rate": 5.824639828570943e-06, "loss": 0.3868, "step": 683 }, { "epoch": 1.5099337748344372, "grad_norm": 0.15234258770942688, "learning_rate": 5.811969213075749e-06, "loss": 0.3695, "step": 684 }, { "epoch": 1.5121412803532008, "grad_norm": 0.1671096384525299, "learning_rate": 5.799293239784783e-06, "loss": 0.3821, "step": 685 }, { "epoch": 1.5143487858719646, "grad_norm": 0.14161601662635803, "learning_rate": 5.786611992340719e-06, "loss": 0.3657, "step": 686 }, { "epoch": 1.5165562913907285, "grad_norm": 0.16359050571918488, "learning_rate": 5.773925554421033e-06, "loss": 0.3788, "step": 687 }, { "epoch": 1.5187637969094923, "grad_norm": 0.14034205675125122, "learning_rate": 5.761234009737449e-06, "loss": 0.3779, "step": 688 }, { "epoch": 1.520971302428256, "grad_norm": 0.14073529839515686, "learning_rate": 5.74853744203539e-06, "loss": 0.3756, "step": 689 }, { "epoch": 1.5231788079470199, "grad_norm": 0.14516787230968475, "learning_rate": 5.735835935093421e-06, "loss": 0.3886, "step": 690 }, { "epoch": 1.5253863134657837, "grad_norm": 0.1415787935256958, "learning_rate": 5.723129572722701e-06, "loss": 0.375, "step": 691 }, { "epoch": 1.5275938189845475, "grad_norm": 0.13726189732551575, "learning_rate": 5.710418438766428e-06, "loss": 0.3735, "step": 692 }, { "epoch": 1.5298013245033113, "grad_norm": 0.14881715178489685, "learning_rate": 5.6977026170992834e-06, "loss": 0.3735, "step": 693 }, { "epoch": 1.532008830022075, "grad_norm": 0.1551184505224228, "learning_rate": 5.68498219162688e-06, "loss": 0.3815, "step": 694 }, { "epoch": 1.534216335540839, "grad_norm": 0.15148243308067322, "learning_rate": 5.672257246285212e-06, "loss": 0.3736, "step": 695 }, { "epoch": 1.5364238410596025, "grad_norm": 0.15206126868724823, "learning_rate": 5.659527865040097e-06, "loss": 0.3771, "step": 696 }, { "epoch": 1.5386313465783665, "grad_norm": 0.14559738337993622, "learning_rate": 5.6467941318866214e-06, "loss": 0.3742, "step": 697 }, { "epoch": 1.5408388520971301, "grad_norm": 0.13176165521144867, "learning_rate": 5.634056130848589e-06, "loss": 0.3715, "step": 698 }, { "epoch": 1.5430463576158941, "grad_norm": 0.14761999249458313, "learning_rate": 5.621313945977964e-06, "loss": 0.377, "step": 699 }, { "epoch": 1.5452538631346577, "grad_norm": 0.12837889790534973, "learning_rate": 5.60856766135432e-06, "loss": 0.371, "step": 700 }, { "epoch": 1.5474613686534218, "grad_norm": 0.1346467137336731, "learning_rate": 5.595817361084283e-06, "loss": 0.3757, "step": 701 }, { "epoch": 1.5496688741721854, "grad_norm": 0.129667267203331, "learning_rate": 5.583063129300971e-06, "loss": 0.3821, "step": 702 }, { "epoch": 1.5518763796909494, "grad_norm": 0.1456080973148346, "learning_rate": 5.570305050163453e-06, "loss": 0.3659, "step": 703 }, { "epoch": 1.554083885209713, "grad_norm": 0.13144616782665253, "learning_rate": 5.557543207856175e-06, "loss": 0.3755, "step": 704 }, { "epoch": 1.5562913907284768, "grad_norm": 0.13695232570171356, "learning_rate": 5.544777686588424e-06, "loss": 0.3704, "step": 705 }, { "epoch": 1.5584988962472406, "grad_norm": 0.1521306186914444, "learning_rate": 5.532008570593756e-06, "loss": 0.3829, "step": 706 }, { "epoch": 1.5607064017660044, "grad_norm": 0.1419302225112915, "learning_rate": 5.519235944129448e-06, "loss": 0.371, "step": 707 }, { "epoch": 1.5629139072847682, "grad_norm": 0.14487622678279877, "learning_rate": 5.506459891475944e-06, "loss": 0.3785, "step": 708 }, { "epoch": 1.565121412803532, "grad_norm": 0.15528716146945953, "learning_rate": 5.4936804969362924e-06, "loss": 0.3853, "step": 709 }, { "epoch": 1.5673289183222958, "grad_norm": 0.14845651388168335, "learning_rate": 5.480897844835595e-06, "loss": 0.3853, "step": 710 }, { "epoch": 1.5695364238410596, "grad_norm": 0.13595354557037354, "learning_rate": 5.4681120195204466e-06, "loss": 0.3718, "step": 711 }, { "epoch": 1.5717439293598234, "grad_norm": 0.15339161455631256, "learning_rate": 5.455323105358384e-06, "loss": 0.3791, "step": 712 }, { "epoch": 1.5739514348785872, "grad_norm": 0.1527385264635086, "learning_rate": 5.442531186737322e-06, "loss": 0.3703, "step": 713 }, { "epoch": 1.576158940397351, "grad_norm": 0.13763274252414703, "learning_rate": 5.429736348065003e-06, "loss": 0.3638, "step": 714 }, { "epoch": 1.5783664459161146, "grad_norm": 0.16397210955619812, "learning_rate": 5.416938673768437e-06, "loss": 0.3761, "step": 715 }, { "epoch": 1.5805739514348787, "grad_norm": 0.14585243165493011, "learning_rate": 5.404138248293346e-06, "loss": 0.3762, "step": 716 }, { "epoch": 1.5827814569536423, "grad_norm": 0.13906413316726685, "learning_rate": 5.3913351561036e-06, "loss": 0.3669, "step": 717 }, { "epoch": 1.5849889624724063, "grad_norm": 0.13377633690834045, "learning_rate": 5.378529481680672e-06, "loss": 0.3721, "step": 718 }, { "epoch": 1.5871964679911699, "grad_norm": 0.14944018423557281, "learning_rate": 5.365721309523072e-06, "loss": 0.3739, "step": 719 }, { "epoch": 1.589403973509934, "grad_norm": 0.14180952310562134, "learning_rate": 5.352910724145789e-06, "loss": 0.373, "step": 720 }, { "epoch": 1.5916114790286975, "grad_norm": 0.1462966352701187, "learning_rate": 5.340097810079741e-06, "loss": 0.3768, "step": 721 }, { "epoch": 1.5938189845474615, "grad_norm": 0.12976975739002228, "learning_rate": 5.327282651871205e-06, "loss": 0.3638, "step": 722 }, { "epoch": 1.596026490066225, "grad_norm": 0.15462516248226166, "learning_rate": 5.3144653340812715e-06, "loss": 0.3854, "step": 723 }, { "epoch": 1.598233995584989, "grad_norm": 0.13364708423614502, "learning_rate": 5.301645941285278e-06, "loss": 0.3862, "step": 724 }, { "epoch": 1.6004415011037527, "grad_norm": 0.14875876903533936, "learning_rate": 5.288824558072257e-06, "loss": 0.3847, "step": 725 }, { "epoch": 1.6026490066225165, "grad_norm": 0.13750705122947693, "learning_rate": 5.276001269044369e-06, "loss": 0.3792, "step": 726 }, { "epoch": 1.6048565121412803, "grad_norm": 0.13552795350551605, "learning_rate": 5.263176158816355e-06, "loss": 0.3824, "step": 727 }, { "epoch": 1.6070640176600441, "grad_norm": 0.13398821651935577, "learning_rate": 5.250349312014976e-06, "loss": 0.3728, "step": 728 }, { "epoch": 1.609271523178808, "grad_norm": 0.14047983288764954, "learning_rate": 5.237520813278443e-06, "loss": 0.3833, "step": 729 }, { "epoch": 1.6114790286975718, "grad_norm": 0.1419335901737213, "learning_rate": 5.224690747255875e-06, "loss": 0.3738, "step": 730 }, { "epoch": 1.6136865342163356, "grad_norm": 0.1423855870962143, "learning_rate": 5.211859198606729e-06, "loss": 0.3702, "step": 731 }, { "epoch": 1.6158940397350994, "grad_norm": 0.14920298755168915, "learning_rate": 5.199026252000245e-06, "loss": 0.3719, "step": 732 }, { "epoch": 1.6181015452538632, "grad_norm": 0.13715742528438568, "learning_rate": 5.186191992114892e-06, "loss": 0.3766, "step": 733 }, { "epoch": 1.6203090507726268, "grad_norm": 0.142774298787117, "learning_rate": 5.173356503637799e-06, "loss": 0.3693, "step": 734 }, { "epoch": 1.6225165562913908, "grad_norm": 0.12615923583507538, "learning_rate": 5.160519871264204e-06, "loss": 0.375, "step": 735 }, { "epoch": 1.6247240618101544, "grad_norm": 0.14882232248783112, "learning_rate": 5.147682179696893e-06, "loss": 0.3746, "step": 736 }, { "epoch": 1.6269315673289184, "grad_norm": 0.12539036571979523, "learning_rate": 5.134843513645642e-06, "loss": 0.3786, "step": 737 }, { "epoch": 1.629139072847682, "grad_norm": 0.14418825507164001, "learning_rate": 5.122003957826657e-06, "loss": 0.3693, "step": 738 }, { "epoch": 1.631346578366446, "grad_norm": 0.14610478281974792, "learning_rate": 5.109163596962013e-06, "loss": 0.3759, "step": 739 }, { "epoch": 1.6335540838852096, "grad_norm": 0.13532301783561707, "learning_rate": 5.096322515779101e-06, "loss": 0.373, "step": 740 }, { "epoch": 1.6357615894039736, "grad_norm": 0.14194665849208832, "learning_rate": 5.083480799010061e-06, "loss": 0.3762, "step": 741 }, { "epoch": 1.6379690949227372, "grad_norm": 0.12951509654521942, "learning_rate": 5.070638531391229e-06, "loss": 0.3723, "step": 742 }, { "epoch": 1.6401766004415013, "grad_norm": 0.14456267654895782, "learning_rate": 5.057795797662577e-06, "loss": 0.3666, "step": 743 }, { "epoch": 1.6423841059602649, "grad_norm": 0.13711507618427277, "learning_rate": 5.044952682567148e-06, "loss": 0.3749, "step": 744 }, { "epoch": 1.6445916114790287, "grad_norm": 0.12987838685512543, "learning_rate": 5.032109270850507e-06, "loss": 0.3779, "step": 745 }, { "epoch": 1.6467991169977925, "grad_norm": 0.1495376080274582, "learning_rate": 5.019265647260173e-06, "loss": 0.3704, "step": 746 }, { "epoch": 1.6490066225165563, "grad_norm": 0.13476036489009857, "learning_rate": 5.006421896545064e-06, "loss": 0.3663, "step": 747 }, { "epoch": 1.65121412803532, "grad_norm": 0.1335316300392151, "learning_rate": 4.993578103454938e-06, "loss": 0.3843, "step": 748 }, { "epoch": 1.653421633554084, "grad_norm": 0.13593193888664246, "learning_rate": 4.980734352739829e-06, "loss": 0.3707, "step": 749 }, { "epoch": 1.6556291390728477, "grad_norm": 0.13324475288391113, "learning_rate": 4.967890729149494e-06, "loss": 0.3746, "step": 750 }, { "epoch": 1.6578366445916115, "grad_norm": 0.12398661673069, "learning_rate": 4.955047317432854e-06, "loss": 0.3783, "step": 751 }, { "epoch": 1.6600441501103753, "grad_norm": 0.12614096701145172, "learning_rate": 4.942204202337425e-06, "loss": 0.3806, "step": 752 }, { "epoch": 1.6622516556291391, "grad_norm": 0.14331743121147156, "learning_rate": 4.929361468608773e-06, "loss": 0.3833, "step": 753 }, { "epoch": 1.664459161147903, "grad_norm": 0.1342383772134781, "learning_rate": 4.9165192009899414e-06, "loss": 0.3799, "step": 754 }, { "epoch": 1.6666666666666665, "grad_norm": 0.13297194242477417, "learning_rate": 4.9036774842209e-06, "loss": 0.3702, "step": 755 }, { "epoch": 1.6688741721854305, "grad_norm": 0.12575410306453705, "learning_rate": 4.890836403037988e-06, "loss": 0.364, "step": 756 }, { "epoch": 1.6710816777041941, "grad_norm": 0.1317722052335739, "learning_rate": 4.877996042173345e-06, "loss": 0.3782, "step": 757 }, { "epoch": 1.6732891832229582, "grad_norm": 0.13363225758075714, "learning_rate": 4.865156486354359e-06, "loss": 0.3804, "step": 758 }, { "epoch": 1.6754966887417218, "grad_norm": 0.14083701372146606, "learning_rate": 4.852317820303107e-06, "loss": 0.38, "step": 759 }, { "epoch": 1.6777041942604858, "grad_norm": 0.1284651756286621, "learning_rate": 4.839480128735798e-06, "loss": 0.3655, "step": 760 }, { "epoch": 1.6799116997792494, "grad_norm": 0.1379299908876419, "learning_rate": 4.826643496362202e-06, "loss": 0.3779, "step": 761 }, { "epoch": 1.6821192052980134, "grad_norm": 0.13508039712905884, "learning_rate": 4.81380800788511e-06, "loss": 0.3785, "step": 762 }, { "epoch": 1.684326710816777, "grad_norm": 0.14211615920066833, "learning_rate": 4.800973747999757e-06, "loss": 0.3648, "step": 763 }, { "epoch": 1.6865342163355408, "grad_norm": 0.1293489634990692, "learning_rate": 4.788140801393273e-06, "loss": 0.3663, "step": 764 }, { "epoch": 1.6887417218543046, "grad_norm": 0.13812755048274994, "learning_rate": 4.775309252744126e-06, "loss": 0.3643, "step": 765 }, { "epoch": 1.6909492273730684, "grad_norm": 0.13112276792526245, "learning_rate": 4.762479186721559e-06, "loss": 0.3703, "step": 766 }, { "epoch": 1.6931567328918322, "grad_norm": 0.12749050557613373, "learning_rate": 4.7496506879850264e-06, "loss": 0.3844, "step": 767 }, { "epoch": 1.695364238410596, "grad_norm": 0.1374199390411377, "learning_rate": 4.736823841183645e-06, "loss": 0.3689, "step": 768 }, { "epoch": 1.6975717439293598, "grad_norm": 0.12659025192260742, "learning_rate": 4.723998730955633e-06, "loss": 0.3606, "step": 769 }, { "epoch": 1.6997792494481236, "grad_norm": 0.11276005953550339, "learning_rate": 4.711175441927746e-06, "loss": 0.3722, "step": 770 }, { "epoch": 1.7019867549668874, "grad_norm": 0.12801696360111237, "learning_rate": 4.698354058714724e-06, "loss": 0.3789, "step": 771 }, { "epoch": 1.7041942604856513, "grad_norm": 0.12910126149654388, "learning_rate": 4.685534665918731e-06, "loss": 0.3669, "step": 772 }, { "epoch": 1.706401766004415, "grad_norm": 0.12605293095111847, "learning_rate": 4.672717348128796e-06, "loss": 0.377, "step": 773 }, { "epoch": 1.7086092715231787, "grad_norm": 0.12655854225158691, "learning_rate": 4.65990218992026e-06, "loss": 0.3738, "step": 774 }, { "epoch": 1.7108167770419427, "grad_norm": 0.1326448917388916, "learning_rate": 4.647089275854212e-06, "loss": 0.3847, "step": 775 }, { "epoch": 1.7130242825607063, "grad_norm": 0.12570975720882416, "learning_rate": 4.63427869047693e-06, "loss": 0.3652, "step": 776 }, { "epoch": 1.7152317880794703, "grad_norm": 0.13106000423431396, "learning_rate": 4.621470518319329e-06, "loss": 0.3781, "step": 777 }, { "epoch": 1.7174392935982339, "grad_norm": 0.12707674503326416, "learning_rate": 4.608664843896402e-06, "loss": 0.367, "step": 778 }, { "epoch": 1.719646799116998, "grad_norm": 0.1342884600162506, "learning_rate": 4.595861751706656e-06, "loss": 0.3827, "step": 779 }, { "epoch": 1.7218543046357615, "grad_norm": 0.15550127625465393, "learning_rate": 4.583061326231564e-06, "loss": 0.384, "step": 780 }, { "epoch": 1.7240618101545255, "grad_norm": 0.1267244517803192, "learning_rate": 4.570263651934998e-06, "loss": 0.38, "step": 781 }, { "epoch": 1.7262693156732891, "grad_norm": 0.13444717228412628, "learning_rate": 4.557468813262678e-06, "loss": 0.3856, "step": 782 }, { "epoch": 1.728476821192053, "grad_norm": 0.12585113942623138, "learning_rate": 4.544676894641617e-06, "loss": 0.3603, "step": 783 }, { "epoch": 1.7306843267108167, "grad_norm": 0.14303074777126312, "learning_rate": 4.531887980479555e-06, "loss": 0.3769, "step": 784 }, { "epoch": 1.7328918322295805, "grad_norm": 0.14289291203022003, "learning_rate": 4.519102155164407e-06, "loss": 0.3679, "step": 785 }, { "epoch": 1.7350993377483444, "grad_norm": 0.13349799811840057, "learning_rate": 4.506319503063708e-06, "loss": 0.3702, "step": 786 }, { "epoch": 1.7373068432671082, "grad_norm": 0.15247680246829987, "learning_rate": 4.493540108524057e-06, "loss": 0.3713, "step": 787 }, { "epoch": 1.739514348785872, "grad_norm": 0.14266639947891235, "learning_rate": 4.480764055870553e-06, "loss": 0.3742, "step": 788 }, { "epoch": 1.7417218543046358, "grad_norm": 0.13838379085063934, "learning_rate": 4.467991429406247e-06, "loss": 0.3857, "step": 789 }, { "epoch": 1.7439293598233996, "grad_norm": 0.13494673371315002, "learning_rate": 4.455222313411578e-06, "loss": 0.3783, "step": 790 }, { "epoch": 1.7461368653421634, "grad_norm": 0.12723763287067413, "learning_rate": 4.442456792143825e-06, "loss": 0.3657, "step": 791 }, { "epoch": 1.7483443708609272, "grad_norm": 0.1302330642938614, "learning_rate": 4.4296949498365496e-06, "loss": 0.3671, "step": 792 }, { "epoch": 1.7505518763796908, "grad_norm": 0.12404655665159225, "learning_rate": 4.416936870699031e-06, "loss": 0.3755, "step": 793 }, { "epoch": 1.7527593818984548, "grad_norm": 0.13801135122776031, "learning_rate": 4.40418263891572e-06, "loss": 0.377, "step": 794 }, { "epoch": 1.7549668874172184, "grad_norm": 0.1327562779188156, "learning_rate": 4.391432338645681e-06, "loss": 0.3649, "step": 795 }, { "epoch": 1.7571743929359824, "grad_norm": 0.13369016349315643, "learning_rate": 4.378686054022037e-06, "loss": 0.3882, "step": 796 }, { "epoch": 1.759381898454746, "grad_norm": 0.14249202609062195, "learning_rate": 4.365943869151412e-06, "loss": 0.3763, "step": 797 }, { "epoch": 1.76158940397351, "grad_norm": 0.1298934519290924, "learning_rate": 4.35320586811338e-06, "loss": 0.3742, "step": 798 }, { "epoch": 1.7637969094922736, "grad_norm": 0.14034999907016754, "learning_rate": 4.3404721349599044e-06, "loss": 0.3685, "step": 799 }, { "epoch": 1.7660044150110377, "grad_norm": 0.12729597091674805, "learning_rate": 4.327742753714788e-06, "loss": 0.3742, "step": 800 }, { "epoch": 1.7682119205298013, "grad_norm": 0.13787223398685455, "learning_rate": 4.315017808373121e-06, "loss": 0.3718, "step": 801 }, { "epoch": 1.7704194260485653, "grad_norm": 0.13648386299610138, "learning_rate": 4.302297382900718e-06, "loss": 0.3722, "step": 802 }, { "epoch": 1.7726269315673289, "grad_norm": 0.12655648589134216, "learning_rate": 4.289581561233574e-06, "loss": 0.3667, "step": 803 }, { "epoch": 1.7748344370860927, "grad_norm": 0.12654156982898712, "learning_rate": 4.276870427277299e-06, "loss": 0.3666, "step": 804 }, { "epoch": 1.7770419426048565, "grad_norm": 0.15781165659427643, "learning_rate": 4.264164064906581e-06, "loss": 0.3787, "step": 805 }, { "epoch": 1.7792494481236203, "grad_norm": 0.12070343643426895, "learning_rate": 4.251462557964612e-06, "loss": 0.3897, "step": 806 }, { "epoch": 1.781456953642384, "grad_norm": 0.1305495798587799, "learning_rate": 4.238765990262554e-06, "loss": 0.3622, "step": 807 }, { "epoch": 1.783664459161148, "grad_norm": 0.12424585223197937, "learning_rate": 4.226074445578969e-06, "loss": 0.3732, "step": 808 }, { "epoch": 1.7858719646799117, "grad_norm": 0.12352015823125839, "learning_rate": 4.213388007659281e-06, "loss": 0.367, "step": 809 }, { "epoch": 1.7880794701986755, "grad_norm": 0.1388520896434784, "learning_rate": 4.200706760215219e-06, "loss": 0.3723, "step": 810 }, { "epoch": 1.7902869757174393, "grad_norm": 0.13200843334197998, "learning_rate": 4.188030786924252e-06, "loss": 0.3765, "step": 811 }, { "epoch": 1.7924944812362031, "grad_norm": 0.11911409348249435, "learning_rate": 4.17536017142906e-06, "loss": 0.3694, "step": 812 }, { "epoch": 1.794701986754967, "grad_norm": 0.1263428032398224, "learning_rate": 4.162694997336957e-06, "loss": 0.3656, "step": 813 }, { "epoch": 1.7969094922737305, "grad_norm": 0.125799298286438, "learning_rate": 4.1500353482193646e-06, "loss": 0.3686, "step": 814 }, { "epoch": 1.7991169977924946, "grad_norm": 0.13606040179729462, "learning_rate": 4.137381307611236e-06, "loss": 0.3579, "step": 815 }, { "epoch": 1.8013245033112582, "grad_norm": 0.13091742992401123, "learning_rate": 4.124732959010526e-06, "loss": 0.369, "step": 816 }, { "epoch": 1.8035320088300222, "grad_norm": 0.13721388578414917, "learning_rate": 4.112090385877624e-06, "loss": 0.3673, "step": 817 }, { "epoch": 1.8057395143487858, "grad_norm": 0.12669645249843597, "learning_rate": 4.099453671634811e-06, "loss": 0.3778, "step": 818 }, { "epoch": 1.8079470198675498, "grad_norm": 0.12563163042068481, "learning_rate": 4.086822899665713e-06, "loss": 0.3687, "step": 819 }, { "epoch": 1.8101545253863134, "grad_norm": 0.1383393555879593, "learning_rate": 4.07419815331474e-06, "loss": 0.3787, "step": 820 }, { "epoch": 1.8123620309050774, "grad_norm": 0.1277354508638382, "learning_rate": 4.061579515886544e-06, "loss": 0.3752, "step": 821 }, { "epoch": 1.814569536423841, "grad_norm": 0.12579599022865295, "learning_rate": 4.048967070645465e-06, "loss": 0.3767, "step": 822 }, { "epoch": 1.8167770419426048, "grad_norm": 0.1388178914785385, "learning_rate": 4.03636090081499e-06, "loss": 0.3795, "step": 823 }, { "epoch": 1.8189845474613686, "grad_norm": 0.13451789319515228, "learning_rate": 4.0237610895771895e-06, "loss": 0.3636, "step": 824 }, { "epoch": 1.8211920529801324, "grad_norm": 0.1298052817583084, "learning_rate": 4.011167720072183e-06, "loss": 0.3831, "step": 825 }, { "epoch": 1.8233995584988962, "grad_norm": 0.13578468561172485, "learning_rate": 3.99858087539758e-06, "loss": 0.3787, "step": 826 }, { "epoch": 1.82560706401766, "grad_norm": 0.1320340931415558, "learning_rate": 3.986000638607937e-06, "loss": 0.3697, "step": 827 }, { "epoch": 1.8278145695364238, "grad_norm": 0.13334733247756958, "learning_rate": 3.973427092714212e-06, "loss": 0.361, "step": 828 }, { "epoch": 1.8300220750551877, "grad_norm": 0.12365536391735077, "learning_rate": 3.960860320683205e-06, "loss": 0.3716, "step": 829 }, { "epoch": 1.8322295805739515, "grad_norm": 0.1276884227991104, "learning_rate": 3.948300405437026e-06, "loss": 0.3717, "step": 830 }, { "epoch": 1.8344370860927153, "grad_norm": 0.13205379247665405, "learning_rate": 3.9357474298525356e-06, "loss": 0.3632, "step": 831 }, { "epoch": 1.836644591611479, "grad_norm": 0.12293750792741776, "learning_rate": 3.9232014767608065e-06, "loss": 0.3809, "step": 832 }, { "epoch": 1.8388520971302427, "grad_norm": 0.1293838620185852, "learning_rate": 3.910662628946573e-06, "loss": 0.3782, "step": 833 }, { "epoch": 1.8410596026490067, "grad_norm": 0.1325875222682953, "learning_rate": 3.89813096914768e-06, "loss": 0.3734, "step": 834 }, { "epoch": 1.8432671081677703, "grad_norm": 0.12951777875423431, "learning_rate": 3.8856065800545475e-06, "loss": 0.3828, "step": 835 }, { "epoch": 1.8454746136865343, "grad_norm": 0.12658190727233887, "learning_rate": 3.873089544309616e-06, "loss": 0.3753, "step": 836 }, { "epoch": 1.847682119205298, "grad_norm": 0.12527629733085632, "learning_rate": 3.860579944506811e-06, "loss": 0.3716, "step": 837 }, { "epoch": 1.849889624724062, "grad_norm": 0.12750259041786194, "learning_rate": 3.848077863190985e-06, "loss": 0.3727, "step": 838 }, { "epoch": 1.8520971302428255, "grad_norm": 0.13881564140319824, "learning_rate": 3.835583382857382e-06, "loss": 0.3743, "step": 839 }, { "epoch": 1.8543046357615895, "grad_norm": 0.12845860421657562, "learning_rate": 3.82309658595109e-06, "loss": 0.3692, "step": 840 }, { "epoch": 1.8565121412803531, "grad_norm": 0.12954264879226685, "learning_rate": 3.8106175548664994e-06, "loss": 0.3778, "step": 841 }, { "epoch": 1.858719646799117, "grad_norm": 0.13028046488761902, "learning_rate": 3.7981463719467587e-06, "loss": 0.3825, "step": 842 }, { "epoch": 1.8609271523178808, "grad_norm": 0.1258542686700821, "learning_rate": 3.7856831194832262e-06, "loss": 0.3817, "step": 843 }, { "epoch": 1.8631346578366446, "grad_norm": 0.1192869246006012, "learning_rate": 3.7732278797149324e-06, "loss": 0.3706, "step": 844 }, { "epoch": 1.8653421633554084, "grad_norm": 0.13153821229934692, "learning_rate": 3.7607807348280364e-06, "loss": 0.3626, "step": 845 }, { "epoch": 1.8675496688741722, "grad_norm": 0.12072479724884033, "learning_rate": 3.7483417669552857e-06, "loss": 0.3557, "step": 846 }, { "epoch": 1.869757174392936, "grad_norm": 0.12467250972986221, "learning_rate": 3.735911058175467e-06, "loss": 0.3621, "step": 847 }, { "epoch": 1.8719646799116998, "grad_norm": 0.12176632136106491, "learning_rate": 3.7234886905128724e-06, "loss": 0.3726, "step": 848 }, { "epoch": 1.8741721854304636, "grad_norm": 0.1291275918483734, "learning_rate": 3.711074745936751e-06, "loss": 0.3663, "step": 849 }, { "epoch": 1.8763796909492274, "grad_norm": 0.1401110738515854, "learning_rate": 3.698669306360778e-06, "loss": 0.3763, "step": 850 }, { "epoch": 1.8785871964679912, "grad_norm": 0.12075633555650711, "learning_rate": 3.6862724536425042e-06, "loss": 0.3761, "step": 851 }, { "epoch": 1.8807947019867548, "grad_norm": 0.12913060188293457, "learning_rate": 3.67388426958282e-06, "loss": 0.3735, "step": 852 }, { "epoch": 1.8830022075055188, "grad_norm": 0.13538530468940735, "learning_rate": 3.6615048359254155e-06, "loss": 0.3737, "step": 853 }, { "epoch": 1.8852097130242824, "grad_norm": 0.1396217793226242, "learning_rate": 3.649134234356242e-06, "loss": 0.3765, "step": 854 }, { "epoch": 1.8874172185430464, "grad_norm": 0.12726326286792755, "learning_rate": 3.636772546502973e-06, "loss": 0.3706, "step": 855 }, { "epoch": 1.88962472406181, "grad_norm": 0.12331200391054153, "learning_rate": 3.624419853934461e-06, "loss": 0.3698, "step": 856 }, { "epoch": 1.891832229580574, "grad_norm": 0.12955856323242188, "learning_rate": 3.612076238160209e-06, "loss": 0.3779, "step": 857 }, { "epoch": 1.8940397350993377, "grad_norm": 0.1281704604625702, "learning_rate": 3.5997417806298172e-06, "loss": 0.3711, "step": 858 }, { "epoch": 1.8962472406181017, "grad_norm": 0.13094061613082886, "learning_rate": 3.5874165627324665e-06, "loss": 0.3706, "step": 859 }, { "epoch": 1.8984547461368653, "grad_norm": 0.1244734600186348, "learning_rate": 3.575100665796362e-06, "loss": 0.3701, "step": 860 }, { "epoch": 1.9006622516556293, "grad_norm": 0.13025392591953278, "learning_rate": 3.5627941710882042e-06, "loss": 0.3774, "step": 861 }, { "epoch": 1.9028697571743929, "grad_norm": 0.13566212356090546, "learning_rate": 3.550497159812658e-06, "loss": 0.376, "step": 862 }, { "epoch": 1.9050772626931567, "grad_norm": 0.13212575018405914, "learning_rate": 3.5382097131118052e-06, "loss": 0.3759, "step": 863 }, { "epoch": 1.9072847682119205, "grad_norm": 0.13214178383350372, "learning_rate": 3.525931912064623e-06, "loss": 0.3706, "step": 864 }, { "epoch": 1.9094922737306843, "grad_norm": 0.1290368288755417, "learning_rate": 3.5136638376864333e-06, "loss": 0.3767, "step": 865 }, { "epoch": 1.9116997792494481, "grad_norm": 0.1320725828409195, "learning_rate": 3.5014055709283847e-06, "loss": 0.3704, "step": 866 }, { "epoch": 1.913907284768212, "grad_norm": 0.13395248353481293, "learning_rate": 3.4891571926769e-06, "loss": 0.3774, "step": 867 }, { "epoch": 1.9161147902869757, "grad_norm": 0.12245456129312515, "learning_rate": 3.4769187837531637e-06, "loss": 0.3872, "step": 868 }, { "epoch": 1.9183222958057395, "grad_norm": 0.1327778398990631, "learning_rate": 3.4646904249125713e-06, "loss": 0.3666, "step": 869 }, { "epoch": 1.9205298013245033, "grad_norm": 0.13339614868164062, "learning_rate": 3.4524721968442017e-06, "loss": 0.3782, "step": 870 }, { "epoch": 1.9227373068432672, "grad_norm": 0.12259134650230408, "learning_rate": 3.4402641801702883e-06, "loss": 0.38, "step": 871 }, { "epoch": 1.924944812362031, "grad_norm": 0.12124523520469666, "learning_rate": 3.4280664554456823e-06, "loss": 0.3746, "step": 872 }, { "epoch": 1.9271523178807946, "grad_norm": 0.12291016429662704, "learning_rate": 3.415879103157327e-06, "loss": 0.3697, "step": 873 }, { "epoch": 1.9293598233995586, "grad_norm": 0.12792356312274933, "learning_rate": 3.403702203723718e-06, "loss": 0.3755, "step": 874 }, { "epoch": 1.9315673289183222, "grad_norm": 0.1225101426243782, "learning_rate": 3.3915358374943813e-06, "loss": 0.3821, "step": 875 }, { "epoch": 1.9337748344370862, "grad_norm": 0.1265992522239685, "learning_rate": 3.379380084749336e-06, "loss": 0.3672, "step": 876 }, { "epoch": 1.9359823399558498, "grad_norm": 0.12633873522281647, "learning_rate": 3.3672350256985727e-06, "loss": 0.3766, "step": 877 }, { "epoch": 1.9381898454746138, "grad_norm": 0.1319660246372223, "learning_rate": 3.3551007404815162e-06, "loss": 0.374, "step": 878 }, { "epoch": 1.9403973509933774, "grad_norm": 0.12871594727039337, "learning_rate": 3.3429773091664985e-06, "loss": 0.3673, "step": 879 }, { "epoch": 1.9426048565121414, "grad_norm": 0.13455092906951904, "learning_rate": 3.330864811750235e-06, "loss": 0.3805, "step": 880 }, { "epoch": 1.944812362030905, "grad_norm": 0.13253872096538544, "learning_rate": 3.3187633281572913e-06, "loss": 0.3695, "step": 881 }, { "epoch": 1.9470198675496688, "grad_norm": 0.11975626647472382, "learning_rate": 3.306672938239561e-06, "loss": 0.3767, "step": 882 }, { "epoch": 1.9492273730684326, "grad_norm": 0.12552706897258759, "learning_rate": 3.2945937217757324e-06, "loss": 0.363, "step": 883 }, { "epoch": 1.9514348785871964, "grad_norm": 0.13104894757270813, "learning_rate": 3.282525758470768e-06, "loss": 0.3699, "step": 884 }, { "epoch": 1.9536423841059603, "grad_norm": 0.12783929705619812, "learning_rate": 3.2704691279553725e-06, "loss": 0.3681, "step": 885 }, { "epoch": 1.955849889624724, "grad_norm": 0.12647448480129242, "learning_rate": 3.2584239097854763e-06, "loss": 0.3873, "step": 886 }, { "epoch": 1.9580573951434879, "grad_norm": 0.11794891208410263, "learning_rate": 3.246390183441703e-06, "loss": 0.3689, "step": 887 }, { "epoch": 1.9602649006622517, "grad_norm": 0.1255064755678177, "learning_rate": 3.234368028328845e-06, "loss": 0.3874, "step": 888 }, { "epoch": 1.9624724061810155, "grad_norm": 0.12461333721876144, "learning_rate": 3.2223575237753448e-06, "loss": 0.3719, "step": 889 }, { "epoch": 1.9646799116997793, "grad_norm": 0.13075792789459229, "learning_rate": 3.2103587490327624e-06, "loss": 0.3746, "step": 890 }, { "epoch": 1.966887417218543, "grad_norm": 0.13413941860198975, "learning_rate": 3.198371783275269e-06, "loss": 0.3723, "step": 891 }, { "epoch": 1.9690949227373067, "grad_norm": 0.13534380495548248, "learning_rate": 3.186396705599106e-06, "loss": 0.3762, "step": 892 }, { "epoch": 1.9713024282560707, "grad_norm": 0.12702573835849762, "learning_rate": 3.1744335950220716e-06, "loss": 0.3763, "step": 893 }, { "epoch": 1.9735099337748343, "grad_norm": 0.131902277469635, "learning_rate": 3.162482530483e-06, "loss": 0.3647, "step": 894 }, { "epoch": 1.9757174392935983, "grad_norm": 0.12826719880104065, "learning_rate": 3.1505435908412415e-06, "loss": 0.3705, "step": 895 }, { "epoch": 1.977924944812362, "grad_norm": 0.1235160231590271, "learning_rate": 3.1386168548761377e-06, "loss": 0.3806, "step": 896 }, { "epoch": 1.980132450331126, "grad_norm": 0.13026514649391174, "learning_rate": 3.126702401286503e-06, "loss": 0.3714, "step": 897 }, { "epoch": 1.9823399558498895, "grad_norm": 0.12384995073080063, "learning_rate": 3.11480030869011e-06, "loss": 0.3775, "step": 898 }, { "epoch": 1.9845474613686536, "grad_norm": 0.11622393131256104, "learning_rate": 3.1029106556231615e-06, "loss": 0.3685, "step": 899 }, { "epoch": 1.9867549668874172, "grad_norm": 0.12467402219772339, "learning_rate": 3.0910335205397834e-06, "loss": 0.3753, "step": 900 }, { "epoch": 1.9889624724061812, "grad_norm": 0.12134755402803421, "learning_rate": 3.0791689818114988e-06, "loss": 0.3763, "step": 901 }, { "epoch": 1.9911699779249448, "grad_norm": 0.118636354804039, "learning_rate": 3.067317117726715e-06, "loss": 0.3779, "step": 902 }, { "epoch": 1.9933774834437086, "grad_norm": 0.12621286511421204, "learning_rate": 3.055478006490199e-06, "loss": 0.3755, "step": 903 }, { "epoch": 1.9955849889624724, "grad_norm": 0.13427864015102386, "learning_rate": 3.043651726222578e-06, "loss": 0.3677, "step": 904 }, { "epoch": 1.9977924944812362, "grad_norm": 0.1349465250968933, "learning_rate": 3.0318383549598085e-06, "loss": 0.3821, "step": 905 }, { "epoch": 2.0, "grad_norm": 0.1383267641067505, "learning_rate": 3.020037970652664e-06, "loss": 0.3605, "step": 906 }, { "epoch": 2.0022075055187636, "grad_norm": 0.1400349736213684, "learning_rate": 3.00825065116623e-06, "loss": 0.3587, "step": 907 }, { "epoch": 2.0044150110375276, "grad_norm": 0.12129946798086166, "learning_rate": 2.996476474279375e-06, "loss": 0.3564, "step": 908 }, { "epoch": 2.006622516556291, "grad_norm": 0.13566596806049347, "learning_rate": 2.9847155176842547e-06, "loss": 0.3628, "step": 909 }, { "epoch": 2.0088300220750552, "grad_norm": 0.12823748588562012, "learning_rate": 2.9729678589857852e-06, "loss": 0.3589, "step": 910 }, { "epoch": 2.011037527593819, "grad_norm": 0.1251477748155594, "learning_rate": 2.9612335757011362e-06, "loss": 0.3583, "step": 911 }, { "epoch": 2.013245033112583, "grad_norm": 0.1308368593454361, "learning_rate": 2.94951274525922e-06, "loss": 0.354, "step": 912 }, { "epoch": 2.0154525386313464, "grad_norm": 0.1357547789812088, "learning_rate": 2.9378054450001836e-06, "loss": 0.3664, "step": 913 }, { "epoch": 2.0176600441501105, "grad_norm": 0.13018201291561127, "learning_rate": 2.9261117521748904e-06, "loss": 0.3627, "step": 914 }, { "epoch": 2.019867549668874, "grad_norm": 0.12550939619541168, "learning_rate": 2.914431743944414e-06, "loss": 0.3518, "step": 915 }, { "epoch": 2.022075055187638, "grad_norm": 0.12784212827682495, "learning_rate": 2.902765497379534e-06, "loss": 0.3547, "step": 916 }, { "epoch": 2.0242825607064017, "grad_norm": 0.13502328097820282, "learning_rate": 2.8911130894602198e-06, "loss": 0.3648, "step": 917 }, { "epoch": 2.0264900662251657, "grad_norm": 0.12056614458560944, "learning_rate": 2.8794745970751308e-06, "loss": 0.3612, "step": 918 }, { "epoch": 2.0286975717439293, "grad_norm": 0.12976831197738647, "learning_rate": 2.8678500970210977e-06, "loss": 0.3619, "step": 919 }, { "epoch": 2.0309050772626933, "grad_norm": 0.13974756002426147, "learning_rate": 2.8562396660026304e-06, "loss": 0.3585, "step": 920 }, { "epoch": 2.033112582781457, "grad_norm": 0.1298479437828064, "learning_rate": 2.8446433806313966e-06, "loss": 0.3533, "step": 921 }, { "epoch": 2.035320088300221, "grad_norm": 0.13592906296253204, "learning_rate": 2.83306131742573e-06, "loss": 0.36, "step": 922 }, { "epoch": 2.0375275938189845, "grad_norm": 0.13170795142650604, "learning_rate": 2.821493552810114e-06, "loss": 0.3535, "step": 923 }, { "epoch": 2.0397350993377485, "grad_norm": 0.14562806487083435, "learning_rate": 2.809940163114685e-06, "loss": 0.3737, "step": 924 }, { "epoch": 2.041942604856512, "grad_norm": 0.12969909608364105, "learning_rate": 2.7984012245747237e-06, "loss": 0.3597, "step": 925 }, { "epoch": 2.044150110375276, "grad_norm": 0.13986288011074066, "learning_rate": 2.786876813330158e-06, "loss": 0.3608, "step": 926 }, { "epoch": 2.0463576158940397, "grad_norm": 0.13949252665042877, "learning_rate": 2.7753670054250583e-06, "loss": 0.3507, "step": 927 }, { "epoch": 2.0485651214128033, "grad_norm": 0.12819348275661469, "learning_rate": 2.763871876807129e-06, "loss": 0.356, "step": 928 }, { "epoch": 2.0507726269315674, "grad_norm": 0.13626167178153992, "learning_rate": 2.7523915033272163e-06, "loss": 0.3534, "step": 929 }, { "epoch": 2.052980132450331, "grad_norm": 0.14214330911636353, "learning_rate": 2.740925960738802e-06, "loss": 0.3627, "step": 930 }, { "epoch": 2.055187637969095, "grad_norm": 0.1371086835861206, "learning_rate": 2.72947532469751e-06, "loss": 0.3594, "step": 931 }, { "epoch": 2.0573951434878586, "grad_norm": 0.15132765471935272, "learning_rate": 2.7180396707606023e-06, "loss": 0.3487, "step": 932 }, { "epoch": 2.0596026490066226, "grad_norm": 0.1387225240468979, "learning_rate": 2.7066190743864774e-06, "loss": 0.3625, "step": 933 }, { "epoch": 2.061810154525386, "grad_norm": 0.13822317123413086, "learning_rate": 2.695213610934179e-06, "loss": 0.3542, "step": 934 }, { "epoch": 2.06401766004415, "grad_norm": 0.1481810212135315, "learning_rate": 2.6838233556628932e-06, "loss": 0.3543, "step": 935 }, { "epoch": 2.066225165562914, "grad_norm": 0.1475333869457245, "learning_rate": 2.672448383731461e-06, "loss": 0.3587, "step": 936 }, { "epoch": 2.068432671081678, "grad_norm": 0.13259293138980865, "learning_rate": 2.661088770197868e-06, "loss": 0.3531, "step": 937 }, { "epoch": 2.0706401766004414, "grad_norm": 0.129195436835289, "learning_rate": 2.6497445900187635e-06, "loss": 0.3557, "step": 938 }, { "epoch": 2.0728476821192054, "grad_norm": 0.12847961485385895, "learning_rate": 2.6384159180489542e-06, "loss": 0.3477, "step": 939 }, { "epoch": 2.075055187637969, "grad_norm": 0.13196007907390594, "learning_rate": 2.6271028290409216e-06, "loss": 0.3584, "step": 940 }, { "epoch": 2.077262693156733, "grad_norm": 0.1364007443189621, "learning_rate": 2.6158053976443164e-06, "loss": 0.3601, "step": 941 }, { "epoch": 2.0794701986754967, "grad_norm": 0.1247749775648117, "learning_rate": 2.604523698405477e-06, "loss": 0.3647, "step": 942 }, { "epoch": 2.0816777041942607, "grad_norm": 0.12793247401714325, "learning_rate": 2.5932578057669273e-06, "loss": 0.3656, "step": 943 }, { "epoch": 2.0838852097130243, "grad_norm": 0.12500843405723572, "learning_rate": 2.5820077940668975e-06, "loss": 0.3511, "step": 944 }, { "epoch": 2.0860927152317883, "grad_norm": 0.13672660291194916, "learning_rate": 2.570773737538825e-06, "loss": 0.3709, "step": 945 }, { "epoch": 2.088300220750552, "grad_norm": 0.1374640017747879, "learning_rate": 2.559555710310866e-06, "loss": 0.3528, "step": 946 }, { "epoch": 2.0905077262693155, "grad_norm": 0.13424260914325714, "learning_rate": 2.5483537864054055e-06, "loss": 0.354, "step": 947 }, { "epoch": 2.0927152317880795, "grad_norm": 0.13996686041355133, "learning_rate": 2.537168039738569e-06, "loss": 0.3655, "step": 948 }, { "epoch": 2.094922737306843, "grad_norm": 0.12471663951873779, "learning_rate": 2.525998544119743e-06, "loss": 0.3606, "step": 949 }, { "epoch": 2.097130242825607, "grad_norm": 0.13321642577648163, "learning_rate": 2.514845373251078e-06, "loss": 0.3477, "step": 950 }, { "epoch": 2.0993377483443707, "grad_norm": 0.1407744139432907, "learning_rate": 2.5037086007270017e-06, "loss": 0.3566, "step": 951 }, { "epoch": 2.1015452538631347, "grad_norm": 0.1319989413022995, "learning_rate": 2.4925883000337407e-06, "loss": 0.3511, "step": 952 }, { "epoch": 2.1037527593818983, "grad_norm": 0.12347118556499481, "learning_rate": 2.4814845445488288e-06, "loss": 0.3609, "step": 953 }, { "epoch": 2.1059602649006623, "grad_norm": 0.15142813324928284, "learning_rate": 2.4703974075406316e-06, "loss": 0.3581, "step": 954 }, { "epoch": 2.108167770419426, "grad_norm": 0.12996384501457214, "learning_rate": 2.4593269621678502e-06, "loss": 0.3586, "step": 955 }, { "epoch": 2.11037527593819, "grad_norm": 0.12407524138689041, "learning_rate": 2.448273281479052e-06, "loss": 0.3561, "step": 956 }, { "epoch": 2.1125827814569536, "grad_norm": 0.12946733832359314, "learning_rate": 2.437236438412177e-06, "loss": 0.3637, "step": 957 }, { "epoch": 2.1147902869757176, "grad_norm": 0.14091919362545013, "learning_rate": 2.426216505794066e-06, "loss": 0.3572, "step": 958 }, { "epoch": 2.116997792494481, "grad_norm": 0.13770008087158203, "learning_rate": 2.4152135563399743e-06, "loss": 0.3617, "step": 959 }, { "epoch": 2.119205298013245, "grad_norm": 0.1215134859085083, "learning_rate": 2.404227662653089e-06, "loss": 0.3613, "step": 960 }, { "epoch": 2.121412803532009, "grad_norm": 0.13981877267360687, "learning_rate": 2.3932588972240642e-06, "loss": 0.3601, "step": 961 }, { "epoch": 2.123620309050773, "grad_norm": 0.13256755471229553, "learning_rate": 2.3823073324305222e-06, "loss": 0.3551, "step": 962 }, { "epoch": 2.1258278145695364, "grad_norm": 0.14095091819763184, "learning_rate": 2.3713730405365965e-06, "loss": 0.3671, "step": 963 }, { "epoch": 2.1280353200883004, "grad_norm": 0.11968285590410233, "learning_rate": 2.3604560936924363e-06, "loss": 0.3558, "step": 964 }, { "epoch": 2.130242825607064, "grad_norm": 0.13734926283359528, "learning_rate": 2.3495565639337446e-06, "loss": 0.3653, "step": 965 }, { "epoch": 2.1324503311258276, "grad_norm": 0.13910116255283356, "learning_rate": 2.3386745231812923e-06, "loss": 0.3559, "step": 966 }, { "epoch": 2.1346578366445916, "grad_norm": 0.13142208755016327, "learning_rate": 2.3278100432404554e-06, "loss": 0.3571, "step": 967 }, { "epoch": 2.136865342163355, "grad_norm": 0.12986791133880615, "learning_rate": 2.3169631958007316e-06, "loss": 0.3613, "step": 968 }, { "epoch": 2.1390728476821192, "grad_norm": 0.12400130182504654, "learning_rate": 2.3061340524352687e-06, "loss": 0.3596, "step": 969 }, { "epoch": 2.141280353200883, "grad_norm": 0.12488456070423126, "learning_rate": 2.295322684600395e-06, "loss": 0.3571, "step": 970 }, { "epoch": 2.143487858719647, "grad_norm": 0.13525961339473724, "learning_rate": 2.284529163635144e-06, "loss": 0.352, "step": 971 }, { "epoch": 2.1456953642384105, "grad_norm": 0.14086809754371643, "learning_rate": 2.273753560760793e-06, "loss": 0.3529, "step": 972 }, { "epoch": 2.1479028697571745, "grad_norm": 0.13002382218837738, "learning_rate": 2.262995947080378e-06, "loss": 0.3628, "step": 973 }, { "epoch": 2.150110375275938, "grad_norm": 0.13714134693145752, "learning_rate": 2.252256393578241e-06, "loss": 0.3602, "step": 974 }, { "epoch": 2.152317880794702, "grad_norm": 0.1353500783443451, "learning_rate": 2.2415349711195455e-06, "loss": 0.3562, "step": 975 }, { "epoch": 2.1545253863134657, "grad_norm": 0.14100465178489685, "learning_rate": 2.230831750449825e-06, "loss": 0.3562, "step": 976 }, { "epoch": 2.1567328918322297, "grad_norm": 0.11746090650558472, "learning_rate": 2.2201468021945024e-06, "loss": 0.3679, "step": 977 }, { "epoch": 2.1589403973509933, "grad_norm": 0.13000161945819855, "learning_rate": 2.209480196858429e-06, "loss": 0.3579, "step": 978 }, { "epoch": 2.1611479028697573, "grad_norm": 0.14291274547576904, "learning_rate": 2.198832004825427e-06, "loss": 0.366, "step": 979 }, { "epoch": 2.163355408388521, "grad_norm": 0.1327233463525772, "learning_rate": 2.1882022963578087e-06, "loss": 0.3576, "step": 980 }, { "epoch": 2.165562913907285, "grad_norm": 0.12592221796512604, "learning_rate": 2.177591141595931e-06, "loss": 0.3554, "step": 981 }, { "epoch": 2.1677704194260485, "grad_norm": 0.13044218719005585, "learning_rate": 2.166998610557718e-06, "loss": 0.3489, "step": 982 }, { "epoch": 2.1699779249448126, "grad_norm": 0.14413036406040192, "learning_rate": 2.1564247731382063e-06, "loss": 0.3653, "step": 983 }, { "epoch": 2.172185430463576, "grad_norm": 0.12830859422683716, "learning_rate": 2.145869699109081e-06, "loss": 0.3566, "step": 984 }, { "epoch": 2.1743929359823397, "grad_norm": 0.12629762291908264, "learning_rate": 2.1353334581182193e-06, "loss": 0.3581, "step": 985 }, { "epoch": 2.1766004415011038, "grad_norm": 0.12835480272769928, "learning_rate": 2.1248161196892295e-06, "loss": 0.3655, "step": 986 }, { "epoch": 2.1788079470198674, "grad_norm": 0.12383104115724564, "learning_rate": 2.1143177532209855e-06, "loss": 0.3564, "step": 987 }, { "epoch": 2.1810154525386314, "grad_norm": 0.1490948498249054, "learning_rate": 2.1038384279871786e-06, "loss": 0.37, "step": 988 }, { "epoch": 2.183222958057395, "grad_norm": 0.12230035662651062, "learning_rate": 2.0933782131358516e-06, "loss": 0.3549, "step": 989 }, { "epoch": 2.185430463576159, "grad_norm": 0.13511480391025543, "learning_rate": 2.082937177688952e-06, "loss": 0.3576, "step": 990 }, { "epoch": 2.1876379690949226, "grad_norm": 0.135064497590065, "learning_rate": 2.0725153905418726e-06, "loss": 0.3544, "step": 991 }, { "epoch": 2.1898454746136866, "grad_norm": 0.12062691897153854, "learning_rate": 2.0621129204629907e-06, "loss": 0.3523, "step": 992 }, { "epoch": 2.19205298013245, "grad_norm": 0.14147672057151794, "learning_rate": 2.0517298360932202e-06, "loss": 0.3583, "step": 993 }, { "epoch": 2.1942604856512142, "grad_norm": 0.12477768212556839, "learning_rate": 2.041366205945563e-06, "loss": 0.3604, "step": 994 }, { "epoch": 2.196467991169978, "grad_norm": 0.14072567224502563, "learning_rate": 2.0310220984046467e-06, "loss": 0.3693, "step": 995 }, { "epoch": 2.198675496688742, "grad_norm": 0.13882459700107574, "learning_rate": 2.020697581726279e-06, "loss": 0.3604, "step": 996 }, { "epoch": 2.2008830022075054, "grad_norm": 0.12994110584259033, "learning_rate": 2.010392724037002e-06, "loss": 0.3589, "step": 997 }, { "epoch": 2.2030905077262695, "grad_norm": 0.13540343940258026, "learning_rate": 2.0001075933336302e-06, "loss": 0.3664, "step": 998 }, { "epoch": 2.205298013245033, "grad_norm": 0.12710383534431458, "learning_rate": 1.9898422574828163e-06, "loss": 0.3606, "step": 999 }, { "epoch": 2.207505518763797, "grad_norm": 0.13436554372310638, "learning_rate": 1.979596784220591e-06, "loss": 0.3583, "step": 1000 }, { "epoch": 2.2097130242825607, "grad_norm": 0.1312997192144394, "learning_rate": 1.969371241151923e-06, "loss": 0.3485, "step": 1001 }, { "epoch": 2.2119205298013247, "grad_norm": 0.12719684839248657, "learning_rate": 1.9591656957502696e-06, "loss": 0.3646, "step": 1002 }, { "epoch": 2.2141280353200883, "grad_norm": 0.11936060339212418, "learning_rate": 1.9489802153571373e-06, "loss": 0.3578, "step": 1003 }, { "epoch": 2.216335540838852, "grad_norm": 0.11757774651050568, "learning_rate": 1.938814867181632e-06, "loss": 0.3633, "step": 1004 }, { "epoch": 2.218543046357616, "grad_norm": 0.12061762064695358, "learning_rate": 1.928669718300013e-06, "loss": 0.3539, "step": 1005 }, { "epoch": 2.2207505518763795, "grad_norm": 0.11802493780851364, "learning_rate": 1.9185448356552575e-06, "loss": 0.3519, "step": 1006 }, { "epoch": 2.2229580573951435, "grad_norm": 0.12381359189748764, "learning_rate": 1.9084402860566128e-06, "loss": 0.3528, "step": 1007 }, { "epoch": 2.225165562913907, "grad_norm": 0.12773284316062927, "learning_rate": 1.8983561361791608e-06, "loss": 0.3618, "step": 1008 }, { "epoch": 2.227373068432671, "grad_norm": 0.1377752125263214, "learning_rate": 1.8882924525633778e-06, "loss": 0.36, "step": 1009 }, { "epoch": 2.2295805739514347, "grad_norm": 0.13043536245822906, "learning_rate": 1.8782493016146868e-06, "loss": 0.3588, "step": 1010 }, { "epoch": 2.2317880794701987, "grad_norm": 0.12282190471887589, "learning_rate": 1.8682267496030276e-06, "loss": 0.3661, "step": 1011 }, { "epoch": 2.2339955849889623, "grad_norm": 0.11887135356664658, "learning_rate": 1.8582248626624217e-06, "loss": 0.3585, "step": 1012 }, { "epoch": 2.2362030905077264, "grad_norm": 0.13011059165000916, "learning_rate": 1.8482437067905268e-06, "loss": 0.3648, "step": 1013 }, { "epoch": 2.23841059602649, "grad_norm": 0.12632228434085846, "learning_rate": 1.8382833478482066e-06, "loss": 0.3711, "step": 1014 }, { "epoch": 2.240618101545254, "grad_norm": 0.13252626359462738, "learning_rate": 1.8283438515590996e-06, "loss": 0.3691, "step": 1015 }, { "epoch": 2.2428256070640176, "grad_norm": 0.12210851162672043, "learning_rate": 1.8184252835091764e-06, "loss": 0.36, "step": 1016 }, { "epoch": 2.2450331125827816, "grad_norm": 0.1259019374847412, "learning_rate": 1.8085277091463188e-06, "loss": 0.3507, "step": 1017 }, { "epoch": 2.247240618101545, "grad_norm": 0.11658161133527756, "learning_rate": 1.798651193779875e-06, "loss": 0.3512, "step": 1018 }, { "epoch": 2.249448123620309, "grad_norm": 0.12686073780059814, "learning_rate": 1.788795802580236e-06, "loss": 0.3554, "step": 1019 }, { "epoch": 2.251655629139073, "grad_norm": 0.13531404733657837, "learning_rate": 1.7789616005784077e-06, "loss": 0.3569, "step": 1020 }, { "epoch": 2.253863134657837, "grad_norm": 0.12338366359472275, "learning_rate": 1.7691486526655782e-06, "loss": 0.361, "step": 1021 }, { "epoch": 2.2560706401766004, "grad_norm": 0.1204574853181839, "learning_rate": 1.7593570235926883e-06, "loss": 0.3665, "step": 1022 }, { "epoch": 2.258278145695364, "grad_norm": 0.12831008434295654, "learning_rate": 1.7495867779700053e-06, "loss": 0.3617, "step": 1023 }, { "epoch": 2.260485651214128, "grad_norm": 0.1243155300617218, "learning_rate": 1.7398379802666993e-06, "loss": 0.3485, "step": 1024 }, { "epoch": 2.262693156732892, "grad_norm": 0.12370171397924423, "learning_rate": 1.7301106948104123e-06, "loss": 0.363, "step": 1025 }, { "epoch": 2.2649006622516556, "grad_norm": 0.12362192571163177, "learning_rate": 1.7204049857868433e-06, "loss": 0.3553, "step": 1026 }, { "epoch": 2.2671081677704192, "grad_norm": 0.1303797960281372, "learning_rate": 1.7107209172393158e-06, "loss": 0.3582, "step": 1027 }, { "epoch": 2.2693156732891833, "grad_norm": 0.12913931906223297, "learning_rate": 1.701058553068357e-06, "loss": 0.3621, "step": 1028 }, { "epoch": 2.271523178807947, "grad_norm": 0.13231627643108368, "learning_rate": 1.6914179570312767e-06, "loss": 0.3529, "step": 1029 }, { "epoch": 2.273730684326711, "grad_norm": 0.12977519631385803, "learning_rate": 1.6817991927417516e-06, "loss": 0.3619, "step": 1030 }, { "epoch": 2.2759381898454745, "grad_norm": 0.12845906615257263, "learning_rate": 1.672202323669397e-06, "loss": 0.3613, "step": 1031 }, { "epoch": 2.2781456953642385, "grad_norm": 0.1263485550880432, "learning_rate": 1.662627413139351e-06, "loss": 0.3608, "step": 1032 }, { "epoch": 2.280353200883002, "grad_norm": 0.12105315178632736, "learning_rate": 1.6530745243318646e-06, "loss": 0.3694, "step": 1033 }, { "epoch": 2.282560706401766, "grad_norm": 0.12585538625717163, "learning_rate": 1.64354372028187e-06, "loss": 0.3674, "step": 1034 }, { "epoch": 2.2847682119205297, "grad_norm": 0.13143764436244965, "learning_rate": 1.634035063878579e-06, "loss": 0.3654, "step": 1035 }, { "epoch": 2.2869757174392937, "grad_norm": 0.1330743134021759, "learning_rate": 1.6245486178650582e-06, "loss": 0.3579, "step": 1036 }, { "epoch": 2.2891832229580573, "grad_norm": 0.12009730935096741, "learning_rate": 1.6150844448378178e-06, "loss": 0.3531, "step": 1037 }, { "epoch": 2.2913907284768213, "grad_norm": 0.13864392042160034, "learning_rate": 1.6056426072464015e-06, "loss": 0.3628, "step": 1038 }, { "epoch": 2.293598233995585, "grad_norm": 0.13580797612667084, "learning_rate": 1.5962231673929735e-06, "loss": 0.3722, "step": 1039 }, { "epoch": 2.295805739514349, "grad_norm": 0.12375160306692123, "learning_rate": 1.586826187431902e-06, "loss": 0.364, "step": 1040 }, { "epoch": 2.2980132450331126, "grad_norm": 0.12487831711769104, "learning_rate": 1.5774517293693558e-06, "loss": 0.3592, "step": 1041 }, { "epoch": 2.300220750551876, "grad_norm": 0.14687076210975647, "learning_rate": 1.5680998550628912e-06, "loss": 0.3506, "step": 1042 }, { "epoch": 2.30242825607064, "grad_norm": 0.12587517499923706, "learning_rate": 1.558770626221044e-06, "loss": 0.3577, "step": 1043 }, { "epoch": 2.304635761589404, "grad_norm": 0.13661305606365204, "learning_rate": 1.5494641044029268e-06, "loss": 0.3516, "step": 1044 }, { "epoch": 2.306843267108168, "grad_norm": 0.1377173811197281, "learning_rate": 1.5401803510178197e-06, "loss": 0.3603, "step": 1045 }, { "epoch": 2.3090507726269314, "grad_norm": 0.13497798144817352, "learning_rate": 1.5309194273247612e-06, "loss": 0.3545, "step": 1046 }, { "epoch": 2.3112582781456954, "grad_norm": 0.12721391022205353, "learning_rate": 1.5216813944321473e-06, "loss": 0.3447, "step": 1047 }, { "epoch": 2.313465783664459, "grad_norm": 0.12541592121124268, "learning_rate": 1.5124663132973338e-06, "loss": 0.3637, "step": 1048 }, { "epoch": 2.315673289183223, "grad_norm": 0.13365043699741364, "learning_rate": 1.5032742447262228e-06, "loss": 0.3535, "step": 1049 }, { "epoch": 2.3178807947019866, "grad_norm": 0.13111183047294617, "learning_rate": 1.4941052493728731e-06, "loss": 0.3528, "step": 1050 }, { "epoch": 2.3200883002207506, "grad_norm": 0.1259627640247345, "learning_rate": 1.48495938773909e-06, "loss": 0.3599, "step": 1051 }, { "epoch": 2.322295805739514, "grad_norm": 0.12294916808605194, "learning_rate": 1.4758367201740303e-06, "loss": 0.3604, "step": 1052 }, { "epoch": 2.3245033112582782, "grad_norm": 0.12867586314678192, "learning_rate": 1.46673730687381e-06, "loss": 0.3649, "step": 1053 }, { "epoch": 2.326710816777042, "grad_norm": 0.12218500673770905, "learning_rate": 1.4576612078810953e-06, "loss": 0.3637, "step": 1054 }, { "epoch": 2.328918322295806, "grad_norm": 0.12343169003725052, "learning_rate": 1.448608483084713e-06, "loss": 0.3552, "step": 1055 }, { "epoch": 2.3311258278145695, "grad_norm": 0.12273656576871872, "learning_rate": 1.4395791922192575e-06, "loss": 0.3613, "step": 1056 }, { "epoch": 2.3333333333333335, "grad_norm": 0.1312197744846344, "learning_rate": 1.430573394864695e-06, "loss": 0.3603, "step": 1057 }, { "epoch": 2.335540838852097, "grad_norm": 0.13838644325733185, "learning_rate": 1.4215911504459645e-06, "loss": 0.3642, "step": 1058 }, { "epoch": 2.337748344370861, "grad_norm": 0.11605527251958847, "learning_rate": 1.412632518232594e-06, "loss": 0.3626, "step": 1059 }, { "epoch": 2.3399558498896247, "grad_norm": 0.13165433704853058, "learning_rate": 1.4036975573383028e-06, "loss": 0.3636, "step": 1060 }, { "epoch": 2.3421633554083887, "grad_norm": 0.1277674436569214, "learning_rate": 1.3947863267206174e-06, "loss": 0.3517, "step": 1061 }, { "epoch": 2.3443708609271523, "grad_norm": 0.13928262889385223, "learning_rate": 1.3858988851804816e-06, "loss": 0.3649, "step": 1062 }, { "epoch": 2.3465783664459163, "grad_norm": 0.13180677592754364, "learning_rate": 1.3770352913618613e-06, "loss": 0.3506, "step": 1063 }, { "epoch": 2.34878587196468, "grad_norm": 0.13746923208236694, "learning_rate": 1.3681956037513656e-06, "loss": 0.3583, "step": 1064 }, { "epoch": 2.3509933774834435, "grad_norm": 0.13763363659381866, "learning_rate": 1.3593798806778546e-06, "loss": 0.354, "step": 1065 }, { "epoch": 2.3532008830022075, "grad_norm": 0.12797822058200836, "learning_rate": 1.3505881803120647e-06, "loss": 0.3719, "step": 1066 }, { "epoch": 2.355408388520971, "grad_norm": 0.12430882453918457, "learning_rate": 1.341820560666211e-06, "loss": 0.3622, "step": 1067 }, { "epoch": 2.357615894039735, "grad_norm": 0.13003040850162506, "learning_rate": 1.3330770795936172e-06, "loss": 0.3644, "step": 1068 }, { "epoch": 2.3598233995584987, "grad_norm": 0.13834795355796814, "learning_rate": 1.3243577947883223e-06, "loss": 0.3539, "step": 1069 }, { "epoch": 2.3620309050772628, "grad_norm": 0.12837594747543335, "learning_rate": 1.315662763784712e-06, "loss": 0.366, "step": 1070 }, { "epoch": 2.3642384105960264, "grad_norm": 0.12611278891563416, "learning_rate": 1.3069920439571277e-06, "loss": 0.3567, "step": 1071 }, { "epoch": 2.3664459161147904, "grad_norm": 0.1504361927509308, "learning_rate": 1.2983456925194953e-06, "loss": 0.3539, "step": 1072 }, { "epoch": 2.368653421633554, "grad_norm": 0.13299378752708435, "learning_rate": 1.2897237665249429e-06, "loss": 0.3612, "step": 1073 }, { "epoch": 2.370860927152318, "grad_norm": 0.13157807290554047, "learning_rate": 1.2811263228654308e-06, "loss": 0.3652, "step": 1074 }, { "epoch": 2.3730684326710816, "grad_norm": 0.11843698471784592, "learning_rate": 1.2725534182713717e-06, "loss": 0.3664, "step": 1075 }, { "epoch": 2.3752759381898456, "grad_norm": 0.1275041550397873, "learning_rate": 1.2640051093112532e-06, "loss": 0.3595, "step": 1076 }, { "epoch": 2.377483443708609, "grad_norm": 0.12319236993789673, "learning_rate": 1.25548145239127e-06, "loss": 0.3458, "step": 1077 }, { "epoch": 2.3796909492273732, "grad_norm": 0.1297508329153061, "learning_rate": 1.2469825037549493e-06, "loss": 0.3599, "step": 1078 }, { "epoch": 2.381898454746137, "grad_norm": 0.13835854828357697, "learning_rate": 1.2385083194827818e-06, "loss": 0.3631, "step": 1079 }, { "epoch": 2.384105960264901, "grad_norm": 0.14326325058937073, "learning_rate": 1.2300589554918502e-06, "loss": 0.3612, "step": 1080 }, { "epoch": 2.3863134657836644, "grad_norm": 0.13561151921749115, "learning_rate": 1.221634467535458e-06, "loss": 0.3623, "step": 1081 }, { "epoch": 2.3885209713024285, "grad_norm": 0.12151821702718735, "learning_rate": 1.2132349112027636e-06, "loss": 0.347, "step": 1082 }, { "epoch": 2.390728476821192, "grad_norm": 0.12261461466550827, "learning_rate": 1.204860341918414e-06, "loss": 0.3615, "step": 1083 }, { "epoch": 2.3929359823399556, "grad_norm": 0.1293371468782425, "learning_rate": 1.1965108149421812e-06, "loss": 0.3611, "step": 1084 }, { "epoch": 2.3951434878587197, "grad_norm": 0.12331968545913696, "learning_rate": 1.1881863853685904e-06, "loss": 0.3635, "step": 1085 }, { "epoch": 2.3973509933774833, "grad_norm": 0.13170567154884338, "learning_rate": 1.1798871081265672e-06, "loss": 0.3698, "step": 1086 }, { "epoch": 2.3995584988962473, "grad_norm": 0.1309044063091278, "learning_rate": 1.1716130379790613e-06, "loss": 0.3474, "step": 1087 }, { "epoch": 2.401766004415011, "grad_norm": 0.12268619239330292, "learning_rate": 1.1633642295227005e-06, "loss": 0.3634, "step": 1088 }, { "epoch": 2.403973509933775, "grad_norm": 0.12474343925714493, "learning_rate": 1.155140737187418e-06, "loss": 0.3651, "step": 1089 }, { "epoch": 2.4061810154525385, "grad_norm": 0.13622671365737915, "learning_rate": 1.1469426152360974e-06, "loss": 0.3626, "step": 1090 }, { "epoch": 2.4083885209713025, "grad_norm": 0.1363033503293991, "learning_rate": 1.138769917764221e-06, "loss": 0.3608, "step": 1091 }, { "epoch": 2.410596026490066, "grad_norm": 0.12666958570480347, "learning_rate": 1.1306226986994989e-06, "loss": 0.3571, "step": 1092 }, { "epoch": 2.41280353200883, "grad_norm": 0.13340440392494202, "learning_rate": 1.1225010118015306e-06, "loss": 0.3602, "step": 1093 }, { "epoch": 2.4150110375275937, "grad_norm": 0.14224812388420105, "learning_rate": 1.1144049106614335e-06, "loss": 0.3525, "step": 1094 }, { "epoch": 2.4172185430463577, "grad_norm": 0.13309240341186523, "learning_rate": 1.1063344487015e-06, "loss": 0.3595, "step": 1095 }, { "epoch": 2.4194260485651213, "grad_norm": 0.1310422122478485, "learning_rate": 1.098289679174841e-06, "loss": 0.3674, "step": 1096 }, { "epoch": 2.4216335540838854, "grad_norm": 0.12578153610229492, "learning_rate": 1.090270655165036e-06, "loss": 0.3543, "step": 1097 }, { "epoch": 2.423841059602649, "grad_norm": 0.13076968491077423, "learning_rate": 1.082277429585784e-06, "loss": 0.3724, "step": 1098 }, { "epoch": 2.426048565121413, "grad_norm": 0.14538316428661346, "learning_rate": 1.074310055180549e-06, "loss": 0.3539, "step": 1099 }, { "epoch": 2.4282560706401766, "grad_norm": 0.13270893692970276, "learning_rate": 1.0663685845222177e-06, "loss": 0.3637, "step": 1100 }, { "epoch": 2.4304635761589406, "grad_norm": 0.13810580968856812, "learning_rate": 1.0584530700127478e-06, "loss": 0.3526, "step": 1101 }, { "epoch": 2.432671081677704, "grad_norm": 0.1298057585954666, "learning_rate": 1.0505635638828288e-06, "loss": 0.3551, "step": 1102 }, { "epoch": 2.4348785871964678, "grad_norm": 0.13741736114025116, "learning_rate": 1.0427001181915298e-06, "loss": 0.3512, "step": 1103 }, { "epoch": 2.437086092715232, "grad_norm": 0.1301170289516449, "learning_rate": 1.034862784825963e-06, "loss": 0.3568, "step": 1104 }, { "epoch": 2.439293598233996, "grad_norm": 0.12320411205291748, "learning_rate": 1.0270516155009336e-06, "loss": 0.3701, "step": 1105 }, { "epoch": 2.4415011037527594, "grad_norm": 0.1281418800354004, "learning_rate": 1.0192666617586072e-06, "loss": 0.3657, "step": 1106 }, { "epoch": 2.443708609271523, "grad_norm": 0.14360150694847107, "learning_rate": 1.0115079749681628e-06, "loss": 0.3598, "step": 1107 }, { "epoch": 2.445916114790287, "grad_norm": 0.13790085911750793, "learning_rate": 1.0037756063254555e-06, "loss": 0.3612, "step": 1108 }, { "epoch": 2.4481236203090506, "grad_norm": 0.1285860389471054, "learning_rate": 9.960696068526826e-07, "loss": 0.3615, "step": 1109 }, { "epoch": 2.4503311258278146, "grad_norm": 0.1286444067955017, "learning_rate": 9.883900273980408e-07, "loss": 0.3587, "step": 1110 }, { "epoch": 2.4525386313465782, "grad_norm": 0.12481298297643661, "learning_rate": 9.807369186353965e-07, "loss": 0.356, "step": 1111 }, { "epoch": 2.4547461368653423, "grad_norm": 0.1254754513502121, "learning_rate": 9.731103310639473e-07, "loss": 0.3575, "step": 1112 }, { "epoch": 2.456953642384106, "grad_norm": 0.12286875396966934, "learning_rate": 9.655103150078892e-07, "loss": 0.3608, "step": 1113 }, { "epoch": 2.45916114790287, "grad_norm": 0.11990831047296524, "learning_rate": 9.57936920616086e-07, "loss": 0.3586, "step": 1114 }, { "epoch": 2.4613686534216335, "grad_norm": 0.11828765273094177, "learning_rate": 9.503901978617392e-07, "loss": 0.3621, "step": 1115 }, { "epoch": 2.4635761589403975, "grad_norm": 0.13093435764312744, "learning_rate": 9.428701965420572e-07, "loss": 0.361, "step": 1116 }, { "epoch": 2.465783664459161, "grad_norm": 0.12574456632137299, "learning_rate": 9.353769662779249e-07, "loss": 0.3556, "step": 1117 }, { "epoch": 2.467991169977925, "grad_norm": 0.136729896068573, "learning_rate": 9.279105565135777e-07, "loss": 0.3661, "step": 1118 }, { "epoch": 2.4701986754966887, "grad_norm": 0.12174253910779953, "learning_rate": 9.204710165162751e-07, "loss": 0.3487, "step": 1119 }, { "epoch": 2.4724061810154527, "grad_norm": 0.13417866826057434, "learning_rate": 9.130583953759781e-07, "loss": 0.358, "step": 1120 }, { "epoch": 2.4746136865342163, "grad_norm": 0.14360179007053375, "learning_rate": 9.056727420050227e-07, "loss": 0.3628, "step": 1121 }, { "epoch": 2.47682119205298, "grad_norm": 0.12731443345546722, "learning_rate": 8.983141051377953e-07, "loss": 0.3603, "step": 1122 }, { "epoch": 2.479028697571744, "grad_norm": 0.13204246759414673, "learning_rate": 8.909825333304134e-07, "loss": 0.355, "step": 1123 }, { "epoch": 2.481236203090508, "grad_norm": 0.1328221708536148, "learning_rate": 8.836780749604096e-07, "loss": 0.3541, "step": 1124 }, { "epoch": 2.4834437086092715, "grad_norm": 0.13670921325683594, "learning_rate": 8.764007782264022e-07, "loss": 0.3649, "step": 1125 }, { "epoch": 2.485651214128035, "grad_norm": 0.124913290143013, "learning_rate": 8.691506911477848e-07, "loss": 0.3474, "step": 1126 }, { "epoch": 2.487858719646799, "grad_norm": 0.12466172128915787, "learning_rate": 8.619278615644106e-07, "loss": 0.3559, "step": 1127 }, { "epoch": 2.4900662251655628, "grad_norm": 0.13907819986343384, "learning_rate": 8.547323371362682e-07, "loss": 0.3544, "step": 1128 }, { "epoch": 2.492273730684327, "grad_norm": 0.13403339684009552, "learning_rate": 8.475641653431782e-07, "loss": 0.3621, "step": 1129 }, { "epoch": 2.4944812362030904, "grad_norm": 0.13107538223266602, "learning_rate": 8.404233934844707e-07, "loss": 0.3629, "step": 1130 }, { "epoch": 2.4966887417218544, "grad_norm": 0.12382876127958298, "learning_rate": 8.333100686786766e-07, "loss": 0.3425, "step": 1131 }, { "epoch": 2.498896247240618, "grad_norm": 0.1340240240097046, "learning_rate": 8.262242378632179e-07, "loss": 0.3637, "step": 1132 }, { "epoch": 2.501103752759382, "grad_norm": 0.12560687959194183, "learning_rate": 8.191659477940972e-07, "loss": 0.3694, "step": 1133 }, { "epoch": 2.5033112582781456, "grad_norm": 0.13509927690029144, "learning_rate": 8.121352450455899e-07, "loss": 0.3588, "step": 1134 }, { "epoch": 2.5055187637969096, "grad_norm": 0.13506671786308289, "learning_rate": 8.051321760099334e-07, "loss": 0.3657, "step": 1135 }, { "epoch": 2.507726269315673, "grad_norm": 0.12824004888534546, "learning_rate": 7.981567868970252e-07, "loss": 0.3541, "step": 1136 }, { "epoch": 2.5099337748344372, "grad_norm": 0.12654347717761993, "learning_rate": 7.91209123734114e-07, "loss": 0.3628, "step": 1137 }, { "epoch": 2.512141280353201, "grad_norm": 0.13879314064979553, "learning_rate": 7.84289232365501e-07, "loss": 0.3631, "step": 1138 }, { "epoch": 2.514348785871965, "grad_norm": 0.11704026162624359, "learning_rate": 7.773971584522355e-07, "loss": 0.3589, "step": 1139 }, { "epoch": 2.5165562913907285, "grad_norm": 0.12439953535795212, "learning_rate": 7.705329474718093e-07, "loss": 0.3537, "step": 1140 }, { "epoch": 2.518763796909492, "grad_norm": 0.13436546921730042, "learning_rate": 7.636966447178601e-07, "loss": 0.356, "step": 1141 }, { "epoch": 2.520971302428256, "grad_norm": 0.12255270779132843, "learning_rate": 7.568882952998762e-07, "loss": 0.3568, "step": 1142 }, { "epoch": 2.52317880794702, "grad_norm": 0.12233200669288635, "learning_rate": 7.501079441428927e-07, "loss": 0.3538, "step": 1143 }, { "epoch": 2.5253863134657837, "grad_norm": 0.12473509460687637, "learning_rate": 7.433556359871968e-07, "loss": 0.3711, "step": 1144 }, { "epoch": 2.5275938189845473, "grad_norm": 0.13673733174800873, "learning_rate": 7.366314153880361e-07, "loss": 0.3613, "step": 1145 }, { "epoch": 2.5298013245033113, "grad_norm": 0.1345418095588684, "learning_rate": 7.299353267153192e-07, "loss": 0.3632, "step": 1146 }, { "epoch": 2.5320088300220753, "grad_norm": 0.1333194524049759, "learning_rate": 7.232674141533274e-07, "loss": 0.3653, "step": 1147 }, { "epoch": 2.534216335540839, "grad_norm": 0.13697415590286255, "learning_rate": 7.166277217004214e-07, "loss": 0.3571, "step": 1148 }, { "epoch": 2.5364238410596025, "grad_norm": 0.13772854208946228, "learning_rate": 7.100162931687476e-07, "loss": 0.3737, "step": 1149 }, { "epoch": 2.5386313465783665, "grad_norm": 0.1227809488773346, "learning_rate": 7.034331721839566e-07, "loss": 0.3622, "step": 1150 }, { "epoch": 2.54083885209713, "grad_norm": 0.12854185700416565, "learning_rate": 6.968784021849106e-07, "loss": 0.3612, "step": 1151 }, { "epoch": 2.543046357615894, "grad_norm": 0.11819867044687271, "learning_rate": 6.903520264233954e-07, "loss": 0.3625, "step": 1152 }, { "epoch": 2.5452538631346577, "grad_norm": 0.1528642773628235, "learning_rate": 6.838540879638367e-07, "loss": 0.3644, "step": 1153 }, { "epoch": 2.5474613686534218, "grad_norm": 0.1466141641139984, "learning_rate": 6.773846296830167e-07, "loss": 0.3681, "step": 1154 }, { "epoch": 2.5496688741721854, "grad_norm": 0.13698391616344452, "learning_rate": 6.7094369426979e-07, "loss": 0.3569, "step": 1155 }, { "epoch": 2.5518763796909494, "grad_norm": 0.1320303976535797, "learning_rate": 6.645313242248042e-07, "loss": 0.3531, "step": 1156 }, { "epoch": 2.554083885209713, "grad_norm": 0.14429907500743866, "learning_rate": 6.581475618602174e-07, "loss": 0.3587, "step": 1157 }, { "epoch": 2.556291390728477, "grad_norm": 0.12878787517547607, "learning_rate": 6.517924492994182e-07, "loss": 0.3648, "step": 1158 }, { "epoch": 2.5584988962472406, "grad_norm": 0.13368941843509674, "learning_rate": 6.454660284767477e-07, "loss": 0.3569, "step": 1159 }, { "epoch": 2.560706401766004, "grad_norm": 0.13522258400917053, "learning_rate": 6.391683411372279e-07, "loss": 0.3624, "step": 1160 }, { "epoch": 2.562913907284768, "grad_norm": 0.1288643628358841, "learning_rate": 6.328994288362783e-07, "loss": 0.3571, "step": 1161 }, { "epoch": 2.5651214128035322, "grad_norm": 0.13528789579868317, "learning_rate": 6.266593329394471e-07, "loss": 0.3614, "step": 1162 }, { "epoch": 2.567328918322296, "grad_norm": 0.11771933734416962, "learning_rate": 6.204480946221386e-07, "loss": 0.3503, "step": 1163 }, { "epoch": 2.5695364238410594, "grad_norm": 0.13160669803619385, "learning_rate": 6.142657548693364e-07, "loss": 0.3662, "step": 1164 }, { "epoch": 2.5717439293598234, "grad_norm": 0.13713331520557404, "learning_rate": 6.081123544753404e-07, "loss": 0.3474, "step": 1165 }, { "epoch": 2.5739514348785875, "grad_norm": 0.12216666340827942, "learning_rate": 6.019879340434904e-07, "loss": 0.3414, "step": 1166 }, { "epoch": 2.576158940397351, "grad_norm": 0.12497063726186752, "learning_rate": 5.958925339859001e-07, "loss": 0.3501, "step": 1167 }, { "epoch": 2.5783664459161146, "grad_norm": 0.1288682073354721, "learning_rate": 5.898261945231965e-07, "loss": 0.3518, "step": 1168 }, { "epoch": 2.5805739514348787, "grad_norm": 0.13969705998897552, "learning_rate": 5.837889556842469e-07, "loss": 0.3595, "step": 1169 }, { "epoch": 2.5827814569536423, "grad_norm": 0.13971775770187378, "learning_rate": 5.777808573058969e-07, "loss": 0.3635, "step": 1170 }, { "epoch": 2.5849889624724063, "grad_norm": 0.12668545544147491, "learning_rate": 5.718019390327084e-07, "loss": 0.3519, "step": 1171 }, { "epoch": 2.58719646799117, "grad_norm": 0.13675172626972198, "learning_rate": 5.658522403166989e-07, "loss": 0.363, "step": 1172 }, { "epoch": 2.589403973509934, "grad_norm": 0.145385280251503, "learning_rate": 5.599318004170778e-07, "loss": 0.3561, "step": 1173 }, { "epoch": 2.5916114790286975, "grad_norm": 0.14911803603172302, "learning_rate": 5.540406583999925e-07, "loss": 0.3578, "step": 1174 }, { "epoch": 2.5938189845474615, "grad_norm": 0.12694287300109863, "learning_rate": 5.481788531382671e-07, "loss": 0.364, "step": 1175 }, { "epoch": 2.596026490066225, "grad_norm": 0.13262110948562622, "learning_rate": 5.423464233111448e-07, "loss": 0.3718, "step": 1176 }, { "epoch": 2.598233995584989, "grad_norm": 0.13582631945610046, "learning_rate": 5.365434074040343e-07, "loss": 0.3654, "step": 1177 }, { "epoch": 2.6004415011037527, "grad_norm": 0.13755717873573303, "learning_rate": 5.307698437082598e-07, "loss": 0.3522, "step": 1178 }, { "epoch": 2.6026490066225163, "grad_norm": 0.1361360102891922, "learning_rate": 5.250257703207984e-07, "loss": 0.3683, "step": 1179 }, { "epoch": 2.6048565121412803, "grad_norm": 0.13162481784820557, "learning_rate": 5.193112251440407e-07, "loss": 0.3559, "step": 1180 }, { "epoch": 2.6070640176600444, "grad_norm": 0.1336871236562729, "learning_rate": 5.136262458855312e-07, "loss": 0.3539, "step": 1181 }, { "epoch": 2.609271523178808, "grad_norm": 0.12142772227525711, "learning_rate": 5.079708700577229e-07, "loss": 0.3633, "step": 1182 }, { "epoch": 2.6114790286975715, "grad_norm": 0.12398800998926163, "learning_rate": 5.023451349777331e-07, "loss": 0.3666, "step": 1183 }, { "epoch": 2.6136865342163356, "grad_norm": 0.1366734653711319, "learning_rate": 4.967490777670903e-07, "loss": 0.3552, "step": 1184 }, { "epoch": 2.6158940397350996, "grad_norm": 0.13601034879684448, "learning_rate": 4.911827353514947e-07, "loss": 0.3606, "step": 1185 }, { "epoch": 2.618101545253863, "grad_norm": 0.13659435510635376, "learning_rate": 4.856461444605732e-07, "loss": 0.356, "step": 1186 }, { "epoch": 2.6203090507726268, "grad_norm": 0.13636133074760437, "learning_rate": 4.801393416276368e-07, "loss": 0.3591, "step": 1187 }, { "epoch": 2.622516556291391, "grad_norm": 0.1251705437898636, "learning_rate": 4.7466236318943816e-07, "loss": 0.3557, "step": 1188 }, { "epoch": 2.6247240618101544, "grad_norm": 0.14416959881782532, "learning_rate": 4.692152452859333e-07, "loss": 0.3587, "step": 1189 }, { "epoch": 2.6269315673289184, "grad_norm": 0.15531152486801147, "learning_rate": 4.637980238600437e-07, "loss": 0.3494, "step": 1190 }, { "epoch": 2.629139072847682, "grad_norm": 0.1295311003923416, "learning_rate": 4.584107346574168e-07, "loss": 0.3595, "step": 1191 }, { "epoch": 2.631346578366446, "grad_norm": 0.13844533264636993, "learning_rate": 4.530534132261932e-07, "loss": 0.3589, "step": 1192 }, { "epoch": 2.6335540838852096, "grad_norm": 0.13652655482292175, "learning_rate": 4.477260949167711e-07, "loss": 0.3492, "step": 1193 }, { "epoch": 2.6357615894039736, "grad_norm": 0.14151211082935333, "learning_rate": 4.4242881488157083e-07, "loss": 0.3565, "step": 1194 }, { "epoch": 2.6379690949227372, "grad_norm": 0.13526977598667145, "learning_rate": 4.371616080748037e-07, "loss": 0.355, "step": 1195 }, { "epoch": 2.6401766004415013, "grad_norm": 0.13757188618183136, "learning_rate": 4.319245092522456e-07, "loss": 0.3547, "step": 1196 }, { "epoch": 2.642384105960265, "grad_norm": 0.1314619481563568, "learning_rate": 4.2671755297100047e-07, "loss": 0.36, "step": 1197 }, { "epoch": 2.6445916114790284, "grad_norm": 0.13309867680072784, "learning_rate": 4.215407735892796e-07, "loss": 0.3578, "step": 1198 }, { "epoch": 2.6467991169977925, "grad_norm": 0.14237742125988007, "learning_rate": 4.1639420526616845e-07, "loss": 0.3567, "step": 1199 }, { "epoch": 2.6490066225165565, "grad_norm": 0.13557758927345276, "learning_rate": 4.1127788196140437e-07, "loss": 0.3627, "step": 1200 }, { "epoch": 2.65121412803532, "grad_norm": 0.14701923727989197, "learning_rate": 4.061918374351559e-07, "loss": 0.3518, "step": 1201 }, { "epoch": 2.6534216335540837, "grad_norm": 0.1429191380739212, "learning_rate": 4.0113610524779246e-07, "loss": 0.3579, "step": 1202 }, { "epoch": 2.6556291390728477, "grad_norm": 0.13261815905570984, "learning_rate": 3.961107187596669e-07, "loss": 0.3558, "step": 1203 }, { "epoch": 2.6578366445916117, "grad_norm": 0.15215714275836945, "learning_rate": 3.911157111308983e-07, "loss": 0.3582, "step": 1204 }, { "epoch": 2.6600441501103753, "grad_norm": 0.134977787733078, "learning_rate": 3.861511153211489e-07, "loss": 0.3484, "step": 1205 }, { "epoch": 2.662251655629139, "grad_norm": 0.13735589385032654, "learning_rate": 3.81216964089407e-07, "loss": 0.3633, "step": 1206 }, { "epoch": 2.664459161147903, "grad_norm": 0.13418880105018616, "learning_rate": 3.763132899937721e-07, "loss": 0.3646, "step": 1207 }, { "epoch": 2.6666666666666665, "grad_norm": 0.13318879902362823, "learning_rate": 3.7144012539123973e-07, "loss": 0.3449, "step": 1208 }, { "epoch": 2.6688741721854305, "grad_norm": 0.14231936633586884, "learning_rate": 3.665975024374879e-07, "loss": 0.3661, "step": 1209 }, { "epoch": 2.671081677704194, "grad_norm": 0.13167575001716614, "learning_rate": 3.617854530866671e-07, "loss": 0.3627, "step": 1210 }, { "epoch": 2.673289183222958, "grad_norm": 0.13379094004631042, "learning_rate": 3.5700400909118306e-07, "loss": 0.3585, "step": 1211 }, { "epoch": 2.6754966887417218, "grad_norm": 0.1270161271095276, "learning_rate": 3.522532020014946e-07, "loss": 0.3571, "step": 1212 }, { "epoch": 2.677704194260486, "grad_norm": 0.1238294392824173, "learning_rate": 3.475330631659002e-07, "loss": 0.3653, "step": 1213 }, { "epoch": 2.6799116997792494, "grad_norm": 0.12466707080602646, "learning_rate": 3.4284362373033566e-07, "loss": 0.3535, "step": 1214 }, { "epoch": 2.6821192052980134, "grad_norm": 0.12770843505859375, "learning_rate": 3.3818491463816385e-07, "loss": 0.3573, "step": 1215 }, { "epoch": 2.684326710816777, "grad_norm": 0.1300242394208908, "learning_rate": 3.335569666299748e-07, "loss": 0.3566, "step": 1216 }, { "epoch": 2.6865342163355406, "grad_norm": 0.1325581669807434, "learning_rate": 3.289598102433794e-07, "loss": 0.3575, "step": 1217 }, { "epoch": 2.6887417218543046, "grad_norm": 0.1356075257062912, "learning_rate": 3.2439347581280865e-07, "loss": 0.3614, "step": 1218 }, { "epoch": 2.6909492273730686, "grad_norm": 0.12728764116764069, "learning_rate": 3.19857993469318e-07, "loss": 0.3586, "step": 1219 }, { "epoch": 2.693156732891832, "grad_norm": 0.1246216669678688, "learning_rate": 3.1535339314038015e-07, "loss": 0.3581, "step": 1220 }, { "epoch": 2.695364238410596, "grad_norm": 0.1401343196630478, "learning_rate": 3.108797045496942e-07, "loss": 0.3611, "step": 1221 }, { "epoch": 2.69757174392936, "grad_norm": 0.1308797299861908, "learning_rate": 3.0643695721698783e-07, "loss": 0.3664, "step": 1222 }, { "epoch": 2.699779249448124, "grad_norm": 0.12849809229373932, "learning_rate": 3.0202518045782337e-07, "loss": 0.3605, "step": 1223 }, { "epoch": 2.7019867549668874, "grad_norm": 0.128408744931221, "learning_rate": 2.9764440338340083e-07, "loss": 0.3583, "step": 1224 }, { "epoch": 2.704194260485651, "grad_norm": 0.1296570599079132, "learning_rate": 2.9329465490036844e-07, "loss": 0.3528, "step": 1225 }, { "epoch": 2.706401766004415, "grad_norm": 0.14551834762096405, "learning_rate": 2.8897596371063153e-07, "loss": 0.3497, "step": 1226 }, { "epoch": 2.7086092715231787, "grad_norm": 0.13115178048610687, "learning_rate": 2.846883583111648e-07, "loss": 0.3594, "step": 1227 }, { "epoch": 2.7108167770419427, "grad_norm": 0.12846940755844116, "learning_rate": 2.804318669938233e-07, "loss": 0.3583, "step": 1228 }, { "epoch": 2.7130242825607063, "grad_norm": 0.1380581110715866, "learning_rate": 2.762065178451517e-07, "loss": 0.3585, "step": 1229 }, { "epoch": 2.7152317880794703, "grad_norm": 0.12574802339076996, "learning_rate": 2.7201233874620534e-07, "loss": 0.3598, "step": 1230 }, { "epoch": 2.717439293598234, "grad_norm": 0.13319121301174164, "learning_rate": 2.678493573723612e-07, "loss": 0.3669, "step": 1231 }, { "epoch": 2.719646799116998, "grad_norm": 0.12967029213905334, "learning_rate": 2.6371760119314026e-07, "loss": 0.3637, "step": 1232 }, { "epoch": 2.7218543046357615, "grad_norm": 0.13314829766750336, "learning_rate": 2.596170974720202e-07, "loss": 0.3505, "step": 1233 }, { "epoch": 2.7240618101545255, "grad_norm": 0.13859589397907257, "learning_rate": 2.5554787326626194e-07, "loss": 0.3508, "step": 1234 }, { "epoch": 2.726269315673289, "grad_norm": 0.13054342567920685, "learning_rate": 2.515099554267247e-07, "loss": 0.3459, "step": 1235 }, { "epoch": 2.7284768211920527, "grad_norm": 0.13782578706741333, "learning_rate": 2.4750337059769425e-07, "loss": 0.3653, "step": 1236 }, { "epoch": 2.7306843267108167, "grad_norm": 0.1294957846403122, "learning_rate": 2.4352814521670375e-07, "loss": 0.3614, "step": 1237 }, { "epoch": 2.7328918322295808, "grad_norm": 0.1325952708721161, "learning_rate": 2.3958430551436095e-07, "loss": 0.3638, "step": 1238 }, { "epoch": 2.7350993377483444, "grad_norm": 0.12079501897096634, "learning_rate": 2.3567187751417475e-07, "loss": 0.3605, "step": 1239 }, { "epoch": 2.737306843267108, "grad_norm": 0.13967706263065338, "learning_rate": 2.3179088703238096e-07, "loss": 0.3721, "step": 1240 }, { "epoch": 2.739514348785872, "grad_norm": 0.133636474609375, "learning_rate": 2.2794135967777908e-07, "loss": 0.3655, "step": 1241 }, { "epoch": 2.741721854304636, "grad_norm": 0.13323010504245758, "learning_rate": 2.2412332085155364e-07, "loss": 0.3531, "step": 1242 }, { "epoch": 2.7439293598233996, "grad_norm": 0.13812537491321564, "learning_rate": 2.2033679574711365e-07, "loss": 0.3453, "step": 1243 }, { "epoch": 2.746136865342163, "grad_norm": 0.12806718051433563, "learning_rate": 2.1658180934992333e-07, "loss": 0.3632, "step": 1244 }, { "epoch": 2.748344370860927, "grad_norm": 0.13933822512626648, "learning_rate": 2.1285838643733958e-07, "loss": 0.3595, "step": 1245 }, { "epoch": 2.750551876379691, "grad_norm": 0.12600216269493103, "learning_rate": 2.0916655157844634e-07, "loss": 0.352, "step": 1246 }, { "epoch": 2.752759381898455, "grad_norm": 0.12791283428668976, "learning_rate": 2.0550632913389213e-07, "loss": 0.3556, "step": 1247 }, { "epoch": 2.7549668874172184, "grad_norm": 0.13090188801288605, "learning_rate": 2.0187774325573174e-07, "loss": 0.3632, "step": 1248 }, { "epoch": 2.7571743929359824, "grad_norm": 0.1253674477338791, "learning_rate": 1.9828081788726307e-07, "loss": 0.3499, "step": 1249 }, { "epoch": 2.759381898454746, "grad_norm": 0.13384896516799927, "learning_rate": 1.9471557676287501e-07, "loss": 0.3556, "step": 1250 }, { "epoch": 2.76158940397351, "grad_norm": 0.13008341193199158, "learning_rate": 1.9118204340788426e-07, "loss": 0.3521, "step": 1251 }, { "epoch": 2.7637969094922736, "grad_norm": 0.14128050208091736, "learning_rate": 1.8768024113838546e-07, "loss": 0.3629, "step": 1252 }, { "epoch": 2.7660044150110377, "grad_norm": 0.13070915639400482, "learning_rate": 1.8421019306109288e-07, "loss": 0.3576, "step": 1253 }, { "epoch": 2.7682119205298013, "grad_norm": 0.13155661523342133, "learning_rate": 1.8077192207319072e-07, "loss": 0.3614, "step": 1254 }, { "epoch": 2.7704194260485653, "grad_norm": 0.14038477838039398, "learning_rate": 1.773654508621825e-07, "loss": 0.354, "step": 1255 }, { "epoch": 2.772626931567329, "grad_norm": 0.13292455673217773, "learning_rate": 1.7399080190573903e-07, "loss": 0.3635, "step": 1256 }, { "epoch": 2.774834437086093, "grad_norm": 0.1393938958644867, "learning_rate": 1.7064799747155248e-07, "loss": 0.3502, "step": 1257 }, { "epoch": 2.7770419426048565, "grad_norm": 0.13566245138645172, "learning_rate": 1.6733705961718694e-07, "loss": 0.353, "step": 1258 }, { "epoch": 2.77924944812362, "grad_norm": 0.1313687562942505, "learning_rate": 1.640580101899353e-07, "loss": 0.3519, "step": 1259 }, { "epoch": 2.781456953642384, "grad_norm": 0.13328541815280914, "learning_rate": 1.6081087082667314e-07, "loss": 0.3513, "step": 1260 }, { "epoch": 2.783664459161148, "grad_norm": 0.13640496134757996, "learning_rate": 1.575956629537173e-07, "loss": 0.3559, "step": 1261 }, { "epoch": 2.7858719646799117, "grad_norm": 0.13169561326503754, "learning_rate": 1.5441240778668321e-07, "loss": 0.3589, "step": 1262 }, { "epoch": 2.7880794701986753, "grad_norm": 0.1369987428188324, "learning_rate": 1.5126112633034761e-07, "loss": 0.368, "step": 1263 }, { "epoch": 2.7902869757174393, "grad_norm": 0.12467605620622635, "learning_rate": 1.4814183937850668e-07, "loss": 0.3563, "step": 1264 }, { "epoch": 2.7924944812362034, "grad_norm": 0.1307917833328247, "learning_rate": 1.4505456751383985e-07, "loss": 0.3516, "step": 1265 }, { "epoch": 2.794701986754967, "grad_norm": 0.14058281481266022, "learning_rate": 1.4199933110777553e-07, "loss": 0.3546, "step": 1266 }, { "epoch": 2.7969094922737305, "grad_norm": 0.12925602495670319, "learning_rate": 1.389761503203535e-07, "loss": 0.3615, "step": 1267 }, { "epoch": 2.7991169977924946, "grad_norm": 0.1409684419631958, "learning_rate": 1.3598504510009602e-07, "loss": 0.3611, "step": 1268 }, { "epoch": 2.801324503311258, "grad_norm": 0.13529111444950104, "learning_rate": 1.3302603518387358e-07, "loss": 0.3556, "step": 1269 }, { "epoch": 2.803532008830022, "grad_norm": 0.12611360847949982, "learning_rate": 1.3009914009677493e-07, "loss": 0.3695, "step": 1270 }, { "epoch": 2.8057395143487858, "grad_norm": 0.1310908943414688, "learning_rate": 1.272043791519778e-07, "loss": 0.3571, "step": 1271 }, { "epoch": 2.80794701986755, "grad_norm": 0.13993841409683228, "learning_rate": 1.2434177145062177e-07, "loss": 0.3572, "step": 1272 }, { "epoch": 2.8101545253863134, "grad_norm": 0.1199953481554985, "learning_rate": 1.215113358816844e-07, "loss": 0.3523, "step": 1273 }, { "epoch": 2.8123620309050774, "grad_norm": 0.12030813843011856, "learning_rate": 1.187130911218537e-07, "loss": 0.3533, "step": 1274 }, { "epoch": 2.814569536423841, "grad_norm": 0.1288137286901474, "learning_rate": 1.1594705563540642e-07, "loss": 0.3587, "step": 1275 }, { "epoch": 2.816777041942605, "grad_norm": 0.1391681730747223, "learning_rate": 1.1321324767408382e-07, "loss": 0.3592, "step": 1276 }, { "epoch": 2.8189845474613686, "grad_norm": 0.12392991036176682, "learning_rate": 1.1051168527697665e-07, "loss": 0.3605, "step": 1277 }, { "epoch": 2.821192052980132, "grad_norm": 0.12354668974876404, "learning_rate": 1.0784238627039977e-07, "loss": 0.3593, "step": 1278 }, { "epoch": 2.8233995584988962, "grad_norm": 0.13385632634162903, "learning_rate": 1.0520536826777783e-07, "loss": 0.3573, "step": 1279 }, { "epoch": 2.8256070640176603, "grad_norm": 0.12551288306713104, "learning_rate": 1.0260064866952968e-07, "loss": 0.3563, "step": 1280 }, { "epoch": 2.827814569536424, "grad_norm": 0.119255430996418, "learning_rate": 1.0002824466295191e-07, "loss": 0.3553, "step": 1281 }, { "epoch": 2.8300220750551874, "grad_norm": 0.12346068769693375, "learning_rate": 9.748817322210558e-08, "loss": 0.3547, "step": 1282 }, { "epoch": 2.8322295805739515, "grad_norm": 0.12130031734704971, "learning_rate": 9.498045110770571e-08, "loss": 0.3476, "step": 1283 }, { "epoch": 2.8344370860927155, "grad_norm": 0.13486534357070923, "learning_rate": 9.250509486700809e-08, "loss": 0.3652, "step": 1284 }, { "epoch": 2.836644591611479, "grad_norm": 0.12220917642116547, "learning_rate": 9.006212083370213e-08, "loss": 0.3548, "step": 1285 }, { "epoch": 2.8388520971302427, "grad_norm": 0.12486063688993454, "learning_rate": 8.765154512780428e-08, "loss": 0.3502, "step": 1286 }, { "epoch": 2.8410596026490067, "grad_norm": 0.11943277716636658, "learning_rate": 8.527338365554749e-08, "loss": 0.3462, "step": 1287 }, { "epoch": 2.8432671081677703, "grad_norm": 0.13910432159900665, "learning_rate": 8.292765210928089e-08, "loss": 0.3585, "step": 1288 }, { "epoch": 2.8454746136865343, "grad_norm": 0.12948773801326752, "learning_rate": 8.061436596736139e-08, "loss": 0.3589, "step": 1289 }, { "epoch": 2.847682119205298, "grad_norm": 0.13374410569667816, "learning_rate": 7.833354049405717e-08, "loss": 0.3505, "step": 1290 }, { "epoch": 2.849889624724062, "grad_norm": 0.1347481608390808, "learning_rate": 7.608519073944165e-08, "loss": 0.3687, "step": 1291 }, { "epoch": 2.8520971302428255, "grad_norm": 0.13521930575370789, "learning_rate": 7.386933153929798e-08, "loss": 0.366, "step": 1292 }, { "epoch": 2.8543046357615895, "grad_norm": 0.13502496480941772, "learning_rate": 7.168597751501972e-08, "loss": 0.366, "step": 1293 }, { "epoch": 2.856512141280353, "grad_norm": 0.1404918134212494, "learning_rate": 6.953514307351306e-08, "loss": 0.357, "step": 1294 }, { "epoch": 2.858719646799117, "grad_norm": 0.12767677009105682, "learning_rate": 6.741684240710477e-08, "loss": 0.3517, "step": 1295 }, { "epoch": 2.8609271523178808, "grad_norm": 0.13126912713050842, "learning_rate": 6.533108949344558e-08, "loss": 0.3564, "step": 1296 }, { "epoch": 2.8631346578366443, "grad_norm": 0.13056641817092896, "learning_rate": 6.32778980954185e-08, "loss": 0.3587, "step": 1297 }, { "epoch": 2.8653421633554084, "grad_norm": 0.13267594575881958, "learning_rate": 6.125728176105129e-08, "loss": 0.3686, "step": 1298 }, { "epoch": 2.8675496688741724, "grad_norm": 0.12520645558834076, "learning_rate": 5.9269253823421855e-08, "loss": 0.3703, "step": 1299 }, { "epoch": 2.869757174392936, "grad_norm": 0.1308923214673996, "learning_rate": 5.731382740057523e-08, "loss": 0.3549, "step": 1300 }, { "epoch": 2.8719646799116996, "grad_norm": 0.13203322887420654, "learning_rate": 5.5391015395432346e-08, "loss": 0.353, "step": 1301 }, { "epoch": 2.8741721854304636, "grad_norm": 0.13061153888702393, "learning_rate": 5.350083049570853e-08, "loss": 0.3556, "step": 1302 }, { "epoch": 2.8763796909492276, "grad_norm": 0.13314051926136017, "learning_rate": 5.164328517382744e-08, "loss": 0.3578, "step": 1303 }, { "epoch": 2.878587196467991, "grad_norm": 0.13481754064559937, "learning_rate": 4.9818391686842214e-08, "loss": 0.362, "step": 1304 }, { "epoch": 2.880794701986755, "grad_norm": 0.12934917211532593, "learning_rate": 4.802616207634947e-08, "loss": 0.3489, "step": 1305 }, { "epoch": 2.883002207505519, "grad_norm": 0.1276032030582428, "learning_rate": 4.626660816841433e-08, "loss": 0.3554, "step": 1306 }, { "epoch": 2.8852097130242824, "grad_norm": 0.12852588295936584, "learning_rate": 4.4539741573489395e-08, "loss": 0.3519, "step": 1307 }, { "epoch": 2.8874172185430464, "grad_norm": 0.1251831203699112, "learning_rate": 4.2845573686339235e-08, "loss": 0.3586, "step": 1308 }, { "epoch": 2.88962472406181, "grad_norm": 0.12635353207588196, "learning_rate": 4.118411568596714e-08, "loss": 0.3521, "step": 1309 }, { "epoch": 2.891832229580574, "grad_norm": 0.13549165427684784, "learning_rate": 3.9555378535537925e-08, "loss": 0.3573, "step": 1310 }, { "epoch": 2.8940397350993377, "grad_norm": 0.12129613012075424, "learning_rate": 3.795937298230801e-08, "loss": 0.3644, "step": 1311 }, { "epoch": 2.8962472406181017, "grad_norm": 0.12583400309085846, "learning_rate": 3.639610955755213e-08, "loss": 0.3628, "step": 1312 }, { "epoch": 2.8984547461368653, "grad_norm": 0.11590792238712311, "learning_rate": 3.486559857649785e-08, "loss": 0.3541, "step": 1313 }, { "epoch": 2.9006622516556293, "grad_norm": 0.1254376322031021, "learning_rate": 3.336785013825339e-08, "loss": 0.3591, "step": 1314 }, { "epoch": 2.902869757174393, "grad_norm": 0.12448057532310486, "learning_rate": 3.190287412574267e-08, "loss": 0.3685, "step": 1315 }, { "epoch": 2.9050772626931565, "grad_norm": 0.12795038521289825, "learning_rate": 3.047068020564037e-08, "loss": 0.3661, "step": 1316 }, { "epoch": 2.9072847682119205, "grad_norm": 0.1275874376296997, "learning_rate": 2.9071277828308654e-08, "loss": 0.3547, "step": 1317 }, { "epoch": 2.9094922737306845, "grad_norm": 0.13140250742435455, "learning_rate": 2.7704676227732764e-08, "loss": 0.3476, "step": 1318 }, { "epoch": 2.911699779249448, "grad_norm": 0.13054263591766357, "learning_rate": 2.637088442146163e-08, "loss": 0.3555, "step": 1319 }, { "epoch": 2.9139072847682117, "grad_norm": 0.14265048503875732, "learning_rate": 2.506991121054847e-08, "loss": 0.3552, "step": 1320 }, { "epoch": 2.9161147902869757, "grad_norm": 0.12008768320083618, "learning_rate": 2.380176517949251e-08, "loss": 0.3555, "step": 1321 }, { "epoch": 2.9183222958057398, "grad_norm": 0.13723327219486237, "learning_rate": 2.256645469618124e-08, "loss": 0.3578, "step": 1322 }, { "epoch": 2.9205298013245033, "grad_norm": 0.13760052621364594, "learning_rate": 2.136398791183658e-08, "loss": 0.3543, "step": 1323 }, { "epoch": 2.922737306843267, "grad_norm": 0.129390150308609, "learning_rate": 2.0194372760961034e-08, "loss": 0.362, "step": 1324 }, { "epoch": 2.924944812362031, "grad_norm": 0.15059252083301544, "learning_rate": 1.905761696128494e-08, "loss": 0.3611, "step": 1325 }, { "epoch": 2.9271523178807946, "grad_norm": 0.13671807944774628, "learning_rate": 1.795372801371431e-08, "loss": 0.3601, "step": 1326 }, { "epoch": 2.9293598233995586, "grad_norm": 0.138703852891922, "learning_rate": 1.6882713202283076e-08, "loss": 0.3525, "step": 1327 }, { "epoch": 2.931567328918322, "grad_norm": 0.12975488603115082, "learning_rate": 1.5844579594105904e-08, "loss": 0.3636, "step": 1328 }, { "epoch": 2.933774834437086, "grad_norm": 0.15756423771381378, "learning_rate": 1.48393340393288e-08, "loss": 0.3586, "step": 1329 }, { "epoch": 2.93598233995585, "grad_norm": 0.13145272433757782, "learning_rate": 1.3866983171084703e-08, "loss": 0.3563, "step": 1330 }, { "epoch": 2.938189845474614, "grad_norm": 0.15646247565746307, "learning_rate": 1.292753340545183e-08, "loss": 0.3567, "step": 1331 }, { "epoch": 2.9403973509933774, "grad_norm": 0.1439492553472519, "learning_rate": 1.2020990941408739e-08, "loss": 0.3678, "step": 1332 }, { "epoch": 2.9426048565121414, "grad_norm": 0.1377602368593216, "learning_rate": 1.1147361760794895e-08, "loss": 0.3547, "step": 1333 }, { "epoch": 2.944812362030905, "grad_norm": 0.14668287336826324, "learning_rate": 1.0306651628270715e-08, "loss": 0.3672, "step": 1334 }, { "epoch": 2.9470198675496686, "grad_norm": 0.12802091240882874, "learning_rate": 9.498866091278702e-09, "loss": 0.3596, "step": 1335 }, { "epoch": 2.9492273730684326, "grad_norm": 0.15803688764572144, "learning_rate": 8.724010480010147e-09, "loss": 0.3586, "step": 1336 }, { "epoch": 2.9514348785871967, "grad_norm": 0.13727080821990967, "learning_rate": 7.982089907364598e-09, "loss": 0.3515, "step": 1337 }, { "epoch": 2.9536423841059603, "grad_norm": 0.14836713671684265, "learning_rate": 7.273109268920997e-09, "loss": 0.3572, "step": 1338 }, { "epoch": 2.955849889624724, "grad_norm": 0.15154053270816803, "learning_rate": 6.597073242902707e-09, "loss": 0.3632, "step": 1339 }, { "epoch": 2.958057395143488, "grad_norm": 0.14268819987773895, "learning_rate": 5.95398629014754e-09, "loss": 0.3737, "step": 1340 }, { "epoch": 2.960264900662252, "grad_norm": 0.13592885434627533, "learning_rate": 5.3438526540777745e-09, "loss": 0.3647, "step": 1341 }, { "epoch": 2.9624724061810155, "grad_norm": 0.15282437205314636, "learning_rate": 4.766676360674072e-09, "loss": 0.3666, "step": 1342 }, { "epoch": 2.964679911699779, "grad_norm": 0.14543361961841583, "learning_rate": 4.22246121844494e-09, "loss": 0.3576, "step": 1343 }, { "epoch": 2.966887417218543, "grad_norm": 0.1356423795223236, "learning_rate": 3.7112108184061966e-09, "loss": 0.3646, "step": 1344 }, { "epoch": 2.9690949227373067, "grad_norm": 0.13516850769519806, "learning_rate": 3.2329285340537696e-09, "loss": 0.3577, "step": 1345 }, { "epoch": 2.9713024282560707, "grad_norm": 0.12816190719604492, "learning_rate": 2.7876175213431557e-09, "loss": 0.3568, "step": 1346 }, { "epoch": 2.9735099337748343, "grad_norm": 0.13490551710128784, "learning_rate": 2.375280718668882e-09, "loss": 0.3548, "step": 1347 }, { "epoch": 2.9757174392935983, "grad_norm": 0.14126001298427582, "learning_rate": 1.995920846843968e-09, "loss": 0.3666, "step": 1348 }, { "epoch": 2.977924944812362, "grad_norm": 0.12279074639081955, "learning_rate": 1.649540409081607e-09, "loss": 0.3562, "step": 1349 }, { "epoch": 2.980132450331126, "grad_norm": 0.12752176821231842, "learning_rate": 1.3361416909812852e-09, "loss": 0.365, "step": 1350 }, { "epoch": 2.9823399558498895, "grad_norm": 0.13065557181835175, "learning_rate": 1.055726760510467e-09, "loss": 0.3615, "step": 1351 }, { "epoch": 2.9845474613686536, "grad_norm": 0.11860226094722748, "learning_rate": 8.082974679929357e-10, "loss": 0.3564, "step": 1352 }, { "epoch": 2.986754966887417, "grad_norm": 0.1291607767343521, "learning_rate": 5.938554460965807e-10, "loss": 0.359, "step": 1353 }, { "epoch": 2.988962472406181, "grad_norm": 0.146859273314476, "learning_rate": 4.12402109820631e-10, "loss": 0.3614, "step": 1354 }, { "epoch": 2.9911699779249448, "grad_norm": 0.12649129331111908, "learning_rate": 2.639386564889934e-10, "loss": 0.3566, "step": 1355 }, { "epoch": 2.993377483443709, "grad_norm": 0.14126034080982208, "learning_rate": 1.4846606574137058e-10, "loss": 0.3608, "step": 1356 }, { "epoch": 2.9955849889624724, "grad_norm": 0.13480675220489502, "learning_rate": 6.59850995254896e-11, "loss": 0.3555, "step": 1357 }, { "epoch": 2.997792494481236, "grad_norm": 0.14647357165813446, "learning_rate": 1.649630209432651e-11, "loss": 0.3643, "step": 1358 }, { "epoch": 3.0, "grad_norm": 0.13931505382061005, "learning_rate": 0.0, "loss": 0.3486, "step": 1359 }, { "epoch": 3.0, "step": 1359, "total_flos": 2.7571726585954304e+16, "train_loss": 0.13500037013037053, "train_runtime": 42654.1727, "train_samples_per_second": 12.208, "train_steps_per_second": 0.032 } ], "logging_steps": 1, "max_steps": 1359, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.7571726585954304e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }