{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1356, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0022123893805309734, "grad_norm": 2.9161036014556885, "learning_rate": 7.352941176470589e-08, "loss": 0.7653, "step": 1 }, { "epoch": 0.004424778761061947, "grad_norm": 2.9344425201416016, "learning_rate": 1.4705882352941178e-07, "loss": 0.7756, "step": 2 }, { "epoch": 0.00663716814159292, "grad_norm": 2.8798561096191406, "learning_rate": 2.2058823529411768e-07, "loss": 0.7784, "step": 3 }, { "epoch": 0.008849557522123894, "grad_norm": 2.898063898086548, "learning_rate": 2.9411764705882356e-07, "loss": 0.7798, "step": 4 }, { "epoch": 0.011061946902654867, "grad_norm": 2.915313959121704, "learning_rate": 3.6764705882352943e-07, "loss": 0.7844, "step": 5 }, { "epoch": 0.01327433628318584, "grad_norm": 2.7517311573028564, "learning_rate": 4.4117647058823536e-07, "loss": 0.7519, "step": 6 }, { "epoch": 0.015486725663716814, "grad_norm": 2.8670687675476074, "learning_rate": 5.147058823529412e-07, "loss": 0.7639, "step": 7 }, { "epoch": 0.017699115044247787, "grad_norm": 2.838557243347168, "learning_rate": 5.882352941176471e-07, "loss": 0.7794, "step": 8 }, { "epoch": 0.01991150442477876, "grad_norm": 2.6696882247924805, "learning_rate": 6.61764705882353e-07, "loss": 0.7539, "step": 9 }, { "epoch": 0.022123893805309734, "grad_norm": 2.657489061355591, "learning_rate": 7.352941176470589e-07, "loss": 0.7437, "step": 10 }, { "epoch": 0.024336283185840708, "grad_norm": 2.6917357444763184, "learning_rate": 8.088235294117648e-07, "loss": 0.763, "step": 11 }, { "epoch": 0.02654867256637168, "grad_norm": 2.2191965579986572, "learning_rate": 8.823529411764707e-07, "loss": 0.7424, "step": 12 }, { "epoch": 0.028761061946902654, "grad_norm": 2.2132158279418945, "learning_rate": 9.558823529411764e-07, "loss": 0.7238, "step": 13 }, { "epoch": 0.030973451327433628, "grad_norm": 2.1964025497436523, "learning_rate": 1.0294117647058825e-06, "loss": 0.7384, "step": 14 }, { "epoch": 0.033185840707964605, "grad_norm": 2.0652434825897217, "learning_rate": 1.1029411764705884e-06, "loss": 0.7366, "step": 15 }, { "epoch": 0.035398230088495575, "grad_norm": 1.6799640655517578, "learning_rate": 1.1764705882352942e-06, "loss": 0.7022, "step": 16 }, { "epoch": 0.03761061946902655, "grad_norm": 1.4400370121002197, "learning_rate": 1.25e-06, "loss": 0.6861, "step": 17 }, { "epoch": 0.03982300884955752, "grad_norm": 1.4010155200958252, "learning_rate": 1.323529411764706e-06, "loss": 0.6904, "step": 18 }, { "epoch": 0.0420353982300885, "grad_norm": 1.3582905530929565, "learning_rate": 1.3970588235294119e-06, "loss": 0.6876, "step": 19 }, { "epoch": 0.04424778761061947, "grad_norm": 1.3372609615325928, "learning_rate": 1.4705882352941177e-06, "loss": 0.6809, "step": 20 }, { "epoch": 0.046460176991150445, "grad_norm": 1.3210690021514893, "learning_rate": 1.5441176470588238e-06, "loss": 0.687, "step": 21 }, { "epoch": 0.048672566371681415, "grad_norm": 1.0419282913208008, "learning_rate": 1.6176470588235297e-06, "loss": 0.6743, "step": 22 }, { "epoch": 0.05088495575221239, "grad_norm": 0.9881656765937805, "learning_rate": 1.6911764705882356e-06, "loss": 0.6483, "step": 23 }, { "epoch": 0.05309734513274336, "grad_norm": 0.9939265251159668, "learning_rate": 1.7647058823529414e-06, "loss": 0.6527, "step": 24 }, { "epoch": 0.05530973451327434, "grad_norm": 0.9658608436584473, "learning_rate": 1.8382352941176473e-06, "loss": 0.6544, "step": 25 }, { "epoch": 0.05752212389380531, "grad_norm": 0.897885262966156, "learning_rate": 1.9117647058823528e-06, "loss": 0.6358, "step": 26 }, { "epoch": 0.059734513274336286, "grad_norm": 0.8456088900566101, "learning_rate": 1.985294117647059e-06, "loss": 0.6403, "step": 27 }, { "epoch": 0.061946902654867256, "grad_norm": 0.8407159447669983, "learning_rate": 2.058823529411765e-06, "loss": 0.6366, "step": 28 }, { "epoch": 0.06415929203539823, "grad_norm": 0.7394031286239624, "learning_rate": 2.132352941176471e-06, "loss": 0.6277, "step": 29 }, { "epoch": 0.06637168141592921, "grad_norm": 0.6359221339225769, "learning_rate": 2.2058823529411767e-06, "loss": 0.6088, "step": 30 }, { "epoch": 0.06858407079646017, "grad_norm": 0.6923512816429138, "learning_rate": 2.2794117647058826e-06, "loss": 0.6072, "step": 31 }, { "epoch": 0.07079646017699115, "grad_norm": 0.6872782707214355, "learning_rate": 2.3529411764705885e-06, "loss": 0.5939, "step": 32 }, { "epoch": 0.07300884955752213, "grad_norm": 0.6121441721916199, "learning_rate": 2.4264705882352943e-06, "loss": 0.5936, "step": 33 }, { "epoch": 0.0752212389380531, "grad_norm": 0.5648632645606995, "learning_rate": 2.5e-06, "loss": 0.5925, "step": 34 }, { "epoch": 0.07743362831858407, "grad_norm": 0.527875542640686, "learning_rate": 2.5735294117647057e-06, "loss": 0.5788, "step": 35 }, { "epoch": 0.07964601769911504, "grad_norm": 0.48248472809791565, "learning_rate": 2.647058823529412e-06, "loss": 0.584, "step": 36 }, { "epoch": 0.08185840707964602, "grad_norm": 0.5289769172668457, "learning_rate": 2.720588235294118e-06, "loss": 0.5811, "step": 37 }, { "epoch": 0.084070796460177, "grad_norm": 0.5214428901672363, "learning_rate": 2.7941176470588237e-06, "loss": 0.5592, "step": 38 }, { "epoch": 0.08628318584070796, "grad_norm": 0.4773428440093994, "learning_rate": 2.8676470588235296e-06, "loss": 0.558, "step": 39 }, { "epoch": 0.08849557522123894, "grad_norm": 0.4300067722797394, "learning_rate": 2.9411764705882355e-06, "loss": 0.5592, "step": 40 }, { "epoch": 0.09070796460176991, "grad_norm": 0.4010540246963501, "learning_rate": 3.0147058823529413e-06, "loss": 0.559, "step": 41 }, { "epoch": 0.09292035398230089, "grad_norm": 0.3986721336841583, "learning_rate": 3.0882352941176476e-06, "loss": 0.5653, "step": 42 }, { "epoch": 0.09513274336283185, "grad_norm": 0.37803158164024353, "learning_rate": 3.161764705882353e-06, "loss": 0.5421, "step": 43 }, { "epoch": 0.09734513274336283, "grad_norm": 0.3909413516521454, "learning_rate": 3.2352941176470594e-06, "loss": 0.5347, "step": 44 }, { "epoch": 0.09955752212389381, "grad_norm": 0.373319536447525, "learning_rate": 3.308823529411765e-06, "loss": 0.5262, "step": 45 }, { "epoch": 0.10176991150442478, "grad_norm": 0.38008248805999756, "learning_rate": 3.382352941176471e-06, "loss": 0.5285, "step": 46 }, { "epoch": 0.10398230088495575, "grad_norm": 0.37302571535110474, "learning_rate": 3.4558823529411766e-06, "loss": 0.5132, "step": 47 }, { "epoch": 0.10619469026548672, "grad_norm": 0.3702549636363983, "learning_rate": 3.529411764705883e-06, "loss": 0.5228, "step": 48 }, { "epoch": 0.1084070796460177, "grad_norm": 0.3243614733219147, "learning_rate": 3.6029411764705883e-06, "loss": 0.5126, "step": 49 }, { "epoch": 0.11061946902654868, "grad_norm": 0.2833242118358612, "learning_rate": 3.6764705882352946e-06, "loss": 0.5061, "step": 50 }, { "epoch": 0.11283185840707964, "grad_norm": 0.27726179361343384, "learning_rate": 3.7500000000000005e-06, "loss": 0.5054, "step": 51 }, { "epoch": 0.11504424778761062, "grad_norm": 0.2611929774284363, "learning_rate": 3.8235294117647055e-06, "loss": 0.5136, "step": 52 }, { "epoch": 0.1172566371681416, "grad_norm": 0.2536194324493408, "learning_rate": 3.897058823529412e-06, "loss": 0.5016, "step": 53 }, { "epoch": 0.11946902654867257, "grad_norm": 0.24586400389671326, "learning_rate": 3.970588235294118e-06, "loss": 0.5017, "step": 54 }, { "epoch": 0.12168141592920353, "grad_norm": 0.22390125691890717, "learning_rate": 4.044117647058824e-06, "loss": 0.5015, "step": 55 }, { "epoch": 0.12389380530973451, "grad_norm": 0.2428012639284134, "learning_rate": 4.11764705882353e-06, "loss": 0.5125, "step": 56 }, { "epoch": 0.1261061946902655, "grad_norm": 0.2092347890138626, "learning_rate": 4.191176470588236e-06, "loss": 0.4978, "step": 57 }, { "epoch": 0.12831858407079647, "grad_norm": 0.2112482786178589, "learning_rate": 4.264705882352942e-06, "loss": 0.5117, "step": 58 }, { "epoch": 0.13053097345132744, "grad_norm": 0.2061765491962433, "learning_rate": 4.3382352941176475e-06, "loss": 0.4917, "step": 59 }, { "epoch": 0.13274336283185842, "grad_norm": 0.20188196003437042, "learning_rate": 4.411764705882353e-06, "loss": 0.4958, "step": 60 }, { "epoch": 0.13495575221238937, "grad_norm": 0.21565476059913635, "learning_rate": 4.485294117647059e-06, "loss": 0.4839, "step": 61 }, { "epoch": 0.13716814159292035, "grad_norm": 0.22550469636917114, "learning_rate": 4.558823529411765e-06, "loss": 0.4919, "step": 62 }, { "epoch": 0.13938053097345132, "grad_norm": 0.20743007957935333, "learning_rate": 4.632352941176471e-06, "loss": 0.4937, "step": 63 }, { "epoch": 0.1415929203539823, "grad_norm": 0.2033708542585373, "learning_rate": 4.705882352941177e-06, "loss": 0.4858, "step": 64 }, { "epoch": 0.14380530973451328, "grad_norm": 0.21126723289489746, "learning_rate": 4.779411764705883e-06, "loss": 0.487, "step": 65 }, { "epoch": 0.14601769911504425, "grad_norm": 0.2027079313993454, "learning_rate": 4.852941176470589e-06, "loss": 0.4975, "step": 66 }, { "epoch": 0.14823008849557523, "grad_norm": 0.18106848001480103, "learning_rate": 4.9264705882352945e-06, "loss": 0.4894, "step": 67 }, { "epoch": 0.1504424778761062, "grad_norm": 0.17369899153709412, "learning_rate": 5e-06, "loss": 0.4825, "step": 68 }, { "epoch": 0.15265486725663716, "grad_norm": 0.1895466446876526, "learning_rate": 5.073529411764706e-06, "loss": 0.4792, "step": 69 }, { "epoch": 0.15486725663716813, "grad_norm": 0.17305336892604828, "learning_rate": 5.147058823529411e-06, "loss": 0.4715, "step": 70 }, { "epoch": 0.1570796460176991, "grad_norm": 0.1629684716463089, "learning_rate": 5.220588235294118e-06, "loss": 0.4696, "step": 71 }, { "epoch": 0.1592920353982301, "grad_norm": 0.1580497771501541, "learning_rate": 5.294117647058824e-06, "loss": 0.4647, "step": 72 }, { "epoch": 0.16150442477876106, "grad_norm": 0.16177481412887573, "learning_rate": 5.36764705882353e-06, "loss": 0.4667, "step": 73 }, { "epoch": 0.16371681415929204, "grad_norm": 0.14660590887069702, "learning_rate": 5.441176470588236e-06, "loss": 0.4702, "step": 74 }, { "epoch": 0.16592920353982302, "grad_norm": 0.16958165168762207, "learning_rate": 5.514705882352942e-06, "loss": 0.4664, "step": 75 }, { "epoch": 0.168141592920354, "grad_norm": 0.18446360528469086, "learning_rate": 5.588235294117647e-06, "loss": 0.4634, "step": 76 }, { "epoch": 0.17035398230088494, "grad_norm": 0.18088003993034363, "learning_rate": 5.661764705882353e-06, "loss": 0.4782, "step": 77 }, { "epoch": 0.17256637168141592, "grad_norm": 0.16989891231060028, "learning_rate": 5.735294117647059e-06, "loss": 0.4751, "step": 78 }, { "epoch": 0.1747787610619469, "grad_norm": 0.15071384608745575, "learning_rate": 5.808823529411766e-06, "loss": 0.4679, "step": 79 }, { "epoch": 0.17699115044247787, "grad_norm": 0.16606192290782928, "learning_rate": 5.882352941176471e-06, "loss": 0.4613, "step": 80 }, { "epoch": 0.17920353982300885, "grad_norm": 0.16193555295467377, "learning_rate": 5.955882352941177e-06, "loss": 0.4656, "step": 81 }, { "epoch": 0.18141592920353983, "grad_norm": 0.14852063357830048, "learning_rate": 6.029411764705883e-06, "loss": 0.4508, "step": 82 }, { "epoch": 0.1836283185840708, "grad_norm": 0.16259585320949554, "learning_rate": 6.102941176470589e-06, "loss": 0.453, "step": 83 }, { "epoch": 0.18584070796460178, "grad_norm": 0.17288607358932495, "learning_rate": 6.176470588235295e-06, "loss": 0.4597, "step": 84 }, { "epoch": 0.18805309734513273, "grad_norm": 0.15836317837238312, "learning_rate": 6.25e-06, "loss": 0.4548, "step": 85 }, { "epoch": 0.1902654867256637, "grad_norm": 0.17005616426467896, "learning_rate": 6.323529411764706e-06, "loss": 0.4608, "step": 86 }, { "epoch": 0.19247787610619468, "grad_norm": 0.148678258061409, "learning_rate": 6.397058823529412e-06, "loss": 0.4599, "step": 87 }, { "epoch": 0.19469026548672566, "grad_norm": 0.16987866163253784, "learning_rate": 6.470588235294119e-06, "loss": 0.4512, "step": 88 }, { "epoch": 0.19690265486725664, "grad_norm": 0.15269418060779572, "learning_rate": 6.544117647058824e-06, "loss": 0.4579, "step": 89 }, { "epoch": 0.19911504424778761, "grad_norm": 0.1564721018075943, "learning_rate": 6.61764705882353e-06, "loss": 0.4577, "step": 90 }, { "epoch": 0.2013274336283186, "grad_norm": 0.15829169750213623, "learning_rate": 6.6911764705882356e-06, "loss": 0.4628, "step": 91 }, { "epoch": 0.20353982300884957, "grad_norm": 0.16252678632736206, "learning_rate": 6.764705882352942e-06, "loss": 0.4508, "step": 92 }, { "epoch": 0.20575221238938052, "grad_norm": 0.16816888749599457, "learning_rate": 6.838235294117648e-06, "loss": 0.4569, "step": 93 }, { "epoch": 0.2079646017699115, "grad_norm": 0.16243158280849457, "learning_rate": 6.911764705882353e-06, "loss": 0.4492, "step": 94 }, { "epoch": 0.21017699115044247, "grad_norm": 0.1554814875125885, "learning_rate": 6.985294117647059e-06, "loss": 0.4501, "step": 95 }, { "epoch": 0.21238938053097345, "grad_norm": 0.15875332057476044, "learning_rate": 7.058823529411766e-06, "loss": 0.4581, "step": 96 }, { "epoch": 0.21460176991150443, "grad_norm": 0.16014939546585083, "learning_rate": 7.132352941176472e-06, "loss": 0.448, "step": 97 }, { "epoch": 0.2168141592920354, "grad_norm": 0.18041987717151642, "learning_rate": 7.205882352941177e-06, "loss": 0.4553, "step": 98 }, { "epoch": 0.21902654867256638, "grad_norm": 0.16102494299411774, "learning_rate": 7.2794117647058826e-06, "loss": 0.4553, "step": 99 }, { "epoch": 0.22123893805309736, "grad_norm": 0.1710263043642044, "learning_rate": 7.352941176470589e-06, "loss": 0.4426, "step": 100 }, { "epoch": 0.2234513274336283, "grad_norm": 0.17998401820659637, "learning_rate": 7.426470588235295e-06, "loss": 0.439, "step": 101 }, { "epoch": 0.22566371681415928, "grad_norm": 0.17915406823158264, "learning_rate": 7.500000000000001e-06, "loss": 0.4467, "step": 102 }, { "epoch": 0.22787610619469026, "grad_norm": 0.16851037740707397, "learning_rate": 7.573529411764706e-06, "loss": 0.4363, "step": 103 }, { "epoch": 0.23008849557522124, "grad_norm": 0.17610934376716614, "learning_rate": 7.647058823529411e-06, "loss": 0.4493, "step": 104 }, { "epoch": 0.2323008849557522, "grad_norm": 0.18994346261024475, "learning_rate": 7.720588235294119e-06, "loss": 0.4485, "step": 105 }, { "epoch": 0.2345132743362832, "grad_norm": 0.1732485294342041, "learning_rate": 7.794117647058825e-06, "loss": 0.4396, "step": 106 }, { "epoch": 0.23672566371681417, "grad_norm": 0.16453640162944794, "learning_rate": 7.86764705882353e-06, "loss": 0.4442, "step": 107 }, { "epoch": 0.23893805309734514, "grad_norm": 0.1798321008682251, "learning_rate": 7.941176470588236e-06, "loss": 0.44, "step": 108 }, { "epoch": 0.2411504424778761, "grad_norm": 0.17025865614414215, "learning_rate": 8.014705882352942e-06, "loss": 0.436, "step": 109 }, { "epoch": 0.24336283185840707, "grad_norm": 0.17872163653373718, "learning_rate": 8.088235294117648e-06, "loss": 0.4393, "step": 110 }, { "epoch": 0.24557522123893805, "grad_norm": 0.17654740810394287, "learning_rate": 8.161764705882354e-06, "loss": 0.4453, "step": 111 }, { "epoch": 0.24778761061946902, "grad_norm": 0.1653042584657669, "learning_rate": 8.23529411764706e-06, "loss": 0.429, "step": 112 }, { "epoch": 0.25, "grad_norm": 0.16623114049434662, "learning_rate": 8.308823529411766e-06, "loss": 0.4378, "step": 113 }, { "epoch": 0.252212389380531, "grad_norm": 0.17289873957633972, "learning_rate": 8.382352941176472e-06, "loss": 0.4386, "step": 114 }, { "epoch": 0.25442477876106195, "grad_norm": 0.17378394305706024, "learning_rate": 8.455882352941177e-06, "loss": 0.4337, "step": 115 }, { "epoch": 0.25663716814159293, "grad_norm": 0.15947799384593964, "learning_rate": 8.529411764705883e-06, "loss": 0.4426, "step": 116 }, { "epoch": 0.2588495575221239, "grad_norm": 0.20766223967075348, "learning_rate": 8.60294117647059e-06, "loss": 0.4345, "step": 117 }, { "epoch": 0.2610619469026549, "grad_norm": 0.19195608794689178, "learning_rate": 8.676470588235295e-06, "loss": 0.4288, "step": 118 }, { "epoch": 0.26327433628318586, "grad_norm": 0.1608441323041916, "learning_rate": 8.750000000000001e-06, "loss": 0.4431, "step": 119 }, { "epoch": 0.26548672566371684, "grad_norm": 0.21266116201877594, "learning_rate": 8.823529411764707e-06, "loss": 0.4312, "step": 120 }, { "epoch": 0.2676991150442478, "grad_norm": 0.21518948674201965, "learning_rate": 8.897058823529413e-06, "loss": 0.4296, "step": 121 }, { "epoch": 0.26991150442477874, "grad_norm": 0.1727646440267563, "learning_rate": 8.970588235294119e-06, "loss": 0.432, "step": 122 }, { "epoch": 0.2721238938053097, "grad_norm": 0.25698140263557434, "learning_rate": 9.044117647058824e-06, "loss": 0.4245, "step": 123 }, { "epoch": 0.2743362831858407, "grad_norm": 0.20740661025047302, "learning_rate": 9.11764705882353e-06, "loss": 0.4407, "step": 124 }, { "epoch": 0.27654867256637167, "grad_norm": 0.19939358532428741, "learning_rate": 9.191176470588236e-06, "loss": 0.4478, "step": 125 }, { "epoch": 0.27876106194690264, "grad_norm": 0.24062572419643402, "learning_rate": 9.264705882352942e-06, "loss": 0.4259, "step": 126 }, { "epoch": 0.2809734513274336, "grad_norm": 0.2283308357000351, "learning_rate": 9.338235294117648e-06, "loss": 0.4377, "step": 127 }, { "epoch": 0.2831858407079646, "grad_norm": 0.20251670479774475, "learning_rate": 9.411764705882354e-06, "loss": 0.4399, "step": 128 }, { "epoch": 0.2853982300884956, "grad_norm": 0.23135575652122498, "learning_rate": 9.48529411764706e-06, "loss": 0.4374, "step": 129 }, { "epoch": 0.28761061946902655, "grad_norm": 0.1677524596452713, "learning_rate": 9.558823529411766e-06, "loss": 0.4382, "step": 130 }, { "epoch": 0.28982300884955753, "grad_norm": 0.19608592987060547, "learning_rate": 9.632352941176471e-06, "loss": 0.4319, "step": 131 }, { "epoch": 0.2920353982300885, "grad_norm": 0.1991558074951172, "learning_rate": 9.705882352941177e-06, "loss": 0.4294, "step": 132 }, { "epoch": 0.2942477876106195, "grad_norm": 0.17177461087703705, "learning_rate": 9.779411764705883e-06, "loss": 0.4349, "step": 133 }, { "epoch": 0.29646017699115046, "grad_norm": 0.1729421466588974, "learning_rate": 9.852941176470589e-06, "loss": 0.4283, "step": 134 }, { "epoch": 0.29867256637168144, "grad_norm": 0.2021639496088028, "learning_rate": 9.926470588235295e-06, "loss": 0.4377, "step": 135 }, { "epoch": 0.3008849557522124, "grad_norm": 0.18062114715576172, "learning_rate": 1e-05, "loss": 0.4427, "step": 136 }, { "epoch": 0.3030973451327434, "grad_norm": 0.16362692415714264, "learning_rate": 9.999983422468849e-06, "loss": 0.4395, "step": 137 }, { "epoch": 0.3053097345132743, "grad_norm": 0.20908989012241364, "learning_rate": 9.999933689985315e-06, "loss": 0.4363, "step": 138 }, { "epoch": 0.3075221238938053, "grad_norm": 0.16169682145118713, "learning_rate": 9.999850802879177e-06, "loss": 0.4303, "step": 139 }, { "epoch": 0.30973451327433627, "grad_norm": 0.17690162360668182, "learning_rate": 9.999734761700061e-06, "loss": 0.4266, "step": 140 }, { "epoch": 0.31194690265486724, "grad_norm": 0.17501935362815857, "learning_rate": 9.99958556721744e-06, "loss": 0.4336, "step": 141 }, { "epoch": 0.3141592920353982, "grad_norm": 0.16887253522872925, "learning_rate": 9.999403220420619e-06, "loss": 0.4351, "step": 142 }, { "epoch": 0.3163716814159292, "grad_norm": 0.1647050976753235, "learning_rate": 9.999187722518747e-06, "loss": 0.426, "step": 143 }, { "epoch": 0.3185840707964602, "grad_norm": 0.18983641266822815, "learning_rate": 9.998939074940788e-06, "loss": 0.4219, "step": 144 }, { "epoch": 0.32079646017699115, "grad_norm": 0.19264842569828033, "learning_rate": 9.99865727933553e-06, "loss": 0.4237, "step": 145 }, { "epoch": 0.3230088495575221, "grad_norm": 0.18118713796138763, "learning_rate": 9.998342337571566e-06, "loss": 0.4174, "step": 146 }, { "epoch": 0.3252212389380531, "grad_norm": 0.22882233560085297, "learning_rate": 9.997994251737276e-06, "loss": 0.4277, "step": 147 }, { "epoch": 0.3274336283185841, "grad_norm": 0.17980162799358368, "learning_rate": 9.99761302414082e-06, "loss": 0.4257, "step": 148 }, { "epoch": 0.32964601769911506, "grad_norm": 0.2404302954673767, "learning_rate": 9.997198657310126e-06, "loss": 0.4305, "step": 149 }, { "epoch": 0.33185840707964603, "grad_norm": 0.19852545857429504, "learning_rate": 9.996751153992861e-06, "loss": 0.4212, "step": 150 }, { "epoch": 0.334070796460177, "grad_norm": 0.22487014532089233, "learning_rate": 9.996270517156431e-06, "loss": 0.4257, "step": 151 }, { "epoch": 0.336283185840708, "grad_norm": 0.22307410836219788, "learning_rate": 9.995756749987942e-06, "loss": 0.4208, "step": 152 }, { "epoch": 0.33849557522123896, "grad_norm": 0.2162560373544693, "learning_rate": 9.995209855894191e-06, "loss": 0.4206, "step": 153 }, { "epoch": 0.3407079646017699, "grad_norm": 0.20444706082344055, "learning_rate": 9.994629838501637e-06, "loss": 0.4295, "step": 154 }, { "epoch": 0.34292035398230086, "grad_norm": 0.1971275806427002, "learning_rate": 9.994016701656384e-06, "loss": 0.4299, "step": 155 }, { "epoch": 0.34513274336283184, "grad_norm": 0.19772182404994965, "learning_rate": 9.993370449424153e-06, "loss": 0.4201, "step": 156 }, { "epoch": 0.3473451327433628, "grad_norm": 0.1802293360233307, "learning_rate": 9.992691086090249e-06, "loss": 0.4254, "step": 157 }, { "epoch": 0.3495575221238938, "grad_norm": 0.1828855574131012, "learning_rate": 9.991978616159535e-06, "loss": 0.4279, "step": 158 }, { "epoch": 0.35176991150442477, "grad_norm": 0.16842177510261536, "learning_rate": 9.991233044356414e-06, "loss": 0.4176, "step": 159 }, { "epoch": 0.35398230088495575, "grad_norm": 0.1759248673915863, "learning_rate": 9.990454375624778e-06, "loss": 0.4266, "step": 160 }, { "epoch": 0.3561946902654867, "grad_norm": 0.18696172535419464, "learning_rate": 9.98964261512799e-06, "loss": 0.4195, "step": 161 }, { "epoch": 0.3584070796460177, "grad_norm": 0.16631002724170685, "learning_rate": 9.988797768248844e-06, "loss": 0.4272, "step": 162 }, { "epoch": 0.3606194690265487, "grad_norm": 0.16732241213321686, "learning_rate": 9.987919840589529e-06, "loss": 0.4195, "step": 163 }, { "epoch": 0.36283185840707965, "grad_norm": 0.18925903737545013, "learning_rate": 9.987008837971595e-06, "loss": 0.4202, "step": 164 }, { "epoch": 0.36504424778761063, "grad_norm": 0.17854173481464386, "learning_rate": 9.986064766435915e-06, "loss": 0.4247, "step": 165 }, { "epoch": 0.3672566371681416, "grad_norm": 0.20543624460697174, "learning_rate": 9.985087632242634e-06, "loss": 0.4163, "step": 166 }, { "epoch": 0.3694690265486726, "grad_norm": 0.17206773161888123, "learning_rate": 9.984077441871144e-06, "loss": 0.4245, "step": 167 }, { "epoch": 0.37168141592920356, "grad_norm": 0.19772297143936157, "learning_rate": 9.98303420202003e-06, "loss": 0.406, "step": 168 }, { "epoch": 0.37389380530973454, "grad_norm": 0.21683964133262634, "learning_rate": 9.981957919607026e-06, "loss": 0.4302, "step": 169 }, { "epoch": 0.37610619469026546, "grad_norm": 0.17397964000701904, "learning_rate": 9.980848601768976e-06, "loss": 0.4155, "step": 170 }, { "epoch": 0.37831858407079644, "grad_norm": 0.19775860011577606, "learning_rate": 9.97970625586178e-06, "loss": 0.4223, "step": 171 }, { "epoch": 0.3805309734513274, "grad_norm": 0.18166320025920868, "learning_rate": 9.978530889460351e-06, "loss": 0.4205, "step": 172 }, { "epoch": 0.3827433628318584, "grad_norm": 0.20108245313167572, "learning_rate": 9.977322510358552e-06, "loss": 0.4262, "step": 173 }, { "epoch": 0.38495575221238937, "grad_norm": 0.17232555150985718, "learning_rate": 9.976081126569164e-06, "loss": 0.4257, "step": 174 }, { "epoch": 0.38716814159292035, "grad_norm": 0.17979106307029724, "learning_rate": 9.97480674632382e-06, "loss": 0.4164, "step": 175 }, { "epoch": 0.3893805309734513, "grad_norm": 0.21929936110973358, "learning_rate": 9.973499378072947e-06, "loss": 0.4156, "step": 176 }, { "epoch": 0.3915929203539823, "grad_norm": 0.19628576934337616, "learning_rate": 9.972159030485722e-06, "loss": 0.4176, "step": 177 }, { "epoch": 0.3938053097345133, "grad_norm": 0.20714552700519562, "learning_rate": 9.970785712450007e-06, "loss": 0.429, "step": 178 }, { "epoch": 0.39601769911504425, "grad_norm": 0.2554994821548462, "learning_rate": 9.96937943307229e-06, "loss": 0.4183, "step": 179 }, { "epoch": 0.39823008849557523, "grad_norm": 0.21966539323329926, "learning_rate": 9.967940201677628e-06, "loss": 0.4148, "step": 180 }, { "epoch": 0.4004424778761062, "grad_norm": 0.19780422747135162, "learning_rate": 9.966468027809582e-06, "loss": 0.4086, "step": 181 }, { "epoch": 0.4026548672566372, "grad_norm": 0.19103385508060455, "learning_rate": 9.964962921230155e-06, "loss": 0.4123, "step": 182 }, { "epoch": 0.40486725663716816, "grad_norm": 0.20896346867084503, "learning_rate": 9.963424891919728e-06, "loss": 0.4138, "step": 183 }, { "epoch": 0.40707964601769914, "grad_norm": 0.1816549003124237, "learning_rate": 9.961853950076992e-06, "loss": 0.4175, "step": 184 }, { "epoch": 0.4092920353982301, "grad_norm": 0.1906682550907135, "learning_rate": 9.960250106118883e-06, "loss": 0.4146, "step": 185 }, { "epoch": 0.41150442477876104, "grad_norm": 0.19309958815574646, "learning_rate": 9.958613370680507e-06, "loss": 0.422, "step": 186 }, { "epoch": 0.413716814159292, "grad_norm": 0.19457145035266876, "learning_rate": 9.956943754615082e-06, "loss": 0.4138, "step": 187 }, { "epoch": 0.415929203539823, "grad_norm": 0.1757512390613556, "learning_rate": 9.955241268993852e-06, "loss": 0.414, "step": 188 }, { "epoch": 0.41814159292035397, "grad_norm": 0.16454994678497314, "learning_rate": 9.953505925106016e-06, "loss": 0.4199, "step": 189 }, { "epoch": 0.42035398230088494, "grad_norm": 0.1722830832004547, "learning_rate": 9.951737734458665e-06, "loss": 0.4219, "step": 190 }, { "epoch": 0.4225663716814159, "grad_norm": 0.1671864539384842, "learning_rate": 9.949936708776692e-06, "loss": 0.4226, "step": 191 }, { "epoch": 0.4247787610619469, "grad_norm": 0.16959279775619507, "learning_rate": 9.94810286000272e-06, "loss": 0.4188, "step": 192 }, { "epoch": 0.4269911504424779, "grad_norm": 0.18185065686702728, "learning_rate": 9.946236200297025e-06, "loss": 0.4225, "step": 193 }, { "epoch": 0.42920353982300885, "grad_norm": 0.1568220853805542, "learning_rate": 9.944336742037451e-06, "loss": 0.4064, "step": 194 }, { "epoch": 0.4314159292035398, "grad_norm": 0.19281834363937378, "learning_rate": 9.942404497819324e-06, "loss": 0.4135, "step": 195 }, { "epoch": 0.4336283185840708, "grad_norm": 0.16964174807071686, "learning_rate": 9.940439480455386e-06, "loss": 0.4196, "step": 196 }, { "epoch": 0.4358407079646018, "grad_norm": 0.17745113372802734, "learning_rate": 9.938441702975689e-06, "loss": 0.4167, "step": 197 }, { "epoch": 0.43805309734513276, "grad_norm": 0.1645510345697403, "learning_rate": 9.936411178627521e-06, "loss": 0.4093, "step": 198 }, { "epoch": 0.44026548672566373, "grad_norm": 0.16442488133907318, "learning_rate": 9.934347920875314e-06, "loss": 0.4248, "step": 199 }, { "epoch": 0.4424778761061947, "grad_norm": 0.17704367637634277, "learning_rate": 9.932251943400554e-06, "loss": 0.4075, "step": 200 }, { "epoch": 0.4446902654867257, "grad_norm": 0.1849382370710373, "learning_rate": 9.930123260101697e-06, "loss": 0.4182, "step": 201 }, { "epoch": 0.4469026548672566, "grad_norm": 0.1769019365310669, "learning_rate": 9.927961885094065e-06, "loss": 0.4044, "step": 202 }, { "epoch": 0.4491150442477876, "grad_norm": 0.17679619789123535, "learning_rate": 9.925767832709765e-06, "loss": 0.4161, "step": 203 }, { "epoch": 0.45132743362831856, "grad_norm": 0.19563840329647064, "learning_rate": 9.923541117497586e-06, "loss": 0.4014, "step": 204 }, { "epoch": 0.45353982300884954, "grad_norm": 0.20635181665420532, "learning_rate": 9.921281754222903e-06, "loss": 0.434, "step": 205 }, { "epoch": 0.4557522123893805, "grad_norm": 0.17097140848636627, "learning_rate": 9.918989757867584e-06, "loss": 0.4147, "step": 206 }, { "epoch": 0.4579646017699115, "grad_norm": 0.17044438421726227, "learning_rate": 9.916665143629881e-06, "loss": 0.4138, "step": 207 }, { "epoch": 0.46017699115044247, "grad_norm": 0.18953591585159302, "learning_rate": 9.914307926924344e-06, "loss": 0.4185, "step": 208 }, { "epoch": 0.46238938053097345, "grad_norm": 0.17236648499965668, "learning_rate": 9.911918123381706e-06, "loss": 0.4079, "step": 209 }, { "epoch": 0.4646017699115044, "grad_norm": 0.15649288892745972, "learning_rate": 9.909495748848783e-06, "loss": 0.4114, "step": 210 }, { "epoch": 0.4668141592920354, "grad_norm": 0.18650922179222107, "learning_rate": 9.907040819388372e-06, "loss": 0.4231, "step": 211 }, { "epoch": 0.4690265486725664, "grad_norm": 0.16785378754138947, "learning_rate": 9.904553351279139e-06, "loss": 0.409, "step": 212 }, { "epoch": 0.47123893805309736, "grad_norm": 0.17052291333675385, "learning_rate": 9.902033361015515e-06, "loss": 0.415, "step": 213 }, { "epoch": 0.47345132743362833, "grad_norm": 0.1584978699684143, "learning_rate": 9.899480865307591e-06, "loss": 0.4014, "step": 214 }, { "epoch": 0.4756637168141593, "grad_norm": 0.16302478313446045, "learning_rate": 9.896895881080995e-06, "loss": 0.4103, "step": 215 }, { "epoch": 0.4778761061946903, "grad_norm": 0.16019950807094574, "learning_rate": 9.89427842547679e-06, "loss": 0.4149, "step": 216 }, { "epoch": 0.48008849557522126, "grad_norm": 0.17266441881656647, "learning_rate": 9.891628515851358e-06, "loss": 0.4079, "step": 217 }, { "epoch": 0.4823008849557522, "grad_norm": 0.17257748544216156, "learning_rate": 9.88894616977628e-06, "loss": 0.4055, "step": 218 }, { "epoch": 0.48451327433628316, "grad_norm": 0.17442859709262848, "learning_rate": 9.88623140503823e-06, "loss": 0.4194, "step": 219 }, { "epoch": 0.48672566371681414, "grad_norm": 0.1760149896144867, "learning_rate": 9.883484239638842e-06, "loss": 0.4103, "step": 220 }, { "epoch": 0.4889380530973451, "grad_norm": 0.16283486783504486, "learning_rate": 9.880704691794608e-06, "loss": 0.4081, "step": 221 }, { "epoch": 0.4911504424778761, "grad_norm": 0.17161579430103302, "learning_rate": 9.877892779936744e-06, "loss": 0.4105, "step": 222 }, { "epoch": 0.49336283185840707, "grad_norm": 0.17453111708164215, "learning_rate": 9.87504852271107e-06, "loss": 0.4186, "step": 223 }, { "epoch": 0.49557522123893805, "grad_norm": 0.17772270739078522, "learning_rate": 9.872171938977895e-06, "loss": 0.4196, "step": 224 }, { "epoch": 0.497787610619469, "grad_norm": 0.1604829877614975, "learning_rate": 9.869263047811877e-06, "loss": 0.4138, "step": 225 }, { "epoch": 0.5, "grad_norm": 0.1991002857685089, "learning_rate": 9.866321868501914e-06, "loss": 0.4094, "step": 226 }, { "epoch": 0.5022123893805309, "grad_norm": 0.1912929266691208, "learning_rate": 9.863348420550998e-06, "loss": 0.4073, "step": 227 }, { "epoch": 0.504424778761062, "grad_norm": 0.1809907853603363, "learning_rate": 9.860342723676105e-06, "loss": 0.417, "step": 228 }, { "epoch": 0.5066371681415929, "grad_norm": 0.20652230083942413, "learning_rate": 9.857304797808043e-06, "loss": 0.4094, "step": 229 }, { "epoch": 0.5088495575221239, "grad_norm": 0.16395345330238342, "learning_rate": 9.85423466309134e-06, "loss": 0.4024, "step": 230 }, { "epoch": 0.5110619469026548, "grad_norm": 0.20369473099708557, "learning_rate": 9.851132339884097e-06, "loss": 0.4139, "step": 231 }, { "epoch": 0.5132743362831859, "grad_norm": 0.169134721159935, "learning_rate": 9.847997848757855e-06, "loss": 0.4113, "step": 232 }, { "epoch": 0.5154867256637168, "grad_norm": 0.19118903577327728, "learning_rate": 9.844831210497468e-06, "loss": 0.4089, "step": 233 }, { "epoch": 0.5176991150442478, "grad_norm": 0.19465133547782898, "learning_rate": 9.84163244610095e-06, "loss": 0.4092, "step": 234 }, { "epoch": 0.5199115044247787, "grad_norm": 0.18870508670806885, "learning_rate": 9.83840157677935e-06, "loss": 0.4108, "step": 235 }, { "epoch": 0.5221238938053098, "grad_norm": 0.18022289872169495, "learning_rate": 9.835138623956603e-06, "loss": 0.4244, "step": 236 }, { "epoch": 0.5243362831858407, "grad_norm": 0.18440207839012146, "learning_rate": 9.831843609269387e-06, "loss": 0.4164, "step": 237 }, { "epoch": 0.5265486725663717, "grad_norm": 0.17794957756996155, "learning_rate": 9.828516554566988e-06, "loss": 0.4069, "step": 238 }, { "epoch": 0.5287610619469026, "grad_norm": 0.1812417358160019, "learning_rate": 9.825157481911146e-06, "loss": 0.4149, "step": 239 }, { "epoch": 0.5309734513274337, "grad_norm": 0.18725480139255524, "learning_rate": 9.821766413575915e-06, "loss": 0.4303, "step": 240 }, { "epoch": 0.5331858407079646, "grad_norm": 0.17442503571510315, "learning_rate": 9.818343372047509e-06, "loss": 0.4036, "step": 241 }, { "epoch": 0.5353982300884956, "grad_norm": 0.21173766255378723, "learning_rate": 9.814888380024161e-06, "loss": 0.4179, "step": 242 }, { "epoch": 0.5376106194690266, "grad_norm": 0.18231187760829926, "learning_rate": 9.811401460415966e-06, "loss": 0.4173, "step": 243 }, { "epoch": 0.5398230088495575, "grad_norm": 0.1679583489894867, "learning_rate": 9.80788263634473e-06, "loss": 0.4115, "step": 244 }, { "epoch": 0.5420353982300885, "grad_norm": 0.16705110669136047, "learning_rate": 9.804331931143822e-06, "loss": 0.4025, "step": 245 }, { "epoch": 0.5442477876106194, "grad_norm": 0.17078952491283417, "learning_rate": 9.80074936835801e-06, "loss": 0.4208, "step": 246 }, { "epoch": 0.5464601769911505, "grad_norm": 0.16935208439826965, "learning_rate": 9.797134971743312e-06, "loss": 0.4036, "step": 247 }, { "epoch": 0.5486725663716814, "grad_norm": 0.1650349199771881, "learning_rate": 9.793488765266838e-06, "loss": 0.4172, "step": 248 }, { "epoch": 0.5508849557522124, "grad_norm": 0.16523662209510803, "learning_rate": 9.789810773106632e-06, "loss": 0.4077, "step": 249 }, { "epoch": 0.5530973451327433, "grad_norm": 0.20408877730369568, "learning_rate": 9.786101019651499e-06, "loss": 0.4043, "step": 250 }, { "epoch": 0.5553097345132744, "grad_norm": 0.1925540715456009, "learning_rate": 9.782359529500867e-06, "loss": 0.4166, "step": 251 }, { "epoch": 0.5575221238938053, "grad_norm": 0.18965932726860046, "learning_rate": 9.7785863274646e-06, "loss": 0.4016, "step": 252 }, { "epoch": 0.5597345132743363, "grad_norm": 0.18967291712760925, "learning_rate": 9.774781438562846e-06, "loss": 0.407, "step": 253 }, { "epoch": 0.5619469026548672, "grad_norm": 0.17891788482666016, "learning_rate": 9.770944888025874e-06, "loss": 0.4078, "step": 254 }, { "epoch": 0.5641592920353983, "grad_norm": 0.1903616040945053, "learning_rate": 9.767076701293898e-06, "loss": 0.4012, "step": 255 }, { "epoch": 0.5663716814159292, "grad_norm": 0.1644955724477768, "learning_rate": 9.763176904016914e-06, "loss": 0.4009, "step": 256 }, { "epoch": 0.5685840707964602, "grad_norm": 0.1887083798646927, "learning_rate": 9.759245522054523e-06, "loss": 0.3965, "step": 257 }, { "epoch": 0.5707964601769911, "grad_norm": 0.16408571600914001, "learning_rate": 9.755282581475769e-06, "loss": 0.4008, "step": 258 }, { "epoch": 0.5730088495575221, "grad_norm": 0.2137821614742279, "learning_rate": 9.751288108558961e-06, "loss": 0.4029, "step": 259 }, { "epoch": 0.5752212389380531, "grad_norm": 0.17689137160778046, "learning_rate": 9.747262129791497e-06, "loss": 0.4045, "step": 260 }, { "epoch": 0.577433628318584, "grad_norm": 0.1932871788740158, "learning_rate": 9.743204671869694e-06, "loss": 0.4082, "step": 261 }, { "epoch": 0.5796460176991151, "grad_norm": 0.2326955646276474, "learning_rate": 9.739115761698607e-06, "loss": 0.4043, "step": 262 }, { "epoch": 0.581858407079646, "grad_norm": 0.22756756842136383, "learning_rate": 9.73499542639185e-06, "loss": 0.4132, "step": 263 }, { "epoch": 0.584070796460177, "grad_norm": 0.18425188958644867, "learning_rate": 9.730843693271413e-06, "loss": 0.4026, "step": 264 }, { "epoch": 0.5862831858407079, "grad_norm": 0.24644114077091217, "learning_rate": 9.726660589867495e-06, "loss": 0.3966, "step": 265 }, { "epoch": 0.588495575221239, "grad_norm": 0.20994241535663605, "learning_rate": 9.722446143918307e-06, "loss": 0.4068, "step": 266 }, { "epoch": 0.5907079646017699, "grad_norm": 0.17565971612930298, "learning_rate": 9.718200383369891e-06, "loss": 0.402, "step": 267 }, { "epoch": 0.5929203539823009, "grad_norm": 0.1879590004682541, "learning_rate": 9.713923336375936e-06, "loss": 0.4109, "step": 268 }, { "epoch": 0.5951327433628318, "grad_norm": 0.16611909866333008, "learning_rate": 9.709615031297598e-06, "loss": 0.3957, "step": 269 }, { "epoch": 0.5973451327433629, "grad_norm": 0.19158808887004852, "learning_rate": 9.705275496703302e-06, "loss": 0.4125, "step": 270 }, { "epoch": 0.5995575221238938, "grad_norm": 0.19456025958061218, "learning_rate": 9.70090476136855e-06, "loss": 0.4079, "step": 271 }, { "epoch": 0.6017699115044248, "grad_norm": 0.17634005844593048, "learning_rate": 9.69650285427575e-06, "loss": 0.4025, "step": 272 }, { "epoch": 0.6039823008849557, "grad_norm": 0.20455583930015564, "learning_rate": 9.692069804613995e-06, "loss": 0.4101, "step": 273 }, { "epoch": 0.6061946902654868, "grad_norm": 0.18841254711151123, "learning_rate": 9.687605641778899e-06, "loss": 0.4026, "step": 274 }, { "epoch": 0.6084070796460177, "grad_norm": 0.19969072937965393, "learning_rate": 9.683110395372379e-06, "loss": 0.4004, "step": 275 }, { "epoch": 0.6106194690265486, "grad_norm": 0.2100944221019745, "learning_rate": 9.678584095202468e-06, "loss": 0.4103, "step": 276 }, { "epoch": 0.6128318584070797, "grad_norm": 0.16795648634433746, "learning_rate": 9.674026771283123e-06, "loss": 0.4046, "step": 277 }, { "epoch": 0.6150442477876106, "grad_norm": 0.19852401316165924, "learning_rate": 9.669438453834014e-06, "loss": 0.4053, "step": 278 }, { "epoch": 0.6172566371681416, "grad_norm": 0.18384751677513123, "learning_rate": 9.664819173280328e-06, "loss": 0.4106, "step": 279 }, { "epoch": 0.6194690265486725, "grad_norm": 0.16662909090518951, "learning_rate": 9.660168960252575e-06, "loss": 0.4158, "step": 280 }, { "epoch": 0.6216814159292036, "grad_norm": 0.22212032973766327, "learning_rate": 9.655487845586378e-06, "loss": 0.4156, "step": 281 }, { "epoch": 0.6238938053097345, "grad_norm": 0.17248587310314178, "learning_rate": 9.650775860322263e-06, "loss": 0.3941, "step": 282 }, { "epoch": 0.6261061946902655, "grad_norm": 0.2107280045747757, "learning_rate": 9.646033035705462e-06, "loss": 0.4202, "step": 283 }, { "epoch": 0.6283185840707964, "grad_norm": 0.16596029698848724, "learning_rate": 9.641259403185706e-06, "loss": 0.4061, "step": 284 }, { "epoch": 0.6305309734513275, "grad_norm": 0.18452270328998566, "learning_rate": 9.636454994417013e-06, "loss": 0.4023, "step": 285 }, { "epoch": 0.6327433628318584, "grad_norm": 0.17428122460842133, "learning_rate": 9.631619841257477e-06, "loss": 0.3974, "step": 286 }, { "epoch": 0.6349557522123894, "grad_norm": 0.18349966406822205, "learning_rate": 9.626753975769054e-06, "loss": 0.4051, "step": 287 }, { "epoch": 0.6371681415929203, "grad_norm": 0.18694283068180084, "learning_rate": 9.621857430217366e-06, "loss": 0.4141, "step": 288 }, { "epoch": 0.6393805309734514, "grad_norm": 0.196182981133461, "learning_rate": 9.616930237071464e-06, "loss": 0.4037, "step": 289 }, { "epoch": 0.6415929203539823, "grad_norm": 0.1949911266565323, "learning_rate": 9.611972429003626e-06, "loss": 0.411, "step": 290 }, { "epoch": 0.6438053097345132, "grad_norm": 0.21375317871570587, "learning_rate": 9.606984038889142e-06, "loss": 0.4025, "step": 291 }, { "epoch": 0.6460176991150443, "grad_norm": 0.2112773060798645, "learning_rate": 9.601965099806085e-06, "loss": 0.3962, "step": 292 }, { "epoch": 0.6482300884955752, "grad_norm": 0.1982385218143463, "learning_rate": 9.596915645035107e-06, "loss": 0.4116, "step": 293 }, { "epoch": 0.6504424778761062, "grad_norm": 0.18961909413337708, "learning_rate": 9.591835708059202e-06, "loss": 0.4016, "step": 294 }, { "epoch": 0.6526548672566371, "grad_norm": 0.205304354429245, "learning_rate": 9.5867253225635e-06, "loss": 0.4107, "step": 295 }, { "epoch": 0.6548672566371682, "grad_norm": 0.1793058067560196, "learning_rate": 9.581584522435025e-06, "loss": 0.4091, "step": 296 }, { "epoch": 0.6570796460176991, "grad_norm": 0.2273833453655243, "learning_rate": 9.57641334176249e-06, "loss": 0.4079, "step": 297 }, { "epoch": 0.6592920353982301, "grad_norm": 0.19448456168174744, "learning_rate": 9.571211814836059e-06, "loss": 0.4126, "step": 298 }, { "epoch": 0.661504424778761, "grad_norm": 0.18413153290748596, "learning_rate": 9.56597997614712e-06, "loss": 0.4023, "step": 299 }, { "epoch": 0.6637168141592921, "grad_norm": 0.20027752220630646, "learning_rate": 9.560717860388061e-06, "loss": 0.4111, "step": 300 }, { "epoch": 0.665929203539823, "grad_norm": 0.1603672206401825, "learning_rate": 9.555425502452038e-06, "loss": 0.4036, "step": 301 }, { "epoch": 0.668141592920354, "grad_norm": 0.18299739062786102, "learning_rate": 9.550102937432743e-06, "loss": 0.3942, "step": 302 }, { "epoch": 0.6703539823008849, "grad_norm": 0.1947639435529709, "learning_rate": 9.544750200624169e-06, "loss": 0.4054, "step": 303 }, { "epoch": 0.672566371681416, "grad_norm": 0.16402120888233185, "learning_rate": 9.539367327520382e-06, "loss": 0.4048, "step": 304 }, { "epoch": 0.6747787610619469, "grad_norm": 0.17655643820762634, "learning_rate": 9.533954353815279e-06, "loss": 0.4068, "step": 305 }, { "epoch": 0.6769911504424779, "grad_norm": 0.1705683320760727, "learning_rate": 9.528511315402358e-06, "loss": 0.3953, "step": 306 }, { "epoch": 0.6792035398230089, "grad_norm": 0.16182245314121246, "learning_rate": 9.523038248374474e-06, "loss": 0.408, "step": 307 }, { "epoch": 0.6814159292035398, "grad_norm": 0.20871901512145996, "learning_rate": 9.517535189023602e-06, "loss": 0.4031, "step": 308 }, { "epoch": 0.6836283185840708, "grad_norm": 0.13933664560317993, "learning_rate": 9.512002173840597e-06, "loss": 0.405, "step": 309 }, { "epoch": 0.6858407079646017, "grad_norm": 0.20702913403511047, "learning_rate": 9.506439239514954e-06, "loss": 0.4012, "step": 310 }, { "epoch": 0.6880530973451328, "grad_norm": 0.1594098061323166, "learning_rate": 9.500846422934557e-06, "loss": 0.3926, "step": 311 }, { "epoch": 0.6902654867256637, "grad_norm": 0.19512666761875153, "learning_rate": 9.495223761185443e-06, "loss": 0.4031, "step": 312 }, { "epoch": 0.6924778761061947, "grad_norm": 0.14987067878246307, "learning_rate": 9.489571291551553e-06, "loss": 0.4138, "step": 313 }, { "epoch": 0.6946902654867256, "grad_norm": 0.18819770216941833, "learning_rate": 9.483889051514483e-06, "loss": 0.4022, "step": 314 }, { "epoch": 0.6969026548672567, "grad_norm": 0.16024132072925568, "learning_rate": 9.478177078753236e-06, "loss": 0.4044, "step": 315 }, { "epoch": 0.6991150442477876, "grad_norm": 0.18151608109474182, "learning_rate": 9.472435411143979e-06, "loss": 0.4081, "step": 316 }, { "epoch": 0.7013274336283186, "grad_norm": 0.17617082595825195, "learning_rate": 9.466664086759777e-06, "loss": 0.4134, "step": 317 }, { "epoch": 0.7035398230088495, "grad_norm": 0.17612871527671814, "learning_rate": 9.460863143870355e-06, "loss": 0.4089, "step": 318 }, { "epoch": 0.7057522123893806, "grad_norm": 0.17769289016723633, "learning_rate": 9.45503262094184e-06, "loss": 0.4047, "step": 319 }, { "epoch": 0.7079646017699115, "grad_norm": 0.1573716551065445, "learning_rate": 9.4491725566365e-06, "loss": 0.4029, "step": 320 }, { "epoch": 0.7101769911504425, "grad_norm": 0.17037460207939148, "learning_rate": 9.443282989812494e-06, "loss": 0.4059, "step": 321 }, { "epoch": 0.7123893805309734, "grad_norm": 0.1654326617717743, "learning_rate": 9.437363959523613e-06, "loss": 0.4099, "step": 322 }, { "epoch": 0.7146017699115044, "grad_norm": 0.1739991158246994, "learning_rate": 9.431415505019024e-06, "loss": 0.3916, "step": 323 }, { "epoch": 0.7168141592920354, "grad_norm": 0.156654492020607, "learning_rate": 9.425437665742998e-06, "loss": 0.3912, "step": 324 }, { "epoch": 0.7190265486725663, "grad_norm": 0.16226203739643097, "learning_rate": 9.419430481334663e-06, "loss": 0.3995, "step": 325 }, { "epoch": 0.7212389380530974, "grad_norm": 0.14988061785697937, "learning_rate": 9.413393991627737e-06, "loss": 0.3959, "step": 326 }, { "epoch": 0.7234513274336283, "grad_norm": 0.1781560331583023, "learning_rate": 9.407328236650257e-06, "loss": 0.3957, "step": 327 }, { "epoch": 0.7256637168141593, "grad_norm": 0.17818742990493774, "learning_rate": 9.401233256624318e-06, "loss": 0.3949, "step": 328 }, { "epoch": 0.7278761061946902, "grad_norm": 0.1621919423341751, "learning_rate": 9.395109091965808e-06, "loss": 0.3978, "step": 329 }, { "epoch": 0.7300884955752213, "grad_norm": 0.1838427484035492, "learning_rate": 9.388955783284141e-06, "loss": 0.3998, "step": 330 }, { "epoch": 0.7323008849557522, "grad_norm": 0.17227084934711456, "learning_rate": 9.382773371381986e-06, "loss": 0.3989, "step": 331 }, { "epoch": 0.7345132743362832, "grad_norm": 0.18801696598529816, "learning_rate": 9.376561897254987e-06, "loss": 0.3956, "step": 332 }, { "epoch": 0.7367256637168141, "grad_norm": 0.15474942326545715, "learning_rate": 9.370321402091514e-06, "loss": 0.4035, "step": 333 }, { "epoch": 0.7389380530973452, "grad_norm": 0.16465848684310913, "learning_rate": 9.36405192727236e-06, "loss": 0.4016, "step": 334 }, { "epoch": 0.7411504424778761, "grad_norm": 0.19732873141765594, "learning_rate": 9.357753514370497e-06, "loss": 0.4038, "step": 335 }, { "epoch": 0.7433628318584071, "grad_norm": 0.16492311656475067, "learning_rate": 9.351426205150778e-06, "loss": 0.4015, "step": 336 }, { "epoch": 0.745575221238938, "grad_norm": 0.18818548321723938, "learning_rate": 9.345070041569666e-06, "loss": 0.399, "step": 337 }, { "epoch": 0.7477876106194691, "grad_norm": 0.17031817138195038, "learning_rate": 9.338685065774964e-06, "loss": 0.3937, "step": 338 }, { "epoch": 0.75, "grad_norm": 0.24733419716358185, "learning_rate": 9.332271320105527e-06, "loss": 0.4081, "step": 339 }, { "epoch": 0.7522123893805309, "grad_norm": 0.18530051410198212, "learning_rate": 9.32582884709098e-06, "loss": 0.4007, "step": 340 }, { "epoch": 0.754424778761062, "grad_norm": 0.18777970969676971, "learning_rate": 9.319357689451444e-06, "loss": 0.3989, "step": 341 }, { "epoch": 0.7566371681415929, "grad_norm": 0.18476152420043945, "learning_rate": 9.312857890097243e-06, "loss": 0.399, "step": 342 }, { "epoch": 0.7588495575221239, "grad_norm": 0.17613354325294495, "learning_rate": 9.30632949212863e-06, "loss": 0.4026, "step": 343 }, { "epoch": 0.7610619469026548, "grad_norm": 0.17226579785346985, "learning_rate": 9.299772538835492e-06, "loss": 0.4051, "step": 344 }, { "epoch": 0.7632743362831859, "grad_norm": 0.18831267952919006, "learning_rate": 9.29318707369707e-06, "loss": 0.3927, "step": 345 }, { "epoch": 0.7654867256637168, "grad_norm": 0.15509851276874542, "learning_rate": 9.286573140381663e-06, "loss": 0.3956, "step": 346 }, { "epoch": 0.7676991150442478, "grad_norm": 0.18900151550769806, "learning_rate": 9.279930782746346e-06, "loss": 0.3975, "step": 347 }, { "epoch": 0.7699115044247787, "grad_norm": 0.16996918618679047, "learning_rate": 9.273260044836675e-06, "loss": 0.3972, "step": 348 }, { "epoch": 0.7721238938053098, "grad_norm": 0.17968420684337616, "learning_rate": 9.266560970886397e-06, "loss": 0.4026, "step": 349 }, { "epoch": 0.7743362831858407, "grad_norm": 0.1665457934141159, "learning_rate": 9.259833605317155e-06, "loss": 0.3946, "step": 350 }, { "epoch": 0.7765486725663717, "grad_norm": 0.18998298048973083, "learning_rate": 9.253077992738193e-06, "loss": 0.3964, "step": 351 }, { "epoch": 0.7787610619469026, "grad_norm": 0.17083673179149628, "learning_rate": 9.246294177946062e-06, "loss": 0.4026, "step": 352 }, { "epoch": 0.7809734513274337, "grad_norm": 0.1637433022260666, "learning_rate": 9.239482205924322e-06, "loss": 0.3922, "step": 353 }, { "epoch": 0.7831858407079646, "grad_norm": 0.1800343543291092, "learning_rate": 9.232642121843247e-06, "loss": 0.407, "step": 354 }, { "epoch": 0.7853982300884956, "grad_norm": 0.16393564641475677, "learning_rate": 9.225773971059518e-06, "loss": 0.4057, "step": 355 }, { "epoch": 0.7876106194690266, "grad_norm": 0.18692757189273834, "learning_rate": 9.218877799115929e-06, "loss": 0.3979, "step": 356 }, { "epoch": 0.7898230088495575, "grad_norm": 0.17437385022640228, "learning_rate": 9.21195365174108e-06, "loss": 0.4057, "step": 357 }, { "epoch": 0.7920353982300885, "grad_norm": 0.1898239552974701, "learning_rate": 9.205001574849081e-06, "loss": 0.4039, "step": 358 }, { "epoch": 0.7942477876106194, "grad_norm": 0.1568395495414734, "learning_rate": 9.19802161453924e-06, "loss": 0.4039, "step": 359 }, { "epoch": 0.7964601769911505, "grad_norm": 0.19088448584079742, "learning_rate": 9.191013817095762e-06, "loss": 0.3953, "step": 360 }, { "epoch": 0.7986725663716814, "grad_norm": 0.17273734509944916, "learning_rate": 9.183978228987436e-06, "loss": 0.4104, "step": 361 }, { "epoch": 0.8008849557522124, "grad_norm": 0.17058604955673218, "learning_rate": 9.176914896867335e-06, "loss": 0.398, "step": 362 }, { "epoch": 0.8030973451327433, "grad_norm": 0.18058565258979797, "learning_rate": 9.169823867572505e-06, "loss": 0.4076, "step": 363 }, { "epoch": 0.8053097345132744, "grad_norm": 0.1585603505373001, "learning_rate": 9.162705188123647e-06, "loss": 0.3988, "step": 364 }, { "epoch": 0.8075221238938053, "grad_norm": 0.1937795877456665, "learning_rate": 9.155558905724815e-06, "loss": 0.4081, "step": 365 }, { "epoch": 0.8097345132743363, "grad_norm": 0.1497035026550293, "learning_rate": 9.148385067763094e-06, "loss": 0.3921, "step": 366 }, { "epoch": 0.8119469026548672, "grad_norm": 0.184015691280365, "learning_rate": 9.141183721808298e-06, "loss": 0.3959, "step": 367 }, { "epoch": 0.8141592920353983, "grad_norm": 0.16560928523540497, "learning_rate": 9.133954915612635e-06, "loss": 0.4016, "step": 368 }, { "epoch": 0.8163716814159292, "grad_norm": 0.17013601958751678, "learning_rate": 9.126698697110414e-06, "loss": 0.4052, "step": 369 }, { "epoch": 0.8185840707964602, "grad_norm": 0.16712944209575653, "learning_rate": 9.119415114417709e-06, "loss": 0.3995, "step": 370 }, { "epoch": 0.8207964601769911, "grad_norm": 0.16210488975048065, "learning_rate": 9.112104215832047e-06, "loss": 0.4094, "step": 371 }, { "epoch": 0.8230088495575221, "grad_norm": 0.1718660295009613, "learning_rate": 9.104766049832088e-06, "loss": 0.3852, "step": 372 }, { "epoch": 0.8252212389380531, "grad_norm": 0.1845821589231491, "learning_rate": 9.0974006650773e-06, "loss": 0.4091, "step": 373 }, { "epoch": 0.827433628318584, "grad_norm": 0.15908625721931458, "learning_rate": 9.090008110407646e-06, "loss": 0.4134, "step": 374 }, { "epoch": 0.8296460176991151, "grad_norm": 0.159079447388649, "learning_rate": 9.082588434843244e-06, "loss": 0.4003, "step": 375 }, { "epoch": 0.831858407079646, "grad_norm": 0.1775507628917694, "learning_rate": 9.075141687584056e-06, "loss": 0.4059, "step": 376 }, { "epoch": 0.834070796460177, "grad_norm": 0.1623350828886032, "learning_rate": 9.067667918009559e-06, "loss": 0.3968, "step": 377 }, { "epoch": 0.8362831858407079, "grad_norm": 0.17421099543571472, "learning_rate": 9.060167175678407e-06, "loss": 0.3928, "step": 378 }, { "epoch": 0.838495575221239, "grad_norm": 0.16807816922664642, "learning_rate": 9.05263951032812e-06, "loss": 0.4035, "step": 379 }, { "epoch": 0.8407079646017699, "grad_norm": 0.16292251646518707, "learning_rate": 9.045084971874738e-06, "loss": 0.4017, "step": 380 }, { "epoch": 0.8429203539823009, "grad_norm": 0.16168609261512756, "learning_rate": 9.037503610412502e-06, "loss": 0.3929, "step": 381 }, { "epoch": 0.8451327433628318, "grad_norm": 0.1535169780254364, "learning_rate": 9.02989547621351e-06, "loss": 0.3944, "step": 382 }, { "epoch": 0.8473451327433629, "grad_norm": 0.16936345398426056, "learning_rate": 9.022260619727401e-06, "loss": 0.3989, "step": 383 }, { "epoch": 0.8495575221238938, "grad_norm": 0.1527484804391861, "learning_rate": 9.014599091581e-06, "loss": 0.3952, "step": 384 }, { "epoch": 0.8517699115044248, "grad_norm": 0.17519870400428772, "learning_rate": 9.006910942577995e-06, "loss": 0.401, "step": 385 }, { "epoch": 0.8539823008849557, "grad_norm": 0.1557687520980835, "learning_rate": 8.999196223698599e-06, "loss": 0.4076, "step": 386 }, { "epoch": 0.8561946902654868, "grad_norm": 0.16832038760185242, "learning_rate": 8.991454986099207e-06, "loss": 0.3916, "step": 387 }, { "epoch": 0.8584070796460177, "grad_norm": 0.16987960040569305, "learning_rate": 8.983687281112066e-06, "loss": 0.3921, "step": 388 }, { "epoch": 0.8606194690265486, "grad_norm": 0.1476796269416809, "learning_rate": 8.975893160244921e-06, "loss": 0.3886, "step": 389 }, { "epoch": 0.8628318584070797, "grad_norm": 0.16488701105117798, "learning_rate": 8.968072675180686e-06, "loss": 0.3944, "step": 390 }, { "epoch": 0.8650442477876106, "grad_norm": 0.1628437340259552, "learning_rate": 8.960225877777095e-06, "loss": 0.3867, "step": 391 }, { "epoch": 0.8672566371681416, "grad_norm": 0.1530245542526245, "learning_rate": 8.952352820066359e-06, "loss": 0.4099, "step": 392 }, { "epoch": 0.8694690265486725, "grad_norm": 0.16439391672611237, "learning_rate": 8.944453554254823e-06, "loss": 0.3991, "step": 393 }, { "epoch": 0.8716814159292036, "grad_norm": 0.17328006029129028, "learning_rate": 8.936528132722616e-06, "loss": 0.3924, "step": 394 }, { "epoch": 0.8738938053097345, "grad_norm": 0.16984131932258606, "learning_rate": 8.928576608023305e-06, "loss": 0.4048, "step": 395 }, { "epoch": 0.8761061946902655, "grad_norm": 0.14963462948799133, "learning_rate": 8.920599032883553e-06, "loss": 0.3918, "step": 396 }, { "epoch": 0.8783185840707964, "grad_norm": 0.18854303658008575, "learning_rate": 8.912595460202758e-06, "loss": 0.3981, "step": 397 }, { "epoch": 0.8805309734513275, "grad_norm": 0.19335384666919708, "learning_rate": 8.90456594305271e-06, "loss": 0.3984, "step": 398 }, { "epoch": 0.8827433628318584, "grad_norm": 0.17484664916992188, "learning_rate": 8.896510534677238e-06, "loss": 0.392, "step": 399 }, { "epoch": 0.8849557522123894, "grad_norm": 0.19245024025440216, "learning_rate": 8.888429288491857e-06, "loss": 0.3949, "step": 400 }, { "epoch": 0.8871681415929203, "grad_norm": 0.15718188881874084, "learning_rate": 8.880322258083408e-06, "loss": 0.3973, "step": 401 }, { "epoch": 0.8893805309734514, "grad_norm": 0.17654351890087128, "learning_rate": 8.872189497209712e-06, "loss": 0.3996, "step": 402 }, { "epoch": 0.8915929203539823, "grad_norm": 0.16744081676006317, "learning_rate": 8.864031059799208e-06, "loss": 0.3879, "step": 403 }, { "epoch": 0.8938053097345132, "grad_norm": 0.16980023682117462, "learning_rate": 8.855846999950595e-06, "loss": 0.4029, "step": 404 }, { "epoch": 0.8960176991150443, "grad_norm": 0.17948906123638153, "learning_rate": 8.847637371932478e-06, "loss": 0.3938, "step": 405 }, { "epoch": 0.8982300884955752, "grad_norm": 0.1663169413805008, "learning_rate": 8.839402230183e-06, "loss": 0.404, "step": 406 }, { "epoch": 0.9004424778761062, "grad_norm": 0.1661122441291809, "learning_rate": 8.831141629309492e-06, "loss": 0.3923, "step": 407 }, { "epoch": 0.9026548672566371, "grad_norm": 0.17008569836616516, "learning_rate": 8.822855624088099e-06, "loss": 0.3911, "step": 408 }, { "epoch": 0.9048672566371682, "grad_norm": 0.16437990963459015, "learning_rate": 8.814544269463422e-06, "loss": 0.3952, "step": 409 }, { "epoch": 0.9070796460176991, "grad_norm": 0.18122254312038422, "learning_rate": 8.806207620548165e-06, "loss": 0.3912, "step": 410 }, { "epoch": 0.9092920353982301, "grad_norm": 0.1973719596862793, "learning_rate": 8.797845732622742e-06, "loss": 0.3978, "step": 411 }, { "epoch": 0.911504424778761, "grad_norm": 0.1716318130493164, "learning_rate": 8.789458661134943e-06, "loss": 0.3966, "step": 412 }, { "epoch": 0.9137168141592921, "grad_norm": 0.1809050291776657, "learning_rate": 8.781046461699538e-06, "loss": 0.3933, "step": 413 }, { "epoch": 0.915929203539823, "grad_norm": 0.1711595505475998, "learning_rate": 8.772609190097932e-06, "loss": 0.3912, "step": 414 }, { "epoch": 0.918141592920354, "grad_norm": 0.16632473468780518, "learning_rate": 8.764146902277773e-06, "loss": 0.3851, "step": 415 }, { "epoch": 0.9203539823008849, "grad_norm": 0.15847158432006836, "learning_rate": 8.755659654352599e-06, "loss": 0.3958, "step": 416 }, { "epoch": 0.922566371681416, "grad_norm": 0.16290074586868286, "learning_rate": 8.747147502601458e-06, "loss": 0.3965, "step": 417 }, { "epoch": 0.9247787610619469, "grad_norm": 0.15800520777702332, "learning_rate": 8.738610503468534e-06, "loss": 0.3952, "step": 418 }, { "epoch": 0.9269911504424779, "grad_norm": 0.13968175649642944, "learning_rate": 8.730048713562771e-06, "loss": 0.3935, "step": 419 }, { "epoch": 0.9292035398230089, "grad_norm": 0.15484796464443207, "learning_rate": 8.72146218965751e-06, "loss": 0.396, "step": 420 }, { "epoch": 0.9314159292035398, "grad_norm": 0.14914363622665405, "learning_rate": 8.712850988690094e-06, "loss": 0.4041, "step": 421 }, { "epoch": 0.9336283185840708, "grad_norm": 0.16192984580993652, "learning_rate": 8.704215167761506e-06, "loss": 0.3897, "step": 422 }, { "epoch": 0.9358407079646017, "grad_norm": 0.17002323269844055, "learning_rate": 8.695554784135982e-06, "loss": 0.3902, "step": 423 }, { "epoch": 0.9380530973451328, "grad_norm": 0.15230736136436462, "learning_rate": 8.686869895240631e-06, "loss": 0.398, "step": 424 }, { "epoch": 0.9402654867256637, "grad_norm": 0.2072356790304184, "learning_rate": 8.678160558665063e-06, "loss": 0.4063, "step": 425 }, { "epoch": 0.9424778761061947, "grad_norm": 0.16368243098258972, "learning_rate": 8.669426832160997e-06, "loss": 0.3988, "step": 426 }, { "epoch": 0.9446902654867256, "grad_norm": 0.16881467401981354, "learning_rate": 8.66066877364188e-06, "loss": 0.3986, "step": 427 }, { "epoch": 0.9469026548672567, "grad_norm": 0.17835210263729095, "learning_rate": 8.651886441182509e-06, "loss": 0.3969, "step": 428 }, { "epoch": 0.9491150442477876, "grad_norm": 0.15917813777923584, "learning_rate": 8.64307989301864e-06, "loss": 0.3793, "step": 429 }, { "epoch": 0.9513274336283186, "grad_norm": 0.1801290661096573, "learning_rate": 8.634249187546601e-06, "loss": 0.3959, "step": 430 }, { "epoch": 0.9535398230088495, "grad_norm": 0.16084758937358856, "learning_rate": 8.625394383322914e-06, "loss": 0.4085, "step": 431 }, { "epoch": 0.9557522123893806, "grad_norm": 0.17122162878513336, "learning_rate": 8.616515539063894e-06, "loss": 0.3996, "step": 432 }, { "epoch": 0.9579646017699115, "grad_norm": 0.16651393473148346, "learning_rate": 8.60761271364527e-06, "loss": 0.4013, "step": 433 }, { "epoch": 0.9601769911504425, "grad_norm": 0.15684399008750916, "learning_rate": 8.598685966101783e-06, "loss": 0.3875, "step": 434 }, { "epoch": 0.9623893805309734, "grad_norm": 0.15404202044010162, "learning_rate": 8.589735355626814e-06, "loss": 0.3929, "step": 435 }, { "epoch": 0.9646017699115044, "grad_norm": 0.1662074774503708, "learning_rate": 8.580760941571968e-06, "loss": 0.3957, "step": 436 }, { "epoch": 0.9668141592920354, "grad_norm": 0.15725825726985931, "learning_rate": 8.571762783446696e-06, "loss": 0.3914, "step": 437 }, { "epoch": 0.9690265486725663, "grad_norm": 0.1751137673854828, "learning_rate": 8.562740940917901e-06, "loss": 0.4001, "step": 438 }, { "epoch": 0.9712389380530974, "grad_norm": 0.1850391924381256, "learning_rate": 8.55369547380953e-06, "loss": 0.39, "step": 439 }, { "epoch": 0.9734513274336283, "grad_norm": 0.16561464965343475, "learning_rate": 8.544626442102188e-06, "loss": 0.3903, "step": 440 }, { "epoch": 0.9756637168141593, "grad_norm": 0.16364803910255432, "learning_rate": 8.535533905932739e-06, "loss": 0.3862, "step": 441 }, { "epoch": 0.9778761061946902, "grad_norm": 0.1534828543663025, "learning_rate": 8.526417925593901e-06, "loss": 0.3986, "step": 442 }, { "epoch": 0.9800884955752213, "grad_norm": 0.17045795917510986, "learning_rate": 8.517278561533857e-06, "loss": 0.393, "step": 443 }, { "epoch": 0.9823008849557522, "grad_norm": 0.16603973507881165, "learning_rate": 8.50811587435584e-06, "loss": 0.3915, "step": 444 }, { "epoch": 0.9845132743362832, "grad_norm": 0.17057141661643982, "learning_rate": 8.498929924817745e-06, "loss": 0.3839, "step": 445 }, { "epoch": 0.9867256637168141, "grad_norm": 0.16480037569999695, "learning_rate": 8.489720773831717e-06, "loss": 0.3931, "step": 446 }, { "epoch": 0.9889380530973452, "grad_norm": 0.15113525092601776, "learning_rate": 8.480488482463753e-06, "loss": 0.4149, "step": 447 }, { "epoch": 0.9911504424778761, "grad_norm": 0.17928218841552734, "learning_rate": 8.471233111933291e-06, "loss": 0.3984, "step": 448 }, { "epoch": 0.9933628318584071, "grad_norm": 0.1759202629327774, "learning_rate": 8.461954723612807e-06, "loss": 0.3989, "step": 449 }, { "epoch": 0.995575221238938, "grad_norm": 0.15595725178718567, "learning_rate": 8.45265337902741e-06, "loss": 0.3943, "step": 450 }, { "epoch": 0.9977876106194691, "grad_norm": 0.17772260308265686, "learning_rate": 8.443329139854434e-06, "loss": 0.3925, "step": 451 }, { "epoch": 1.0, "grad_norm": 0.17362035810947418, "learning_rate": 8.433982067923021e-06, "loss": 0.3919, "step": 452 }, { "epoch": 1.002212389380531, "grad_norm": 0.1918303370475769, "learning_rate": 8.424612225213726e-06, "loss": 0.3747, "step": 453 }, { "epoch": 1.0044247787610618, "grad_norm": 0.1567537784576416, "learning_rate": 8.41521967385809e-06, "loss": 0.3705, "step": 454 }, { "epoch": 1.0066371681415929, "grad_norm": 0.20494352281093597, "learning_rate": 8.405804476138239e-06, "loss": 0.3764, "step": 455 }, { "epoch": 1.008849557522124, "grad_norm": 0.1830655187368393, "learning_rate": 8.396366694486466e-06, "loss": 0.3826, "step": 456 }, { "epoch": 1.011061946902655, "grad_norm": 0.17687253654003143, "learning_rate": 8.386906391484819e-06, "loss": 0.3913, "step": 457 }, { "epoch": 1.0132743362831858, "grad_norm": 0.21283870935440063, "learning_rate": 8.377423629864686e-06, "loss": 0.3771, "step": 458 }, { "epoch": 1.0154867256637168, "grad_norm": 0.1914159059524536, "learning_rate": 8.367918472506375e-06, "loss": 0.3738, "step": 459 }, { "epoch": 1.0176991150442478, "grad_norm": 0.17665503919124603, "learning_rate": 8.358390982438706e-06, "loss": 0.3784, "step": 460 }, { "epoch": 1.0199115044247788, "grad_norm": 0.1868385225534439, "learning_rate": 8.348841222838579e-06, "loss": 0.3876, "step": 461 }, { "epoch": 1.0221238938053097, "grad_norm": 0.17088955640792847, "learning_rate": 8.339269257030576e-06, "loss": 0.3828, "step": 462 }, { "epoch": 1.0243362831858407, "grad_norm": 0.174570694565773, "learning_rate": 8.329675148486518e-06, "loss": 0.3789, "step": 463 }, { "epoch": 1.0265486725663717, "grad_norm": 0.16205543279647827, "learning_rate": 8.32005896082506e-06, "loss": 0.3784, "step": 464 }, { "epoch": 1.0287610619469028, "grad_norm": 0.18252155184745789, "learning_rate": 8.310420757811258e-06, "loss": 0.3797, "step": 465 }, { "epoch": 1.0309734513274336, "grad_norm": 0.2198188453912735, "learning_rate": 8.30076060335616e-06, "loss": 0.3797, "step": 466 }, { "epoch": 1.0331858407079646, "grad_norm": 0.15900884568691254, "learning_rate": 8.291078561516368e-06, "loss": 0.3737, "step": 467 }, { "epoch": 1.0353982300884956, "grad_norm": 0.21332651376724243, "learning_rate": 8.281374696493628e-06, "loss": 0.3728, "step": 468 }, { "epoch": 1.0376106194690267, "grad_norm": 0.18220320343971252, "learning_rate": 8.271649072634381e-06, "loss": 0.3669, "step": 469 }, { "epoch": 1.0398230088495575, "grad_norm": 0.16027416288852692, "learning_rate": 8.261901754429367e-06, "loss": 0.3791, "step": 470 }, { "epoch": 1.0420353982300885, "grad_norm": 0.19321398437023163, "learning_rate": 8.25213280651317e-06, "loss": 0.3803, "step": 471 }, { "epoch": 1.0442477876106195, "grad_norm": 0.14267107844352722, "learning_rate": 8.24234229366381e-06, "loss": 0.3769, "step": 472 }, { "epoch": 1.0464601769911503, "grad_norm": 0.16832539439201355, "learning_rate": 8.232530280802296e-06, "loss": 0.3792, "step": 473 }, { "epoch": 1.0486725663716814, "grad_norm": 0.15776118636131287, "learning_rate": 8.222696832992208e-06, "loss": 0.3741, "step": 474 }, { "epoch": 1.0508849557522124, "grad_norm": 0.16222868859767914, "learning_rate": 8.212842015439263e-06, "loss": 0.3758, "step": 475 }, { "epoch": 1.0530973451327434, "grad_norm": 0.188007190823555, "learning_rate": 8.202965893490877e-06, "loss": 0.3845, "step": 476 }, { "epoch": 1.0553097345132743, "grad_norm": 0.17090556025505066, "learning_rate": 8.193068532635737e-06, "loss": 0.3944, "step": 477 }, { "epoch": 1.0575221238938053, "grad_norm": 0.17572391033172607, "learning_rate": 8.18314999850337e-06, "loss": 0.3736, "step": 478 }, { "epoch": 1.0597345132743363, "grad_norm": 0.1866428256034851, "learning_rate": 8.173210356863696e-06, "loss": 0.3803, "step": 479 }, { "epoch": 1.0619469026548674, "grad_norm": 0.15502867102622986, "learning_rate": 8.163249673626603e-06, "loss": 0.375, "step": 480 }, { "epoch": 1.0641592920353982, "grad_norm": 0.1912487894296646, "learning_rate": 8.153268014841507e-06, "loss": 0.3697, "step": 481 }, { "epoch": 1.0663716814159292, "grad_norm": 0.179610013961792, "learning_rate": 8.143265446696909e-06, "loss": 0.3882, "step": 482 }, { "epoch": 1.0685840707964602, "grad_norm": 0.1993560791015625, "learning_rate": 8.133242035519968e-06, "loss": 0.3825, "step": 483 }, { "epoch": 1.0707964601769913, "grad_norm": 0.153557687997818, "learning_rate": 8.123197847776043e-06, "loss": 0.3835, "step": 484 }, { "epoch": 1.073008849557522, "grad_norm": 0.1919250637292862, "learning_rate": 8.113132950068272e-06, "loss": 0.3769, "step": 485 }, { "epoch": 1.075221238938053, "grad_norm": 0.19700811803340912, "learning_rate": 8.103047409137114e-06, "loss": 0.384, "step": 486 }, { "epoch": 1.0774336283185841, "grad_norm": 0.16748955845832825, "learning_rate": 8.09294129185992e-06, "loss": 0.3746, "step": 487 }, { "epoch": 1.079646017699115, "grad_norm": 0.19604726135730743, "learning_rate": 8.082814665250476e-06, "loss": 0.38, "step": 488 }, { "epoch": 1.081858407079646, "grad_norm": 0.16203738749027252, "learning_rate": 8.072667596458573e-06, "loss": 0.3702, "step": 489 }, { "epoch": 1.084070796460177, "grad_norm": 0.1730918437242508, "learning_rate": 8.062500152769547e-06, "loss": 0.3838, "step": 490 }, { "epoch": 1.086283185840708, "grad_norm": 0.1520654857158661, "learning_rate": 8.052312401603848e-06, "loss": 0.3749, "step": 491 }, { "epoch": 1.0884955752212389, "grad_norm": 0.1743493527173996, "learning_rate": 8.042104410516576e-06, "loss": 0.3961, "step": 492 }, { "epoch": 1.0907079646017699, "grad_norm": 0.1602836549282074, "learning_rate": 8.03187624719705e-06, "loss": 0.39, "step": 493 }, { "epoch": 1.092920353982301, "grad_norm": 0.15004900097846985, "learning_rate": 8.021627979468348e-06, "loss": 0.3783, "step": 494 }, { "epoch": 1.095132743362832, "grad_norm": 0.17503239214420319, "learning_rate": 8.01135967528686e-06, "loss": 0.3767, "step": 495 }, { "epoch": 1.0973451327433628, "grad_norm": 0.15622718632221222, "learning_rate": 8.001071402741843e-06, "loss": 0.3758, "step": 496 }, { "epoch": 1.0995575221238938, "grad_norm": 0.16066192090511322, "learning_rate": 7.990763230054953e-06, "loss": 0.3781, "step": 497 }, { "epoch": 1.1017699115044248, "grad_norm": 0.1617273986339569, "learning_rate": 7.980435225579819e-06, "loss": 0.3826, "step": 498 }, { "epoch": 1.1039823008849559, "grad_norm": 0.16359879076480865, "learning_rate": 7.970087457801563e-06, "loss": 0.3831, "step": 499 }, { "epoch": 1.1061946902654867, "grad_norm": 0.1616351306438446, "learning_rate": 7.959719995336364e-06, "loss": 0.3809, "step": 500 }, { "epoch": 1.1084070796460177, "grad_norm": 0.14531832933425903, "learning_rate": 7.949332906930995e-06, "loss": 0.3766, "step": 501 }, { "epoch": 1.1106194690265487, "grad_norm": 0.17033183574676514, "learning_rate": 7.938926261462366e-06, "loss": 0.3779, "step": 502 }, { "epoch": 1.1128318584070795, "grad_norm": 0.16740630567073822, "learning_rate": 7.928500127937075e-06, "loss": 0.3788, "step": 503 }, { "epoch": 1.1150442477876106, "grad_norm": 0.15742437541484833, "learning_rate": 7.918054575490943e-06, "loss": 0.3778, "step": 504 }, { "epoch": 1.1172566371681416, "grad_norm": 0.17109255492687225, "learning_rate": 7.90758967338856e-06, "loss": 0.3772, "step": 505 }, { "epoch": 1.1194690265486726, "grad_norm": 0.15209320187568665, "learning_rate": 7.897105491022819e-06, "loss": 0.3812, "step": 506 }, { "epoch": 1.1216814159292035, "grad_norm": 0.16745400428771973, "learning_rate": 7.886602097914466e-06, "loss": 0.3747, "step": 507 }, { "epoch": 1.1238938053097345, "grad_norm": 0.1631808876991272, "learning_rate": 7.876079563711631e-06, "loss": 0.3791, "step": 508 }, { "epoch": 1.1261061946902655, "grad_norm": 0.15150891244411469, "learning_rate": 7.86553795818937e-06, "loss": 0.382, "step": 509 }, { "epoch": 1.1283185840707965, "grad_norm": 0.17548954486846924, "learning_rate": 7.854977351249199e-06, "loss": 0.3692, "step": 510 }, { "epoch": 1.1305309734513274, "grad_norm": 0.15667495131492615, "learning_rate": 7.844397812918637e-06, "loss": 0.3848, "step": 511 }, { "epoch": 1.1327433628318584, "grad_norm": 0.1640884280204773, "learning_rate": 7.833799413350732e-06, "loss": 0.3812, "step": 512 }, { "epoch": 1.1349557522123894, "grad_norm": 0.1685960441827774, "learning_rate": 7.823182222823603e-06, "loss": 0.3851, "step": 513 }, { "epoch": 1.1371681415929205, "grad_norm": 0.16258198022842407, "learning_rate": 7.812546311739976e-06, "loss": 0.376, "step": 514 }, { "epoch": 1.1393805309734513, "grad_norm": 0.18257202208042145, "learning_rate": 7.801891750626706e-06, "loss": 0.3805, "step": 515 }, { "epoch": 1.1415929203539823, "grad_norm": 0.18744589388370514, "learning_rate": 7.791218610134324e-06, "loss": 0.38, "step": 516 }, { "epoch": 1.1438053097345133, "grad_norm": 0.17237654328346252, "learning_rate": 7.780526961036556e-06, "loss": 0.377, "step": 517 }, { "epoch": 1.1460176991150441, "grad_norm": 0.18574164807796478, "learning_rate": 7.769816874229862e-06, "loss": 0.3794, "step": 518 }, { "epoch": 1.1482300884955752, "grad_norm": 0.1659688502550125, "learning_rate": 7.759088420732958e-06, "loss": 0.3836, "step": 519 }, { "epoch": 1.1504424778761062, "grad_norm": 0.18307393789291382, "learning_rate": 7.748341671686355e-06, "loss": 0.3849, "step": 520 }, { "epoch": 1.1526548672566372, "grad_norm": 0.17125365138053894, "learning_rate": 7.737576698351878e-06, "loss": 0.3718, "step": 521 }, { "epoch": 1.154867256637168, "grad_norm": 0.1512458175420761, "learning_rate": 7.726793572112203e-06, "loss": 0.363, "step": 522 }, { "epoch": 1.157079646017699, "grad_norm": 0.16603504121303558, "learning_rate": 7.715992364470371e-06, "loss": 0.3753, "step": 523 }, { "epoch": 1.1592920353982301, "grad_norm": 0.15968966484069824, "learning_rate": 7.705173147049326e-06, "loss": 0.3811, "step": 524 }, { "epoch": 1.1615044247787611, "grad_norm": 0.1410856544971466, "learning_rate": 7.694335991591431e-06, "loss": 0.3915, "step": 525 }, { "epoch": 1.163716814159292, "grad_norm": 0.14456801116466522, "learning_rate": 7.683480969958005e-06, "loss": 0.3813, "step": 526 }, { "epoch": 1.165929203539823, "grad_norm": 0.15530700981616974, "learning_rate": 7.672608154128824e-06, "loss": 0.3871, "step": 527 }, { "epoch": 1.168141592920354, "grad_norm": 0.14297249913215637, "learning_rate": 7.66171761620167e-06, "loss": 0.3807, "step": 528 }, { "epoch": 1.170353982300885, "grad_norm": 0.13215981423854828, "learning_rate": 7.650809428391834e-06, "loss": 0.379, "step": 529 }, { "epoch": 1.1725663716814159, "grad_norm": 0.15017202496528625, "learning_rate": 7.63988366303165e-06, "loss": 0.3731, "step": 530 }, { "epoch": 1.174778761061947, "grad_norm": 0.15315097570419312, "learning_rate": 7.628940392569995e-06, "loss": 0.376, "step": 531 }, { "epoch": 1.176991150442478, "grad_norm": 0.15044665336608887, "learning_rate": 7.61797968957184e-06, "loss": 0.3816, "step": 532 }, { "epoch": 1.1792035398230087, "grad_norm": 0.15934717655181885, "learning_rate": 7.607001626717738e-06, "loss": 0.3741, "step": 533 }, { "epoch": 1.1814159292035398, "grad_norm": 0.14937704801559448, "learning_rate": 7.596006276803365e-06, "loss": 0.3849, "step": 534 }, { "epoch": 1.1836283185840708, "grad_norm": 0.14274275302886963, "learning_rate": 7.58499371273902e-06, "loss": 0.3697, "step": 535 }, { "epoch": 1.1858407079646018, "grad_norm": 0.14679598808288574, "learning_rate": 7.5739640075491546e-06, "loss": 0.3769, "step": 536 }, { "epoch": 1.1880530973451326, "grad_norm": 0.1407579779624939, "learning_rate": 7.562917234371879e-06, "loss": 0.3774, "step": 537 }, { "epoch": 1.1902654867256637, "grad_norm": 0.14973343908786774, "learning_rate": 7.551853466458486e-06, "loss": 0.3842, "step": 538 }, { "epoch": 1.1924778761061947, "grad_norm": 0.17621995508670807, "learning_rate": 7.540772777172958e-06, "loss": 0.3706, "step": 539 }, { "epoch": 1.1946902654867257, "grad_norm": 0.17068910598754883, "learning_rate": 7.529675239991483e-06, "loss": 0.3837, "step": 540 }, { "epoch": 1.1969026548672566, "grad_norm": 0.1598779559135437, "learning_rate": 7.518560928501969e-06, "loss": 0.365, "step": 541 }, { "epoch": 1.1991150442477876, "grad_norm": 0.1503140926361084, "learning_rate": 7.507429916403553e-06, "loss": 0.3714, "step": 542 }, { "epoch": 1.2013274336283186, "grad_norm": 0.16745896637439728, "learning_rate": 7.496282277506115e-06, "loss": 0.3772, "step": 543 }, { "epoch": 1.2035398230088497, "grad_norm": 0.1463339626789093, "learning_rate": 7.48511808572979e-06, "loss": 0.3675, "step": 544 }, { "epoch": 1.2057522123893805, "grad_norm": 0.16788248717784882, "learning_rate": 7.473937415104471e-06, "loss": 0.3768, "step": 545 }, { "epoch": 1.2079646017699115, "grad_norm": 0.1352882832288742, "learning_rate": 7.462740339769323e-06, "loss": 0.3761, "step": 546 }, { "epoch": 1.2101769911504425, "grad_norm": 0.15344858169555664, "learning_rate": 7.451526933972294e-06, "loss": 0.377, "step": 547 }, { "epoch": 1.2123893805309733, "grad_norm": 0.16328397393226624, "learning_rate": 7.440297272069615e-06, "loss": 0.3877, "step": 548 }, { "epoch": 1.2146017699115044, "grad_norm": 0.16266027092933655, "learning_rate": 7.429051428525318e-06, "loss": 0.378, "step": 549 }, { "epoch": 1.2168141592920354, "grad_norm": 0.15220795571804047, "learning_rate": 7.417789477910728e-06, "loss": 0.3779, "step": 550 }, { "epoch": 1.2190265486725664, "grad_norm": 0.15698575973510742, "learning_rate": 7.406511494903982e-06, "loss": 0.3748, "step": 551 }, { "epoch": 1.2212389380530975, "grad_norm": 0.17737598717212677, "learning_rate": 7.395217554289524e-06, "loss": 0.3764, "step": 552 }, { "epoch": 1.2234513274336283, "grad_norm": 0.1444133073091507, "learning_rate": 7.383907730957618e-06, "loss": 0.3822, "step": 553 }, { "epoch": 1.2256637168141593, "grad_norm": 0.16357633471488953, "learning_rate": 7.372582099903841e-06, "loss": 0.3647, "step": 554 }, { "epoch": 1.2278761061946903, "grad_norm": 0.14997902512550354, "learning_rate": 7.361240736228594e-06, "loss": 0.3735, "step": 555 }, { "epoch": 1.2300884955752212, "grad_norm": 0.12847259640693665, "learning_rate": 7.349883715136601e-06, "loss": 0.368, "step": 556 }, { "epoch": 1.2323008849557522, "grad_norm": 0.15516282618045807, "learning_rate": 7.3385111119364105e-06, "loss": 0.3819, "step": 557 }, { "epoch": 1.2345132743362832, "grad_norm": 0.16119252145290375, "learning_rate": 7.327123002039897e-06, "loss": 0.3759, "step": 558 }, { "epoch": 1.2367256637168142, "grad_norm": 0.15966367721557617, "learning_rate": 7.315719460961757e-06, "loss": 0.3806, "step": 559 }, { "epoch": 1.238938053097345, "grad_norm": 0.16308777034282684, "learning_rate": 7.304300564319013e-06, "loss": 0.3883, "step": 560 }, { "epoch": 1.241150442477876, "grad_norm": 0.14293457567691803, "learning_rate": 7.292866387830515e-06, "loss": 0.3748, "step": 561 }, { "epoch": 1.2433628318584071, "grad_norm": 0.17340628802776337, "learning_rate": 7.281417007316427e-06, "loss": 0.3832, "step": 562 }, { "epoch": 1.245575221238938, "grad_norm": 0.15052464604377747, "learning_rate": 7.269952498697734e-06, "loss": 0.3714, "step": 563 }, { "epoch": 1.247787610619469, "grad_norm": 0.18350891768932343, "learning_rate": 7.258472937995736e-06, "loss": 0.3789, "step": 564 }, { "epoch": 1.25, "grad_norm": 0.16274268925189972, "learning_rate": 7.246978401331543e-06, "loss": 0.3766, "step": 565 }, { "epoch": 1.252212389380531, "grad_norm": 0.1731940507888794, "learning_rate": 7.235468964925571e-06, "loss": 0.3794, "step": 566 }, { "epoch": 1.254424778761062, "grad_norm": 0.145197331905365, "learning_rate": 7.223944705097035e-06, "loss": 0.3771, "step": 567 }, { "epoch": 1.2566371681415929, "grad_norm": 0.14675655961036682, "learning_rate": 7.212405698263446e-06, "loss": 0.3639, "step": 568 }, { "epoch": 1.258849557522124, "grad_norm": 0.14033639430999756, "learning_rate": 7.200852020940102e-06, "loss": 0.3658, "step": 569 }, { "epoch": 1.261061946902655, "grad_norm": 0.15826043486595154, "learning_rate": 7.189283749739584e-06, "loss": 0.3772, "step": 570 }, { "epoch": 1.2632743362831858, "grad_norm": 0.14263145625591278, "learning_rate": 7.177700961371239e-06, "loss": 0.3824, "step": 571 }, { "epoch": 1.2654867256637168, "grad_norm": 0.16336825489997864, "learning_rate": 7.1661037326406825e-06, "loss": 0.3687, "step": 572 }, { "epoch": 1.2676991150442478, "grad_norm": 0.148985356092453, "learning_rate": 7.154492140449283e-06, "loss": 0.3687, "step": 573 }, { "epoch": 1.2699115044247788, "grad_norm": 0.14789626002311707, "learning_rate": 7.142866261793651e-06, "loss": 0.3791, "step": 574 }, { "epoch": 1.2721238938053097, "grad_norm": 0.13839952647686005, "learning_rate": 7.1312261737651354e-06, "loss": 0.3785, "step": 575 }, { "epoch": 1.2743362831858407, "grad_norm": 0.15205146372318268, "learning_rate": 7.119571953549305e-06, "loss": 0.3842, "step": 576 }, { "epoch": 1.2765486725663717, "grad_norm": 0.14949627220630646, "learning_rate": 7.107903678425436e-06, "loss": 0.3729, "step": 577 }, { "epoch": 1.2787610619469025, "grad_norm": 0.14976395666599274, "learning_rate": 7.09622142576601e-06, "loss": 0.3747, "step": 578 }, { "epoch": 1.2809734513274336, "grad_norm": 0.15081347525119781, "learning_rate": 7.084525273036187e-06, "loss": 0.3937, "step": 579 }, { "epoch": 1.2831858407079646, "grad_norm": 0.16921454668045044, "learning_rate": 7.072815297793303e-06, "loss": 0.3827, "step": 580 }, { "epoch": 1.2853982300884956, "grad_norm": 0.16270343959331512, "learning_rate": 7.061091577686349e-06, "loss": 0.3797, "step": 581 }, { "epoch": 1.2876106194690267, "grad_norm": 0.15536241233348846, "learning_rate": 7.0493541904554605e-06, "loss": 0.3727, "step": 582 }, { "epoch": 1.2898230088495575, "grad_norm": 0.1545332819223404, "learning_rate": 7.037603213931397e-06, "loss": 0.3826, "step": 583 }, { "epoch": 1.2920353982300885, "grad_norm": 0.1420442909002304, "learning_rate": 7.025838726035032e-06, "loss": 0.3786, "step": 584 }, { "epoch": 1.2942477876106195, "grad_norm": 0.15322910249233246, "learning_rate": 7.014060804776831e-06, "loss": 0.3814, "step": 585 }, { "epoch": 1.2964601769911503, "grad_norm": 0.1383352428674698, "learning_rate": 7.002269528256334e-06, "loss": 0.3838, "step": 586 }, { "epoch": 1.2986725663716814, "grad_norm": 0.15499426424503326, "learning_rate": 6.990464974661644e-06, "loss": 0.3803, "step": 587 }, { "epoch": 1.3008849557522124, "grad_norm": 0.13908816874027252, "learning_rate": 6.978647222268904e-06, "loss": 0.3804, "step": 588 }, { "epoch": 1.3030973451327434, "grad_norm": 0.15479522943496704, "learning_rate": 6.9668163494417775e-06, "loss": 0.3747, "step": 589 }, { "epoch": 1.3053097345132743, "grad_norm": 0.139928936958313, "learning_rate": 6.954972434630928e-06, "loss": 0.3735, "step": 590 }, { "epoch": 1.3075221238938053, "grad_norm": 0.1611955463886261, "learning_rate": 6.943115556373503e-06, "loss": 0.3719, "step": 591 }, { "epoch": 1.3097345132743363, "grad_norm": 0.16825099289417267, "learning_rate": 6.93124579329261e-06, "loss": 0.3768, "step": 592 }, { "epoch": 1.3119469026548671, "grad_norm": 0.15945830941200256, "learning_rate": 6.919363224096797e-06, "loss": 0.3772, "step": 593 }, { "epoch": 1.3141592920353982, "grad_norm": 0.17050500214099884, "learning_rate": 6.907467927579528e-06, "loss": 0.3768, "step": 594 }, { "epoch": 1.3163716814159292, "grad_norm": 0.15621913969516754, "learning_rate": 6.8955599826186606e-06, "loss": 0.3966, "step": 595 }, { "epoch": 1.3185840707964602, "grad_norm": 0.15994904935359955, "learning_rate": 6.883639468175926e-06, "loss": 0.3798, "step": 596 }, { "epoch": 1.3207964601769913, "grad_norm": 0.14877241849899292, "learning_rate": 6.871706463296407e-06, "loss": 0.3649, "step": 597 }, { "epoch": 1.323008849557522, "grad_norm": 0.1415458768606186, "learning_rate": 6.859761047108007e-06, "loss": 0.3721, "step": 598 }, { "epoch": 1.325221238938053, "grad_norm": 0.15687789022922516, "learning_rate": 6.847803298820927e-06, "loss": 0.3731, "step": 599 }, { "epoch": 1.3274336283185841, "grad_norm": 0.14746886491775513, "learning_rate": 6.835833297727148e-06, "loss": 0.3756, "step": 600 }, { "epoch": 1.329646017699115, "grad_norm": 0.16469167172908783, "learning_rate": 6.823851123199894e-06, "loss": 0.3807, "step": 601 }, { "epoch": 1.331858407079646, "grad_norm": 0.14631612598896027, "learning_rate": 6.811856854693114e-06, "loss": 0.3753, "step": 602 }, { "epoch": 1.334070796460177, "grad_norm": 0.13799001276493073, "learning_rate": 6.799850571740955e-06, "loss": 0.3635, "step": 603 }, { "epoch": 1.336283185840708, "grad_norm": 0.14614099264144897, "learning_rate": 6.787832353957225e-06, "loss": 0.3819, "step": 604 }, { "epoch": 1.338495575221239, "grad_norm": 0.14514301717281342, "learning_rate": 6.775802281034876e-06, "loss": 0.3881, "step": 605 }, { "epoch": 1.3407079646017699, "grad_norm": 0.149121955037117, "learning_rate": 6.763760432745475e-06, "loss": 0.3737, "step": 606 }, { "epoch": 1.342920353982301, "grad_norm": 0.14145825803279877, "learning_rate": 6.751706888938665e-06, "loss": 0.3786, "step": 607 }, { "epoch": 1.3451327433628317, "grad_norm": 0.15232518315315247, "learning_rate": 6.739641729541645e-06, "loss": 0.3735, "step": 608 }, { "epoch": 1.3473451327433628, "grad_norm": 0.15066224336624146, "learning_rate": 6.72756503455864e-06, "loss": 0.3678, "step": 609 }, { "epoch": 1.3495575221238938, "grad_norm": 0.14223569631576538, "learning_rate": 6.715476884070362e-06, "loss": 0.3834, "step": 610 }, { "epoch": 1.3517699115044248, "grad_norm": 0.13839849829673767, "learning_rate": 6.703377358233489e-06, "loss": 0.3744, "step": 611 }, { "epoch": 1.3539823008849559, "grad_norm": 0.14612992107868195, "learning_rate": 6.691266537280128e-06, "loss": 0.3845, "step": 612 }, { "epoch": 1.3561946902654867, "grad_norm": 0.14740517735481262, "learning_rate": 6.679144501517283e-06, "loss": 0.3871, "step": 613 }, { "epoch": 1.3584070796460177, "grad_norm": 0.13791871070861816, "learning_rate": 6.667011331326324e-06, "loss": 0.3644, "step": 614 }, { "epoch": 1.3606194690265487, "grad_norm": 0.15612854063510895, "learning_rate": 6.654867107162454e-06, "loss": 0.3771, "step": 615 }, { "epoch": 1.3628318584070795, "grad_norm": 0.14010357856750488, "learning_rate": 6.6427119095541745e-06, "loss": 0.3658, "step": 616 }, { "epoch": 1.3650442477876106, "grad_norm": 0.14571578800678253, "learning_rate": 6.6305458191027525e-06, "loss": 0.3761, "step": 617 }, { "epoch": 1.3672566371681416, "grad_norm": 0.1459769606590271, "learning_rate": 6.618368916481686e-06, "loss": 0.3761, "step": 618 }, { "epoch": 1.3694690265486726, "grad_norm": 0.154235377907753, "learning_rate": 6.606181282436166e-06, "loss": 0.3721, "step": 619 }, { "epoch": 1.3716814159292037, "grad_norm": 0.1487996131181717, "learning_rate": 6.593982997782549e-06, "loss": 0.3735, "step": 620 }, { "epoch": 1.3738938053097345, "grad_norm": 0.1267055720090866, "learning_rate": 6.58177414340781e-06, "loss": 0.3798, "step": 621 }, { "epoch": 1.3761061946902655, "grad_norm": 0.18154090642929077, "learning_rate": 6.569554800269014e-06, "loss": 0.3767, "step": 622 }, { "epoch": 1.3783185840707963, "grad_norm": 0.14018799364566803, "learning_rate": 6.557325049392781e-06, "loss": 0.3635, "step": 623 }, { "epoch": 1.3805309734513274, "grad_norm": 0.18179184198379517, "learning_rate": 6.545084971874738e-06, "loss": 0.3787, "step": 624 }, { "epoch": 1.3827433628318584, "grad_norm": 0.13475830852985382, "learning_rate": 6.532834648878993e-06, "loss": 0.3721, "step": 625 }, { "epoch": 1.3849557522123894, "grad_norm": 0.16122294962406158, "learning_rate": 6.520574161637591e-06, "loss": 0.3839, "step": 626 }, { "epoch": 1.3871681415929205, "grad_norm": 0.13646018505096436, "learning_rate": 6.5083035914499736e-06, "loss": 0.3766, "step": 627 }, { "epoch": 1.3893805309734513, "grad_norm": 0.17333966493606567, "learning_rate": 6.496023019682447e-06, "loss": 0.3701, "step": 628 }, { "epoch": 1.3915929203539823, "grad_norm": 0.14083491265773773, "learning_rate": 6.483732527767633e-06, "loss": 0.3803, "step": 629 }, { "epoch": 1.3938053097345133, "grad_norm": 0.16099579632282257, "learning_rate": 6.4714321972039395e-06, "loss": 0.3743, "step": 630 }, { "epoch": 1.3960176991150441, "grad_norm": 0.1497550904750824, "learning_rate": 6.459122109555011e-06, "loss": 0.3649, "step": 631 }, { "epoch": 1.3982300884955752, "grad_norm": 0.16616369783878326, "learning_rate": 6.4468023464491906e-06, "loss": 0.3748, "step": 632 }, { "epoch": 1.4004424778761062, "grad_norm": 0.15206293761730194, "learning_rate": 6.434472989578983e-06, "loss": 0.3732, "step": 633 }, { "epoch": 1.4026548672566372, "grad_norm": 0.15567412972450256, "learning_rate": 6.422134120700506e-06, "loss": 0.3752, "step": 634 }, { "epoch": 1.4048672566371683, "grad_norm": 0.1447472721338272, "learning_rate": 6.409785821632952e-06, "loss": 0.3775, "step": 635 }, { "epoch": 1.407079646017699, "grad_norm": 0.1411687433719635, "learning_rate": 6.397428174258048e-06, "loss": 0.3827, "step": 636 }, { "epoch": 1.4092920353982301, "grad_norm": 0.17297664284706116, "learning_rate": 6.385061260519507e-06, "loss": 0.3826, "step": 637 }, { "epoch": 1.411504424778761, "grad_norm": 0.1365060657262802, "learning_rate": 6.3726851624224875e-06, "loss": 0.3818, "step": 638 }, { "epoch": 1.413716814159292, "grad_norm": 0.15566328167915344, "learning_rate": 6.360299962033051e-06, "loss": 0.3718, "step": 639 }, { "epoch": 1.415929203539823, "grad_norm": 0.13370470702648163, "learning_rate": 6.347905741477613e-06, "loss": 0.3708, "step": 640 }, { "epoch": 1.418141592920354, "grad_norm": 0.15631158649921417, "learning_rate": 6.335502582942409e-06, "loss": 0.381, "step": 641 }, { "epoch": 1.420353982300885, "grad_norm": 0.15640929341316223, "learning_rate": 6.323090568672935e-06, "loss": 0.3673, "step": 642 }, { "epoch": 1.4225663716814159, "grad_norm": 0.13334697484970093, "learning_rate": 6.310669780973414e-06, "loss": 0.3827, "step": 643 }, { "epoch": 1.424778761061947, "grad_norm": 0.1540810763835907, "learning_rate": 6.298240302206242e-06, "loss": 0.3776, "step": 644 }, { "epoch": 1.426991150442478, "grad_norm": 0.1316145956516266, "learning_rate": 6.285802214791448e-06, "loss": 0.3844, "step": 645 }, { "epoch": 1.4292035398230087, "grad_norm": 0.14876599609851837, "learning_rate": 6.273355601206143e-06, "loss": 0.3691, "step": 646 }, { "epoch": 1.4314159292035398, "grad_norm": 0.1518520712852478, "learning_rate": 6.260900543983982e-06, "loss": 0.378, "step": 647 }, { "epoch": 1.4336283185840708, "grad_norm": 0.1387074887752533, "learning_rate": 6.2484371257146e-06, "loss": 0.3791, "step": 648 }, { "epoch": 1.4358407079646018, "grad_norm": 0.16339018940925598, "learning_rate": 6.235965429043082e-06, "loss": 0.3655, "step": 649 }, { "epoch": 1.4380530973451329, "grad_norm": 0.15789134800434113, "learning_rate": 6.223485536669401e-06, "loss": 0.3773, "step": 650 }, { "epoch": 1.4402654867256637, "grad_norm": 0.145791694521904, "learning_rate": 6.210997531347879e-06, "loss": 0.3766, "step": 651 }, { "epoch": 1.4424778761061947, "grad_norm": 0.16261202096939087, "learning_rate": 6.1985014958866386e-06, "loss": 0.3708, "step": 652 }, { "epoch": 1.4446902654867257, "grad_norm": 0.14985325932502747, "learning_rate": 6.185997513147043e-06, "loss": 0.3868, "step": 653 }, { "epoch": 1.4469026548672566, "grad_norm": 0.1505439281463623, "learning_rate": 6.173485666043159e-06, "loss": 0.3803, "step": 654 }, { "epoch": 1.4491150442477876, "grad_norm": 0.13515354692935944, "learning_rate": 6.160966037541201e-06, "loss": 0.3704, "step": 655 }, { "epoch": 1.4513274336283186, "grad_norm": 0.14188633859157562, "learning_rate": 6.148438710658979e-06, "loss": 0.3761, "step": 656 }, { "epoch": 1.4535398230088497, "grad_norm": 0.15125387907028198, "learning_rate": 6.135903768465356e-06, "loss": 0.3812, "step": 657 }, { "epoch": 1.4557522123893805, "grad_norm": 0.13983094692230225, "learning_rate": 6.123361294079691e-06, "loss": 0.3757, "step": 658 }, { "epoch": 1.4579646017699115, "grad_norm": 0.13178189098834991, "learning_rate": 6.110811370671286e-06, "loss": 0.3757, "step": 659 }, { "epoch": 1.4601769911504425, "grad_norm": 0.13429959118366241, "learning_rate": 6.098254081458839e-06, "loss": 0.378, "step": 660 }, { "epoch": 1.4623893805309733, "grad_norm": 0.14196458458900452, "learning_rate": 6.085689509709893e-06, "loss": 0.3817, "step": 661 }, { "epoch": 1.4646017699115044, "grad_norm": 0.13938362896442413, "learning_rate": 6.07311773874028e-06, "loss": 0.366, "step": 662 }, { "epoch": 1.4668141592920354, "grad_norm": 0.13635776937007904, "learning_rate": 6.060538851913568e-06, "loss": 0.3853, "step": 663 }, { "epoch": 1.4690265486725664, "grad_norm": 0.15285412967205048, "learning_rate": 6.047952932640513e-06, "loss": 0.3793, "step": 664 }, { "epoch": 1.4712389380530975, "grad_norm": 0.14070792496204376, "learning_rate": 6.035360064378504e-06, "loss": 0.3772, "step": 665 }, { "epoch": 1.4734513274336283, "grad_norm": 0.14458097517490387, "learning_rate": 6.022760330631006e-06, "loss": 0.3866, "step": 666 }, { "epoch": 1.4756637168141593, "grad_norm": 0.15231792628765106, "learning_rate": 6.01015381494701e-06, "loss": 0.3714, "step": 667 }, { "epoch": 1.4778761061946903, "grad_norm": 0.16281698644161224, "learning_rate": 5.997540600920479e-06, "loss": 0.3855, "step": 668 }, { "epoch": 1.4800884955752212, "grad_norm": 0.1445644348859787, "learning_rate": 5.984920772189793e-06, "loss": 0.3861, "step": 669 }, { "epoch": 1.4823008849557522, "grad_norm": 0.15494373440742493, "learning_rate": 5.972294412437194e-06, "loss": 0.3752, "step": 670 }, { "epoch": 1.4845132743362832, "grad_norm": 0.1489405632019043, "learning_rate": 5.959661605388229e-06, "loss": 0.3674, "step": 671 }, { "epoch": 1.4867256637168142, "grad_norm": 0.13303735852241516, "learning_rate": 5.947022434811202e-06, "loss": 0.3668, "step": 672 }, { "epoch": 1.488938053097345, "grad_norm": 0.16998596489429474, "learning_rate": 5.934376984516608e-06, "loss": 0.3856, "step": 673 }, { "epoch": 1.491150442477876, "grad_norm": 0.1437830924987793, "learning_rate": 5.921725338356585e-06, "loss": 0.3686, "step": 674 }, { "epoch": 1.4933628318584071, "grad_norm": 0.15047383308410645, "learning_rate": 5.909067580224359e-06, "loss": 0.3714, "step": 675 }, { "epoch": 1.495575221238938, "grad_norm": 0.15299569070339203, "learning_rate": 5.896403794053679e-06, "loss": 0.368, "step": 676 }, { "epoch": 1.497787610619469, "grad_norm": 0.14748515188694, "learning_rate": 5.883734063818272e-06, "loss": 0.3693, "step": 677 }, { "epoch": 1.5, "grad_norm": 0.1613292545080185, "learning_rate": 5.871058473531273e-06, "loss": 0.3722, "step": 678 }, { "epoch": 1.502212389380531, "grad_norm": 0.14528776705265045, "learning_rate": 5.858377107244683e-06, "loss": 0.372, "step": 679 }, { "epoch": 1.504424778761062, "grad_norm": 0.15699909627437592, "learning_rate": 5.845690049048799e-06, "loss": 0.3737, "step": 680 }, { "epoch": 1.5066371681415929, "grad_norm": 0.15869326889514923, "learning_rate": 5.83299738307166e-06, "loss": 0.3801, "step": 681 }, { "epoch": 1.508849557522124, "grad_norm": 0.1448500007390976, "learning_rate": 5.820299193478496e-06, "loss": 0.3838, "step": 682 }, { "epoch": 1.5110619469026547, "grad_norm": 0.16063882410526276, "learning_rate": 5.807595564471157e-06, "loss": 0.3716, "step": 683 }, { "epoch": 1.5132743362831858, "grad_norm": 0.14111746847629547, "learning_rate": 5.794886580287565e-06, "loss": 0.3769, "step": 684 }, { "epoch": 1.5154867256637168, "grad_norm": 0.16199442744255066, "learning_rate": 5.782172325201155e-06, "loss": 0.374, "step": 685 }, { "epoch": 1.5176991150442478, "grad_norm": 0.14267931878566742, "learning_rate": 5.76945288352031e-06, "loss": 0.369, "step": 686 }, { "epoch": 1.5199115044247788, "grad_norm": 0.14546652138233185, "learning_rate": 5.756728339587806e-06, "loss": 0.3732, "step": 687 }, { "epoch": 1.5221238938053099, "grad_norm": 0.17248912155628204, "learning_rate": 5.743998777780252e-06, "loss": 0.3788, "step": 688 }, { "epoch": 1.5243362831858407, "grad_norm": 0.16043299436569214, "learning_rate": 5.731264282507531e-06, "loss": 0.3834, "step": 689 }, { "epoch": 1.5265486725663717, "grad_norm": 0.1646987795829773, "learning_rate": 5.71852493821224e-06, "loss": 0.3778, "step": 690 }, { "epoch": 1.5287610619469025, "grad_norm": 0.1607351303100586, "learning_rate": 5.7057808293691305e-06, "loss": 0.3764, "step": 691 }, { "epoch": 1.5309734513274336, "grad_norm": 0.16537661850452423, "learning_rate": 5.6930320404845475e-06, "loss": 0.3757, "step": 692 }, { "epoch": 1.5331858407079646, "grad_norm": 0.16083936393260956, "learning_rate": 5.680278656095868e-06, "loss": 0.3768, "step": 693 }, { "epoch": 1.5353982300884956, "grad_norm": 0.1676769256591797, "learning_rate": 5.6675207607709426e-06, "loss": 0.3799, "step": 694 }, { "epoch": 1.5376106194690267, "grad_norm": 0.1764930635690689, "learning_rate": 5.654758439107533e-06, "loss": 0.3736, "step": 695 }, { "epoch": 1.5398230088495575, "grad_norm": 0.14429877698421478, "learning_rate": 5.641991775732756e-06, "loss": 0.3689, "step": 696 }, { "epoch": 1.5420353982300885, "grad_norm": 0.17751586437225342, "learning_rate": 5.629220855302513e-06, "loss": 0.3798, "step": 697 }, { "epoch": 1.5442477876106193, "grad_norm": 0.15950946509838104, "learning_rate": 5.6164457625009386e-06, "loss": 0.3776, "step": 698 }, { "epoch": 1.5464601769911503, "grad_norm": 0.1793259084224701, "learning_rate": 5.60366658203983e-06, "loss": 0.3851, "step": 699 }, { "epoch": 1.5486725663716814, "grad_norm": 0.1719927191734314, "learning_rate": 5.590883398658095e-06, "loss": 0.3795, "step": 700 }, { "epoch": 1.5508849557522124, "grad_norm": 0.15601767599582672, "learning_rate": 5.5780962971211795e-06, "loss": 0.3781, "step": 701 }, { "epoch": 1.5530973451327434, "grad_norm": 0.15840476751327515, "learning_rate": 5.565305362220515e-06, "loss": 0.3767, "step": 702 }, { "epoch": 1.5553097345132745, "grad_norm": 0.16765564680099487, "learning_rate": 5.552510678772949e-06, "loss": 0.3825, "step": 703 }, { "epoch": 1.5575221238938053, "grad_norm": 0.15824854373931885, "learning_rate": 5.539712331620186e-06, "loss": 0.3646, "step": 704 }, { "epoch": 1.5597345132743363, "grad_norm": 0.18552853167057037, "learning_rate": 5.526910405628227e-06, "loss": 0.3633, "step": 705 }, { "epoch": 1.5619469026548671, "grad_norm": 0.14328765869140625, "learning_rate": 5.514104985686802e-06, "loss": 0.3725, "step": 706 }, { "epoch": 1.5641592920353982, "grad_norm": 0.18817439675331116, "learning_rate": 5.501296156708812e-06, "loss": 0.3777, "step": 707 }, { "epoch": 1.5663716814159292, "grad_norm": 0.13829733431339264, "learning_rate": 5.488484003629759e-06, "loss": 0.3648, "step": 708 }, { "epoch": 1.5685840707964602, "grad_norm": 0.1701672077178955, "learning_rate": 5.475668611407191e-06, "loss": 0.3723, "step": 709 }, { "epoch": 1.5707964601769913, "grad_norm": 0.14328286051750183, "learning_rate": 5.462850065020133e-06, "loss": 0.3709, "step": 710 }, { "epoch": 1.573008849557522, "grad_norm": 0.1651555746793747, "learning_rate": 5.4500284494685275e-06, "loss": 0.3755, "step": 711 }, { "epoch": 1.575221238938053, "grad_norm": 0.16392961144447327, "learning_rate": 5.437203849772664e-06, "loss": 0.3764, "step": 712 }, { "epoch": 1.577433628318584, "grad_norm": 0.14715544879436493, "learning_rate": 5.424376350972625e-06, "loss": 0.3737, "step": 713 }, { "epoch": 1.579646017699115, "grad_norm": 0.16308023035526276, "learning_rate": 5.411546038127715e-06, "loss": 0.3772, "step": 714 }, { "epoch": 1.581858407079646, "grad_norm": 0.18041308224201202, "learning_rate": 5.398712996315898e-06, "loss": 0.3732, "step": 715 }, { "epoch": 1.584070796460177, "grad_norm": 0.1642719805240631, "learning_rate": 5.385877310633233e-06, "loss": 0.3868, "step": 716 }, { "epoch": 1.586283185840708, "grad_norm": 0.17112895846366882, "learning_rate": 5.373039066193312e-06, "loss": 0.3741, "step": 717 }, { "epoch": 1.588495575221239, "grad_norm": 0.1910698413848877, "learning_rate": 5.360198348126696e-06, "loss": 0.3843, "step": 718 }, { "epoch": 1.5907079646017699, "grad_norm": 0.16182135045528412, "learning_rate": 5.347355241580344e-06, "loss": 0.3707, "step": 719 }, { "epoch": 1.592920353982301, "grad_norm": 0.15010400116443634, "learning_rate": 5.334509831717058e-06, "loss": 0.3743, "step": 720 }, { "epoch": 1.5951327433628317, "grad_norm": 0.1402837634086609, "learning_rate": 5.321662203714909e-06, "loss": 0.3779, "step": 721 }, { "epoch": 1.5973451327433628, "grad_norm": 0.14762374758720398, "learning_rate": 5.308812442766679e-06, "loss": 0.3666, "step": 722 }, { "epoch": 1.5995575221238938, "grad_norm": 0.1661965399980545, "learning_rate": 5.295960634079292e-06, "loss": 0.3711, "step": 723 }, { "epoch": 1.6017699115044248, "grad_norm": 0.1451561152935028, "learning_rate": 5.283106862873253e-06, "loss": 0.3804, "step": 724 }, { "epoch": 1.6039823008849559, "grad_norm": 0.16984255611896515, "learning_rate": 5.270251214382078e-06, "loss": 0.3844, "step": 725 }, { "epoch": 1.606194690265487, "grad_norm": 0.14532634615898132, "learning_rate": 5.257393773851733e-06, "loss": 0.3637, "step": 726 }, { "epoch": 1.6084070796460177, "grad_norm": 0.149654358625412, "learning_rate": 5.244534626540067e-06, "loss": 0.3722, "step": 727 }, { "epoch": 1.6106194690265485, "grad_norm": 0.14878515899181366, "learning_rate": 5.231673857716244e-06, "loss": 0.3775, "step": 728 }, { "epoch": 1.6128318584070795, "grad_norm": 0.1508331298828125, "learning_rate": 5.218811552660184e-06, "loss": 0.372, "step": 729 }, { "epoch": 1.6150442477876106, "grad_norm": 0.14183960855007172, "learning_rate": 5.205947796661991e-06, "loss": 0.3736, "step": 730 }, { "epoch": 1.6172566371681416, "grad_norm": 0.14773423969745636, "learning_rate": 5.193082675021393e-06, "loss": 0.3699, "step": 731 }, { "epoch": 1.6194690265486726, "grad_norm": 0.1459917426109314, "learning_rate": 5.1802162730471704e-06, "loss": 0.3742, "step": 732 }, { "epoch": 1.6216814159292037, "grad_norm": 0.14351701736450195, "learning_rate": 5.167348676056596e-06, "loss": 0.375, "step": 733 }, { "epoch": 1.6238938053097345, "grad_norm": 0.14323677122592926, "learning_rate": 5.154479969374865e-06, "loss": 0.381, "step": 734 }, { "epoch": 1.6261061946902655, "grad_norm": 0.14793291687965393, "learning_rate": 5.1416102383345315e-06, "loss": 0.3756, "step": 735 }, { "epoch": 1.6283185840707963, "grad_norm": 0.14177453517913818, "learning_rate": 5.1287395682749444e-06, "loss": 0.3766, "step": 736 }, { "epoch": 1.6305309734513274, "grad_norm": 0.1743677854537964, "learning_rate": 5.115868044541674e-06, "loss": 0.3707, "step": 737 }, { "epoch": 1.6327433628318584, "grad_norm": 0.14689166843891144, "learning_rate": 5.102995752485956e-06, "loss": 0.3672, "step": 738 }, { "epoch": 1.6349557522123894, "grad_norm": 0.14640650153160095, "learning_rate": 5.090122777464121e-06, "loss": 0.3774, "step": 739 }, { "epoch": 1.6371681415929205, "grad_norm": 0.1713770627975464, "learning_rate": 5.077249204837026e-06, "loss": 0.3658, "step": 740 }, { "epoch": 1.6393805309734515, "grad_norm": 0.1488184779882431, "learning_rate": 5.064375119969491e-06, "loss": 0.3702, "step": 741 }, { "epoch": 1.6415929203539823, "grad_norm": 0.14465579390525818, "learning_rate": 5.051500608229734e-06, "loss": 0.3652, "step": 742 }, { "epoch": 1.643805309734513, "grad_norm": 0.1728200763463974, "learning_rate": 5.038625754988802e-06, "loss": 0.3762, "step": 743 }, { "epoch": 1.6460176991150441, "grad_norm": 0.14261655509471893, "learning_rate": 5.025750645620004e-06, "loss": 0.3812, "step": 744 }, { "epoch": 1.6482300884955752, "grad_norm": 0.14125920832157135, "learning_rate": 5.012875365498357e-06, "loss": 0.3794, "step": 745 }, { "epoch": 1.6504424778761062, "grad_norm": 0.15553857386112213, "learning_rate": 5e-06, "loss": 0.3711, "step": 746 }, { "epoch": 1.6526548672566372, "grad_norm": 0.14529219269752502, "learning_rate": 4.9871246345016445e-06, "loss": 0.3743, "step": 747 }, { "epoch": 1.6548672566371683, "grad_norm": 0.13652414083480835, "learning_rate": 4.974249354379997e-06, "loss": 0.3773, "step": 748 }, { "epoch": 1.657079646017699, "grad_norm": 0.1763504445552826, "learning_rate": 4.961374245011201e-06, "loss": 0.3791, "step": 749 }, { "epoch": 1.6592920353982301, "grad_norm": 0.14034205675125122, "learning_rate": 4.948499391770268e-06, "loss": 0.3848, "step": 750 }, { "epoch": 1.661504424778761, "grad_norm": 0.14464423060417175, "learning_rate": 4.93562488003051e-06, "loss": 0.3716, "step": 751 }, { "epoch": 1.663716814159292, "grad_norm": 0.17310793697834015, "learning_rate": 4.922750795162974e-06, "loss": 0.3722, "step": 752 }, { "epoch": 1.665929203539823, "grad_norm": 0.14925870299339294, "learning_rate": 4.909877222535879e-06, "loss": 0.3732, "step": 753 }, { "epoch": 1.668141592920354, "grad_norm": 0.14656174182891846, "learning_rate": 4.897004247514044e-06, "loss": 0.3779, "step": 754 }, { "epoch": 1.670353982300885, "grad_norm": 0.17445078492164612, "learning_rate": 4.884131955458327e-06, "loss": 0.375, "step": 755 }, { "epoch": 1.672566371681416, "grad_norm": 0.17037415504455566, "learning_rate": 4.871260431725058e-06, "loss": 0.3852, "step": 756 }, { "epoch": 1.674778761061947, "grad_norm": 0.15467847883701324, "learning_rate": 4.858389761665469e-06, "loss": 0.3716, "step": 757 }, { "epoch": 1.676991150442478, "grad_norm": 0.1854114532470703, "learning_rate": 4.845520030625136e-06, "loss": 0.379, "step": 758 }, { "epoch": 1.6792035398230087, "grad_norm": 0.14063851535320282, "learning_rate": 4.832651323943406e-06, "loss": 0.3794, "step": 759 }, { "epoch": 1.6814159292035398, "grad_norm": 0.15527218580245972, "learning_rate": 4.819783726952831e-06, "loss": 0.3817, "step": 760 }, { "epoch": 1.6836283185840708, "grad_norm": 0.14957115054130554, "learning_rate": 4.806917324978608e-06, "loss": 0.3724, "step": 761 }, { "epoch": 1.6858407079646018, "grad_norm": 0.13799116015434265, "learning_rate": 4.79405220333801e-06, "loss": 0.3667, "step": 762 }, { "epoch": 1.6880530973451329, "grad_norm": 0.13975948095321655, "learning_rate": 4.781188447339817e-06, "loss": 0.3765, "step": 763 }, { "epoch": 1.6902654867256637, "grad_norm": 0.1460770219564438, "learning_rate": 4.768326142283757e-06, "loss": 0.3709, "step": 764 }, { "epoch": 1.6924778761061947, "grad_norm": 0.1463456004858017, "learning_rate": 4.755465373459934e-06, "loss": 0.383, "step": 765 }, { "epoch": 1.6946902654867255, "grad_norm": 0.14482566714286804, "learning_rate": 4.742606226148268e-06, "loss": 0.3766, "step": 766 }, { "epoch": 1.6969026548672566, "grad_norm": 0.15000230073928833, "learning_rate": 4.7297487856179224e-06, "loss": 0.3657, "step": 767 }, { "epoch": 1.6991150442477876, "grad_norm": 0.15064841508865356, "learning_rate": 4.716893137126748e-06, "loss": 0.369, "step": 768 }, { "epoch": 1.7013274336283186, "grad_norm": 0.14898055791854858, "learning_rate": 4.704039365920709e-06, "loss": 0.3826, "step": 769 }, { "epoch": 1.7035398230088497, "grad_norm": 0.1572868973016739, "learning_rate": 4.691187557233323e-06, "loss": 0.3764, "step": 770 }, { "epoch": 1.7057522123893807, "grad_norm": 0.12325187027454376, "learning_rate": 4.678337796285093e-06, "loss": 0.3803, "step": 771 }, { "epoch": 1.7079646017699115, "grad_norm": 0.1512259542942047, "learning_rate": 4.665490168282943e-06, "loss": 0.377, "step": 772 }, { "epoch": 1.7101769911504425, "grad_norm": 0.1448717713356018, "learning_rate": 4.6526447584196575e-06, "loss": 0.3773, "step": 773 }, { "epoch": 1.7123893805309733, "grad_norm": 0.16558457911014557, "learning_rate": 4.639801651873305e-06, "loss": 0.3772, "step": 774 }, { "epoch": 1.7146017699115044, "grad_norm": 0.14137940108776093, "learning_rate": 4.6269609338066875e-06, "loss": 0.3747, "step": 775 }, { "epoch": 1.7168141592920354, "grad_norm": 0.14484670758247375, "learning_rate": 4.614122689366769e-06, "loss": 0.3832, "step": 776 }, { "epoch": 1.7190265486725664, "grad_norm": 0.1538202315568924, "learning_rate": 4.601287003684104e-06, "loss": 0.377, "step": 777 }, { "epoch": 1.7212389380530975, "grad_norm": 0.15519271790981293, "learning_rate": 4.588453961872286e-06, "loss": 0.3653, "step": 778 }, { "epoch": 1.7234513274336283, "grad_norm": 0.1556425839662552, "learning_rate": 4.575623649027376e-06, "loss": 0.3729, "step": 779 }, { "epoch": 1.7256637168141593, "grad_norm": 0.15341098606586456, "learning_rate": 4.562796150227337e-06, "loss": 0.3722, "step": 780 }, { "epoch": 1.7278761061946901, "grad_norm": 0.1618175208568573, "learning_rate": 4.549971550531474e-06, "loss": 0.3715, "step": 781 }, { "epoch": 1.7300884955752212, "grad_norm": 0.1501229852437973, "learning_rate": 4.537149934979869e-06, "loss": 0.3617, "step": 782 }, { "epoch": 1.7323008849557522, "grad_norm": 0.1526263803243637, "learning_rate": 4.524331388592812e-06, "loss": 0.3816, "step": 783 }, { "epoch": 1.7345132743362832, "grad_norm": 0.14755792915821075, "learning_rate": 4.511515996370244e-06, "loss": 0.3719, "step": 784 }, { "epoch": 1.7367256637168142, "grad_norm": 0.16713263094425201, "learning_rate": 4.498703843291189e-06, "loss": 0.3833, "step": 785 }, { "epoch": 1.7389380530973453, "grad_norm": 0.14121297001838684, "learning_rate": 4.485895014313198e-06, "loss": 0.3729, "step": 786 }, { "epoch": 1.741150442477876, "grad_norm": 0.1821858286857605, "learning_rate": 4.4730895943717735e-06, "loss": 0.3866, "step": 787 }, { "epoch": 1.7433628318584071, "grad_norm": 0.150221586227417, "learning_rate": 4.460287668379815e-06, "loss": 0.3715, "step": 788 }, { "epoch": 1.745575221238938, "grad_norm": 0.16274835169315338, "learning_rate": 4.447489321227052e-06, "loss": 0.3703, "step": 789 }, { "epoch": 1.747787610619469, "grad_norm": 0.14972063899040222, "learning_rate": 4.434694637779486e-06, "loss": 0.3592, "step": 790 }, { "epoch": 1.75, "grad_norm": 0.15914012491703033, "learning_rate": 4.421903702878822e-06, "loss": 0.3754, "step": 791 }, { "epoch": 1.752212389380531, "grad_norm": 0.1527598351240158, "learning_rate": 4.409116601341908e-06, "loss": 0.3754, "step": 792 }, { "epoch": 1.754424778761062, "grad_norm": 0.17452898621559143, "learning_rate": 4.396333417960172e-06, "loss": 0.3781, "step": 793 }, { "epoch": 1.7566371681415929, "grad_norm": 0.1623452603816986, "learning_rate": 4.383554237499064e-06, "loss": 0.3787, "step": 794 }, { "epoch": 1.758849557522124, "grad_norm": 0.14795322716236115, "learning_rate": 4.37077914469749e-06, "loss": 0.3723, "step": 795 }, { "epoch": 1.7610619469026547, "grad_norm": 0.18765774369239807, "learning_rate": 4.358008224267245e-06, "loss": 0.3643, "step": 796 }, { "epoch": 1.7632743362831858, "grad_norm": 0.1436903327703476, "learning_rate": 4.345241560892467e-06, "loss": 0.3757, "step": 797 }, { "epoch": 1.7654867256637168, "grad_norm": 0.17049679160118103, "learning_rate": 4.332479239229059e-06, "loss": 0.3775, "step": 798 }, { "epoch": 1.7676991150442478, "grad_norm": 0.15550793707370758, "learning_rate": 4.319721343904133e-06, "loss": 0.3825, "step": 799 }, { "epoch": 1.7699115044247788, "grad_norm": 0.1573866903781891, "learning_rate": 4.306967959515454e-06, "loss": 0.3696, "step": 800 }, { "epoch": 1.7721238938053099, "grad_norm": 0.16475972533226013, "learning_rate": 4.29421917063087e-06, "loss": 0.3789, "step": 801 }, { "epoch": 1.7743362831858407, "grad_norm": 0.13793496787548065, "learning_rate": 4.2814750617877615e-06, "loss": 0.3718, "step": 802 }, { "epoch": 1.7765486725663717, "grad_norm": 0.1852518618106842, "learning_rate": 4.268735717492472e-06, "loss": 0.3672, "step": 803 }, { "epoch": 1.7787610619469025, "grad_norm": 0.14557519555091858, "learning_rate": 4.256001222219751e-06, "loss": 0.3802, "step": 804 }, { "epoch": 1.7809734513274336, "grad_norm": 0.16771559417247772, "learning_rate": 4.243271660412197e-06, "loss": 0.365, "step": 805 }, { "epoch": 1.7831858407079646, "grad_norm": 0.15564660727977753, "learning_rate": 4.230547116479691e-06, "loss": 0.3758, "step": 806 }, { "epoch": 1.7853982300884956, "grad_norm": 0.18923735618591309, "learning_rate": 4.217827674798845e-06, "loss": 0.3865, "step": 807 }, { "epoch": 1.7876106194690267, "grad_norm": 0.14485545456409454, "learning_rate": 4.2051134197124354e-06, "loss": 0.3771, "step": 808 }, { "epoch": 1.7898230088495575, "grad_norm": 0.17624424397945404, "learning_rate": 4.192404435528844e-06, "loss": 0.3785, "step": 809 }, { "epoch": 1.7920353982300885, "grad_norm": 0.16202057898044586, "learning_rate": 4.179700806521506e-06, "loss": 0.3815, "step": 810 }, { "epoch": 1.7942477876106193, "grad_norm": 0.13739414513111115, "learning_rate": 4.167002616928341e-06, "loss": 0.369, "step": 811 }, { "epoch": 1.7964601769911503, "grad_norm": 0.1619684100151062, "learning_rate": 4.154309950951203e-06, "loss": 0.3843, "step": 812 }, { "epoch": 1.7986725663716814, "grad_norm": 0.14789964258670807, "learning_rate": 4.141622892755318e-06, "loss": 0.3842, "step": 813 }, { "epoch": 1.8008849557522124, "grad_norm": 0.16962085664272308, "learning_rate": 4.128941526468728e-06, "loss": 0.3713, "step": 814 }, { "epoch": 1.8030973451327434, "grad_norm": 0.16202415525913239, "learning_rate": 4.116265936181731e-06, "loss": 0.3629, "step": 815 }, { "epoch": 1.8053097345132745, "grad_norm": 0.1684337556362152, "learning_rate": 4.103596205946323e-06, "loss": 0.383, "step": 816 }, { "epoch": 1.8075221238938053, "grad_norm": 0.14394833147525787, "learning_rate": 4.090932419775642e-06, "loss": 0.3619, "step": 817 }, { "epoch": 1.8097345132743363, "grad_norm": 0.15720577538013458, "learning_rate": 4.078274661643415e-06, "loss": 0.3705, "step": 818 }, { "epoch": 1.8119469026548671, "grad_norm": 0.1556907743215561, "learning_rate": 4.065623015483394e-06, "loss": 0.3794, "step": 819 }, { "epoch": 1.8141592920353982, "grad_norm": 0.16392920911312103, "learning_rate": 4.0529775651888e-06, "loss": 0.3671, "step": 820 }, { "epoch": 1.8163716814159292, "grad_norm": 0.13912513852119446, "learning_rate": 4.040338394611772e-06, "loss": 0.3662, "step": 821 }, { "epoch": 1.8185840707964602, "grad_norm": 0.16647394001483917, "learning_rate": 4.027705587562808e-06, "loss": 0.3716, "step": 822 }, { "epoch": 1.8207964601769913, "grad_norm": 0.15679362416267395, "learning_rate": 4.015079227810208e-06, "loss": 0.3732, "step": 823 }, { "epoch": 1.823008849557522, "grad_norm": 0.16492421925067902, "learning_rate": 4.002459399079523e-06, "loss": 0.3749, "step": 824 }, { "epoch": 1.825221238938053, "grad_norm": 0.1708700805902481, "learning_rate": 3.9898461850529925e-06, "loss": 0.3775, "step": 825 }, { "epoch": 1.827433628318584, "grad_norm": 0.14519424736499786, "learning_rate": 3.977239669368998e-06, "loss": 0.3708, "step": 826 }, { "epoch": 1.829646017699115, "grad_norm": 0.17130383849143982, "learning_rate": 3.964639935621498e-06, "loss": 0.361, "step": 827 }, { "epoch": 1.831858407079646, "grad_norm": 0.1383930891752243, "learning_rate": 3.952047067359488e-06, "loss": 0.3771, "step": 828 }, { "epoch": 1.834070796460177, "grad_norm": 0.16033463180065155, "learning_rate": 3.939461148086434e-06, "loss": 0.378, "step": 829 }, { "epoch": 1.836283185840708, "grad_norm": 0.1625954806804657, "learning_rate": 3.926882261259723e-06, "loss": 0.3696, "step": 830 }, { "epoch": 1.838495575221239, "grad_norm": 0.14928433299064636, "learning_rate": 3.9143104902901085e-06, "loss": 0.3771, "step": 831 }, { "epoch": 1.8407079646017699, "grad_norm": 0.14306089282035828, "learning_rate": 3.901745918541162e-06, "loss": 0.3773, "step": 832 }, { "epoch": 1.842920353982301, "grad_norm": 0.1791837066411972, "learning_rate": 3.889188629328716e-06, "loss": 0.3767, "step": 833 }, { "epoch": 1.8451327433628317, "grad_norm": 0.14630135893821716, "learning_rate": 3.876638705920312e-06, "loss": 0.3726, "step": 834 }, { "epoch": 1.8473451327433628, "grad_norm": 0.15009824931621552, "learning_rate": 3.864096231534645e-06, "loss": 0.3656, "step": 835 }, { "epoch": 1.8495575221238938, "grad_norm": 0.14151619374752045, "learning_rate": 3.851561289341023e-06, "loss": 0.3777, "step": 836 }, { "epoch": 1.8517699115044248, "grad_norm": 0.15849147737026215, "learning_rate": 3.839033962458802e-06, "loss": 0.3706, "step": 837 }, { "epoch": 1.8539823008849559, "grad_norm": 0.14633609354496002, "learning_rate": 3.826514333956843e-06, "loss": 0.3643, "step": 838 }, { "epoch": 1.856194690265487, "grad_norm": 0.1464887112379074, "learning_rate": 3.8140024868529585e-06, "loss": 0.3688, "step": 839 }, { "epoch": 1.8584070796460177, "grad_norm": 0.16905274987220764, "learning_rate": 3.8014985041133627e-06, "loss": 0.3862, "step": 840 }, { "epoch": 1.8606194690265485, "grad_norm": 0.1511407345533371, "learning_rate": 3.789002468652121e-06, "loss": 0.3659, "step": 841 }, { "epoch": 1.8628318584070795, "grad_norm": 0.13423565030097961, "learning_rate": 3.7765144633306006e-06, "loss": 0.3716, "step": 842 }, { "epoch": 1.8650442477876106, "grad_norm": 0.1653309166431427, "learning_rate": 3.76403457095692e-06, "loss": 0.3645, "step": 843 }, { "epoch": 1.8672566371681416, "grad_norm": 0.1364339143037796, "learning_rate": 3.7515628742854006e-06, "loss": 0.3665, "step": 844 }, { "epoch": 1.8694690265486726, "grad_norm": 0.16295693814754486, "learning_rate": 3.7390994560160187e-06, "loss": 0.3732, "step": 845 }, { "epoch": 1.8716814159292037, "grad_norm": 0.13839292526245117, "learning_rate": 3.726644398793857e-06, "loss": 0.3687, "step": 846 }, { "epoch": 1.8738938053097345, "grad_norm": 0.15138795971870422, "learning_rate": 3.714197785208554e-06, "loss": 0.3795, "step": 847 }, { "epoch": 1.8761061946902655, "grad_norm": 0.14297796785831451, "learning_rate": 3.701759697793761e-06, "loss": 0.3879, "step": 848 }, { "epoch": 1.8783185840707963, "grad_norm": 0.15279804170131683, "learning_rate": 3.689330219026588e-06, "loss": 0.3725, "step": 849 }, { "epoch": 1.8805309734513274, "grad_norm": 0.1471881866455078, "learning_rate": 3.6769094313270647e-06, "loss": 0.3709, "step": 850 }, { "epoch": 1.8827433628318584, "grad_norm": 0.153276726603508, "learning_rate": 3.6644974170575907e-06, "loss": 0.3666, "step": 851 }, { "epoch": 1.8849557522123894, "grad_norm": 0.13241925835609436, "learning_rate": 3.652094258522387e-06, "loss": 0.3726, "step": 852 }, { "epoch": 1.8871681415929205, "grad_norm": 0.14609576761722565, "learning_rate": 3.6397000379669513e-06, "loss": 0.3795, "step": 853 }, { "epoch": 1.8893805309734515, "grad_norm": 0.13605985045433044, "learning_rate": 3.627314837577514e-06, "loss": 0.3688, "step": 854 }, { "epoch": 1.8915929203539823, "grad_norm": 0.14954599738121033, "learning_rate": 3.6149387394804946e-06, "loss": 0.3654, "step": 855 }, { "epoch": 1.893805309734513, "grad_norm": 0.1630832552909851, "learning_rate": 3.6025718257419532e-06, "loss": 0.3837, "step": 856 }, { "epoch": 1.8960176991150441, "grad_norm": 0.15406356751918793, "learning_rate": 3.590214178367049e-06, "loss": 0.3705, "step": 857 }, { "epoch": 1.8982300884955752, "grad_norm": 0.14090083539485931, "learning_rate": 3.5778658792994957e-06, "loss": 0.3754, "step": 858 }, { "epoch": 1.9004424778761062, "grad_norm": 0.1341380476951599, "learning_rate": 3.565527010421019e-06, "loss": 0.3688, "step": 859 }, { "epoch": 1.9026548672566372, "grad_norm": 0.17473238706588745, "learning_rate": 3.5531976535508107e-06, "loss": 0.3743, "step": 860 }, { "epoch": 1.9048672566371683, "grad_norm": 0.14425218105316162, "learning_rate": 3.540877890444989e-06, "loss": 0.3699, "step": 861 }, { "epoch": 1.907079646017699, "grad_norm": 0.1527353972196579, "learning_rate": 3.528567802796061e-06, "loss": 0.3648, "step": 862 }, { "epoch": 1.9092920353982301, "grad_norm": 0.142634317278862, "learning_rate": 3.5162674722323677e-06, "loss": 0.3673, "step": 863 }, { "epoch": 1.911504424778761, "grad_norm": 0.19563040137290955, "learning_rate": 3.5039769803175545e-06, "loss": 0.3703, "step": 864 }, { "epoch": 1.913716814159292, "grad_norm": 0.16823236644268036, "learning_rate": 3.4916964085500277e-06, "loss": 0.3762, "step": 865 }, { "epoch": 1.915929203539823, "grad_norm": 0.15324640274047852, "learning_rate": 3.4794258383624115e-06, "loss": 0.3818, "step": 866 }, { "epoch": 1.918141592920354, "grad_norm": 0.15772469341754913, "learning_rate": 3.4671653511210086e-06, "loss": 0.3761, "step": 867 }, { "epoch": 1.920353982300885, "grad_norm": 0.15935544669628143, "learning_rate": 3.4549150281252635e-06, "loss": 0.3738, "step": 868 }, { "epoch": 1.922566371681416, "grad_norm": 0.15252330899238586, "learning_rate": 3.442674950607221e-06, "loss": 0.3828, "step": 869 }, { "epoch": 1.924778761061947, "grad_norm": 0.1612301915884018, "learning_rate": 3.430445199730987e-06, "loss": 0.3762, "step": 870 }, { "epoch": 1.926991150442478, "grad_norm": 0.16883090138435364, "learning_rate": 3.4182258565921933e-06, "loss": 0.371, "step": 871 }, { "epoch": 1.9292035398230087, "grad_norm": 0.14610210061073303, "learning_rate": 3.406017002217452e-06, "loss": 0.3766, "step": 872 }, { "epoch": 1.9314159292035398, "grad_norm": 0.15019983053207397, "learning_rate": 3.393818717563834e-06, "loss": 0.3646, "step": 873 }, { "epoch": 1.9336283185840708, "grad_norm": 0.1793092042207718, "learning_rate": 3.3816310835183153e-06, "loss": 0.3751, "step": 874 }, { "epoch": 1.9358407079646018, "grad_norm": 0.1681625097990036, "learning_rate": 3.369454180897248e-06, "loss": 0.3729, "step": 875 }, { "epoch": 1.9380530973451329, "grad_norm": 0.13819968700408936, "learning_rate": 3.3572880904458267e-06, "loss": 0.3644, "step": 876 }, { "epoch": 1.9402654867256637, "grad_norm": 0.14391469955444336, "learning_rate": 3.345132892837547e-06, "loss": 0.374, "step": 877 }, { "epoch": 1.9424778761061947, "grad_norm": 0.1573283076286316, "learning_rate": 3.332988668673677e-06, "loss": 0.3648, "step": 878 }, { "epoch": 1.9446902654867255, "grad_norm": 0.1618097424507141, "learning_rate": 3.320855498482718e-06, "loss": 0.37, "step": 879 }, { "epoch": 1.9469026548672566, "grad_norm": 0.14835193753242493, "learning_rate": 3.308733462719873e-06, "loss": 0.3658, "step": 880 }, { "epoch": 1.9491150442477876, "grad_norm": 0.16873405873775482, "learning_rate": 3.2966226417665125e-06, "loss": 0.3727, "step": 881 }, { "epoch": 1.9513274336283186, "grad_norm": 0.17028756439685822, "learning_rate": 3.2845231159296404e-06, "loss": 0.3718, "step": 882 }, { "epoch": 1.9535398230088497, "grad_norm": 0.1492159515619278, "learning_rate": 3.2724349654413612e-06, "loss": 0.3769, "step": 883 }, { "epoch": 1.9557522123893807, "grad_norm": 0.1774129718542099, "learning_rate": 3.2603582704583547e-06, "loss": 0.3734, "step": 884 }, { "epoch": 1.9579646017699115, "grad_norm": 0.16402071714401245, "learning_rate": 3.2482931110613358e-06, "loss": 0.3791, "step": 885 }, { "epoch": 1.9601769911504425, "grad_norm": 0.16409912705421448, "learning_rate": 3.236239567254526e-06, "loss": 0.3689, "step": 886 }, { "epoch": 1.9623893805309733, "grad_norm": 0.15910659730434418, "learning_rate": 3.224197718965124e-06, "loss": 0.367, "step": 887 }, { "epoch": 1.9646017699115044, "grad_norm": 0.16524411737918854, "learning_rate": 3.2121676460427765e-06, "loss": 0.3767, "step": 888 }, { "epoch": 1.9668141592920354, "grad_norm": 0.1652010977268219, "learning_rate": 3.2001494282590466e-06, "loss": 0.3797, "step": 889 }, { "epoch": 1.9690265486725664, "grad_norm": 0.17844292521476746, "learning_rate": 3.188143145306888e-06, "loss": 0.3864, "step": 890 }, { "epoch": 1.9712389380530975, "grad_norm": 0.1592842936515808, "learning_rate": 3.176148876800109e-06, "loss": 0.3792, "step": 891 }, { "epoch": 1.9734513274336283, "grad_norm": 0.1572018712759018, "learning_rate": 3.164166702272855e-06, "loss": 0.3733, "step": 892 }, { "epoch": 1.9756637168141593, "grad_norm": 0.15086941421031952, "learning_rate": 3.1521967011790753e-06, "loss": 0.3731, "step": 893 }, { "epoch": 1.9778761061946901, "grad_norm": 0.15987545251846313, "learning_rate": 3.140238952891994e-06, "loss": 0.3718, "step": 894 }, { "epoch": 1.9800884955752212, "grad_norm": 0.19604964554309845, "learning_rate": 3.1282935367035935e-06, "loss": 0.3693, "step": 895 }, { "epoch": 1.9823008849557522, "grad_norm": 0.14786747097969055, "learning_rate": 3.116360531824074e-06, "loss": 0.3707, "step": 896 }, { "epoch": 1.9845132743362832, "grad_norm": 0.15440818667411804, "learning_rate": 3.1044400173813415e-06, "loss": 0.3782, "step": 897 }, { "epoch": 1.9867256637168142, "grad_norm": 0.1537942886352539, "learning_rate": 3.0925320724204743e-06, "loss": 0.3859, "step": 898 }, { "epoch": 1.9889380530973453, "grad_norm": 0.14648617804050446, "learning_rate": 3.080636775903205e-06, "loss": 0.3701, "step": 899 }, { "epoch": 1.991150442477876, "grad_norm": 0.14667072892189026, "learning_rate": 3.068754206707392e-06, "loss": 0.37, "step": 900 }, { "epoch": 1.9933628318584071, "grad_norm": 0.1474722921848297, "learning_rate": 3.056884443626499e-06, "loss": 0.3815, "step": 901 }, { "epoch": 1.995575221238938, "grad_norm": 0.14442405104637146, "learning_rate": 3.0450275653690743e-06, "loss": 0.376, "step": 902 }, { "epoch": 1.997787610619469, "grad_norm": 0.1478486955165863, "learning_rate": 3.033183650558225e-06, "loss": 0.3692, "step": 903 }, { "epoch": 2.0, "grad_norm": 0.15566988289356232, "learning_rate": 3.021352777731096e-06, "loss": 0.3595, "step": 904 }, { "epoch": 2.002212389380531, "grad_norm": 0.1503528654575348, "learning_rate": 3.0095350253383567e-06, "loss": 0.3539, "step": 905 }, { "epoch": 2.004424778761062, "grad_norm": 0.14453503489494324, "learning_rate": 2.997730471743667e-06, "loss": 0.3616, "step": 906 }, { "epoch": 2.006637168141593, "grad_norm": 0.14768028259277344, "learning_rate": 2.985939195223171e-06, "loss": 0.3451, "step": 907 }, { "epoch": 2.0088495575221237, "grad_norm": 0.17169658839702606, "learning_rate": 2.9741612739649694e-06, "loss": 0.3583, "step": 908 }, { "epoch": 2.0110619469026547, "grad_norm": 0.1610146015882492, "learning_rate": 2.9623967860686035e-06, "loss": 0.3563, "step": 909 }, { "epoch": 2.0132743362831858, "grad_norm": 0.1743859350681305, "learning_rate": 2.9506458095445408e-06, "loss": 0.3538, "step": 910 }, { "epoch": 2.015486725663717, "grad_norm": 0.16596271097660065, "learning_rate": 2.9389084223136523e-06, "loss": 0.3465, "step": 911 }, { "epoch": 2.017699115044248, "grad_norm": 0.16805611550807953, "learning_rate": 2.9271847022066992e-06, "loss": 0.3568, "step": 912 }, { "epoch": 2.019911504424779, "grad_norm": 0.2030407339334488, "learning_rate": 2.915474726963815e-06, "loss": 0.3584, "step": 913 }, { "epoch": 2.02212389380531, "grad_norm": 0.1419757902622223, "learning_rate": 2.903778574233992e-06, "loss": 0.3556, "step": 914 }, { "epoch": 2.024336283185841, "grad_norm": 0.19129696488380432, "learning_rate": 2.892096321574564e-06, "loss": 0.3629, "step": 915 }, { "epoch": 2.0265486725663715, "grad_norm": 0.14490734040737152, "learning_rate": 2.880428046450697e-06, "loss": 0.3446, "step": 916 }, { "epoch": 2.0287610619469025, "grad_norm": 0.19220589101314545, "learning_rate": 2.8687738262348645e-06, "loss": 0.3534, "step": 917 }, { "epoch": 2.0309734513274336, "grad_norm": 0.1297258734703064, "learning_rate": 2.8571337382063503e-06, "loss": 0.3588, "step": 918 }, { "epoch": 2.0331858407079646, "grad_norm": 0.18526272475719452, "learning_rate": 2.845507859550718e-06, "loss": 0.3543, "step": 919 }, { "epoch": 2.0353982300884956, "grad_norm": 0.13665147125720978, "learning_rate": 2.8338962673593196e-06, "loss": 0.3411, "step": 920 }, { "epoch": 2.0376106194690267, "grad_norm": 0.15971770882606506, "learning_rate": 2.822299038628762e-06, "loss": 0.3489, "step": 921 }, { "epoch": 2.0398230088495577, "grad_norm": 0.14711672067642212, "learning_rate": 2.810716250260418e-06, "loss": 0.355, "step": 922 }, { "epoch": 2.0420353982300883, "grad_norm": 0.15557117760181427, "learning_rate": 2.799147979059898e-06, "loss": 0.3549, "step": 923 }, { "epoch": 2.0442477876106193, "grad_norm": 0.1388644129037857, "learning_rate": 2.787594301736556e-06, "loss": 0.3549, "step": 924 }, { "epoch": 2.0464601769911503, "grad_norm": 0.1850692480802536, "learning_rate": 2.7760552949029683e-06, "loss": 0.3578, "step": 925 }, { "epoch": 2.0486725663716814, "grad_norm": 0.15951766073703766, "learning_rate": 2.7645310350744296e-06, "loss": 0.3405, "step": 926 }, { "epoch": 2.0508849557522124, "grad_norm": 0.16866743564605713, "learning_rate": 2.7530215986684584e-06, "loss": 0.3567, "step": 927 }, { "epoch": 2.0530973451327434, "grad_norm": 0.15932750701904297, "learning_rate": 2.741527062004264e-06, "loss": 0.3509, "step": 928 }, { "epoch": 2.0553097345132745, "grad_norm": 0.15171128511428833, "learning_rate": 2.7300475013022666e-06, "loss": 0.362, "step": 929 }, { "epoch": 2.0575221238938055, "grad_norm": 0.15653663873672485, "learning_rate": 2.7185829926835728e-06, "loss": 0.362, "step": 930 }, { "epoch": 2.059734513274336, "grad_norm": 0.14654697477817535, "learning_rate": 2.7071336121694856e-06, "loss": 0.3612, "step": 931 }, { "epoch": 2.061946902654867, "grad_norm": 0.14019131660461426, "learning_rate": 2.695699435680986e-06, "loss": 0.3618, "step": 932 }, { "epoch": 2.064159292035398, "grad_norm": 0.16346430778503418, "learning_rate": 2.6842805390382453e-06, "loss": 0.3476, "step": 933 }, { "epoch": 2.066371681415929, "grad_norm": 0.12734098732471466, "learning_rate": 2.6728769979601044e-06, "loss": 0.3576, "step": 934 }, { "epoch": 2.0685840707964602, "grad_norm": 0.15421155095100403, "learning_rate": 2.661488888063591e-06, "loss": 0.3605, "step": 935 }, { "epoch": 2.0707964601769913, "grad_norm": 0.14653554558753967, "learning_rate": 2.6501162848634023e-06, "loss": 0.3525, "step": 936 }, { "epoch": 2.0730088495575223, "grad_norm": 0.15689359605312347, "learning_rate": 2.6387592637714062e-06, "loss": 0.3531, "step": 937 }, { "epoch": 2.0752212389380533, "grad_norm": 0.1751505434513092, "learning_rate": 2.6274179000961604e-06, "loss": 0.3489, "step": 938 }, { "epoch": 2.077433628318584, "grad_norm": 0.1466556340456009, "learning_rate": 2.6160922690423818e-06, "loss": 0.3594, "step": 939 }, { "epoch": 2.079646017699115, "grad_norm": 0.17935603857040405, "learning_rate": 2.6047824457104766e-06, "loss": 0.3536, "step": 940 }, { "epoch": 2.081858407079646, "grad_norm": 0.13906565308570862, "learning_rate": 2.5934885050960183e-06, "loss": 0.366, "step": 941 }, { "epoch": 2.084070796460177, "grad_norm": 0.16516318917274475, "learning_rate": 2.5822105220892733e-06, "loss": 0.3612, "step": 942 }, { "epoch": 2.086283185840708, "grad_norm": 0.1419246345758438, "learning_rate": 2.5709485714746827e-06, "loss": 0.3484, "step": 943 }, { "epoch": 2.088495575221239, "grad_norm": 0.1455831378698349, "learning_rate": 2.559702727930386e-06, "loss": 0.3614, "step": 944 }, { "epoch": 2.09070796460177, "grad_norm": 0.15348093211650848, "learning_rate": 2.548473066027709e-06, "loss": 0.3595, "step": 945 }, { "epoch": 2.0929203539823007, "grad_norm": 0.14580413699150085, "learning_rate": 2.537259660230679e-06, "loss": 0.3576, "step": 946 }, { "epoch": 2.0951327433628317, "grad_norm": 0.16142421960830688, "learning_rate": 2.5260625848955324e-06, "loss": 0.3595, "step": 947 }, { "epoch": 2.0973451327433628, "grad_norm": 0.1431436985731125, "learning_rate": 2.5148819142702095e-06, "loss": 0.361, "step": 948 }, { "epoch": 2.099557522123894, "grad_norm": 0.18138659000396729, "learning_rate": 2.5037177224938846e-06, "loss": 0.3591, "step": 949 }, { "epoch": 2.101769911504425, "grad_norm": 0.1631578505039215, "learning_rate": 2.492570083596447e-06, "loss": 0.3576, "step": 950 }, { "epoch": 2.103982300884956, "grad_norm": 0.1554289162158966, "learning_rate": 2.4814390714980325e-06, "loss": 0.347, "step": 951 }, { "epoch": 2.106194690265487, "grad_norm": 0.15791471302509308, "learning_rate": 2.4703247600085173e-06, "loss": 0.3632, "step": 952 }, { "epoch": 2.1084070796460175, "grad_norm": 0.1508740782737732, "learning_rate": 2.459227222827043e-06, "loss": 0.3548, "step": 953 }, { "epoch": 2.1106194690265485, "grad_norm": 0.1677217036485672, "learning_rate": 2.4481465335415162e-06, "loss": 0.3544, "step": 954 }, { "epoch": 2.1128318584070795, "grad_norm": 0.13547132909297943, "learning_rate": 2.437082765628122e-06, "loss": 0.3438, "step": 955 }, { "epoch": 2.1150442477876106, "grad_norm": 0.15041135251522064, "learning_rate": 2.426035992450848e-06, "loss": 0.3618, "step": 956 }, { "epoch": 2.1172566371681416, "grad_norm": 0.16150957345962524, "learning_rate": 2.4150062872609812e-06, "loss": 0.3565, "step": 957 }, { "epoch": 2.1194690265486726, "grad_norm": 0.16388878226280212, "learning_rate": 2.4039937231966374e-06, "loss": 0.3576, "step": 958 }, { "epoch": 2.1216814159292037, "grad_norm": 0.1527947336435318, "learning_rate": 2.3929983732822607e-06, "loss": 0.3599, "step": 959 }, { "epoch": 2.1238938053097347, "grad_norm": 0.1690537929534912, "learning_rate": 2.3820203104281616e-06, "loss": 0.3512, "step": 960 }, { "epoch": 2.1261061946902653, "grad_norm": 0.14079472422599792, "learning_rate": 2.3710596074300045e-06, "loss": 0.3641, "step": 961 }, { "epoch": 2.1283185840707963, "grad_norm": 0.14791461825370789, "learning_rate": 2.360116336968353e-06, "loss": 0.3528, "step": 962 }, { "epoch": 2.1305309734513274, "grad_norm": 0.1540174037218094, "learning_rate": 2.3491905716081668e-06, "loss": 0.3546, "step": 963 }, { "epoch": 2.1327433628318584, "grad_norm": 0.14626696705818176, "learning_rate": 2.3382823837983314e-06, "loss": 0.3511, "step": 964 }, { "epoch": 2.1349557522123894, "grad_norm": 0.15072183310985565, "learning_rate": 2.327391845871179e-06, "loss": 0.3621, "step": 965 }, { "epoch": 2.1371681415929205, "grad_norm": 0.14171431958675385, "learning_rate": 2.316519030041998e-06, "loss": 0.3499, "step": 966 }, { "epoch": 2.1393805309734515, "grad_norm": 0.134790301322937, "learning_rate": 2.3056640084085707e-06, "loss": 0.3564, "step": 967 }, { "epoch": 2.1415929203539825, "grad_norm": 0.1500406712293625, "learning_rate": 2.2948268529506768e-06, "loss": 0.3546, "step": 968 }, { "epoch": 2.143805309734513, "grad_norm": 0.1394817978143692, "learning_rate": 2.284007635529632e-06, "loss": 0.3521, "step": 969 }, { "epoch": 2.146017699115044, "grad_norm": 0.16197897493839264, "learning_rate": 2.2732064278877975e-06, "loss": 0.3573, "step": 970 }, { "epoch": 2.148230088495575, "grad_norm": 0.14796555042266846, "learning_rate": 2.2624233016481224e-06, "loss": 0.3676, "step": 971 }, { "epoch": 2.150442477876106, "grad_norm": 0.13100607693195343, "learning_rate": 2.2516583283136474e-06, "loss": 0.3575, "step": 972 }, { "epoch": 2.1526548672566372, "grad_norm": 0.1775142401456833, "learning_rate": 2.2409115792670434e-06, "loss": 0.3584, "step": 973 }, { "epoch": 2.1548672566371683, "grad_norm": 0.17847150564193726, "learning_rate": 2.2301831257701405e-06, "loss": 0.3592, "step": 974 }, { "epoch": 2.1570796460176993, "grad_norm": 0.14856776595115662, "learning_rate": 2.2194730389634444e-06, "loss": 0.3571, "step": 975 }, { "epoch": 2.15929203539823, "grad_norm": 0.17716825008392334, "learning_rate": 2.2087813898656775e-06, "loss": 0.3626, "step": 976 }, { "epoch": 2.161504424778761, "grad_norm": 0.13400061428546906, "learning_rate": 2.1981082493732945e-06, "loss": 0.3591, "step": 977 }, { "epoch": 2.163716814159292, "grad_norm": 0.14946198463439941, "learning_rate": 2.187453688260027e-06, "loss": 0.3581, "step": 978 }, { "epoch": 2.165929203539823, "grad_norm": 0.15970373153686523, "learning_rate": 2.176817777176398e-06, "loss": 0.357, "step": 979 }, { "epoch": 2.168141592920354, "grad_norm": 0.1397649049758911, "learning_rate": 2.166200586649272e-06, "loss": 0.3541, "step": 980 }, { "epoch": 2.170353982300885, "grad_norm": 0.1510685831308365, "learning_rate": 2.1556021870813653e-06, "loss": 0.357, "step": 981 }, { "epoch": 2.172566371681416, "grad_norm": 0.15586382150650024, "learning_rate": 2.1450226487508017e-06, "loss": 0.3425, "step": 982 }, { "epoch": 2.1747787610619467, "grad_norm": 0.14433304965496063, "learning_rate": 2.134462041810632e-06, "loss": 0.3635, "step": 983 }, { "epoch": 2.1769911504424777, "grad_norm": 0.16394202411174774, "learning_rate": 2.12392043628837e-06, "loss": 0.3509, "step": 984 }, { "epoch": 2.1792035398230087, "grad_norm": 0.1602417677640915, "learning_rate": 2.1133979020855357e-06, "loss": 0.3595, "step": 985 }, { "epoch": 2.1814159292035398, "grad_norm": 0.14802055060863495, "learning_rate": 2.102894508977182e-06, "loss": 0.3518, "step": 986 }, { "epoch": 2.183628318584071, "grad_norm": 0.13788118958473206, "learning_rate": 2.0924103266114422e-06, "loss": 0.3537, "step": 987 }, { "epoch": 2.185840707964602, "grad_norm": 0.16340523958206177, "learning_rate": 2.081945424509057e-06, "loss": 0.3588, "step": 988 }, { "epoch": 2.188053097345133, "grad_norm": 0.14665019512176514, "learning_rate": 2.0714998720629264e-06, "loss": 0.3485, "step": 989 }, { "epoch": 2.190265486725664, "grad_norm": 0.14981283247470856, "learning_rate": 2.061073738537635e-06, "loss": 0.3615, "step": 990 }, { "epoch": 2.1924778761061945, "grad_norm": 0.13499760627746582, "learning_rate": 2.0506670930690074e-06, "loss": 0.3551, "step": 991 }, { "epoch": 2.1946902654867255, "grad_norm": 0.15379314124584198, "learning_rate": 2.040280004663637e-06, "loss": 0.3471, "step": 992 }, { "epoch": 2.1969026548672566, "grad_norm": 0.14276005327701569, "learning_rate": 2.0299125421984367e-06, "loss": 0.3565, "step": 993 }, { "epoch": 2.1991150442477876, "grad_norm": 0.15052393078804016, "learning_rate": 2.0195647744201826e-06, "loss": 0.3552, "step": 994 }, { "epoch": 2.2013274336283186, "grad_norm": 0.16427332162857056, "learning_rate": 2.0092367699450466e-06, "loss": 0.3588, "step": 995 }, { "epoch": 2.2035398230088497, "grad_norm": 0.14771218597888947, "learning_rate": 1.9989285972581595e-06, "loss": 0.3583, "step": 996 }, { "epoch": 2.2057522123893807, "grad_norm": 0.13689512014389038, "learning_rate": 1.9886403247131395e-06, "loss": 0.3525, "step": 997 }, { "epoch": 2.2079646017699117, "grad_norm": 0.15630333125591278, "learning_rate": 1.9783720205316535e-06, "loss": 0.3492, "step": 998 }, { "epoch": 2.2101769911504423, "grad_norm": 0.13555830717086792, "learning_rate": 1.9681237528029513e-06, "loss": 0.3368, "step": 999 }, { "epoch": 2.2123893805309733, "grad_norm": 0.16636361181735992, "learning_rate": 1.957895589483426e-06, "loss": 0.3607, "step": 1000 }, { "epoch": 2.2146017699115044, "grad_norm": 0.1358216553926468, "learning_rate": 1.947687598396154e-06, "loss": 0.3587, "step": 1001 }, { "epoch": 2.2168141592920354, "grad_norm": 0.15513230860233307, "learning_rate": 1.9374998472304523e-06, "loss": 0.3537, "step": 1002 }, { "epoch": 2.2190265486725664, "grad_norm": 0.16106431186199188, "learning_rate": 1.927332403541428e-06, "loss": 0.3582, "step": 1003 }, { "epoch": 2.2212389380530975, "grad_norm": 0.16195151209831238, "learning_rate": 1.9171853347495234e-06, "loss": 0.356, "step": 1004 }, { "epoch": 2.2234513274336285, "grad_norm": 0.1492871195077896, "learning_rate": 1.9070587081400815e-06, "loss": 0.3611, "step": 1005 }, { "epoch": 2.225663716814159, "grad_norm": 0.15143686532974243, "learning_rate": 1.896952590862886e-06, "loss": 0.3542, "step": 1006 }, { "epoch": 2.22787610619469, "grad_norm": 0.1420651376247406, "learning_rate": 1.8868670499317298e-06, "loss": 0.3589, "step": 1007 }, { "epoch": 2.230088495575221, "grad_norm": 0.15692518651485443, "learning_rate": 1.8768021522239576e-06, "loss": 0.3606, "step": 1008 }, { "epoch": 2.232300884955752, "grad_norm": 0.14444686472415924, "learning_rate": 1.8667579644800344e-06, "loss": 0.3584, "step": 1009 }, { "epoch": 2.234513274336283, "grad_norm": 0.13391245901584625, "learning_rate": 1.856734553303091e-06, "loss": 0.3522, "step": 1010 }, { "epoch": 2.2367256637168142, "grad_norm": 0.15556755661964417, "learning_rate": 1.8467319851584952e-06, "loss": 0.3591, "step": 1011 }, { "epoch": 2.2389380530973453, "grad_norm": 0.14630340039730072, "learning_rate": 1.8367503263733983e-06, "loss": 0.3556, "step": 1012 }, { "epoch": 2.241150442477876, "grad_norm": 0.13597136735916138, "learning_rate": 1.8267896431363048e-06, "loss": 0.3519, "step": 1013 }, { "epoch": 2.243362831858407, "grad_norm": 0.14077752828598022, "learning_rate": 1.8168500014966316e-06, "loss": 0.349, "step": 1014 }, { "epoch": 2.245575221238938, "grad_norm": 0.15375056862831116, "learning_rate": 1.8069314673642624e-06, "loss": 0.3575, "step": 1015 }, { "epoch": 2.247787610619469, "grad_norm": 0.1405712366104126, "learning_rate": 1.7970341065091246e-06, "loss": 0.3543, "step": 1016 }, { "epoch": 2.25, "grad_norm": 0.14993660151958466, "learning_rate": 1.7871579845607378e-06, "loss": 0.3564, "step": 1017 }, { "epoch": 2.252212389380531, "grad_norm": 0.14628523588180542, "learning_rate": 1.7773031670077934e-06, "loss": 0.3488, "step": 1018 }, { "epoch": 2.254424778761062, "grad_norm": 0.1479613482952118, "learning_rate": 1.7674697191977053e-06, "loss": 0.3619, "step": 1019 }, { "epoch": 2.256637168141593, "grad_norm": 0.14518770575523376, "learning_rate": 1.757657706336192e-06, "loss": 0.3695, "step": 1020 }, { "epoch": 2.258849557522124, "grad_norm": 0.14480751752853394, "learning_rate": 1.7478671934868302e-06, "loss": 0.3556, "step": 1021 }, { "epoch": 2.2610619469026547, "grad_norm": 0.15581965446472168, "learning_rate": 1.7380982455706353e-06, "loss": 0.3559, "step": 1022 }, { "epoch": 2.2632743362831858, "grad_norm": 0.14438295364379883, "learning_rate": 1.72835092736562e-06, "loss": 0.3697, "step": 1023 }, { "epoch": 2.265486725663717, "grad_norm": 0.1663781851530075, "learning_rate": 1.7186253035063738e-06, "loss": 0.3517, "step": 1024 }, { "epoch": 2.267699115044248, "grad_norm": 0.14601710438728333, "learning_rate": 1.7089214384836322e-06, "loss": 0.3467, "step": 1025 }, { "epoch": 2.269911504424779, "grad_norm": 0.15470395982265472, "learning_rate": 1.699239396643841e-06, "loss": 0.3529, "step": 1026 }, { "epoch": 2.27212389380531, "grad_norm": 0.15689963102340698, "learning_rate": 1.6895792421887437e-06, "loss": 0.3575, "step": 1027 }, { "epoch": 2.274336283185841, "grad_norm": 0.1612994223833084, "learning_rate": 1.6799410391749416e-06, "loss": 0.356, "step": 1028 }, { "epoch": 2.2765486725663715, "grad_norm": 0.15020525455474854, "learning_rate": 1.670324851513483e-06, "loss": 0.3403, "step": 1029 }, { "epoch": 2.2787610619469025, "grad_norm": 0.16059263050556183, "learning_rate": 1.6607307429694237e-06, "loss": 0.3604, "step": 1030 }, { "epoch": 2.2809734513274336, "grad_norm": 0.1457761973142624, "learning_rate": 1.6511587771614208e-06, "loss": 0.3573, "step": 1031 }, { "epoch": 2.2831858407079646, "grad_norm": 0.14527295529842377, "learning_rate": 1.6416090175612958e-06, "loss": 0.3563, "step": 1032 }, { "epoch": 2.2853982300884956, "grad_norm": 0.145360067486763, "learning_rate": 1.6320815274936269e-06, "loss": 0.356, "step": 1033 }, { "epoch": 2.2876106194690267, "grad_norm": 0.14848755300045013, "learning_rate": 1.622576370135317e-06, "loss": 0.3522, "step": 1034 }, { "epoch": 2.2898230088495577, "grad_norm": 0.1375618577003479, "learning_rate": 1.613093608515181e-06, "loss": 0.3513, "step": 1035 }, { "epoch": 2.2920353982300883, "grad_norm": 0.13908326625823975, "learning_rate": 1.6036333055135345e-06, "loss": 0.3604, "step": 1036 }, { "epoch": 2.2942477876106193, "grad_norm": 0.15508785843849182, "learning_rate": 1.5941955238617612e-06, "loss": 0.3604, "step": 1037 }, { "epoch": 2.2964601769911503, "grad_norm": 0.14073608815670013, "learning_rate": 1.5847803261419109e-06, "loss": 0.3624, "step": 1038 }, { "epoch": 2.2986725663716814, "grad_norm": 0.1389189213514328, "learning_rate": 1.575387774786274e-06, "loss": 0.3512, "step": 1039 }, { "epoch": 2.3008849557522124, "grad_norm": 0.1345841884613037, "learning_rate": 1.5660179320769792e-06, "loss": 0.3625, "step": 1040 }, { "epoch": 2.3030973451327434, "grad_norm": 0.14056994020938873, "learning_rate": 1.556670860145567e-06, "loss": 0.3559, "step": 1041 }, { "epoch": 2.3053097345132745, "grad_norm": 0.13505183160305023, "learning_rate": 1.5473466209725907e-06, "loss": 0.3525, "step": 1042 }, { "epoch": 2.307522123893805, "grad_norm": 0.13815952837467194, "learning_rate": 1.5380452763871951e-06, "loss": 0.3681, "step": 1043 }, { "epoch": 2.309734513274336, "grad_norm": 0.1454608291387558, "learning_rate": 1.5287668880667107e-06, "loss": 0.3542, "step": 1044 }, { "epoch": 2.311946902654867, "grad_norm": 0.1534843146800995, "learning_rate": 1.5195115175362485e-06, "loss": 0.3575, "step": 1045 }, { "epoch": 2.314159292035398, "grad_norm": 0.16305556893348694, "learning_rate": 1.5102792261682813e-06, "loss": 0.3586, "step": 1046 }, { "epoch": 2.316371681415929, "grad_norm": 0.16194356977939606, "learning_rate": 1.5010700751822555e-06, "loss": 0.351, "step": 1047 }, { "epoch": 2.3185840707964602, "grad_norm": 0.17614226043224335, "learning_rate": 1.4918841256441603e-06, "loss": 0.3603, "step": 1048 }, { "epoch": 2.3207964601769913, "grad_norm": 0.1571769267320633, "learning_rate": 1.4827214384661447e-06, "loss": 0.3493, "step": 1049 }, { "epoch": 2.3230088495575223, "grad_norm": 0.1548829972743988, "learning_rate": 1.473582074406099e-06, "loss": 0.363, "step": 1050 }, { "epoch": 2.3252212389380533, "grad_norm": 0.1630624383687973, "learning_rate": 1.4644660940672628e-06, "loss": 0.348, "step": 1051 }, { "epoch": 2.327433628318584, "grad_norm": 0.18259887397289276, "learning_rate": 1.455373557897814e-06, "loss": 0.3582, "step": 1052 }, { "epoch": 2.329646017699115, "grad_norm": 0.1488334983587265, "learning_rate": 1.4463045261904718e-06, "loss": 0.3559, "step": 1053 }, { "epoch": 2.331858407079646, "grad_norm": 0.16422435641288757, "learning_rate": 1.4372590590821012e-06, "loss": 0.3586, "step": 1054 }, { "epoch": 2.334070796460177, "grad_norm": 0.13306331634521484, "learning_rate": 1.4282372165533042e-06, "loss": 0.3465, "step": 1055 }, { "epoch": 2.336283185840708, "grad_norm": 0.1893102079629898, "learning_rate": 1.4192390584280347e-06, "loss": 0.3616, "step": 1056 }, { "epoch": 2.338495575221239, "grad_norm": 0.15167655050754547, "learning_rate": 1.4102646443731866e-06, "loss": 0.3515, "step": 1057 }, { "epoch": 2.34070796460177, "grad_norm": 0.17918401956558228, "learning_rate": 1.4013140338982168e-06, "loss": 0.3634, "step": 1058 }, { "epoch": 2.3429203539823007, "grad_norm": 0.152436301112175, "learning_rate": 1.392387286354731e-06, "loss": 0.3541, "step": 1059 }, { "epoch": 2.3451327433628317, "grad_norm": 0.20484967529773712, "learning_rate": 1.3834844609361064e-06, "loss": 0.3529, "step": 1060 }, { "epoch": 2.3473451327433628, "grad_norm": 0.15942007303237915, "learning_rate": 1.3746056166770872e-06, "loss": 0.3576, "step": 1061 }, { "epoch": 2.349557522123894, "grad_norm": 0.17121823132038116, "learning_rate": 1.3657508124533992e-06, "loss": 0.3464, "step": 1062 }, { "epoch": 2.351769911504425, "grad_norm": 0.16647647321224213, "learning_rate": 1.3569201069813626e-06, "loss": 0.36, "step": 1063 }, { "epoch": 2.353982300884956, "grad_norm": 0.14579357206821442, "learning_rate": 1.3481135588174926e-06, "loss": 0.3429, "step": 1064 }, { "epoch": 2.356194690265487, "grad_norm": 0.15810784697532654, "learning_rate": 1.3393312263581222e-06, "loss": 0.3518, "step": 1065 }, { "epoch": 2.3584070796460175, "grad_norm": 0.15230241417884827, "learning_rate": 1.330573167839005e-06, "loss": 0.3682, "step": 1066 }, { "epoch": 2.3606194690265485, "grad_norm": 0.15005375444889069, "learning_rate": 1.3218394413349389e-06, "loss": 0.3484, "step": 1067 }, { "epoch": 2.3628318584070795, "grad_norm": 0.13220618665218353, "learning_rate": 1.3131301047593685e-06, "loss": 0.3527, "step": 1068 }, { "epoch": 2.3650442477876106, "grad_norm": 0.15161649882793427, "learning_rate": 1.3044452158640197e-06, "loss": 0.3679, "step": 1069 }, { "epoch": 2.3672566371681416, "grad_norm": 0.14612559974193573, "learning_rate": 1.2957848322384959e-06, "loss": 0.3595, "step": 1070 }, { "epoch": 2.3694690265486726, "grad_norm": 0.14127154648303986, "learning_rate": 1.2871490113099066e-06, "loss": 0.3536, "step": 1071 }, { "epoch": 2.3716814159292037, "grad_norm": 0.13997747004032135, "learning_rate": 1.2785378103424917e-06, "loss": 0.3549, "step": 1072 }, { "epoch": 2.3738938053097347, "grad_norm": 0.13644693791866302, "learning_rate": 1.2699512864372287e-06, "loss": 0.3596, "step": 1073 }, { "epoch": 2.3761061946902653, "grad_norm": 0.14478059113025665, "learning_rate": 1.2613894965314682e-06, "loss": 0.3588, "step": 1074 }, { "epoch": 2.3783185840707963, "grad_norm": 0.15023688971996307, "learning_rate": 1.2528524973985424e-06, "loss": 0.3606, "step": 1075 }, { "epoch": 2.3805309734513274, "grad_norm": 0.14657047390937805, "learning_rate": 1.2443403456474017e-06, "loss": 0.3647, "step": 1076 }, { "epoch": 2.3827433628318584, "grad_norm": 0.1500697135925293, "learning_rate": 1.2358530977222276e-06, "loss": 0.3551, "step": 1077 }, { "epoch": 2.3849557522123894, "grad_norm": 0.14721229672431946, "learning_rate": 1.2273908099020703e-06, "loss": 0.3595, "step": 1078 }, { "epoch": 2.3871681415929205, "grad_norm": 0.15616820752620697, "learning_rate": 1.218953538300462e-06, "loss": 0.3498, "step": 1079 }, { "epoch": 2.3893805309734515, "grad_norm": 0.15405981242656708, "learning_rate": 1.2105413388650577e-06, "loss": 0.3482, "step": 1080 }, { "epoch": 2.3915929203539825, "grad_norm": 0.14757871627807617, "learning_rate": 1.2021542673772584e-06, "loss": 0.3592, "step": 1081 }, { "epoch": 2.393805309734513, "grad_norm": 0.16278573870658875, "learning_rate": 1.193792379451837e-06, "loss": 0.3515, "step": 1082 }, { "epoch": 2.396017699115044, "grad_norm": 0.15779441595077515, "learning_rate": 1.1854557305365783e-06, "loss": 0.351, "step": 1083 }, { "epoch": 2.398230088495575, "grad_norm": 0.16500075161457062, "learning_rate": 1.1771443759119028e-06, "loss": 0.3538, "step": 1084 }, { "epoch": 2.400442477876106, "grad_norm": 0.15716195106506348, "learning_rate": 1.1688583706905099e-06, "loss": 0.3522, "step": 1085 }, { "epoch": 2.4026548672566372, "grad_norm": 0.1632101982831955, "learning_rate": 1.1605977698170001e-06, "loss": 0.3612, "step": 1086 }, { "epoch": 2.4048672566371683, "grad_norm": 0.14858639240264893, "learning_rate": 1.1523626280675237e-06, "loss": 0.3575, "step": 1087 }, { "epoch": 2.4070796460176993, "grad_norm": 0.16127295792102814, "learning_rate": 1.1441530000494055e-06, "loss": 0.3557, "step": 1088 }, { "epoch": 2.40929203539823, "grad_norm": 0.15704701840877533, "learning_rate": 1.135968940200794e-06, "loss": 0.3547, "step": 1089 }, { "epoch": 2.411504424778761, "grad_norm": 0.14434731006622314, "learning_rate": 1.1278105027902898e-06, "loss": 0.3561, "step": 1090 }, { "epoch": 2.413716814159292, "grad_norm": 0.17251616716384888, "learning_rate": 1.1196777419165927e-06, "loss": 0.3646, "step": 1091 }, { "epoch": 2.415929203539823, "grad_norm": 0.17024458944797516, "learning_rate": 1.1115707115081448e-06, "loss": 0.3536, "step": 1092 }, { "epoch": 2.418141592920354, "grad_norm": 0.14864687621593475, "learning_rate": 1.1034894653227618e-06, "loss": 0.3526, "step": 1093 }, { "epoch": 2.420353982300885, "grad_norm": 0.18286971747875214, "learning_rate": 1.095434056947291e-06, "loss": 0.356, "step": 1094 }, { "epoch": 2.422566371681416, "grad_norm": 0.1528891921043396, "learning_rate": 1.0874045397972433e-06, "loss": 0.352, "step": 1095 }, { "epoch": 2.4247787610619467, "grad_norm": 0.15347851812839508, "learning_rate": 1.0794009671164484e-06, "loss": 0.3591, "step": 1096 }, { "epoch": 2.4269911504424777, "grad_norm": 0.16304871439933777, "learning_rate": 1.0714233919766953e-06, "loss": 0.3557, "step": 1097 }, { "epoch": 2.4292035398230087, "grad_norm": 0.13769623637199402, "learning_rate": 1.0634718672773863e-06, "loss": 0.3534, "step": 1098 }, { "epoch": 2.4314159292035398, "grad_norm": 0.1583521068096161, "learning_rate": 1.055546445745178e-06, "loss": 0.3591, "step": 1099 }, { "epoch": 2.433628318584071, "grad_norm": 0.17005863785743713, "learning_rate": 1.0476471799336424e-06, "loss": 0.3532, "step": 1100 }, { "epoch": 2.435840707964602, "grad_norm": 0.1480191946029663, "learning_rate": 1.0397741222229057e-06, "loss": 0.3556, "step": 1101 }, { "epoch": 2.438053097345133, "grad_norm": 0.16332688927650452, "learning_rate": 1.0319273248193145e-06, "loss": 0.3595, "step": 1102 }, { "epoch": 2.440265486725664, "grad_norm": 0.17268969118595123, "learning_rate": 1.0241068397550807e-06, "loss": 0.3522, "step": 1103 }, { "epoch": 2.442477876106195, "grad_norm": 0.15803387761116028, "learning_rate": 1.0163127188879352e-06, "loss": 0.3526, "step": 1104 }, { "epoch": 2.4446902654867255, "grad_norm": 0.15174445509910583, "learning_rate": 1.008545013900794e-06, "loss": 0.3521, "step": 1105 }, { "epoch": 2.4469026548672566, "grad_norm": 0.1590649038553238, "learning_rate": 1.0008037763014033e-06, "loss": 0.3529, "step": 1106 }, { "epoch": 2.4491150442477876, "grad_norm": 0.15736490488052368, "learning_rate": 9.930890574220076e-07, "loss": 0.3549, "step": 1107 }, { "epoch": 2.4513274336283186, "grad_norm": 0.16248001158237457, "learning_rate": 9.85400908419002e-07, "loss": 0.3503, "step": 1108 }, { "epoch": 2.4535398230088497, "grad_norm": 0.16862428188323975, "learning_rate": 9.77739380272601e-07, "loss": 0.3513, "step": 1109 }, { "epoch": 2.4557522123893807, "grad_norm": 0.16417399048805237, "learning_rate": 9.701045237864898e-07, "loss": 0.3505, "step": 1110 }, { "epoch": 2.4579646017699117, "grad_norm": 0.14535874128341675, "learning_rate": 9.624963895874995e-07, "loss": 0.3623, "step": 1111 }, { "epoch": 2.4601769911504423, "grad_norm": 0.1544649451971054, "learning_rate": 9.549150281252633e-07, "loss": 0.3524, "step": 1112 }, { "epoch": 2.4623893805309733, "grad_norm": 0.157157301902771, "learning_rate": 9.473604896718808e-07, "loss": 0.3542, "step": 1113 }, { "epoch": 2.4646017699115044, "grad_norm": 0.16022609174251556, "learning_rate": 9.398328243215937e-07, "loss": 0.3486, "step": 1114 }, { "epoch": 2.4668141592920354, "grad_norm": 0.14643922448158264, "learning_rate": 9.323320819904419e-07, "loss": 0.3588, "step": 1115 }, { "epoch": 2.4690265486725664, "grad_norm": 0.15757547318935394, "learning_rate": 9.248583124159438e-07, "loss": 0.3619, "step": 1116 }, { "epoch": 2.4712389380530975, "grad_norm": 0.14214399456977844, "learning_rate": 9.174115651567561e-07, "loss": 0.3559, "step": 1117 }, { "epoch": 2.4734513274336285, "grad_norm": 0.1586948037147522, "learning_rate": 9.099918895923554e-07, "loss": 0.3504, "step": 1118 }, { "epoch": 2.475663716814159, "grad_norm": 0.14957429468631744, "learning_rate": 9.025993349226997e-07, "loss": 0.3714, "step": 1119 }, { "epoch": 2.47787610619469, "grad_norm": 0.15308675169944763, "learning_rate": 8.952339501679142e-07, "loss": 0.3639, "step": 1120 }, { "epoch": 2.480088495575221, "grad_norm": 0.1536092609167099, "learning_rate": 8.878957841679542e-07, "loss": 0.3543, "step": 1121 }, { "epoch": 2.482300884955752, "grad_norm": 0.14460043609142303, "learning_rate": 8.805848855822918e-07, "loss": 0.3573, "step": 1122 }, { "epoch": 2.484513274336283, "grad_norm": 0.1362440139055252, "learning_rate": 8.733013028895864e-07, "loss": 0.3557, "step": 1123 }, { "epoch": 2.4867256637168142, "grad_norm": 0.15375365316867828, "learning_rate": 8.660450843873647e-07, "loss": 0.3555, "step": 1124 }, { "epoch": 2.4889380530973453, "grad_norm": 0.14579614996910095, "learning_rate": 8.588162781917042e-07, "loss": 0.358, "step": 1125 }, { "epoch": 2.491150442477876, "grad_norm": 0.18101723492145538, "learning_rate": 8.516149322369055e-07, "loss": 0.3595, "step": 1126 }, { "epoch": 2.493362831858407, "grad_norm": 0.16425468027591705, "learning_rate": 8.444410942751863e-07, "loss": 0.3459, "step": 1127 }, { "epoch": 2.495575221238938, "grad_norm": 0.17149604856967926, "learning_rate": 8.372948118763536e-07, "loss": 0.3601, "step": 1128 }, { "epoch": 2.497787610619469, "grad_norm": 0.1419076919555664, "learning_rate": 8.301761324274965e-07, "loss": 0.3441, "step": 1129 }, { "epoch": 2.5, "grad_norm": 0.14477266371250153, "learning_rate": 8.230851031326653e-07, "loss": 0.3539, "step": 1130 }, { "epoch": 2.502212389380531, "grad_norm": 0.1530475616455078, "learning_rate": 8.160217710125661e-07, "loss": 0.3502, "step": 1131 }, { "epoch": 2.504424778761062, "grad_norm": 0.14788125455379486, "learning_rate": 8.089861829042406e-07, "loss": 0.3591, "step": 1132 }, { "epoch": 2.5066371681415927, "grad_norm": 0.15901798009872437, "learning_rate": 8.019783854607593e-07, "loss": 0.3551, "step": 1133 }, { "epoch": 2.508849557522124, "grad_norm": 0.1555824875831604, "learning_rate": 7.949984251509185e-07, "loss": 0.3448, "step": 1134 }, { "epoch": 2.5110619469026547, "grad_norm": 0.15857146680355072, "learning_rate": 7.880463482589196e-07, "loss": 0.3433, "step": 1135 }, { "epoch": 2.5132743362831858, "grad_norm": 0.16051232814788818, "learning_rate": 7.811222008840719e-07, "loss": 0.3563, "step": 1136 }, { "epoch": 2.515486725663717, "grad_norm": 0.13366073369979858, "learning_rate": 7.742260289404819e-07, "loss": 0.3552, "step": 1137 }, { "epoch": 2.517699115044248, "grad_norm": 0.16636747121810913, "learning_rate": 7.673578781567537e-07, "loss": 0.3549, "step": 1138 }, { "epoch": 2.519911504424779, "grad_norm": 0.16012783348560333, "learning_rate": 7.605177940756774e-07, "loss": 0.3505, "step": 1139 }, { "epoch": 2.52212389380531, "grad_norm": 0.1498229205608368, "learning_rate": 7.537058220539395e-07, "loss": 0.3579, "step": 1140 }, { "epoch": 2.524336283185841, "grad_norm": 0.17187750339508057, "learning_rate": 7.469220072618094e-07, "loss": 0.3548, "step": 1141 }, { "epoch": 2.5265486725663715, "grad_norm": 0.15075713396072388, "learning_rate": 7.401663946828469e-07, "loss": 0.3568, "step": 1142 }, { "epoch": 2.5287610619469025, "grad_norm": 0.18497484922409058, "learning_rate": 7.334390291136051e-07, "loss": 0.3579, "step": 1143 }, { "epoch": 2.5309734513274336, "grad_norm": 0.1425379067659378, "learning_rate": 7.267399551633253e-07, "loss": 0.3569, "step": 1144 }, { "epoch": 2.5331858407079646, "grad_norm": 0.17833374440670013, "learning_rate": 7.200692172536555e-07, "loss": 0.3476, "step": 1145 }, { "epoch": 2.5353982300884956, "grad_norm": 0.15675534307956696, "learning_rate": 7.13426859618338e-07, "loss": 0.3683, "step": 1146 }, { "epoch": 2.5376106194690267, "grad_norm": 0.16931159794330597, "learning_rate": 7.068129263029316e-07, "loss": 0.3558, "step": 1147 }, { "epoch": 2.5398230088495577, "grad_norm": 0.1581314355134964, "learning_rate": 7.002274611645083e-07, "loss": 0.36, "step": 1148 }, { "epoch": 2.5420353982300883, "grad_norm": 0.1542506217956543, "learning_rate": 6.936705078713713e-07, "loss": 0.3541, "step": 1149 }, { "epoch": 2.5442477876106193, "grad_norm": 0.16385367512702942, "learning_rate": 6.871421099027586e-07, "loss": 0.3587, "step": 1150 }, { "epoch": 2.5464601769911503, "grad_norm": 0.14884746074676514, "learning_rate": 6.806423105485576e-07, "loss": 0.3542, "step": 1151 }, { "epoch": 2.5486725663716814, "grad_norm": 0.15693628787994385, "learning_rate": 6.741711529090212e-07, "loss": 0.3573, "step": 1152 }, { "epoch": 2.5508849557522124, "grad_norm": 0.13773605227470398, "learning_rate": 6.677286798944743e-07, "loss": 0.3537, "step": 1153 }, { "epoch": 2.5530973451327434, "grad_norm": 0.15046106278896332, "learning_rate": 6.61314934225037e-07, "loss": 0.3533, "step": 1154 }, { "epoch": 2.5553097345132745, "grad_norm": 0.15596453845500946, "learning_rate": 6.549299584303343e-07, "loss": 0.3541, "step": 1155 }, { "epoch": 2.557522123893805, "grad_norm": 0.1392877697944641, "learning_rate": 6.485737948492237e-07, "loss": 0.3671, "step": 1156 }, { "epoch": 2.5597345132743365, "grad_norm": 0.16301456093788147, "learning_rate": 6.422464856295035e-07, "loss": 0.3551, "step": 1157 }, { "epoch": 2.561946902654867, "grad_norm": 0.14623309671878815, "learning_rate": 6.359480727276407e-07, "loss": 0.3566, "step": 1158 }, { "epoch": 2.564159292035398, "grad_norm": 0.15128614008426666, "learning_rate": 6.296785979084891e-07, "loss": 0.3573, "step": 1159 }, { "epoch": 2.566371681415929, "grad_norm": 0.1874575912952423, "learning_rate": 6.234381027450132e-07, "loss": 0.3497, "step": 1160 }, { "epoch": 2.5685840707964602, "grad_norm": 0.15378333628177643, "learning_rate": 6.172266286180162e-07, "loss": 0.3462, "step": 1161 }, { "epoch": 2.5707964601769913, "grad_norm": 0.1589403748512268, "learning_rate": 6.110442167158592e-07, "loss": 0.3496, "step": 1162 }, { "epoch": 2.573008849557522, "grad_norm": 0.1497460901737213, "learning_rate": 6.048909080341936e-07, "loss": 0.3509, "step": 1163 }, { "epoch": 2.5752212389380533, "grad_norm": 0.15763600170612335, "learning_rate": 5.987667433756844e-07, "loss": 0.3511, "step": 1164 }, { "epoch": 2.577433628318584, "grad_norm": 0.16659215092658997, "learning_rate": 5.926717633497453e-07, "loss": 0.364, "step": 1165 }, { "epoch": 2.579646017699115, "grad_norm": 0.15350916981697083, "learning_rate": 5.866060083722624e-07, "loss": 0.3604, "step": 1166 }, { "epoch": 2.581858407079646, "grad_norm": 0.1593090444803238, "learning_rate": 5.805695186653365e-07, "loss": 0.3554, "step": 1167 }, { "epoch": 2.584070796460177, "grad_norm": 0.13886038959026337, "learning_rate": 5.745623342570039e-07, "loss": 0.3567, "step": 1168 }, { "epoch": 2.586283185840708, "grad_norm": 0.17093704640865326, "learning_rate": 5.685844949809777e-07, "loss": 0.3569, "step": 1169 }, { "epoch": 2.588495575221239, "grad_norm": 0.11451248824596405, "learning_rate": 5.626360404763875e-07, "loss": 0.3468, "step": 1170 }, { "epoch": 2.59070796460177, "grad_norm": 0.15386511385440826, "learning_rate": 5.567170101875074e-07, "loss": 0.3639, "step": 1171 }, { "epoch": 2.5929203539823007, "grad_norm": 0.14580263197422028, "learning_rate": 5.508274433635019e-07, "loss": 0.3555, "step": 1172 }, { "epoch": 2.5951327433628317, "grad_norm": 0.14180666208267212, "learning_rate": 5.449673790581611e-07, "loss": 0.3593, "step": 1173 }, { "epoch": 2.5973451327433628, "grad_norm": 0.13817204535007477, "learning_rate": 5.391368561296456e-07, "loss": 0.3412, "step": 1174 }, { "epoch": 2.599557522123894, "grad_norm": 0.14034853875637054, "learning_rate": 5.333359132402238e-07, "loss": 0.3539, "step": 1175 }, { "epoch": 2.601769911504425, "grad_norm": 0.14221332967281342, "learning_rate": 5.275645888560233e-07, "loss": 0.3541, "step": 1176 }, { "epoch": 2.603982300884956, "grad_norm": 0.15998932719230652, "learning_rate": 5.218229212467635e-07, "loss": 0.3696, "step": 1177 }, { "epoch": 2.606194690265487, "grad_norm": 0.15716540813446045, "learning_rate": 5.161109484855182e-07, "loss": 0.3518, "step": 1178 }, { "epoch": 2.6084070796460175, "grad_norm": 0.13059420883655548, "learning_rate": 5.104287084484489e-07, "loss": 0.3468, "step": 1179 }, { "epoch": 2.6106194690265485, "grad_norm": 0.15905797481536865, "learning_rate": 5.047762388145582e-07, "loss": 0.3515, "step": 1180 }, { "epoch": 2.6128318584070795, "grad_norm": 0.1637641042470932, "learning_rate": 4.991535770654449e-07, "loss": 0.3589, "step": 1181 }, { "epoch": 2.6150442477876106, "grad_norm": 0.14523540437221527, "learning_rate": 4.935607604850473e-07, "loss": 0.3532, "step": 1182 }, { "epoch": 2.6172566371681416, "grad_norm": 0.15000000596046448, "learning_rate": 4.879978261594037e-07, "loss": 0.3557, "step": 1183 }, { "epoch": 2.6194690265486726, "grad_norm": 0.155274897813797, "learning_rate": 4.824648109763991e-07, "loss": 0.3578, "step": 1184 }, { "epoch": 2.6216814159292037, "grad_norm": 0.1417737454175949, "learning_rate": 4.769617516255276e-07, "loss": 0.3631, "step": 1185 }, { "epoch": 2.6238938053097343, "grad_norm": 0.1490320861339569, "learning_rate": 4.71488684597643e-07, "loss": 0.3526, "step": 1186 }, { "epoch": 2.6261061946902657, "grad_norm": 0.1615634262561798, "learning_rate": 4.660456461847224e-07, "loss": 0.3572, "step": 1187 }, { "epoch": 2.6283185840707963, "grad_norm": 0.14165712893009186, "learning_rate": 4.606326724796195e-07, "loss": 0.3653, "step": 1188 }, { "epoch": 2.6305309734513274, "grad_norm": 0.14493530988693237, "learning_rate": 4.5524979937583203e-07, "loss": 0.3527, "step": 1189 }, { "epoch": 2.6327433628318584, "grad_norm": 0.13274651765823364, "learning_rate": 4.498970625672588e-07, "loss": 0.3544, "step": 1190 }, { "epoch": 2.6349557522123894, "grad_norm": 0.16449962556362152, "learning_rate": 4.445744975479627e-07, "loss": 0.3589, "step": 1191 }, { "epoch": 2.6371681415929205, "grad_norm": 0.1559862345457077, "learning_rate": 4.392821396119407e-07, "loss": 0.3507, "step": 1192 }, { "epoch": 2.6393805309734515, "grad_norm": 0.15669386088848114, "learning_rate": 4.3402002385288133e-07, "loss": 0.3505, "step": 1193 }, { "epoch": 2.6415929203539825, "grad_norm": 0.17834536731243134, "learning_rate": 4.2878818516394307e-07, "loss": 0.3508, "step": 1194 }, { "epoch": 2.643805309734513, "grad_norm": 0.1579374521970749, "learning_rate": 4.235866582375114e-07, "loss": 0.3627, "step": 1195 }, { "epoch": 2.646017699115044, "grad_norm": 0.16911913454532623, "learning_rate": 4.184154775649768e-07, "loss": 0.36, "step": 1196 }, { "epoch": 2.648230088495575, "grad_norm": 0.17118600010871887, "learning_rate": 4.132746774365021e-07, "loss": 0.3571, "step": 1197 }, { "epoch": 2.650442477876106, "grad_norm": 0.1715717762708664, "learning_rate": 4.0816429194079857e-07, "loss": 0.361, "step": 1198 }, { "epoch": 2.6526548672566372, "grad_norm": 0.17212989926338196, "learning_rate": 4.030843549648944e-07, "loss": 0.3517, "step": 1199 }, { "epoch": 2.6548672566371683, "grad_norm": 0.17287541925907135, "learning_rate": 3.9803490019391545e-07, "loss": 0.354, "step": 1200 }, { "epoch": 2.6570796460176993, "grad_norm": 0.18660354614257812, "learning_rate": 3.930159611108603e-07, "loss": 0.3559, "step": 1201 }, { "epoch": 2.65929203539823, "grad_norm": 0.13577118515968323, "learning_rate": 3.880275709963749e-07, "loss": 0.3612, "step": 1202 }, { "epoch": 2.661504424778761, "grad_norm": 0.17823448777198792, "learning_rate": 3.8306976292853794e-07, "loss": 0.3542, "step": 1203 }, { "epoch": 2.663716814159292, "grad_norm": 0.14912089705467224, "learning_rate": 3.7814256978263465e-07, "loss": 0.3525, "step": 1204 }, { "epoch": 2.665929203539823, "grad_norm": 0.16290871798992157, "learning_rate": 3.7324602423094603e-07, "loss": 0.3601, "step": 1205 }, { "epoch": 2.668141592920354, "grad_norm": 0.15941615402698517, "learning_rate": 3.683801587425251e-07, "loss": 0.3627, "step": 1206 }, { "epoch": 2.670353982300885, "grad_norm": 0.14429469406604767, "learning_rate": 3.635450055829881e-07, "loss": 0.3604, "step": 1207 }, { "epoch": 2.672566371681416, "grad_norm": 0.15333802998065948, "learning_rate": 3.587405968142943e-07, "loss": 0.3507, "step": 1208 }, { "epoch": 2.6747787610619467, "grad_norm": 0.1445317417383194, "learning_rate": 3.539669642945387e-07, "loss": 0.3567, "step": 1209 }, { "epoch": 2.676991150442478, "grad_norm": 0.16064175963401794, "learning_rate": 3.4922413967773905e-07, "loss": 0.3623, "step": 1210 }, { "epoch": 2.6792035398230087, "grad_norm": 0.1633787751197815, "learning_rate": 3.4451215441362264e-07, "loss": 0.3509, "step": 1211 }, { "epoch": 2.6814159292035398, "grad_norm": 0.14114798605442047, "learning_rate": 3.3983103974742415e-07, "loss": 0.3658, "step": 1212 }, { "epoch": 2.683628318584071, "grad_norm": 0.15633976459503174, "learning_rate": 3.351808267196721e-07, "loss": 0.3668, "step": 1213 }, { "epoch": 2.685840707964602, "grad_norm": 0.14381296932697296, "learning_rate": 3.3056154616598777e-07, "loss": 0.3581, "step": 1214 }, { "epoch": 2.688053097345133, "grad_norm": 0.14406733214855194, "learning_rate": 3.2597322871687766e-07, "loss": 0.3543, "step": 1215 }, { "epoch": 2.6902654867256635, "grad_norm": 0.156460240483284, "learning_rate": 3.214159047975324e-07, "loss": 0.352, "step": 1216 }, { "epoch": 2.692477876106195, "grad_norm": 0.1374759078025818, "learning_rate": 3.1688960462762263e-07, "loss": 0.3601, "step": 1217 }, { "epoch": 2.6946902654867255, "grad_norm": 0.16220209002494812, "learning_rate": 3.1239435822110253e-07, "loss": 0.3614, "step": 1218 }, { "epoch": 2.6969026548672566, "grad_norm": 0.15912321209907532, "learning_rate": 3.079301953860059e-07, "loss": 0.3573, "step": 1219 }, { "epoch": 2.6991150442477876, "grad_norm": 0.16917406022548676, "learning_rate": 3.034971457242525e-07, "loss": 0.3601, "step": 1220 }, { "epoch": 2.7013274336283186, "grad_norm": 0.16305102407932281, "learning_rate": 2.990952386314505e-07, "loss": 0.3594, "step": 1221 }, { "epoch": 2.7035398230088497, "grad_norm": 0.1461646407842636, "learning_rate": 2.947245032967e-07, "loss": 0.3517, "step": 1222 }, { "epoch": 2.7057522123893807, "grad_norm": 0.16588905453681946, "learning_rate": 2.9038496870240187e-07, "loss": 0.3564, "step": 1223 }, { "epoch": 2.7079646017699117, "grad_norm": 0.14253613352775574, "learning_rate": 2.860766636240636e-07, "loss": 0.3633, "step": 1224 }, { "epoch": 2.7101769911504423, "grad_norm": 0.15777863562107086, "learning_rate": 2.817996166301107e-07, "loss": 0.3465, "step": 1225 }, { "epoch": 2.7123893805309733, "grad_norm": 0.1424233764410019, "learning_rate": 2.7755385608169374e-07, "loss": 0.3523, "step": 1226 }, { "epoch": 2.7146017699115044, "grad_norm": 0.1586133986711502, "learning_rate": 2.733394101325054e-07, "loss": 0.3579, "step": 1227 }, { "epoch": 2.7168141592920354, "grad_norm": 0.14557451009750366, "learning_rate": 2.691563067285874e-07, "loss": 0.3539, "step": 1228 }, { "epoch": 2.7190265486725664, "grad_norm": 0.15484808385372162, "learning_rate": 2.65004573608153e-07, "loss": 0.3577, "step": 1229 }, { "epoch": 2.7212389380530975, "grad_norm": 0.14703123271465302, "learning_rate": 2.608842383013943e-07, "loss": 0.3586, "step": 1230 }, { "epoch": 2.7234513274336285, "grad_norm": 0.1443336308002472, "learning_rate": 2.567953281303059e-07, "loss": 0.3571, "step": 1231 }, { "epoch": 2.725663716814159, "grad_norm": 0.14558246731758118, "learning_rate": 2.527378702085037e-07, "loss": 0.3584, "step": 1232 }, { "epoch": 2.72787610619469, "grad_norm": 0.14799098670482635, "learning_rate": 2.4871189144104025e-07, "loss": 0.3612, "step": 1233 }, { "epoch": 2.730088495575221, "grad_norm": 0.14273111522197723, "learning_rate": 2.447174185242324e-07, "loss": 0.3557, "step": 1234 }, { "epoch": 2.732300884955752, "grad_norm": 0.15998686850070953, "learning_rate": 2.407544779454779e-07, "loss": 0.3445, "step": 1235 }, { "epoch": 2.734513274336283, "grad_norm": 0.16166724264621735, "learning_rate": 2.368230959830875e-07, "loss": 0.3381, "step": 1236 }, { "epoch": 2.7367256637168142, "grad_norm": 0.15504410862922668, "learning_rate": 2.32923298706102e-07, "loss": 0.3548, "step": 1237 }, { "epoch": 2.7389380530973453, "grad_norm": 0.15463854372501373, "learning_rate": 2.2905511197412634e-07, "loss": 0.354, "step": 1238 }, { "epoch": 2.741150442477876, "grad_norm": 0.15159274637699127, "learning_rate": 2.2521856143715492e-07, "loss": 0.353, "step": 1239 }, { "epoch": 2.7433628318584073, "grad_norm": 0.16764990985393524, "learning_rate": 2.2141367253540325e-07, "loss": 0.3575, "step": 1240 }, { "epoch": 2.745575221238938, "grad_norm": 0.15683040022850037, "learning_rate": 2.1764047049913528e-07, "loss": 0.3505, "step": 1241 }, { "epoch": 2.747787610619469, "grad_norm": 0.15731599926948547, "learning_rate": 2.1389898034850086e-07, "loss": 0.3592, "step": 1242 }, { "epoch": 2.75, "grad_norm": 0.15384253859519958, "learning_rate": 2.1018922689336973e-07, "loss": 0.3599, "step": 1243 }, { "epoch": 2.752212389380531, "grad_norm": 0.1472090780735016, "learning_rate": 2.0651123473316103e-07, "loss": 0.358, "step": 1244 }, { "epoch": 2.754424778761062, "grad_norm": 0.144203320145607, "learning_rate": 2.0286502825668852e-07, "loss": 0.3519, "step": 1245 }, { "epoch": 2.7566371681415927, "grad_norm": 0.1510661542415619, "learning_rate": 1.992506316419912e-07, "loss": 0.3585, "step": 1246 }, { "epoch": 2.758849557522124, "grad_norm": 0.1556992530822754, "learning_rate": 1.9566806885617906e-07, "loss": 0.3556, "step": 1247 }, { "epoch": 2.7610619469026547, "grad_norm": 0.14370451867580414, "learning_rate": 1.921173636552698e-07, "loss": 0.3595, "step": 1248 }, { "epoch": 2.7632743362831858, "grad_norm": 0.14656773209571838, "learning_rate": 1.8859853958403507e-07, "loss": 0.3483, "step": 1249 }, { "epoch": 2.765486725663717, "grad_norm": 0.13567841053009033, "learning_rate": 1.8511161997584015e-07, "loss": 0.354, "step": 1250 }, { "epoch": 2.767699115044248, "grad_norm": 0.1580413579940796, "learning_rate": 1.8165662795249172e-07, "loss": 0.3644, "step": 1251 }, { "epoch": 2.769911504424779, "grad_norm": 0.14433561265468597, "learning_rate": 1.7823358642408694e-07, "loss": 0.3485, "step": 1252 }, { "epoch": 2.77212389380531, "grad_norm": 0.1432889699935913, "learning_rate": 1.748425180888541e-07, "loss": 0.3645, "step": 1253 }, { "epoch": 2.774336283185841, "grad_norm": 0.14327970147132874, "learning_rate": 1.7148344543301277e-07, "loss": 0.3495, "step": 1254 }, { "epoch": 2.7765486725663715, "grad_norm": 0.1495339274406433, "learning_rate": 1.681563907306133e-07, "loss": 0.3439, "step": 1255 }, { "epoch": 2.7787610619469025, "grad_norm": 0.14024530351161957, "learning_rate": 1.6486137604339813e-07, "loss": 0.3537, "step": 1256 }, { "epoch": 2.7809734513274336, "grad_norm": 0.16141167283058167, "learning_rate": 1.6159842322065022e-07, "loss": 0.3502, "step": 1257 }, { "epoch": 2.7831858407079646, "grad_norm": 0.14094969630241394, "learning_rate": 1.5836755389905035e-07, "loss": 0.3506, "step": 1258 }, { "epoch": 2.7853982300884956, "grad_norm": 0.15461289882659912, "learning_rate": 1.5516878950253333e-07, "loss": 0.3616, "step": 1259 }, { "epoch": 2.7876106194690267, "grad_norm": 0.14529578387737274, "learning_rate": 1.5200215124214491e-07, "loss": 0.3474, "step": 1260 }, { "epoch": 2.7898230088495577, "grad_norm": 0.13706517219543457, "learning_rate": 1.4886766011590449e-07, "loss": 0.3509, "step": 1261 }, { "epoch": 2.7920353982300883, "grad_norm": 0.15313966572284698, "learning_rate": 1.4576533690866035e-07, "loss": 0.3551, "step": 1262 }, { "epoch": 2.7942477876106193, "grad_norm": 0.1422608494758606, "learning_rate": 1.4269520219195753e-07, "loss": 0.3437, "step": 1263 }, { "epoch": 2.7964601769911503, "grad_norm": 0.13179507851600647, "learning_rate": 1.396572763238957e-07, "loss": 0.3562, "step": 1264 }, { "epoch": 2.7986725663716814, "grad_norm": 0.14244070649147034, "learning_rate": 1.3665157944900198e-07, "loss": 0.3641, "step": 1265 }, { "epoch": 2.8008849557522124, "grad_norm": 0.14658492803573608, "learning_rate": 1.3367813149808728e-07, "loss": 0.3567, "step": 1266 }, { "epoch": 2.8030973451327434, "grad_norm": 0.1432352066040039, "learning_rate": 1.3073695218812356e-07, "loss": 0.3483, "step": 1267 }, { "epoch": 2.8053097345132745, "grad_norm": 0.1332673579454422, "learning_rate": 1.278280610221072e-07, "loss": 0.3594, "step": 1268 }, { "epoch": 2.807522123893805, "grad_norm": 0.15009014308452606, "learning_rate": 1.2495147728893043e-07, "loss": 0.3592, "step": 1269 }, { "epoch": 2.8097345132743365, "grad_norm": 0.13184407353401184, "learning_rate": 1.2210722006325782e-07, "loss": 0.3542, "step": 1270 }, { "epoch": 2.811946902654867, "grad_norm": 0.13673901557922363, "learning_rate": 1.192953082053927e-07, "loss": 0.3622, "step": 1271 }, { "epoch": 2.814159292035398, "grad_norm": 0.1729208379983902, "learning_rate": 1.1651576036115942e-07, "loss": 0.3544, "step": 1272 }, { "epoch": 2.816371681415929, "grad_norm": 0.139160618185997, "learning_rate": 1.1376859496177228e-07, "loss": 0.3499, "step": 1273 }, { "epoch": 2.8185840707964602, "grad_norm": 0.15454445779323578, "learning_rate": 1.110538302237213e-07, "loss": 0.3572, "step": 1274 }, { "epoch": 2.8207964601769913, "grad_norm": 0.1475721150636673, "learning_rate": 1.0837148414864329e-07, "loss": 0.3636, "step": 1275 }, { "epoch": 2.823008849557522, "grad_norm": 0.1556055098772049, "learning_rate": 1.0572157452321097e-07, "loss": 0.348, "step": 1276 }, { "epoch": 2.8252212389380533, "grad_norm": 0.14078259468078613, "learning_rate": 1.0310411891900629e-07, "loss": 0.3546, "step": 1277 }, { "epoch": 2.827433628318584, "grad_norm": 0.15832266211509705, "learning_rate": 1.0051913469241003e-07, "loss": 0.3639, "step": 1278 }, { "epoch": 2.829646017699115, "grad_norm": 0.16918309032917023, "learning_rate": 9.796663898448578e-08, "loss": 0.3579, "step": 1279 }, { "epoch": 2.831858407079646, "grad_norm": 0.13976359367370605, "learning_rate": 9.544664872086329e-08, "loss": 0.3632, "step": 1280 }, { "epoch": 2.834070796460177, "grad_norm": 0.14802858233451843, "learning_rate": 9.295918061163034e-08, "loss": 0.3548, "step": 1281 }, { "epoch": 2.836283185840708, "grad_norm": 0.15616321563720703, "learning_rate": 9.050425115121775e-08, "loss": 0.3521, "step": 1282 }, { "epoch": 2.838495575221239, "grad_norm": 0.13561472296714783, "learning_rate": 8.808187661829504e-08, "loss": 0.3601, "step": 1283 }, { "epoch": 2.84070796460177, "grad_norm": 0.16056296229362488, "learning_rate": 8.569207307565664e-08, "loss": 0.3596, "step": 1284 }, { "epoch": 2.8429203539823007, "grad_norm": 0.1590995341539383, "learning_rate": 8.333485637012029e-08, "loss": 0.3573, "step": 1285 }, { "epoch": 2.8451327433628317, "grad_norm": 0.15869586169719696, "learning_rate": 8.101024213241826e-08, "loss": 0.3526, "step": 1286 }, { "epoch": 2.8473451327433628, "grad_norm": 0.15596012771129608, "learning_rate": 7.871824577709797e-08, "loss": 0.3504, "step": 1287 }, { "epoch": 2.849557522123894, "grad_norm": 0.1455894559621811, "learning_rate": 7.645888250241485e-08, "loss": 0.3554, "step": 1288 }, { "epoch": 2.851769911504425, "grad_norm": 0.1496601700782776, "learning_rate": 7.423216729023574e-08, "loss": 0.3666, "step": 1289 }, { "epoch": 2.853982300884956, "grad_norm": 0.13408933579921722, "learning_rate": 7.203811490593626e-08, "loss": 0.3636, "step": 1290 }, { "epoch": 2.856194690265487, "grad_norm": 0.1369679570198059, "learning_rate": 6.987673989830523e-08, "loss": 0.3587, "step": 1291 }, { "epoch": 2.8584070796460175, "grad_norm": 0.15344880521297455, "learning_rate": 6.774805659944761e-08, "loss": 0.3522, "step": 1292 }, { "epoch": 2.8606194690265485, "grad_norm": 0.1609773337841034, "learning_rate": 6.565207912468785e-08, "loss": 0.3535, "step": 1293 }, { "epoch": 2.8628318584070795, "grad_norm": 0.14126433432102203, "learning_rate": 6.358882137248001e-08, "loss": 0.3452, "step": 1294 }, { "epoch": 2.8650442477876106, "grad_norm": 0.13827714323997498, "learning_rate": 6.15582970243117e-08, "loss": 0.3488, "step": 1295 }, { "epoch": 2.8672566371681416, "grad_norm": 0.1625031977891922, "learning_rate": 5.9560519544614725e-08, "loss": 0.3541, "step": 1296 }, { "epoch": 2.8694690265486726, "grad_norm": 0.14330527186393738, "learning_rate": 5.759550218067622e-08, "loss": 0.3535, "step": 1297 }, { "epoch": 2.8716814159292037, "grad_norm": 0.16162008047103882, "learning_rate": 5.566325796255101e-08, "loss": 0.3562, "step": 1298 }, { "epoch": 2.8738938053097343, "grad_norm": 0.1462399661540985, "learning_rate": 5.3763799702975516e-08, "loss": 0.3531, "step": 1299 }, { "epoch": 2.8761061946902657, "grad_norm": 0.14997854828834534, "learning_rate": 5.1897139997280074e-08, "loss": 0.3561, "step": 1300 }, { "epoch": 2.8783185840707963, "grad_norm": 0.1586892157793045, "learning_rate": 5.0063291223308993e-08, "loss": 0.3636, "step": 1301 }, { "epoch": 2.8805309734513274, "grad_norm": 0.1484476923942566, "learning_rate": 4.826226554133562e-08, "loss": 0.3637, "step": 1302 }, { "epoch": 2.8827433628318584, "grad_norm": 0.1511233150959015, "learning_rate": 4.649407489398461e-08, "loss": 0.3618, "step": 1303 }, { "epoch": 2.8849557522123894, "grad_norm": 0.14198242127895355, "learning_rate": 4.4758731006149804e-08, "loss": 0.3595, "step": 1304 }, { "epoch": 2.8871681415929205, "grad_norm": 0.15620240569114685, "learning_rate": 4.305624538491815e-08, "loss": 0.3599, "step": 1305 }, { "epoch": 2.8893805309734515, "grad_norm": 0.15184152126312256, "learning_rate": 4.138662931949255e-08, "loss": 0.3552, "step": 1306 }, { "epoch": 2.8915929203539825, "grad_norm": 0.14311784505844116, "learning_rate": 3.974989388111861e-08, "loss": 0.3568, "step": 1307 }, { "epoch": 2.893805309734513, "grad_norm": 0.14970344305038452, "learning_rate": 3.814604992300908e-08, "loss": 0.3564, "step": 1308 }, { "epoch": 2.896017699115044, "grad_norm": 0.13278964161872864, "learning_rate": 3.657510808027343e-08, "loss": 0.362, "step": 1309 }, { "epoch": 2.898230088495575, "grad_norm": 0.15549008548259735, "learning_rate": 3.503707876984674e-08, "loss": 0.353, "step": 1310 }, { "epoch": 2.900442477876106, "grad_norm": 0.15713003277778625, "learning_rate": 3.353197219041981e-08, "loss": 0.3599, "step": 1311 }, { "epoch": 2.9026548672566372, "grad_norm": 0.17100383341312408, "learning_rate": 3.205979832237416e-08, "loss": 0.36, "step": 1312 }, { "epoch": 2.9048672566371683, "grad_norm": 0.15952634811401367, "learning_rate": 3.062056692771154e-08, "loss": 0.3562, "step": 1313 }, { "epoch": 2.9070796460176993, "grad_norm": 0.13878712058067322, "learning_rate": 2.9214287549995114e-08, "loss": 0.3486, "step": 1314 }, { "epoch": 2.90929203539823, "grad_norm": 0.14025112986564636, "learning_rate": 2.7840969514279503e-08, "loss": 0.3565, "step": 1315 }, { "epoch": 2.911504424778761, "grad_norm": 0.15756577253341675, "learning_rate": 2.6500621927054716e-08, "loss": 0.3475, "step": 1316 }, { "epoch": 2.913716814159292, "grad_norm": 0.14622363448143005, "learning_rate": 2.519325367618175e-08, "loss": 0.3631, "step": 1317 }, { "epoch": 2.915929203539823, "grad_norm": 0.15228000283241272, "learning_rate": 2.3918873430835986e-08, "loss": 0.3489, "step": 1318 }, { "epoch": 2.918141592920354, "grad_norm": 0.1473086029291153, "learning_rate": 2.2677489641448335e-08, "loss": 0.3459, "step": 1319 }, { "epoch": 2.920353982300885, "grad_norm": 0.13773617148399353, "learning_rate": 2.1469110539650283e-08, "loss": 0.3674, "step": 1320 }, { "epoch": 2.922566371681416, "grad_norm": 0.15642863512039185, "learning_rate": 2.0293744138219495e-08, "loss": 0.3493, "step": 1321 }, { "epoch": 2.9247787610619467, "grad_norm": 0.1433519721031189, "learning_rate": 1.9151398231024297e-08, "loss": 0.3517, "step": 1322 }, { "epoch": 2.926991150442478, "grad_norm": 0.14097705483436584, "learning_rate": 1.8042080392974837e-08, "loss": 0.3654, "step": 1323 }, { "epoch": 2.9292035398230087, "grad_norm": 0.15193448960781097, "learning_rate": 1.6965797979971442e-08, "loss": 0.3556, "step": 1324 }, { "epoch": 2.9314159292035398, "grad_norm": 0.14747443795204163, "learning_rate": 1.5922558128856903e-08, "loss": 0.3521, "step": 1325 }, { "epoch": 2.933628318584071, "grad_norm": 0.16591234505176544, "learning_rate": 1.4912367757366485e-08, "loss": 0.3628, "step": 1326 }, { "epoch": 2.935840707964602, "grad_norm": 0.15506120026111603, "learning_rate": 1.3935233564086326e-08, "loss": 0.3529, "step": 1327 }, { "epoch": 2.938053097345133, "grad_norm": 0.16547641158103943, "learning_rate": 1.2991162028405113e-08, "loss": 0.3603, "step": 1328 }, { "epoch": 2.9402654867256635, "grad_norm": 0.17579355835914612, "learning_rate": 1.2080159410471914e-08, "loss": 0.3652, "step": 1329 }, { "epoch": 2.942477876106195, "grad_norm": 0.13282181322574615, "learning_rate": 1.1202231751157866e-08, "loss": 0.3529, "step": 1330 }, { "epoch": 2.9446902654867255, "grad_norm": 0.14515455067157745, "learning_rate": 1.0357384872011767e-08, "loss": 0.3637, "step": 1331 }, { "epoch": 2.9469026548672566, "grad_norm": 0.1463669091463089, "learning_rate": 9.545624375223439e-09, "loss": 0.3551, "step": 1332 }, { "epoch": 2.9491150442477876, "grad_norm": 0.1455363929271698, "learning_rate": 8.766955643587094e-09, "loss": 0.3561, "step": 1333 }, { "epoch": 2.9513274336283186, "grad_norm": 0.15577441453933716, "learning_rate": 8.021383840465247e-09, "loss": 0.3589, "step": 1334 }, { "epoch": 2.9535398230088497, "grad_norm": 0.14496655762195587, "learning_rate": 7.308913909752635e-09, "loss": 0.3571, "step": 1335 }, { "epoch": 2.9557522123893807, "grad_norm": 0.15788689255714417, "learning_rate": 6.629550575847355e-09, "loss": 0.3575, "step": 1336 }, { "epoch": 2.9579646017699117, "grad_norm": 0.1579669564962387, "learning_rate": 5.983298343615884e-09, "loss": 0.3494, "step": 1337 }, { "epoch": 2.9601769911504423, "grad_norm": 0.15941593050956726, "learning_rate": 5.3701614983647786e-09, "loss": 0.3584, "step": 1338 }, { "epoch": 2.9623893805309733, "grad_norm": 0.16959086060523987, "learning_rate": 4.7901441058118006e-09, "loss": 0.3517, "step": 1339 }, { "epoch": 2.9646017699115044, "grad_norm": 0.16718046367168427, "learning_rate": 4.243250012059275e-09, "loss": 0.3584, "step": 1340 }, { "epoch": 2.9668141592920354, "grad_norm": 0.1667267084121704, "learning_rate": 3.729482843569665e-09, "loss": 0.3652, "step": 1341 }, { "epoch": 2.9690265486725664, "grad_norm": 0.15754228830337524, "learning_rate": 3.2488460071389285e-09, "loss": 0.3578, "step": 1342 }, { "epoch": 2.9712389380530975, "grad_norm": 0.15564315021038055, "learning_rate": 2.801342689875974e-09, "loss": 0.3509, "step": 1343 }, { "epoch": 2.9734513274336285, "grad_norm": 0.16392084956169128, "learning_rate": 2.3869758591810177e-09, "loss": 0.3574, "step": 1344 }, { "epoch": 2.975663716814159, "grad_norm": 0.147927388548851, "learning_rate": 2.005748262725593e-09, "loss": 0.3557, "step": 1345 }, { "epoch": 2.97787610619469, "grad_norm": 0.14830511808395386, "learning_rate": 1.657662428434792e-09, "loss": 0.3524, "step": 1346 }, { "epoch": 2.980088495575221, "grad_norm": 0.15716874599456787, "learning_rate": 1.342720664469499e-09, "loss": 0.3549, "step": 1347 }, { "epoch": 2.982300884955752, "grad_norm": 0.17521563172340393, "learning_rate": 1.0609250592130693e-09, "loss": 0.3629, "step": 1348 }, { "epoch": 2.984513274336283, "grad_norm": 0.15542881190776825, "learning_rate": 8.122774812552303e-10, "loss": 0.3501, "step": 1349 }, { "epoch": 2.9867256637168142, "grad_norm": 0.20576362311840057, "learning_rate": 5.967795793820896e-10, "loss": 0.3627, "step": 1350 }, { "epoch": 2.9889380530973453, "grad_norm": 0.16061358153820038, "learning_rate": 4.1443278256170227e-10, "loss": 0.3524, "step": 1351 }, { "epoch": 2.991150442477876, "grad_norm": 0.1754293441772461, "learning_rate": 2.6523829993907456e-10, "loss": 0.3595, "step": 1352 }, { "epoch": 2.9933628318584073, "grad_norm": 0.17124155163764954, "learning_rate": 1.4919712082339665e-10, "loss": 0.3549, "step": 1353 }, { "epoch": 2.995575221238938, "grad_norm": 0.14956091344356537, "learning_rate": 6.63100146863771e-11, "loss": 0.353, "step": 1354 }, { "epoch": 2.997787610619469, "grad_norm": 0.1708720326423645, "learning_rate": 1.6577531152806027e-11, "loss": 0.3521, "step": 1355 }, { "epoch": 3.0, "grad_norm": 0.17063909769058228, "learning_rate": 0.0, "loss": 0.3461, "step": 1356 }, { "epoch": 3.0, "step": 1356, "total_flos": 2.751475469831373e+16, "train_loss": 0.20257837947315172, "train_runtime": 61703.9023, "train_samples_per_second": 8.428, "train_steps_per_second": 0.022 } ], "logging_steps": 1, "max_steps": 1356, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.751475469831373e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }