{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.986425339366516, "eval_steps": 500, "global_step": 550, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00904977375565611, "grad_norm": 6.353778491475079, "learning_rate": 1.4545454545454546e-06, "loss": 0.8197, "step": 1 }, { "epoch": 0.01809954751131222, "grad_norm": 6.38646932272591, "learning_rate": 2.9090909090909093e-06, "loss": 0.8224, "step": 2 }, { "epoch": 0.027149321266968326, "grad_norm": 6.053910996051301, "learning_rate": 4.363636363636364e-06, "loss": 0.8034, "step": 3 }, { "epoch": 0.03619909502262444, "grad_norm": 4.366009246926635, "learning_rate": 5.8181818181818185e-06, "loss": 0.774, "step": 4 }, { "epoch": 0.04524886877828054, "grad_norm": 2.274884898169855, "learning_rate": 7.272727272727273e-06, "loss": 0.7292, "step": 5 }, { "epoch": 0.05429864253393665, "grad_norm": 1.9029591404903814, "learning_rate": 8.727272727272728e-06, "loss": 0.7138, "step": 6 }, { "epoch": 0.06334841628959276, "grad_norm": 4.535776001491079, "learning_rate": 1.0181818181818182e-05, "loss": 0.7317, "step": 7 }, { "epoch": 0.07239819004524888, "grad_norm": 5.273781078723551, "learning_rate": 1.1636363636363637e-05, "loss": 0.7227, "step": 8 }, { "epoch": 0.08144796380090498, "grad_norm": 6.613983727436227, "learning_rate": 1.3090909090909092e-05, "loss": 0.7225, "step": 9 }, { "epoch": 0.09049773755656108, "grad_norm": 5.322076805028839, "learning_rate": 1.4545454545454546e-05, "loss": 0.7032, "step": 10 }, { "epoch": 0.09954751131221719, "grad_norm": 2.6848510839250563, "learning_rate": 1.6000000000000003e-05, "loss": 0.6662, "step": 11 }, { "epoch": 0.1085972850678733, "grad_norm": 2.501883063773779, "learning_rate": 1.7454545454545456e-05, "loss": 0.6361, "step": 12 }, { "epoch": 0.11764705882352941, "grad_norm": 2.4408014518675407, "learning_rate": 1.8909090909090912e-05, "loss": 0.6328, "step": 13 }, { "epoch": 0.12669683257918551, "grad_norm": 1.3581412540415263, "learning_rate": 2.0363636363636365e-05, "loss": 0.6038, "step": 14 }, { "epoch": 0.13574660633484162, "grad_norm": 1.3171074276840706, "learning_rate": 2.1818181818181818e-05, "loss": 0.5949, "step": 15 }, { "epoch": 0.14479638009049775, "grad_norm": 1.1346220985487154, "learning_rate": 2.3272727272727274e-05, "loss": 0.5812, "step": 16 }, { "epoch": 0.15384615384615385, "grad_norm": 0.9805713190517853, "learning_rate": 2.4727272727272727e-05, "loss": 0.5777, "step": 17 }, { "epoch": 0.16289592760180996, "grad_norm": 0.981975629498607, "learning_rate": 2.6181818181818183e-05, "loss": 0.5671, "step": 18 }, { "epoch": 0.17194570135746606, "grad_norm": 0.8633430106624941, "learning_rate": 2.763636363636364e-05, "loss": 0.5617, "step": 19 }, { "epoch": 0.18099547511312217, "grad_norm": 0.9927559005340078, "learning_rate": 2.9090909090909093e-05, "loss": 0.5547, "step": 20 }, { "epoch": 0.19004524886877827, "grad_norm": 0.7908715300138798, "learning_rate": 3.054545454545455e-05, "loss": 0.5392, "step": 21 }, { "epoch": 0.19909502262443438, "grad_norm": 1.077249549339691, "learning_rate": 3.2000000000000005e-05, "loss": 0.5373, "step": 22 }, { "epoch": 0.2081447963800905, "grad_norm": 1.3602867857746606, "learning_rate": 3.345454545454546e-05, "loss": 0.534, "step": 23 }, { "epoch": 0.2171945701357466, "grad_norm": 0.687608913929421, "learning_rate": 3.490909090909091e-05, "loss": 0.5286, "step": 24 }, { "epoch": 0.22624434389140272, "grad_norm": 1.2223237721495865, "learning_rate": 3.6363636363636364e-05, "loss": 0.5301, "step": 25 }, { "epoch": 0.23529411764705882, "grad_norm": 0.8580084887459775, "learning_rate": 3.7818181818181824e-05, "loss": 0.5214, "step": 26 }, { "epoch": 0.24434389140271492, "grad_norm": 1.1071155887269823, "learning_rate": 3.927272727272728e-05, "loss": 0.5115, "step": 27 }, { "epoch": 0.25339366515837103, "grad_norm": 0.7195369247916135, "learning_rate": 4.072727272727273e-05, "loss": 0.5095, "step": 28 }, { "epoch": 0.26244343891402716, "grad_norm": 1.5961441486786672, "learning_rate": 4.218181818181818e-05, "loss": 0.5087, "step": 29 }, { "epoch": 0.27149321266968324, "grad_norm": 0.827814164011704, "learning_rate": 4.3636363636363636e-05, "loss": 0.5111, "step": 30 }, { "epoch": 0.28054298642533937, "grad_norm": 1.1252915824734908, "learning_rate": 4.509090909090909e-05, "loss": 0.5072, "step": 31 }, { "epoch": 0.2895927601809955, "grad_norm": 1.1219477837842433, "learning_rate": 4.654545454545455e-05, "loss": 0.5046, "step": 32 }, { "epoch": 0.2986425339366516, "grad_norm": 1.505289650732032, "learning_rate": 4.8e-05, "loss": 0.503, "step": 33 }, { "epoch": 0.3076923076923077, "grad_norm": 0.8491008498263081, "learning_rate": 4.9454545454545454e-05, "loss": 0.5011, "step": 34 }, { "epoch": 0.3167420814479638, "grad_norm": 1.4306190204877898, "learning_rate": 5.0909090909090914e-05, "loss": 0.4895, "step": 35 }, { "epoch": 0.3257918552036199, "grad_norm": 1.0851621930114523, "learning_rate": 5.236363636363637e-05, "loss": 0.4956, "step": 36 }, { "epoch": 0.334841628959276, "grad_norm": 1.206194741973144, "learning_rate": 5.381818181818182e-05, "loss": 0.4887, "step": 37 }, { "epoch": 0.3438914027149321, "grad_norm": 1.4311749390727002, "learning_rate": 5.527272727272728e-05, "loss": 0.487, "step": 38 }, { "epoch": 0.35294117647058826, "grad_norm": 0.738265754189619, "learning_rate": 5.672727272727273e-05, "loss": 0.4854, "step": 39 }, { "epoch": 0.36199095022624433, "grad_norm": 1.6935271715239177, "learning_rate": 5.8181818181818185e-05, "loss": 0.4943, "step": 40 }, { "epoch": 0.37104072398190047, "grad_norm": 0.8407873740347103, "learning_rate": 5.9636363636363645e-05, "loss": 0.4846, "step": 41 }, { "epoch": 0.38009049773755654, "grad_norm": 1.168749658810784, "learning_rate": 6.10909090909091e-05, "loss": 0.4765, "step": 42 }, { "epoch": 0.3891402714932127, "grad_norm": 0.9365669815926128, "learning_rate": 6.254545454545456e-05, "loss": 0.4865, "step": 43 }, { "epoch": 0.39819004524886875, "grad_norm": 1.481731917951899, "learning_rate": 6.400000000000001e-05, "loss": 0.4858, "step": 44 }, { "epoch": 0.4072398190045249, "grad_norm": 1.52181493703478, "learning_rate": 6.545454545454546e-05, "loss": 0.4949, "step": 45 }, { "epoch": 0.416289592760181, "grad_norm": 1.449133028804705, "learning_rate": 6.690909090909092e-05, "loss": 0.4799, "step": 46 }, { "epoch": 0.4253393665158371, "grad_norm": 1.4529455783010232, "learning_rate": 6.836363636363637e-05, "loss": 0.4779, "step": 47 }, { "epoch": 0.4343891402714932, "grad_norm": 1.3548957795298309, "learning_rate": 6.981818181818182e-05, "loss": 0.4776, "step": 48 }, { "epoch": 0.4434389140271493, "grad_norm": 1.3963891315975814, "learning_rate": 7.127272727272728e-05, "loss": 0.4731, "step": 49 }, { "epoch": 0.45248868778280543, "grad_norm": 1.0657728232300334, "learning_rate": 7.272727272727273e-05, "loss": 0.4702, "step": 50 }, { "epoch": 0.46153846153846156, "grad_norm": 1.2659845812441222, "learning_rate": 7.418181818181818e-05, "loss": 0.4853, "step": 51 }, { "epoch": 0.47058823529411764, "grad_norm": 1.0364505292451895, "learning_rate": 7.563636363636365e-05, "loss": 0.4663, "step": 52 }, { "epoch": 0.4796380090497738, "grad_norm": 1.188728427839461, "learning_rate": 7.70909090909091e-05, "loss": 0.4819, "step": 53 }, { "epoch": 0.48868778280542985, "grad_norm": 1.3147320829735498, "learning_rate": 7.854545454545455e-05, "loss": 0.4765, "step": 54 }, { "epoch": 0.497737556561086, "grad_norm": 1.1857487638939694, "learning_rate": 8e-05, "loss": 0.4764, "step": 55 }, { "epoch": 0.5067873303167421, "grad_norm": 1.4561796125999034, "learning_rate": 7.999919440291627e-05, "loss": 0.4853, "step": 56 }, { "epoch": 0.5158371040723982, "grad_norm": 1.3437982531985815, "learning_rate": 7.999677764411438e-05, "loss": 0.4766, "step": 57 }, { "epoch": 0.5248868778280543, "grad_norm": 0.7865846274351967, "learning_rate": 7.999274982094104e-05, "loss": 0.4697, "step": 58 }, { "epoch": 0.5339366515837104, "grad_norm": 1.7177406402851725, "learning_rate": 7.998711109563637e-05, "loss": 0.4753, "step": 59 }, { "epoch": 0.5429864253393665, "grad_norm": 0.765099828673138, "learning_rate": 7.997986169532741e-05, "loss": 0.4646, "step": 60 }, { "epoch": 0.5520361990950227, "grad_norm": 1.291319953299644, "learning_rate": 7.997100191201896e-05, "loss": 0.4719, "step": 61 }, { "epoch": 0.5610859728506787, "grad_norm": 1.0231054776331459, "learning_rate": 7.996053210258176e-05, "loss": 0.4597, "step": 62 }, { "epoch": 0.5701357466063348, "grad_norm": 1.020678257497354, "learning_rate": 7.994845268873825e-05, "loss": 0.4631, "step": 63 }, { "epoch": 0.579185520361991, "grad_norm": 1.0920667405954658, "learning_rate": 7.993476415704543e-05, "loss": 0.4557, "step": 64 }, { "epoch": 0.5882352941176471, "grad_norm": 1.1128002216774169, "learning_rate": 7.991946705887539e-05, "loss": 0.4601, "step": 65 }, { "epoch": 0.5972850678733032, "grad_norm": 1.0096600697154985, "learning_rate": 7.990256201039297e-05, "loss": 0.4616, "step": 66 }, { "epoch": 0.6063348416289592, "grad_norm": 1.0630268901035373, "learning_rate": 7.98840496925311e-05, "loss": 0.4536, "step": 67 }, { "epoch": 0.6153846153846154, "grad_norm": 0.6564221253831782, "learning_rate": 7.986393085096324e-05, "loss": 0.4483, "step": 68 }, { "epoch": 0.6244343891402715, "grad_norm": 0.6260742925219016, "learning_rate": 7.984220629607336e-05, "loss": 0.4508, "step": 69 }, { "epoch": 0.6334841628959276, "grad_norm": 0.8273325354500644, "learning_rate": 7.981887690292339e-05, "loss": 0.4494, "step": 70 }, { "epoch": 0.6425339366515838, "grad_norm": 0.884333850023049, "learning_rate": 7.979394361121789e-05, "loss": 0.4519, "step": 71 }, { "epoch": 0.6515837104072398, "grad_norm": 1.0182608804991282, "learning_rate": 7.97674074252662e-05, "loss": 0.4555, "step": 72 }, { "epoch": 0.6606334841628959, "grad_norm": 0.8109090124549581, "learning_rate": 7.9739269413942e-05, "loss": 0.4521, "step": 73 }, { "epoch": 0.669683257918552, "grad_norm": 0.9807135848274302, "learning_rate": 7.970953071064036e-05, "loss": 0.4531, "step": 74 }, { "epoch": 0.6787330316742082, "grad_norm": 1.504171669224647, "learning_rate": 7.967819251323182e-05, "loss": 0.4705, "step": 75 }, { "epoch": 0.6877828054298643, "grad_norm": 0.5907918802273645, "learning_rate": 7.964525608401445e-05, "loss": 0.4488, "step": 76 }, { "epoch": 0.6968325791855203, "grad_norm": 1.198484065683406, "learning_rate": 7.961072274966282e-05, "loss": 0.4551, "step": 77 }, { "epoch": 0.7058823529411765, "grad_norm": 0.7873070678087877, "learning_rate": 7.957459390117458e-05, "loss": 0.4457, "step": 78 }, { "epoch": 0.7149321266968326, "grad_norm": 0.8039969398610661, "learning_rate": 7.95368709938145e-05, "loss": 0.454, "step": 79 }, { "epoch": 0.7239819004524887, "grad_norm": 0.7842737782633313, "learning_rate": 7.949755554705577e-05, "loss": 0.441, "step": 80 }, { "epoch": 0.7330316742081447, "grad_norm": 0.8912309653170191, "learning_rate": 7.945664914451888e-05, "loss": 0.4442, "step": 81 }, { "epoch": 0.7420814479638009, "grad_norm": 0.7795917187202375, "learning_rate": 7.941415343390773e-05, "loss": 0.4406, "step": 82 }, { "epoch": 0.751131221719457, "grad_norm": 0.6205090396508431, "learning_rate": 7.937007012694335e-05, "loss": 0.4507, "step": 83 }, { "epoch": 0.7601809954751131, "grad_norm": 0.8937106249538119, "learning_rate": 7.932440099929493e-05, "loss": 0.4532, "step": 84 }, { "epoch": 0.7692307692307693, "grad_norm": 0.6165097160255976, "learning_rate": 7.927714789050826e-05, "loss": 0.4454, "step": 85 }, { "epoch": 0.7782805429864253, "grad_norm": 0.6003524566440082, "learning_rate": 7.92283127039317e-05, "loss": 0.4367, "step": 86 }, { "epoch": 0.7873303167420814, "grad_norm": 0.6906315651004887, "learning_rate": 7.917789740663941e-05, "loss": 0.4332, "step": 87 }, { "epoch": 0.7963800904977375, "grad_norm": 0.515950260791789, "learning_rate": 7.912590402935223e-05, "loss": 0.4345, "step": 88 }, { "epoch": 0.8054298642533937, "grad_norm": 0.5111081573242718, "learning_rate": 7.907233466635582e-05, "loss": 0.4419, "step": 89 }, { "epoch": 0.8144796380090498, "grad_norm": 0.4275091010122135, "learning_rate": 7.90171914754163e-05, "loss": 0.4383, "step": 90 }, { "epoch": 0.8235294117647058, "grad_norm": 0.5029385491118522, "learning_rate": 7.896047667769335e-05, "loss": 0.4348, "step": 91 }, { "epoch": 0.832579185520362, "grad_norm": 0.4078317539372912, "learning_rate": 7.890219255765077e-05, "loss": 0.4293, "step": 92 }, { "epoch": 0.8416289592760181, "grad_norm": 0.4768996453352078, "learning_rate": 7.884234146296442e-05, "loss": 0.4375, "step": 93 }, { "epoch": 0.8506787330316742, "grad_norm": 0.5642712829841973, "learning_rate": 7.878092580442766e-05, "loss": 0.4301, "step": 94 }, { "epoch": 0.8597285067873304, "grad_norm": 0.7926376614480052, "learning_rate": 7.871794805585427e-05, "loss": 0.4377, "step": 95 }, { "epoch": 0.8687782805429864, "grad_norm": 0.8704449632643672, "learning_rate": 7.865341075397874e-05, "loss": 0.4303, "step": 96 }, { "epoch": 0.8778280542986425, "grad_norm": 0.8638439306395513, "learning_rate": 7.858731649835424e-05, "loss": 0.4321, "step": 97 }, { "epoch": 0.8868778280542986, "grad_norm": 1.1122142269322102, "learning_rate": 7.85196679512477e-05, "loss": 0.4369, "step": 98 }, { "epoch": 0.8959276018099548, "grad_norm": 0.8695186582366218, "learning_rate": 7.845046783753276e-05, "loss": 0.4229, "step": 99 }, { "epoch": 0.9049773755656109, "grad_norm": 0.6907557363110941, "learning_rate": 7.837971894457991e-05, "loss": 0.4294, "step": 100 }, { "epoch": 0.9140271493212669, "grad_norm": 0.5776183515910319, "learning_rate": 7.830742412214422e-05, "loss": 0.4276, "step": 101 }, { "epoch": 0.9230769230769231, "grad_norm": 0.45053560734923015, "learning_rate": 7.82335862822506e-05, "loss": 0.4284, "step": 102 }, { "epoch": 0.9321266968325792, "grad_norm": 0.643150601551687, "learning_rate": 7.815820839907651e-05, "loss": 0.4237, "step": 103 }, { "epoch": 0.9411764705882353, "grad_norm": 0.6867059068503916, "learning_rate": 7.808129350883207e-05, "loss": 0.4325, "step": 104 }, { "epoch": 0.9502262443438914, "grad_norm": 0.47092962014394424, "learning_rate": 7.800284470963783e-05, "loss": 0.4282, "step": 105 }, { "epoch": 0.9592760180995475, "grad_norm": 0.5794346362766767, "learning_rate": 7.792286516139999e-05, "loss": 0.426, "step": 106 }, { "epoch": 0.9683257918552036, "grad_norm": 0.5899947636404669, "learning_rate": 7.784135808568308e-05, "loss": 0.4241, "step": 107 }, { "epoch": 0.9773755656108597, "grad_norm": 0.5788528664751798, "learning_rate": 7.775832676558028e-05, "loss": 0.4356, "step": 108 }, { "epoch": 0.9864253393665159, "grad_norm": 0.8755965755346112, "learning_rate": 7.7673774545581e-05, "loss": 0.4366, "step": 109 }, { "epoch": 0.995475113122172, "grad_norm": 0.9177669118629138, "learning_rate": 7.758770483143634e-05, "loss": 0.4329, "step": 110 }, { "epoch": 1.006787330316742, "grad_norm": 0.6446624913715124, "learning_rate": 7.750012109002185e-05, "loss": 0.4137, "step": 111 }, { "epoch": 1.0158371040723981, "grad_norm": 0.5702180405824077, "learning_rate": 7.741102684919787e-05, "loss": 0.414, "step": 112 }, { "epoch": 1.0248868778280542, "grad_norm": 0.604164406050923, "learning_rate": 7.732042569766741e-05, "loss": 0.4059, "step": 113 }, { "epoch": 1.0339366515837105, "grad_norm": 0.578300789859933, "learning_rate": 7.722832128483165e-05, "loss": 0.4072, "step": 114 }, { "epoch": 1.0429864253393666, "grad_norm": 0.7009255889818755, "learning_rate": 7.71347173206429e-05, "loss": 0.4044, "step": 115 }, { "epoch": 1.0520361990950227, "grad_norm": 0.709442615518018, "learning_rate": 7.703961757545522e-05, "loss": 0.4171, "step": 116 }, { "epoch": 1.0610859728506787, "grad_norm": 0.6827354691854699, "learning_rate": 7.694302587987245e-05, "loss": 0.4079, "step": 117 }, { "epoch": 1.0701357466063348, "grad_norm": 0.6818158818500779, "learning_rate": 7.6844946124594e-05, "loss": 0.4126, "step": 118 }, { "epoch": 1.079185520361991, "grad_norm": 0.863785651181626, "learning_rate": 7.674538226025815e-05, "loss": 0.4109, "step": 119 }, { "epoch": 1.088235294117647, "grad_norm": 0.5508478361873977, "learning_rate": 7.664433829728279e-05, "loss": 0.4028, "step": 120 }, { "epoch": 1.0972850678733033, "grad_norm": 0.5954544705810189, "learning_rate": 7.654181830570404e-05, "loss": 0.3969, "step": 121 }, { "epoch": 1.1063348416289593, "grad_norm": 0.5872119406822112, "learning_rate": 7.64378264150122e-05, "loss": 0.3993, "step": 122 }, { "epoch": 1.1153846153846154, "grad_norm": 0.5228235141898725, "learning_rate": 7.633236681398549e-05, "loss": 0.4063, "step": 123 }, { "epoch": 1.1244343891402715, "grad_norm": 0.6996331404201509, "learning_rate": 7.622544375052124e-05, "loss": 0.4076, "step": 124 }, { "epoch": 1.1334841628959276, "grad_norm": 0.5428162110150009, "learning_rate": 7.611706153146486e-05, "loss": 0.4016, "step": 125 }, { "epoch": 1.1425339366515836, "grad_norm": 0.4476576669392731, "learning_rate": 7.600722452243632e-05, "loss": 0.4018, "step": 126 }, { "epoch": 1.1515837104072397, "grad_norm": 0.47501951208322907, "learning_rate": 7.589593714765434e-05, "loss": 0.4003, "step": 127 }, { "epoch": 1.160633484162896, "grad_norm": 0.39864387167073856, "learning_rate": 7.578320388975816e-05, "loss": 0.4043, "step": 128 }, { "epoch": 1.169683257918552, "grad_norm": 0.4415298124574496, "learning_rate": 7.566902928962694e-05, "loss": 0.3996, "step": 129 }, { "epoch": 1.1787330316742082, "grad_norm": 0.3522729758159122, "learning_rate": 7.555341794619695e-05, "loss": 0.402, "step": 130 }, { "epoch": 1.1877828054298643, "grad_norm": 0.37499820674043755, "learning_rate": 7.543637451627623e-05, "loss": 0.3963, "step": 131 }, { "epoch": 1.1968325791855203, "grad_norm": 0.3140691864451784, "learning_rate": 7.531790371435709e-05, "loss": 0.3961, "step": 132 }, { "epoch": 1.2058823529411764, "grad_norm": 0.3383990510694389, "learning_rate": 7.519801031242613e-05, "loss": 0.4026, "step": 133 }, { "epoch": 1.2149321266968327, "grad_norm": 0.3704940782389753, "learning_rate": 7.507669913977213e-05, "loss": 0.3964, "step": 134 }, { "epoch": 1.2239819004524888, "grad_norm": 0.3963648582905255, "learning_rate": 7.49539750827914e-05, "loss": 0.3989, "step": 135 }, { "epoch": 1.2330316742081449, "grad_norm": 0.43661160954834166, "learning_rate": 7.482984308479109e-05, "loss": 0.3992, "step": 136 }, { "epoch": 1.242081447963801, "grad_norm": 0.4760701506797231, "learning_rate": 7.470430814578997e-05, "loss": 0.4038, "step": 137 }, { "epoch": 1.251131221719457, "grad_norm": 0.6080541303280323, "learning_rate": 7.457737532231708e-05, "loss": 0.4017, "step": 138 }, { "epoch": 1.260180995475113, "grad_norm": 0.8992194165468265, "learning_rate": 7.444904972720803e-05, "loss": 0.4021, "step": 139 }, { "epoch": 1.2692307692307692, "grad_norm": 0.9388790657173226, "learning_rate": 7.431933652939909e-05, "loss": 0.4012, "step": 140 }, { "epoch": 1.2782805429864252, "grad_norm": 0.7115639008601973, "learning_rate": 7.418824095371895e-05, "loss": 0.3981, "step": 141 }, { "epoch": 1.2873303167420813, "grad_norm": 0.482309574495901, "learning_rate": 7.405576828067829e-05, "loss": 0.4016, "step": 142 }, { "epoch": 1.2963800904977376, "grad_norm": 0.4429825222163161, "learning_rate": 7.392192384625704e-05, "loss": 0.3992, "step": 143 }, { "epoch": 1.3054298642533937, "grad_norm": 0.46955938360600935, "learning_rate": 7.378671304168955e-05, "loss": 0.3963, "step": 144 }, { "epoch": 1.3144796380090498, "grad_norm": 0.41143211547851066, "learning_rate": 7.365014131324725e-05, "loss": 0.3989, "step": 145 }, { "epoch": 1.3235294117647058, "grad_norm": 0.3655361428229988, "learning_rate": 7.35122141620195e-05, "loss": 0.3981, "step": 146 }, { "epoch": 1.332579185520362, "grad_norm": 0.3675624921847117, "learning_rate": 7.337293714369182e-05, "loss": 0.3888, "step": 147 }, { "epoch": 1.3416289592760182, "grad_norm": 0.38410932117144336, "learning_rate": 7.323231586832219e-05, "loss": 0.3983, "step": 148 }, { "epoch": 1.3506787330316743, "grad_norm": 0.35565241834369704, "learning_rate": 7.30903560001151e-05, "loss": 0.3927, "step": 149 }, { "epoch": 1.3597285067873304, "grad_norm": 0.3699524089900378, "learning_rate": 7.294706325719331e-05, "loss": 0.4009, "step": 150 }, { "epoch": 1.3687782805429864, "grad_norm": 0.37565752487407766, "learning_rate": 7.280244341136765e-05, "loss": 0.3968, "step": 151 }, { "epoch": 1.3778280542986425, "grad_norm": 0.41101581863985015, "learning_rate": 7.26565022879044e-05, "loss": 0.396, "step": 152 }, { "epoch": 1.3868778280542986, "grad_norm": 0.4755720566371037, "learning_rate": 7.250924576529072e-05, "loss": 0.3941, "step": 153 }, { "epoch": 1.3959276018099547, "grad_norm": 0.5478106344364355, "learning_rate": 7.236067977499791e-05, "loss": 0.396, "step": 154 }, { "epoch": 1.4049773755656108, "grad_norm": 0.6612146719144061, "learning_rate": 7.221081030124235e-05, "loss": 0.3927, "step": 155 }, { "epoch": 1.4140271493212668, "grad_norm": 0.7051474533138959, "learning_rate": 7.205964338074462e-05, "loss": 0.3991, "step": 156 }, { "epoch": 1.4230769230769231, "grad_norm": 0.6630123949633135, "learning_rate": 7.190718510248622e-05, "loss": 0.3969, "step": 157 }, { "epoch": 1.4321266968325792, "grad_norm": 0.5841525031103902, "learning_rate": 7.175344160746438e-05, "loss": 0.4009, "step": 158 }, { "epoch": 1.4411764705882353, "grad_norm": 0.5791330941522758, "learning_rate": 7.159841908844465e-05, "loss": 0.3904, "step": 159 }, { "epoch": 1.4502262443438914, "grad_norm": 0.6974262225802738, "learning_rate": 7.144212378971151e-05, "loss": 0.3965, "step": 160 }, { "epoch": 1.4592760180995474, "grad_norm": 0.7970932314626533, "learning_rate": 7.128456200681678e-05, "loss": 0.3913, "step": 161 }, { "epoch": 1.4683257918552037, "grad_norm": 0.7048456834043483, "learning_rate": 7.11257400863261e-05, "loss": 0.3901, "step": 162 }, { "epoch": 1.4773755656108598, "grad_norm": 0.4096523489923981, "learning_rate": 7.096566442556331e-05, "loss": 0.3937, "step": 163 }, { "epoch": 1.4864253393665159, "grad_norm": 0.3554945327024847, "learning_rate": 7.080434147235263e-05, "loss": 0.3892, "step": 164 }, { "epoch": 1.495475113122172, "grad_norm": 0.5448243159096916, "learning_rate": 7.064177772475912e-05, "loss": 0.3985, "step": 165 }, { "epoch": 1.504524886877828, "grad_norm": 0.4087602063081316, "learning_rate": 7.047797973082684e-05, "loss": 0.3953, "step": 166 }, { "epoch": 1.5135746606334841, "grad_norm": 0.27977995475090656, "learning_rate": 7.031295408831508e-05, "loss": 0.3859, "step": 167 }, { "epoch": 1.5226244343891402, "grad_norm": 0.4128081941990591, "learning_rate": 7.014670744443267e-05, "loss": 0.3993, "step": 168 }, { "epoch": 1.5316742081447963, "grad_norm": 0.3728102691478845, "learning_rate": 6.997924649557017e-05, "loss": 0.3924, "step": 169 }, { "epoch": 1.5407239819004523, "grad_norm": 0.2814741945781907, "learning_rate": 6.98105779870302e-05, "loss": 0.396, "step": 170 }, { "epoch": 1.5497737556561086, "grad_norm": 0.3864339523003812, "learning_rate": 6.964070871275567e-05, "loss": 0.3958, "step": 171 }, { "epoch": 1.5588235294117647, "grad_norm": 0.32053481390391325, "learning_rate": 6.94696455150562e-05, "loss": 0.3927, "step": 172 }, { "epoch": 1.5678733031674208, "grad_norm": 0.25239085998303645, "learning_rate": 6.929739528433244e-05, "loss": 0.3931, "step": 173 }, { "epoch": 1.5769230769230769, "grad_norm": 0.3797864909903561, "learning_rate": 6.912396495879857e-05, "loss": 0.3947, "step": 174 }, { "epoch": 1.5859728506787332, "grad_norm": 0.3778896067077122, "learning_rate": 6.89493615242028e-05, "loss": 0.3958, "step": 175 }, { "epoch": 1.5950226244343892, "grad_norm": 0.34652871604736857, "learning_rate": 6.877359201354606e-05, "loss": 0.3875, "step": 176 }, { "epoch": 1.6040723981900453, "grad_norm": 0.42674803575084275, "learning_rate": 6.859666350679854e-05, "loss": 0.3909, "step": 177 }, { "epoch": 1.6131221719457014, "grad_norm": 0.29475704922050416, "learning_rate": 6.841858313061477e-05, "loss": 0.3861, "step": 178 }, { "epoch": 1.6221719457013575, "grad_norm": 0.3605164686190752, "learning_rate": 6.823935805804626e-05, "loss": 0.3946, "step": 179 }, { "epoch": 1.6312217194570136, "grad_norm": 0.3793403076875403, "learning_rate": 6.805899550825285e-05, "loss": 0.3849, "step": 180 }, { "epoch": 1.6402714932126696, "grad_norm": 0.38647041207728616, "learning_rate": 6.787750274621175e-05, "loss": 0.3841, "step": 181 }, { "epoch": 1.6493212669683257, "grad_norm": 0.41289104580852704, "learning_rate": 6.769488708242492e-05, "loss": 0.3901, "step": 182 }, { "epoch": 1.6583710407239818, "grad_norm": 0.4086406307183557, "learning_rate": 6.751115587262469e-05, "loss": 0.3886, "step": 183 }, { "epoch": 1.6674208144796379, "grad_norm": 0.35780257992696674, "learning_rate": 6.732631651747739e-05, "loss": 0.3832, "step": 184 }, { "epoch": 1.6764705882352942, "grad_norm": 0.30719607252305003, "learning_rate": 6.714037646228529e-05, "loss": 0.3904, "step": 185 }, { "epoch": 1.6855203619909502, "grad_norm": 0.38674289134885453, "learning_rate": 6.695334319668672e-05, "loss": 0.3886, "step": 186 }, { "epoch": 1.6945701357466063, "grad_norm": 0.4005595923099328, "learning_rate": 6.676522425435433e-05, "loss": 0.3919, "step": 187 }, { "epoch": 1.7036199095022626, "grad_norm": 0.2349267690302485, "learning_rate": 6.65760272126917e-05, "loss": 0.3883, "step": 188 }, { "epoch": 1.7126696832579187, "grad_norm": 0.3574524568496641, "learning_rate": 6.638575969252806e-05, "loss": 0.3865, "step": 189 }, { "epoch": 1.7217194570135748, "grad_norm": 0.42670317548076125, "learning_rate": 6.619442935781141e-05, "loss": 0.3914, "step": 190 }, { "epoch": 1.7307692307692308, "grad_norm": 0.30707518042373566, "learning_rate": 6.600204391529971e-05, "loss": 0.3865, "step": 191 }, { "epoch": 1.739819004524887, "grad_norm": 0.2622083300455114, "learning_rate": 6.580861111425053e-05, "loss": 0.391, "step": 192 }, { "epoch": 1.748868778280543, "grad_norm": 0.38093733410787994, "learning_rate": 6.56141387461089e-05, "loss": 0.3872, "step": 193 }, { "epoch": 1.757918552036199, "grad_norm": 0.43742318682713477, "learning_rate": 6.541863464419346e-05, "loss": 0.3938, "step": 194 }, { "epoch": 1.7669683257918551, "grad_norm": 0.4418280959306971, "learning_rate": 6.52221066833809e-05, "loss": 0.3848, "step": 195 }, { "epoch": 1.7760180995475112, "grad_norm": 0.41049856835634324, "learning_rate": 6.502456277978887e-05, "loss": 0.3852, "step": 196 }, { "epoch": 1.7850678733031673, "grad_norm": 0.4150478314670342, "learning_rate": 6.482601089045696e-05, "loss": 0.3844, "step": 197 }, { "epoch": 1.7941176470588234, "grad_norm": 0.43894520954893845, "learning_rate": 6.462645901302633e-05, "loss": 0.3814, "step": 198 }, { "epoch": 1.8031674208144797, "grad_norm": 0.41995958530789795, "learning_rate": 6.442591518541753e-05, "loss": 0.381, "step": 199 }, { "epoch": 1.8122171945701357, "grad_norm": 0.3201288359968715, "learning_rate": 6.422438748550667e-05, "loss": 0.3852, "step": 200 }, { "epoch": 1.8212669683257918, "grad_norm": 0.2931969021729461, "learning_rate": 6.402188403080013e-05, "loss": 0.3878, "step": 201 }, { "epoch": 1.8303167420814481, "grad_norm": 0.4066483749649814, "learning_rate": 6.381841297810753e-05, "loss": 0.3834, "step": 202 }, { "epoch": 1.8393665158371042, "grad_norm": 0.40957398047436605, "learning_rate": 6.361398252321321e-05, "loss": 0.3886, "step": 203 }, { "epoch": 1.8484162895927603, "grad_norm": 0.32160755061403157, "learning_rate": 6.340860090054608e-05, "loss": 0.3841, "step": 204 }, { "epoch": 1.8574660633484164, "grad_norm": 0.23090971242916025, "learning_rate": 6.320227638284793e-05, "loss": 0.3842, "step": 205 }, { "epoch": 1.8665158371040724, "grad_norm": 0.32946206927022925, "learning_rate": 6.29950172808403e-05, "loss": 0.3901, "step": 206 }, { "epoch": 1.8755656108597285, "grad_norm": 0.3530001057183439, "learning_rate": 6.278683194288956e-05, "loss": 0.3852, "step": 207 }, { "epoch": 1.8846153846153846, "grad_norm": 0.28218082313802967, "learning_rate": 6.257772875467078e-05, "loss": 0.387, "step": 208 }, { "epoch": 1.8936651583710407, "grad_norm": 0.3059885996215407, "learning_rate": 6.236771613882987e-05, "loss": 0.3883, "step": 209 }, { "epoch": 1.9027149321266967, "grad_norm": 0.34171931460456756, "learning_rate": 6.215680255464442e-05, "loss": 0.3824, "step": 210 }, { "epoch": 1.9117647058823528, "grad_norm": 0.29374439911760764, "learning_rate": 6.194499649768281e-05, "loss": 0.3864, "step": 211 }, { "epoch": 1.9208144796380089, "grad_norm": 0.29144021228496453, "learning_rate": 6.173230649946213e-05, "loss": 0.3836, "step": 212 }, { "epoch": 1.9298642533936652, "grad_norm": 0.29148303948179005, "learning_rate": 6.15187411271045e-05, "loss": 0.3796, "step": 213 }, { "epoch": 1.9389140271493213, "grad_norm": 0.2754963794966634, "learning_rate": 6.130430898299199e-05, "loss": 0.3848, "step": 214 }, { "epoch": 1.9479638009049773, "grad_norm": 0.26672266015023416, "learning_rate": 6.10890187044201e-05, "loss": 0.3836, "step": 215 }, { "epoch": 1.9570135746606336, "grad_norm": 0.3413044280319467, "learning_rate": 6.087287896324984e-05, "loss": 0.385, "step": 216 }, { "epoch": 1.9660633484162897, "grad_norm": 0.4653279793260335, "learning_rate": 6.0655898465558484e-05, "loss": 0.3871, "step": 217 }, { "epoch": 1.9751131221719458, "grad_norm": 0.4607547435763325, "learning_rate": 6.043808595128883e-05, "loss": 0.3827, "step": 218 }, { "epoch": 1.9841628959276019, "grad_norm": 0.3114177233194163, "learning_rate": 6.021945019389719e-05, "loss": 0.391, "step": 219 }, { "epoch": 1.993212669683258, "grad_norm": 0.263615834735127, "learning_rate": 6.000000000000001e-05, "loss": 0.381, "step": 220 }, { "epoch": 2.004524886877828, "grad_norm": 0.3080883305611625, "learning_rate": 5.977974420901908e-05, "loss": 0.3732, "step": 221 }, { "epoch": 2.013574660633484, "grad_norm": 0.2983316325982498, "learning_rate": 5.955869169282556e-05, "loss": 0.3539, "step": 222 }, { "epoch": 2.02262443438914, "grad_norm": 0.35442755044439017, "learning_rate": 5.9336851355382557e-05, "loss": 0.3625, "step": 223 }, { "epoch": 2.0316742081447963, "grad_norm": 0.5038394603992613, "learning_rate": 5.911423213238653e-05, "loss": 0.3565, "step": 224 }, { "epoch": 2.0407239819004523, "grad_norm": 0.6565318273183188, "learning_rate": 5.889084299090732e-05, "loss": 0.3562, "step": 225 }, { "epoch": 2.0497737556561084, "grad_norm": 0.8028771505743133, "learning_rate": 5.866669292902695e-05, "loss": 0.36, "step": 226 }, { "epoch": 2.0588235294117645, "grad_norm": 0.8257109533011188, "learning_rate": 5.844179097547725e-05, "loss": 0.3602, "step": 227 }, { "epoch": 2.067873303167421, "grad_norm": 0.6350182381420444, "learning_rate": 5.821614618927613e-05, "loss": 0.3571, "step": 228 }, { "epoch": 2.076923076923077, "grad_norm": 0.38135287559376224, "learning_rate": 5.798976765936264e-05, "loss": 0.3559, "step": 229 }, { "epoch": 2.085972850678733, "grad_norm": 0.5026969066810225, "learning_rate": 5.776266450423097e-05, "loss": 0.355, "step": 230 }, { "epoch": 2.0950226244343892, "grad_norm": 0.6298538131496503, "learning_rate": 5.75348458715631e-05, "loss": 0.3595, "step": 231 }, { "epoch": 2.1040723981900453, "grad_norm": 0.3867446080865823, "learning_rate": 5.7306320937860336e-05, "loss": 0.352, "step": 232 }, { "epoch": 2.1131221719457014, "grad_norm": 0.3115968628085643, "learning_rate": 5.7077098908073676e-05, "loss": 0.3582, "step": 233 }, { "epoch": 2.1221719457013575, "grad_norm": 0.3939923894024485, "learning_rate": 5.684718901523307e-05, "loss": 0.3533, "step": 234 }, { "epoch": 2.1312217194570136, "grad_norm": 0.25112142036368945, "learning_rate": 5.661660052007547e-05, "loss": 0.3564, "step": 235 }, { "epoch": 2.1402714932126696, "grad_norm": 0.3261992149324723, "learning_rate": 5.6385342710671815e-05, "loss": 0.3518, "step": 236 }, { "epoch": 2.1493212669683257, "grad_norm": 0.32489012611444446, "learning_rate": 5.6153424902053e-05, "loss": 0.3535, "step": 237 }, { "epoch": 2.158371040723982, "grad_norm": 0.24162335661441722, "learning_rate": 5.59208564358345e-05, "loss": 0.3548, "step": 238 }, { "epoch": 2.167420814479638, "grad_norm": 0.3778776749277093, "learning_rate": 5.568764667984022e-05, "loss": 0.3605, "step": 239 }, { "epoch": 2.176470588235294, "grad_norm": 0.33672093596701086, "learning_rate": 5.5453805027725145e-05, "loss": 0.3538, "step": 240 }, { "epoch": 2.1855203619909505, "grad_norm": 0.24527198542362685, "learning_rate": 5.521934089859692e-05, "loss": 0.3497, "step": 241 }, { "epoch": 2.1945701357466065, "grad_norm": 0.3958849461770139, "learning_rate": 5.4984263736636494e-05, "loss": 0.355, "step": 242 }, { "epoch": 2.2036199095022626, "grad_norm": 0.38916418294477323, "learning_rate": 5.4748583010717636e-05, "loss": 0.3557, "step": 243 }, { "epoch": 2.2126696832579187, "grad_norm": 0.36390563994328023, "learning_rate": 5.451230821402564e-05, "loss": 0.362, "step": 244 }, { "epoch": 2.2217194570135748, "grad_norm": 0.3301056539987755, "learning_rate": 5.427544886367488e-05, "loss": 0.3519, "step": 245 }, { "epoch": 2.230769230769231, "grad_norm": 0.215969613599219, "learning_rate": 5.403801450032544e-05, "loss": 0.3551, "step": 246 }, { "epoch": 2.239819004524887, "grad_norm": 0.2378223811885189, "learning_rate": 5.380001468779883e-05, "loss": 0.359, "step": 247 }, { "epoch": 2.248868778280543, "grad_norm": 0.3256287018733439, "learning_rate": 5.356145901269282e-05, "loss": 0.3569, "step": 248 }, { "epoch": 2.257918552036199, "grad_norm": 0.2802481742472565, "learning_rate": 5.3322357083995235e-05, "loss": 0.3584, "step": 249 }, { "epoch": 2.266968325791855, "grad_norm": 0.30050063645860137, "learning_rate": 5.3082718532696874e-05, "loss": 0.3527, "step": 250 }, { "epoch": 2.276018099547511, "grad_norm": 0.337462061586617, "learning_rate": 5.284255301140364e-05, "loss": 0.3508, "step": 251 }, { "epoch": 2.2850678733031673, "grad_norm": 0.22314780810636892, "learning_rate": 5.2601870193947716e-05, "loss": 0.3486, "step": 252 }, { "epoch": 2.2941176470588234, "grad_norm": 0.2251099676952424, "learning_rate": 5.23606797749979e-05, "loss": 0.3521, "step": 253 }, { "epoch": 2.3031674208144794, "grad_norm": 0.2657181581176672, "learning_rate": 5.21189914696691e-05, "loss": 0.3506, "step": 254 }, { "epoch": 2.3122171945701355, "grad_norm": 0.23800570673769336, "learning_rate": 5.1876815013131e-05, "loss": 0.3511, "step": 255 }, { "epoch": 2.321266968325792, "grad_norm": 0.2673484662572871, "learning_rate": 5.163416016021597e-05, "loss": 0.3537, "step": 256 }, { "epoch": 2.330316742081448, "grad_norm": 0.25378700284589706, "learning_rate": 5.1391036685026093e-05, "loss": 0.3492, "step": 257 }, { "epoch": 2.339366515837104, "grad_norm": 0.2198551274983534, "learning_rate": 5.114745438053952e-05, "loss": 0.3522, "step": 258 }, { "epoch": 2.3484162895927603, "grad_norm": 0.22757974892064678, "learning_rate": 5.0903423058215925e-05, "loss": 0.3531, "step": 259 }, { "epoch": 2.3574660633484164, "grad_norm": 0.2162305337190899, "learning_rate": 5.06589525476014e-05, "loss": 0.3509, "step": 260 }, { "epoch": 2.3665158371040724, "grad_norm": 0.22176152775329794, "learning_rate": 5.0414052695932486e-05, "loss": 0.3539, "step": 261 }, { "epoch": 2.3755656108597285, "grad_norm": 0.20465996494402094, "learning_rate": 5.016873336773949e-05, "loss": 0.3513, "step": 262 }, { "epoch": 2.3846153846153846, "grad_norm": 0.19749781398286828, "learning_rate": 4.992300444444916e-05, "loss": 0.3569, "step": 263 }, { "epoch": 2.3936651583710407, "grad_norm": 0.21862557545293346, "learning_rate": 4.967687582398671e-05, "loss": 0.3532, "step": 264 }, { "epoch": 2.4027149321266967, "grad_norm": 0.22129803132569947, "learning_rate": 4.94303574203771e-05, "loss": 0.3546, "step": 265 }, { "epoch": 2.411764705882353, "grad_norm": 0.22177138969097251, "learning_rate": 4.9183459163345644e-05, "loss": 0.3502, "step": 266 }, { "epoch": 2.420814479638009, "grad_norm": 0.19327131140015494, "learning_rate": 4.893619099791817e-05, "loss": 0.3522, "step": 267 }, { "epoch": 2.4298642533936654, "grad_norm": 0.20052109770619267, "learning_rate": 4.868856288402032e-05, "loss": 0.356, "step": 268 }, { "epoch": 2.4389140271493215, "grad_norm": 0.19644855379050266, "learning_rate": 4.8440584796076395e-05, "loss": 0.3534, "step": 269 }, { "epoch": 2.4479638009049776, "grad_norm": 0.20192604708935719, "learning_rate": 4.819226672260763e-05, "loss": 0.3592, "step": 270 }, { "epoch": 2.4570135746606336, "grad_norm": 0.1922532533153157, "learning_rate": 4.794361866582982e-05, "loss": 0.3552, "step": 271 }, { "epoch": 2.4660633484162897, "grad_norm": 0.22962467692658226, "learning_rate": 4.7694650641250446e-05, "loss": 0.3531, "step": 272 }, { "epoch": 2.475113122171946, "grad_norm": 0.21255494327094565, "learning_rate": 4.7445372677265205e-05, "loss": 0.3554, "step": 273 }, { "epoch": 2.484162895927602, "grad_norm": 0.19556932541533853, "learning_rate": 4.719579481475416e-05, "loss": 0.3558, "step": 274 }, { "epoch": 2.493212669683258, "grad_norm": 0.2127502030224185, "learning_rate": 4.694592710667723e-05, "loss": 0.3558, "step": 275 }, { "epoch": 2.502262443438914, "grad_norm": 0.2051192723904285, "learning_rate": 4.6695779617669236e-05, "loss": 0.355, "step": 276 }, { "epoch": 2.51131221719457, "grad_norm": 0.21459679896962008, "learning_rate": 4.6445362423634574e-05, "loss": 0.3547, "step": 277 }, { "epoch": 2.520361990950226, "grad_norm": 0.21249030888822018, "learning_rate": 4.61946856113413e-05, "loss": 0.3524, "step": 278 }, { "epoch": 2.5294117647058822, "grad_norm": 0.18973130323510315, "learning_rate": 4.594375927801487e-05, "loss": 0.3518, "step": 279 }, { "epoch": 2.5384615384615383, "grad_norm": 0.20099116045455737, "learning_rate": 4.5692593530931416e-05, "loss": 0.3551, "step": 280 }, { "epoch": 2.5475113122171944, "grad_norm": 0.2404974178503931, "learning_rate": 4.5441198487010574e-05, "loss": 0.3546, "step": 281 }, { "epoch": 2.5565610859728505, "grad_norm": 0.21416667678544446, "learning_rate": 4.5189584272408074e-05, "loss": 0.3521, "step": 282 }, { "epoch": 2.5656108597285066, "grad_norm": 0.1821577098280573, "learning_rate": 4.493776102210779e-05, "loss": 0.351, "step": 283 }, { "epoch": 2.5746606334841626, "grad_norm": 0.1940123893538777, "learning_rate": 4.468573887951354e-05, "loss": 0.352, "step": 284 }, { "epoch": 2.583710407239819, "grad_norm": 0.21908628404417796, "learning_rate": 4.4433527996040443e-05, "loss": 0.3555, "step": 285 }, { "epoch": 2.5927601809954752, "grad_norm": 0.16351576977199667, "learning_rate": 4.418113853070614e-05, "loss": 0.3534, "step": 286 }, { "epoch": 2.6018099547511313, "grad_norm": 0.18499158266871718, "learning_rate": 4.392858064972149e-05, "loss": 0.3555, "step": 287 }, { "epoch": 2.6108597285067874, "grad_norm": 0.17222994187080978, "learning_rate": 4.3675864526081106e-05, "loss": 0.3522, "step": 288 }, { "epoch": 2.6199095022624435, "grad_norm": 0.17596242704110496, "learning_rate": 4.34230003391536e-05, "loss": 0.3529, "step": 289 }, { "epoch": 2.6289592760180995, "grad_norm": 0.1788336399501724, "learning_rate": 4.316999827427154e-05, "loss": 0.3501, "step": 290 }, { "epoch": 2.6380090497737556, "grad_norm": 0.15674058262172438, "learning_rate": 4.2916868522321235e-05, "loss": 0.3593, "step": 291 }, { "epoch": 2.6470588235294117, "grad_norm": 0.19461132000856615, "learning_rate": 4.266362127933216e-05, "loss": 0.3542, "step": 292 }, { "epoch": 2.6561085972850678, "grad_norm": 0.18265415777078592, "learning_rate": 4.2410266746066345e-05, "loss": 0.3529, "step": 293 }, { "epoch": 2.665158371040724, "grad_norm": 0.19366748651198787, "learning_rate": 4.215681512760744e-05, "loss": 0.3545, "step": 294 }, { "epoch": 2.6742081447963804, "grad_norm": 0.19701470985076372, "learning_rate": 4.19032766329497e-05, "loss": 0.3533, "step": 295 }, { "epoch": 2.6832579185520364, "grad_norm": 0.2051473001306171, "learning_rate": 4.1649661474586694e-05, "loss": 0.3543, "step": 296 }, { "epoch": 2.6923076923076925, "grad_norm": 0.18550660724506887, "learning_rate": 4.139597986810005e-05, "loss": 0.3569, "step": 297 }, { "epoch": 2.7013574660633486, "grad_norm": 0.20456848012158177, "learning_rate": 4.114224203174785e-05, "loss": 0.3496, "step": 298 }, { "epoch": 2.7104072398190047, "grad_norm": 0.1780622844651236, "learning_rate": 4.0888458186053184e-05, "loss": 0.354, "step": 299 }, { "epoch": 2.7194570135746607, "grad_norm": 0.1640758315975294, "learning_rate": 4.063463855339232e-05, "loss": 0.3496, "step": 300 }, { "epoch": 2.728506787330317, "grad_norm": 0.19228477606561048, "learning_rate": 4.0380793357583076e-05, "loss": 0.3491, "step": 301 }, { "epoch": 2.737556561085973, "grad_norm": 0.1755323101374011, "learning_rate": 4.012693282347289e-05, "loss": 0.3564, "step": 302 }, { "epoch": 2.746606334841629, "grad_norm": 0.1809635624979587, "learning_rate": 3.9873067176527114e-05, "loss": 0.3544, "step": 303 }, { "epoch": 2.755656108597285, "grad_norm": 0.18201509342269132, "learning_rate": 3.961920664241694e-05, "loss": 0.3546, "step": 304 }, { "epoch": 2.764705882352941, "grad_norm": 0.17873879422611214, "learning_rate": 3.9365361446607684e-05, "loss": 0.355, "step": 305 }, { "epoch": 2.773755656108597, "grad_norm": 0.21315929069191436, "learning_rate": 3.911154181394682e-05, "loss": 0.3564, "step": 306 }, { "epoch": 2.7828054298642533, "grad_norm": 0.19819053615429819, "learning_rate": 3.885775796825216e-05, "loss": 0.3548, "step": 307 }, { "epoch": 2.7918552036199094, "grad_norm": 0.13490710454922517, "learning_rate": 3.860402013189998e-05, "loss": 0.3485, "step": 308 }, { "epoch": 2.8009049773755654, "grad_norm": 0.20244714474335393, "learning_rate": 3.835033852541332e-05, "loss": 0.3562, "step": 309 }, { "epoch": 2.8099547511312215, "grad_norm": 0.1724672504964484, "learning_rate": 3.809672336705031e-05, "loss": 0.348, "step": 310 }, { "epoch": 2.8190045248868776, "grad_norm": 0.15081349665835933, "learning_rate": 3.784318487239257e-05, "loss": 0.3546, "step": 311 }, { "epoch": 2.8280542986425337, "grad_norm": 0.1783285611622551, "learning_rate": 3.758973325393367e-05, "loss": 0.35, "step": 312 }, { "epoch": 2.83710407239819, "grad_norm": 0.16120590818140387, "learning_rate": 3.7336378720667846e-05, "loss": 0.3537, "step": 313 }, { "epoch": 2.8461538461538463, "grad_norm": 0.20165982856627884, "learning_rate": 3.708313147767878e-05, "loss": 0.3543, "step": 314 }, { "epoch": 2.8552036199095023, "grad_norm": 0.21095988182947403, "learning_rate": 3.683000172572846e-05, "loss": 0.3544, "step": 315 }, { "epoch": 2.8642533936651584, "grad_norm": 0.18625234007151376, "learning_rate": 3.657699966084642e-05, "loss": 0.3531, "step": 316 }, { "epoch": 2.8733031674208145, "grad_norm": 0.22479750462747256, "learning_rate": 3.632413547391891e-05, "loss": 0.3556, "step": 317 }, { "epoch": 2.8823529411764706, "grad_norm": 0.20909188100700477, "learning_rate": 3.6071419350278515e-05, "loss": 0.3471, "step": 318 }, { "epoch": 2.8914027149321266, "grad_norm": 0.18778607505398776, "learning_rate": 3.581886146929387e-05, "loss": 0.3451, "step": 319 }, { "epoch": 2.9004524886877827, "grad_norm": 0.21561728314983256, "learning_rate": 3.556647200395956e-05, "loss": 0.352, "step": 320 }, { "epoch": 2.909502262443439, "grad_norm": 0.20718371330348473, "learning_rate": 3.5314261120486474e-05, "loss": 0.3493, "step": 321 }, { "epoch": 2.918552036199095, "grad_norm": 0.16443205391902374, "learning_rate": 3.5062238977892214e-05, "loss": 0.3508, "step": 322 }, { "epoch": 2.9276018099547514, "grad_norm": 0.19396119036177084, "learning_rate": 3.481041572759193e-05, "loss": 0.355, "step": 323 }, { "epoch": 2.9366515837104075, "grad_norm": 0.18132782101939904, "learning_rate": 3.4558801512989446e-05, "loss": 0.354, "step": 324 }, { "epoch": 2.9457013574660635, "grad_norm": 0.17332578387246808, "learning_rate": 3.4307406469068604e-05, "loss": 0.3542, "step": 325 }, { "epoch": 2.9547511312217196, "grad_norm": 0.15354450735940553, "learning_rate": 3.405624072198514e-05, "loss": 0.3553, "step": 326 }, { "epoch": 2.9638009049773757, "grad_norm": 0.17733336095930732, "learning_rate": 3.3805314388658714e-05, "loss": 0.3507, "step": 327 }, { "epoch": 2.9728506787330318, "grad_norm": 0.16922858743324629, "learning_rate": 3.3554637576365446e-05, "loss": 0.3469, "step": 328 }, { "epoch": 2.981900452488688, "grad_norm": 0.1635466514191799, "learning_rate": 3.330422038233078e-05, "loss": 0.3499, "step": 329 }, { "epoch": 2.990950226244344, "grad_norm": 0.181396866738228, "learning_rate": 3.305407289332279e-05, "loss": 0.3475, "step": 330 }, { "epoch": 3.002262443438914, "grad_norm": 0.17232098683504665, "learning_rate": 3.280420518524585e-05, "loss": 0.3429, "step": 331 }, { "epoch": 3.01131221719457, "grad_norm": 0.2085779039684738, "learning_rate": 3.25546273227348e-05, "loss": 0.3276, "step": 332 }, { "epoch": 3.020361990950226, "grad_norm": 0.20454372196591433, "learning_rate": 3.230534935874958e-05, "loss": 0.322, "step": 333 }, { "epoch": 3.0294117647058822, "grad_norm": 0.21400683996161396, "learning_rate": 3.205638133417019e-05, "loss": 0.3247, "step": 334 }, { "epoch": 3.0384615384615383, "grad_norm": 0.20994330105610337, "learning_rate": 3.180773327739238e-05, "loss": 0.328, "step": 335 }, { "epoch": 3.0475113122171944, "grad_norm": 0.2107968970213985, "learning_rate": 3.155941520392362e-05, "loss": 0.3235, "step": 336 }, { "epoch": 3.0565610859728505, "grad_norm": 0.20164240350868537, "learning_rate": 3.1311437115979696e-05, "loss": 0.3203, "step": 337 }, { "epoch": 3.065610859728507, "grad_norm": 0.20071440677649943, "learning_rate": 3.1063809002081834e-05, "loss": 0.3221, "step": 338 }, { "epoch": 3.074660633484163, "grad_norm": 0.17663930655160845, "learning_rate": 3.0816540836654356e-05, "loss": 0.3226, "step": 339 }, { "epoch": 3.083710407239819, "grad_norm": 0.18106709441769947, "learning_rate": 3.0569642579622905e-05, "loss": 0.3214, "step": 340 }, { "epoch": 3.0927601809954752, "grad_norm": 0.17780441082604515, "learning_rate": 3.0323124176013297e-05, "loss": 0.3225, "step": 341 }, { "epoch": 3.1018099547511313, "grad_norm": 0.17074344590150428, "learning_rate": 3.007699555555086e-05, "loss": 0.3183, "step": 342 }, { "epoch": 3.1108597285067874, "grad_norm": 0.22118329908847872, "learning_rate": 2.9831266632260534e-05, "loss": 0.3221, "step": 343 }, { "epoch": 3.1199095022624435, "grad_norm": 0.1862795261167787, "learning_rate": 2.958594730406752e-05, "loss": 0.3233, "step": 344 }, { "epoch": 3.1289592760180995, "grad_norm": 0.18624339502853474, "learning_rate": 2.9341047452398607e-05, "loss": 0.3232, "step": 345 }, { "epoch": 3.1380090497737556, "grad_norm": 0.19906920288843233, "learning_rate": 2.9096576941784095e-05, "loss": 0.3217, "step": 346 }, { "epoch": 3.1470588235294117, "grad_norm": 0.1759509265715728, "learning_rate": 2.8852545619460495e-05, "loss": 0.3294, "step": 347 }, { "epoch": 3.1561085972850678, "grad_norm": 0.18215420840842134, "learning_rate": 2.860896331497391e-05, "loss": 0.3178, "step": 348 }, { "epoch": 3.165158371040724, "grad_norm": 0.16928396196250736, "learning_rate": 2.8365839839784038e-05, "loss": 0.3208, "step": 349 }, { "epoch": 3.17420814479638, "grad_norm": 0.18758285635894378, "learning_rate": 2.8123184986869022e-05, "loss": 0.3252, "step": 350 }, { "epoch": 3.183257918552036, "grad_norm": 0.1727196365880379, "learning_rate": 2.7881008530330914e-05, "loss": 0.3219, "step": 351 }, { "epoch": 3.1923076923076925, "grad_norm": 0.19928771398379677, "learning_rate": 2.7639320225002108e-05, "loss": 0.3279, "step": 352 }, { "epoch": 3.2013574660633486, "grad_norm": 0.15934892115019736, "learning_rate": 2.7398129806052298e-05, "loss": 0.3211, "step": 353 }, { "epoch": 3.2104072398190047, "grad_norm": 0.1620987722032314, "learning_rate": 2.715744698859637e-05, "loss": 0.3264, "step": 354 }, { "epoch": 3.2194570135746607, "grad_norm": 0.14580042456343553, "learning_rate": 2.691728146730314e-05, "loss": 0.3247, "step": 355 }, { "epoch": 3.228506787330317, "grad_norm": 0.1669926797929679, "learning_rate": 2.6677642916004772e-05, "loss": 0.3219, "step": 356 }, { "epoch": 3.237556561085973, "grad_norm": 0.13912969057523117, "learning_rate": 2.6438540987307174e-05, "loss": 0.3224, "step": 357 }, { "epoch": 3.246606334841629, "grad_norm": 0.17393961594448798, "learning_rate": 2.6199985312201185e-05, "loss": 0.3287, "step": 358 }, { "epoch": 3.255656108597285, "grad_norm": 0.13336325385823727, "learning_rate": 2.5961985499674587e-05, "loss": 0.3195, "step": 359 }, { "epoch": 3.264705882352941, "grad_norm": 0.16625718933722608, "learning_rate": 2.5724551136325132e-05, "loss": 0.3155, "step": 360 }, { "epoch": 3.273755656108597, "grad_norm": 0.1313925578933839, "learning_rate": 2.5487691785974366e-05, "loss": 0.3245, "step": 361 }, { "epoch": 3.2828054298642533, "grad_norm": 0.1580151751002634, "learning_rate": 2.5251416989282377e-05, "loss": 0.3244, "step": 362 }, { "epoch": 3.2918552036199094, "grad_norm": 0.13919373737905993, "learning_rate": 2.501573626336352e-05, "loss": 0.3204, "step": 363 }, { "epoch": 3.3009049773755654, "grad_norm": 0.1394478161996878, "learning_rate": 2.478065910140308e-05, "loss": 0.319, "step": 364 }, { "epoch": 3.3099547511312215, "grad_norm": 0.1422678230130045, "learning_rate": 2.4546194972274852e-05, "loss": 0.3234, "step": 365 }, { "epoch": 3.3190045248868776, "grad_norm": 0.13767311772002824, "learning_rate": 2.431235332015978e-05, "loss": 0.3222, "step": 366 }, { "epoch": 3.328054298642534, "grad_norm": 0.15146521934073406, "learning_rate": 2.4079143564165524e-05, "loss": 0.3196, "step": 367 }, { "epoch": 3.33710407239819, "grad_norm": 0.14554574710226792, "learning_rate": 2.3846575097947015e-05, "loss": 0.3238, "step": 368 }, { "epoch": 3.3461538461538463, "grad_norm": 0.12428809362324093, "learning_rate": 2.361465728932819e-05, "loss": 0.3281, "step": 369 }, { "epoch": 3.3552036199095023, "grad_norm": 0.1359067551518604, "learning_rate": 2.338339947992455e-05, "loss": 0.3197, "step": 370 }, { "epoch": 3.3642533936651584, "grad_norm": 0.12263338859613652, "learning_rate": 2.315281098476694e-05, "loss": 0.3217, "step": 371 }, { "epoch": 3.3733031674208145, "grad_norm": 0.13047474507331128, "learning_rate": 2.2922901091926334e-05, "loss": 0.3244, "step": 372 }, { "epoch": 3.3823529411764706, "grad_norm": 0.12872903204410682, "learning_rate": 2.269367906213966e-05, "loss": 0.3229, "step": 373 }, { "epoch": 3.3914027149321266, "grad_norm": 0.11879478583728734, "learning_rate": 2.24651541284369e-05, "loss": 0.3246, "step": 374 }, { "epoch": 3.4004524886877827, "grad_norm": 0.13996450916284633, "learning_rate": 2.2237335495769035e-05, "loss": 0.3255, "step": 375 }, { "epoch": 3.409502262443439, "grad_norm": 0.12258279755344627, "learning_rate": 2.2010232340637375e-05, "loss": 0.3222, "step": 376 }, { "epoch": 3.418552036199095, "grad_norm": 0.13396529243580008, "learning_rate": 2.1783853810723895e-05, "loss": 0.3188, "step": 377 }, { "epoch": 3.427601809954751, "grad_norm": 0.12827246370155101, "learning_rate": 2.155820902452276e-05, "loss": 0.3233, "step": 378 }, { "epoch": 3.4366515837104075, "grad_norm": 0.13361073513914423, "learning_rate": 2.1333307070973054e-05, "loss": 0.32, "step": 379 }, { "epoch": 3.4457013574660635, "grad_norm": 0.1294537429532072, "learning_rate": 2.11091570090927e-05, "loss": 0.3234, "step": 380 }, { "epoch": 3.4547511312217196, "grad_norm": 0.12477628801726541, "learning_rate": 2.0885767867613485e-05, "loss": 0.3236, "step": 381 }, { "epoch": 3.4638009049773757, "grad_norm": 0.12556744064400813, "learning_rate": 2.0663148644617443e-05, "loss": 0.3193, "step": 382 }, { "epoch": 3.4728506787330318, "grad_norm": 0.12838442646777212, "learning_rate": 2.044130830717445e-05, "loss": 0.3197, "step": 383 }, { "epoch": 3.481900452488688, "grad_norm": 0.11666989370139778, "learning_rate": 2.0220255790980935e-05, "loss": 0.3222, "step": 384 }, { "epoch": 3.490950226244344, "grad_norm": 0.1201654171036639, "learning_rate": 2.0000000000000012e-05, "loss": 0.323, "step": 385 }, { "epoch": 3.5, "grad_norm": 0.12406834396457092, "learning_rate": 1.9780549806102827e-05, "loss": 0.3233, "step": 386 }, { "epoch": 3.509049773755656, "grad_norm": 0.13211561659340326, "learning_rate": 1.9561914048711182e-05, "loss": 0.3328, "step": 387 }, { "epoch": 3.518099547511312, "grad_norm": 0.11794332998469956, "learning_rate": 1.934410153444153e-05, "loss": 0.3206, "step": 388 }, { "epoch": 3.5271493212669682, "grad_norm": 0.11751517121074768, "learning_rate": 1.912712103675017e-05, "loss": 0.3234, "step": 389 }, { "epoch": 3.5361990950226243, "grad_norm": 0.11384664165774953, "learning_rate": 1.8910981295579903e-05, "loss": 0.326, "step": 390 }, { "epoch": 3.5452488687782804, "grad_norm": 0.11619434472524914, "learning_rate": 1.8695691017008005e-05, "loss": 0.328, "step": 391 }, { "epoch": 3.5542986425339365, "grad_norm": 0.11461255200579494, "learning_rate": 1.8481258872895503e-05, "loss": 0.3249, "step": 392 }, { "epoch": 3.5633484162895925, "grad_norm": 0.1269351876097148, "learning_rate": 1.826769350053789e-05, "loss": 0.325, "step": 393 }, { "epoch": 3.5723981900452486, "grad_norm": 0.1262020845365235, "learning_rate": 1.8055003502317213e-05, "loss": 0.3245, "step": 394 }, { "epoch": 3.581447963800905, "grad_norm": 0.1244584292444969, "learning_rate": 1.7843197445355593e-05, "loss": 0.3257, "step": 395 }, { "epoch": 3.590497737556561, "grad_norm": 0.121141018365639, "learning_rate": 1.7632283861170135e-05, "loss": 0.3267, "step": 396 }, { "epoch": 3.5995475113122173, "grad_norm": 0.1199129930925027, "learning_rate": 1.7422271245329244e-05, "loss": 0.3257, "step": 397 }, { "epoch": 3.6085972850678734, "grad_norm": 0.11856999524364402, "learning_rate": 1.721316805711045e-05, "loss": 0.3211, "step": 398 }, { "epoch": 3.6176470588235294, "grad_norm": 0.11909188120229579, "learning_rate": 1.7004982719159712e-05, "loss": 0.3201, "step": 399 }, { "epoch": 3.6266968325791855, "grad_norm": 0.11144474240232284, "learning_rate": 1.679772361715208e-05, "loss": 0.3223, "step": 400 }, { "epoch": 3.6357466063348416, "grad_norm": 0.12526137904233012, "learning_rate": 1.6591399099453952e-05, "loss": 0.3251, "step": 401 }, { "epoch": 3.6447963800904977, "grad_norm": 0.1129996909861758, "learning_rate": 1.6386017476786818e-05, "loss": 0.3264, "step": 402 }, { "epoch": 3.6538461538461537, "grad_norm": 0.11339940885061946, "learning_rate": 1.6181587021892484e-05, "loss": 0.3252, "step": 403 }, { "epoch": 3.66289592760181, "grad_norm": 0.11557131743659978, "learning_rate": 1.5978115969199882e-05, "loss": 0.3241, "step": 404 }, { "epoch": 3.6719457013574663, "grad_norm": 0.12094104342387753, "learning_rate": 1.5775612514493343e-05, "loss": 0.3224, "step": 405 }, { "epoch": 3.6809954751131224, "grad_norm": 0.12996591315042658, "learning_rate": 1.557408481458247e-05, "loss": 0.3214, "step": 406 }, { "epoch": 3.6900452488687785, "grad_norm": 0.10932053351798611, "learning_rate": 1.537354098697367e-05, "loss": 0.3243, "step": 407 }, { "epoch": 3.6990950226244346, "grad_norm": 0.119643634861931, "learning_rate": 1.5173989109543055e-05, "loss": 0.3262, "step": 408 }, { "epoch": 3.7081447963800906, "grad_norm": 0.11650012137085076, "learning_rate": 1.497543722021114e-05, "loss": 0.3248, "step": 409 }, { "epoch": 3.7171945701357467, "grad_norm": 0.11291091689614972, "learning_rate": 1.4777893316619114e-05, "loss": 0.3203, "step": 410 }, { "epoch": 3.726244343891403, "grad_norm": 0.1162086489522257, "learning_rate": 1.4581365355806555e-05, "loss": 0.3243, "step": 411 }, { "epoch": 3.735294117647059, "grad_norm": 0.10927589657003309, "learning_rate": 1.4385861253891111e-05, "loss": 0.327, "step": 412 }, { "epoch": 3.744343891402715, "grad_norm": 0.10801027792504748, "learning_rate": 1.4191388885749487e-05, "loss": 0.3237, "step": 413 }, { "epoch": 3.753393665158371, "grad_norm": 0.11625407101384182, "learning_rate": 1.3997956084700301e-05, "loss": 0.3223, "step": 414 }, { "epoch": 3.762443438914027, "grad_norm": 0.10784892069818107, "learning_rate": 1.3805570642188602e-05, "loss": 0.3258, "step": 415 }, { "epoch": 3.771493212669683, "grad_norm": 0.11300531775753117, "learning_rate": 1.3614240307471942e-05, "loss": 0.3248, "step": 416 }, { "epoch": 3.7805429864253393, "grad_norm": 0.10835333011179236, "learning_rate": 1.34239727873083e-05, "loss": 0.3223, "step": 417 }, { "epoch": 3.7895927601809953, "grad_norm": 0.11188577965419376, "learning_rate": 1.3234775745645684e-05, "loss": 0.321, "step": 418 }, { "epoch": 3.7986425339366514, "grad_norm": 0.12447909695863361, "learning_rate": 1.3046656803313287e-05, "loss": 0.324, "step": 419 }, { "epoch": 3.8076923076923075, "grad_norm": 0.10209002045981093, "learning_rate": 1.2859623537714719e-05, "loss": 0.3227, "step": 420 }, { "epoch": 3.8167420814479636, "grad_norm": 0.11319252778270614, "learning_rate": 1.2673683482522629e-05, "loss": 0.3289, "step": 421 }, { "epoch": 3.8257918552036196, "grad_norm": 0.10456931028449813, "learning_rate": 1.248884412737532e-05, "loss": 0.3236, "step": 422 }, { "epoch": 3.834841628959276, "grad_norm": 0.10976135305835821, "learning_rate": 1.2305112917575092e-05, "loss": 0.3273, "step": 423 }, { "epoch": 3.8438914027149322, "grad_norm": 0.10710488304052018, "learning_rate": 1.2122497253788267e-05, "loss": 0.3224, "step": 424 }, { "epoch": 3.8529411764705883, "grad_norm": 0.10332457574753726, "learning_rate": 1.1941004491747145e-05, "loss": 0.3174, "step": 425 }, { "epoch": 3.8619909502262444, "grad_norm": 0.10300034100616205, "learning_rate": 1.1760641941953744e-05, "loss": 0.3252, "step": 426 }, { "epoch": 3.8710407239819005, "grad_norm": 0.09955598289499694, "learning_rate": 1.1581416869385253e-05, "loss": 0.3223, "step": 427 }, { "epoch": 3.8800904977375565, "grad_norm": 0.10841097410633975, "learning_rate": 1.1403336493201462e-05, "loss": 0.3215, "step": 428 }, { "epoch": 3.8891402714932126, "grad_norm": 0.10174185620822972, "learning_rate": 1.1226407986453963e-05, "loss": 0.3238, "step": 429 }, { "epoch": 3.8981900452488687, "grad_norm": 0.10861737022921748, "learning_rate": 1.1050638475797193e-05, "loss": 0.3256, "step": 430 }, { "epoch": 3.9072398190045248, "grad_norm": 0.10297039570713754, "learning_rate": 1.0876035041201436e-05, "loss": 0.3221, "step": 431 }, { "epoch": 3.916289592760181, "grad_norm": 0.1076446648237558, "learning_rate": 1.070260471566757e-05, "loss": 0.3252, "step": 432 }, { "epoch": 3.9253393665158374, "grad_norm": 0.10925197562208173, "learning_rate": 1.0530354484943798e-05, "loss": 0.3185, "step": 433 }, { "epoch": 3.9343891402714934, "grad_norm": 0.09823521379039317, "learning_rate": 1.0359291287244334e-05, "loss": 0.3241, "step": 434 }, { "epoch": 3.9434389140271495, "grad_norm": 0.11042331281486452, "learning_rate": 1.0189422012969814e-05, "loss": 0.3264, "step": 435 }, { "epoch": 3.9524886877828056, "grad_norm": 0.0998774378258076, "learning_rate": 1.0020753504429845e-05, "loss": 0.3184, "step": 436 }, { "epoch": 3.9615384615384617, "grad_norm": 0.10587358010063602, "learning_rate": 9.85329255556735e-06, "loss": 0.3259, "step": 437 }, { "epoch": 3.9705882352941178, "grad_norm": 0.10052895526823628, "learning_rate": 9.687045911684928e-06, "loss": 0.3239, "step": 438 }, { "epoch": 3.979638009049774, "grad_norm": 0.09541866120111013, "learning_rate": 9.522020269173172e-06, "loss": 0.323, "step": 439 }, { "epoch": 3.98868778280543, "grad_norm": 0.09557886438970163, "learning_rate": 9.358222275240884e-06, "loss": 0.324, "step": 440 }, { "epoch": 3.997737556561086, "grad_norm": 0.10151624518760664, "learning_rate": 9.195658527647371e-06, "loss": 0.3264, "step": 441 }, { "epoch": 4.009049773755656, "grad_norm": 0.16524757813062563, "learning_rate": 9.034335574436701e-06, "loss": 0.3062, "step": 442 }, { "epoch": 4.018099547511312, "grad_norm": 0.12618495103542215, "learning_rate": 8.8742599136739e-06, "loss": 0.3037, "step": 443 }, { "epoch": 4.027149321266968, "grad_norm": 0.12794305399302355, "learning_rate": 8.715437993183235e-06, "loss": 0.3071, "step": 444 }, { "epoch": 4.036199095022624, "grad_norm": 0.1752311726452474, "learning_rate": 8.55787621028851e-06, "loss": 0.3049, "step": 445 }, { "epoch": 4.04524886877828, "grad_norm": 0.13744941353125903, "learning_rate": 8.401580911555353e-06, "loss": 0.3022, "step": 446 }, { "epoch": 4.0542986425339365, "grad_norm": 0.1442214999179444, "learning_rate": 8.24655839253563e-06, "loss": 0.304, "step": 447 }, { "epoch": 4.0633484162895925, "grad_norm": 0.14721326705532867, "learning_rate": 8.09281489751379e-06, "loss": 0.3014, "step": 448 }, { "epoch": 4.072398190045249, "grad_norm": 0.13164956472111614, "learning_rate": 7.940356619255385e-06, "loss": 0.3, "step": 449 }, { "epoch": 4.081447963800905, "grad_norm": 0.14645232900110186, "learning_rate": 7.789189698757656e-06, "loss": 0.3051, "step": 450 }, { "epoch": 4.090497737556561, "grad_norm": 0.13594041957864836, "learning_rate": 7.639320225002106e-06, "loss": 0.3008, "step": 451 }, { "epoch": 4.099547511312217, "grad_norm": 0.12959173947863253, "learning_rate": 7.490754234709285e-06, "loss": 0.3029, "step": 452 }, { "epoch": 4.108597285067873, "grad_norm": 0.1312102043451062, "learning_rate": 7.343497712095619e-06, "loss": 0.3018, "step": 453 }, { "epoch": 4.117647058823529, "grad_norm": 0.12721628165894824, "learning_rate": 7.1975565886323575e-06, "loss": 0.3021, "step": 454 }, { "epoch": 4.126696832579185, "grad_norm": 0.12135804195023794, "learning_rate": 7.052936742806693e-06, "loss": 0.3029, "step": 455 }, { "epoch": 4.135746606334842, "grad_norm": 0.11021717009086096, "learning_rate": 6.909643999884918e-06, "loss": 0.3036, "step": 456 }, { "epoch": 4.144796380090498, "grad_norm": 0.1171461454928136, "learning_rate": 6.767684131677814e-06, "loss": 0.3019, "step": 457 }, { "epoch": 4.153846153846154, "grad_norm": 0.12398145068200955, "learning_rate": 6.627062856308191e-06, "loss": 0.3057, "step": 458 }, { "epoch": 4.16289592760181, "grad_norm": 0.10813946756475247, "learning_rate": 6.487785837980509e-06, "loss": 0.3055, "step": 459 }, { "epoch": 4.171945701357466, "grad_norm": 0.10512317676255085, "learning_rate": 6.349858686752748e-06, "loss": 0.3028, "step": 460 }, { "epoch": 4.180995475113122, "grad_norm": 0.11457252381271353, "learning_rate": 6.213286958310476e-06, "loss": 0.3029, "step": 461 }, { "epoch": 4.1900452488687785, "grad_norm": 0.11257398302051942, "learning_rate": 6.078076153742962e-06, "loss": 0.3042, "step": 462 }, { "epoch": 4.199095022624435, "grad_norm": 0.1068866308205338, "learning_rate": 5.9442317193217245e-06, "loss": 0.3051, "step": 463 }, { "epoch": 4.208144796380091, "grad_norm": 0.09665417595653125, "learning_rate": 5.811759046281062e-06, "loss": 0.3044, "step": 464 }, { "epoch": 4.217194570135747, "grad_norm": 0.10132756213863214, "learning_rate": 5.680663470600918e-06, "loss": 0.3047, "step": 465 }, { "epoch": 4.226244343891403, "grad_norm": 0.10197900247318212, "learning_rate": 5.550950272791977e-06, "loss": 0.3036, "step": 466 }, { "epoch": 4.235294117647059, "grad_norm": 0.10011788827590404, "learning_rate": 5.422624677682935e-06, "loss": 0.3042, "step": 467 }, { "epoch": 4.244343891402715, "grad_norm": 0.09704681985529298, "learning_rate": 5.29569185421003e-06, "loss": 0.2997, "step": 468 }, { "epoch": 4.253393665158371, "grad_norm": 0.0889335395278227, "learning_rate": 5.1701569152089196e-06, "loss": 0.3033, "step": 469 }, { "epoch": 4.262443438914027, "grad_norm": 0.09701179356263233, "learning_rate": 5.046024917208603e-06, "loss": 0.3074, "step": 470 }, { "epoch": 4.271493212669683, "grad_norm": 0.09991090929407241, "learning_rate": 4.923300860227884e-06, "loss": 0.3074, "step": 471 }, { "epoch": 4.280542986425339, "grad_norm": 0.09090875982688397, "learning_rate": 4.801989687573878e-06, "loss": 0.3049, "step": 472 }, { "epoch": 4.289592760180995, "grad_norm": 0.08947832790676682, "learning_rate": 4.6820962856429205e-06, "loss": 0.3039, "step": 473 }, { "epoch": 4.298642533936651, "grad_norm": 0.08534002305718665, "learning_rate": 4.563625483723777e-06, "loss": 0.3079, "step": 474 }, { "epoch": 4.3076923076923075, "grad_norm": 0.08940524333139004, "learning_rate": 4.446582053803066e-06, "loss": 0.2974, "step": 475 }, { "epoch": 4.316742081447964, "grad_norm": 0.09171172616230228, "learning_rate": 4.330970710373063e-06, "loss": 0.3076, "step": 476 }, { "epoch": 4.32579185520362, "grad_norm": 0.08806621206621702, "learning_rate": 4.216796110241852e-06, "loss": 0.3048, "step": 477 }, { "epoch": 4.334841628959276, "grad_norm": 0.08409826711122455, "learning_rate": 4.104062852345671e-06, "loss": 0.3013, "step": 478 }, { "epoch": 4.343891402714932, "grad_norm": 0.0893222760374292, "learning_rate": 3.9927754775636886e-06, "loss": 0.3052, "step": 479 }, { "epoch": 4.352941176470588, "grad_norm": 0.08399072448662723, "learning_rate": 3.882938468535158e-06, "loss": 0.3024, "step": 480 }, { "epoch": 4.361990950226244, "grad_norm": 0.08499472493876709, "learning_rate": 3.7745562494787645e-06, "loss": 0.3022, "step": 481 }, { "epoch": 4.371040723981901, "grad_norm": 0.08145638475145033, "learning_rate": 3.667633186014512e-06, "loss": 0.3054, "step": 482 }, { "epoch": 4.380090497737557, "grad_norm": 0.08565742656060206, "learning_rate": 3.5621735849877957e-06, "loss": 0.3026, "step": 483 }, { "epoch": 4.389140271493213, "grad_norm": 0.08492763740421684, "learning_rate": 3.458181694295961e-06, "loss": 0.3055, "step": 484 }, { "epoch": 4.398190045248869, "grad_norm": 0.08463274662713963, "learning_rate": 3.3556617027172168e-06, "loss": 0.3024, "step": 485 }, { "epoch": 4.407239819004525, "grad_norm": 0.08462399748218868, "learning_rate": 3.2546177397418677e-06, "loss": 0.3066, "step": 486 }, { "epoch": 4.416289592760181, "grad_norm": 0.08376453881287438, "learning_rate": 3.155053875406e-06, "loss": 0.3042, "step": 487 }, { "epoch": 4.425339366515837, "grad_norm": 0.08830270926768068, "learning_rate": 3.0569741201275626e-06, "loss": 0.2987, "step": 488 }, { "epoch": 4.4343891402714934, "grad_norm": 0.08240870254024944, "learning_rate": 2.96038242454479e-06, "loss": 0.3003, "step": 489 }, { "epoch": 4.4434389140271495, "grad_norm": 0.09052455529057644, "learning_rate": 2.8652826793570975e-06, "loss": 0.3028, "step": 490 }, { "epoch": 4.452488687782806, "grad_norm": 0.08455563785960757, "learning_rate": 2.77167871516836e-06, "loss": 0.3073, "step": 491 }, { "epoch": 4.461538461538462, "grad_norm": 0.08127731938352123, "learning_rate": 2.679574302332597e-06, "loss": 0.2999, "step": 492 }, { "epoch": 4.470588235294118, "grad_norm": 0.08359245964188998, "learning_rate": 2.5889731508021363e-06, "loss": 0.3049, "step": 493 }, { "epoch": 4.479638009049774, "grad_norm": 0.08004277394855593, "learning_rate": 2.4998789099781507e-06, "loss": 0.2995, "step": 494 }, { "epoch": 4.48868778280543, "grad_norm": 0.09293016471118876, "learning_rate": 2.4122951685636674e-06, "loss": 0.3084, "step": 495 }, { "epoch": 4.497737556561086, "grad_norm": 0.08857129082558779, "learning_rate": 2.3262254544190154e-06, "loss": 0.3028, "step": 496 }, { "epoch": 4.506787330316742, "grad_norm": 0.0851690989816207, "learning_rate": 2.2416732344197368e-06, "loss": 0.3052, "step": 497 }, { "epoch": 4.515837104072398, "grad_norm": 0.08380605677642579, "learning_rate": 2.1586419143169125e-06, "loss": 0.3071, "step": 498 }, { "epoch": 4.524886877828054, "grad_norm": 0.07847574450950273, "learning_rate": 2.0771348386000233e-06, "loss": 0.3034, "step": 499 }, { "epoch": 4.53393665158371, "grad_norm": 0.08166491411892216, "learning_rate": 1.997155290362187e-06, "loss": 0.3065, "step": 500 }, { "epoch": 4.542986425339366, "grad_norm": 0.07778971565605453, "learning_rate": 1.9187064911679432e-06, "loss": 0.3024, "step": 501 }, { "epoch": 4.552036199095022, "grad_norm": 0.07933626982513363, "learning_rate": 1.841791600923495e-06, "loss": 0.3023, "step": 502 }, { "epoch": 4.5610859728506785, "grad_norm": 0.08136425063541126, "learning_rate": 1.766413717749389e-06, "loss": 0.304, "step": 503 }, { "epoch": 4.570135746606335, "grad_norm": 0.08138859986883137, "learning_rate": 1.6925758778557933e-06, "loss": 0.3055, "step": 504 }, { "epoch": 4.579185520361991, "grad_norm": 0.0787210528579053, "learning_rate": 1.6202810554201099e-06, "loss": 0.3049, "step": 505 }, { "epoch": 4.588235294117647, "grad_norm": 0.08051236562441969, "learning_rate": 1.5495321624672443e-06, "loss": 0.3044, "step": 506 }, { "epoch": 4.597285067873303, "grad_norm": 0.0788996281741959, "learning_rate": 1.4803320487523087e-06, "loss": 0.3057, "step": 507 }, { "epoch": 4.606334841628959, "grad_norm": 0.07764118519403024, "learning_rate": 1.4126835016457752e-06, "loss": 0.3038, "step": 508 }, { "epoch": 4.615384615384615, "grad_norm": 0.08001544469825624, "learning_rate": 1.3465892460212594e-06, "loss": 0.3064, "step": 509 }, { "epoch": 4.624434389140271, "grad_norm": 0.09063120326499433, "learning_rate": 1.2820519441457502e-06, "loss": 0.3047, "step": 510 }, { "epoch": 4.633484162895927, "grad_norm": 0.07794623861050551, "learning_rate": 1.2190741955723495e-06, "loss": 0.3045, "step": 511 }, { "epoch": 4.642533936651584, "grad_norm": 0.07595212127483163, "learning_rate": 1.1576585370355908e-06, "loss": 0.3065, "step": 512 }, { "epoch": 4.65158371040724, "grad_norm": 0.07739705488047151, "learning_rate": 1.0978074423492368e-06, "loss": 0.3014, "step": 513 }, { "epoch": 4.660633484162896, "grad_norm": 0.07685164772052698, "learning_rate": 1.0395233223066614e-06, "loss": 0.2996, "step": 514 }, { "epoch": 4.669683257918552, "grad_norm": 0.08015550295148605, "learning_rate": 9.828085245837183e-07, "loss": 0.307, "step": 515 }, { "epoch": 4.678733031674208, "grad_norm": 0.08113161848829428, "learning_rate": 9.276653336441943e-07, "loss": 0.3091, "step": 516 }, { "epoch": 4.6877828054298645, "grad_norm": 0.08234301018528185, "learning_rate": 8.740959706477725e-07, "loss": 0.3022, "step": 517 }, { "epoch": 4.6968325791855206, "grad_norm": 0.07811510319887344, "learning_rate": 8.221025933605919e-07, "loss": 0.3027, "step": 518 }, { "epoch": 4.705882352941177, "grad_norm": 0.07966253839252949, "learning_rate": 7.716872960683041e-07, "loss": 0.304, "step": 519 }, { "epoch": 4.714932126696833, "grad_norm": 0.07556300199073022, "learning_rate": 7.228521094917318e-07, "loss": 0.303, "step": 520 }, { "epoch": 4.723981900452489, "grad_norm": 0.07949886614317854, "learning_rate": 6.755990007050761e-07, "loss": 0.3061, "step": 521 }, { "epoch": 4.733031674208145, "grad_norm": 0.07841434742669152, "learning_rate": 6.299298730566516e-07, "loss": 0.3068, "step": 522 }, { "epoch": 4.742081447963801, "grad_norm": 0.07541915667043782, "learning_rate": 5.858465660922808e-07, "loss": 0.3066, "step": 523 }, { "epoch": 4.751131221719457, "grad_norm": 0.07733742513960733, "learning_rate": 5.433508554811307e-07, "loss": 0.3057, "step": 524 }, { "epoch": 4.760180995475113, "grad_norm": 0.07836214413259006, "learning_rate": 5.024444529442285e-07, "loss": 0.3031, "step": 525 }, { "epoch": 4.769230769230769, "grad_norm": 0.07953356517050592, "learning_rate": 4.6312900618550317e-07, "loss": 0.3016, "step": 526 }, { "epoch": 4.778280542986425, "grad_norm": 0.07617521054682652, "learning_rate": 4.254060988254205e-07, "loss": 0.2999, "step": 527 }, { "epoch": 4.787330316742081, "grad_norm": 0.07846308080577853, "learning_rate": 3.8927725033718553e-07, "loss": 0.3051, "step": 528 }, { "epoch": 4.796380090497737, "grad_norm": 0.07896885874732473, "learning_rate": 3.547439159855559e-07, "loss": 0.3063, "step": 529 }, { "epoch": 4.8054298642533935, "grad_norm": 0.07878929850813644, "learning_rate": 3.218074867681864e-07, "loss": 0.302, "step": 530 }, { "epoch": 4.8144796380090495, "grad_norm": 0.07613359977307133, "learning_rate": 2.9046928935966056e-07, "loss": 0.2972, "step": 531 }, { "epoch": 4.823529411764706, "grad_norm": 0.07688492198921053, "learning_rate": 2.607305860579912e-07, "loss": 0.3067, "step": 532 }, { "epoch": 4.832579185520362, "grad_norm": 0.07476518062895497, "learning_rate": 2.3259257473381647e-07, "loss": 0.2991, "step": 533 }, { "epoch": 4.841628959276018, "grad_norm": 0.07552831124187401, "learning_rate": 2.060563887821232e-07, "loss": 0.3005, "step": 534 }, { "epoch": 4.850678733031674, "grad_norm": 0.07631228481012646, "learning_rate": 1.8112309707661647e-07, "loss": 0.3035, "step": 535 }, { "epoch": 4.859728506787331, "grad_norm": 0.07467074977475399, "learning_rate": 1.577937039266475e-07, "loss": 0.3056, "step": 536 }, { "epoch": 4.868778280542987, "grad_norm": 0.07431003159382249, "learning_rate": 1.3606914903677492e-07, "loss": 0.2996, "step": 537 }, { "epoch": 4.877828054298643, "grad_norm": 0.07468768960052453, "learning_rate": 1.1595030746890168e-07, "loss": 0.3077, "step": 538 }, { "epoch": 4.886877828054299, "grad_norm": 0.07691379852821048, "learning_rate": 9.74379896070321e-08, "loss": 0.303, "step": 539 }, { "epoch": 4.895927601809955, "grad_norm": 0.07405617978413666, "learning_rate": 8.053294112462696e-08, "loss": 0.3029, "step": 540 }, { "epoch": 4.904977375565611, "grad_norm": 0.07237727103301038, "learning_rate": 6.523584295457408e-08, "loss": 0.3045, "step": 541 }, { "epoch": 4.914027149321267, "grad_norm": 0.07736995275320246, "learning_rate": 5.154731126176149e-08, "loss": 0.3072, "step": 542 }, { "epoch": 4.923076923076923, "grad_norm": 0.07450947266190626, "learning_rate": 3.946789741824386e-08, "loss": 0.2999, "step": 543 }, { "epoch": 4.932126696832579, "grad_norm": 0.07656920155960474, "learning_rate": 2.8998087981055854e-08, "loss": 0.3001, "step": 544 }, { "epoch": 4.9411764705882355, "grad_norm": 0.07427882247951784, "learning_rate": 2.01383046725967e-08, "loss": 0.3011, "step": 545 }, { "epoch": 4.950226244343892, "grad_norm": 0.07493198557289557, "learning_rate": 1.288890436363488e-08, "loss": 0.3073, "step": 546 }, { "epoch": 4.959276018099548, "grad_norm": 0.07595703866942234, "learning_rate": 7.250179058968521e-09, "loss": 0.302, "step": 547 }, { "epoch": 4.968325791855204, "grad_norm": 0.07667669947739308, "learning_rate": 3.222355885625916e-09, "loss": 0.3058, "step": 548 }, { "epoch": 4.97737556561086, "grad_norm": 0.0778877092524117, "learning_rate": 8.055970837395066e-10, "loss": 0.3071, "step": 549 }, { "epoch": 4.986425339366516, "grad_norm": 0.07598122336675614, "learning_rate": 0.0, "loss": 0.3021, "step": 550 }, { "epoch": 4.986425339366516, "step": 550, "total_flos": 1.3210208775492862e+19, "train_loss": 0.37584147680889474, "train_runtime": 27424.4359, "train_samples_per_second": 10.307, "train_steps_per_second": 0.02 } ], "logging_steps": 1, "max_steps": 550, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3210208775492862e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }