diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7917 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.9878183831672205, + "eval_steps": 500, + "global_step": 1125, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.004429678848283499, + "grad_norm": 0.0, + "learning_rate": 0.0, + "loss": 2.1161, + "step": 1 + }, + { + "epoch": 0.008859357696566999, + "grad_norm": 0.0, + "learning_rate": 0.0, + "loss": 2.6201, + "step": 2 + }, + { + "epoch": 0.013289036544850499, + "grad_norm": 0.0, + "learning_rate": 0.0, + "loss": 2.8945, + "step": 3 + }, + { + "epoch": 0.017718715393133997, + "grad_norm": 0.0, + "learning_rate": 0.0, + "loss": 2.1602, + "step": 4 + }, + { + "epoch": 0.0221483942414175, + "grad_norm": 10.605673789978027, + "learning_rate": 2.9411764705882355e-06, + "loss": 3.0746, + "step": 5 + }, + { + "epoch": 0.026578073089700997, + "grad_norm": 9.166430473327637, + "learning_rate": 5.882352941176471e-06, + "loss": 2.4977, + "step": 6 + }, + { + "epoch": 0.031007751937984496, + "grad_norm": 8.228297233581543, + "learning_rate": 8.823529411764707e-06, + "loss": 2.4886, + "step": 7 + }, + { + "epoch": 0.035437430786267994, + "grad_norm": 10.802596092224121, + "learning_rate": 1.1764705882352942e-05, + "loss": 3.153, + "step": 8 + }, + { + "epoch": 0.03986710963455149, + "grad_norm": 7.030856132507324, + "learning_rate": 1.4705882352941177e-05, + "loss": 2.3531, + "step": 9 + }, + { + "epoch": 0.044296788482835, + "grad_norm": 10.649585723876953, + "learning_rate": 1.7647058823529414e-05, + "loss": 2.3243, + "step": 10 + }, + { + "epoch": 0.048726467331118496, + "grad_norm": 7.4812493324279785, + "learning_rate": 2.058823529411765e-05, + "loss": 2.4188, + "step": 11 + }, + { + "epoch": 0.053156146179401995, + "grad_norm": 11.968286514282227, + "learning_rate": 2.3529411764705884e-05, + "loss": 2.5725, + "step": 12 + }, + { + "epoch": 0.05758582502768549, + "grad_norm": 5.769841194152832, + "learning_rate": 2.647058823529412e-05, + "loss": 2.7926, + "step": 13 + }, + { + "epoch": 0.06201550387596899, + "grad_norm": 12.024309158325195, + "learning_rate": 2.9411764705882354e-05, + "loss": 2.9558, + "step": 14 + }, + { + "epoch": 0.0664451827242525, + "grad_norm": 13.791522026062012, + "learning_rate": 3.235294117647059e-05, + "loss": 1.9734, + "step": 15 + }, + { + "epoch": 0.07087486157253599, + "grad_norm": 6.770407199859619, + "learning_rate": 3.529411764705883e-05, + "loss": 2.2124, + "step": 16 + }, + { + "epoch": 0.0753045404208195, + "grad_norm": 6.50883150100708, + "learning_rate": 3.8235294117647055e-05, + "loss": 2.4441, + "step": 17 + }, + { + "epoch": 0.07973421926910298, + "grad_norm": 7.501686096191406, + "learning_rate": 4.11764705882353e-05, + "loss": 2.2666, + "step": 18 + }, + { + "epoch": 0.08416389811738649, + "grad_norm": 6.544267177581787, + "learning_rate": 4.411764705882353e-05, + "loss": 2.3097, + "step": 19 + }, + { + "epoch": 0.08859357696567, + "grad_norm": 7.397747993469238, + "learning_rate": 4.705882352941177e-05, + "loss": 2.4401, + "step": 20 + }, + { + "epoch": 0.09302325581395349, + "grad_norm": 7.936941623687744, + "learning_rate": 5e-05, + "loss": 2.5795, + "step": 21 + }, + { + "epoch": 0.09745293466223699, + "grad_norm": 10.404101371765137, + "learning_rate": 5.294117647058824e-05, + "loss": 2.5834, + "step": 22 + }, + { + "epoch": 0.10188261351052048, + "grad_norm": 9.431268692016602, + "learning_rate": 5.588235294117647e-05, + "loss": 2.5247, + "step": 23 + }, + { + "epoch": 0.10631229235880399, + "grad_norm": 10.473112106323242, + "learning_rate": 5.882352941176471e-05, + "loss": 2.5272, + "step": 24 + }, + { + "epoch": 0.11074197120708748, + "grad_norm": 6.828806400299072, + "learning_rate": 6.176470588235295e-05, + "loss": 2.1176, + "step": 25 + }, + { + "epoch": 0.11517165005537099, + "grad_norm": 5.893914699554443, + "learning_rate": 6.470588235294118e-05, + "loss": 2.6795, + "step": 26 + }, + { + "epoch": 0.11960132890365449, + "grad_norm": 6.441338539123535, + "learning_rate": 6.764705882352942e-05, + "loss": 2.301, + "step": 27 + }, + { + "epoch": 0.12403100775193798, + "grad_norm": 6.272483825683594, + "learning_rate": 7.058823529411765e-05, + "loss": 2.1281, + "step": 28 + }, + { + "epoch": 0.12846068660022147, + "grad_norm": 5.158710956573486, + "learning_rate": 7.352941176470589e-05, + "loss": 2.1881, + "step": 29 + }, + { + "epoch": 0.132890365448505, + "grad_norm": 16.882787704467773, + "learning_rate": 7.647058823529411e-05, + "loss": 2.7238, + "step": 30 + }, + { + "epoch": 0.13732004429678848, + "grad_norm": 12.249922752380371, + "learning_rate": 7.941176470588235e-05, + "loss": 2.2578, + "step": 31 + }, + { + "epoch": 0.14174972314507198, + "grad_norm": 5.460890769958496, + "learning_rate": 8.23529411764706e-05, + "loss": 1.8949, + "step": 32 + }, + { + "epoch": 0.1461794019933555, + "grad_norm": 6.217385768890381, + "learning_rate": 8.529411764705883e-05, + "loss": 2.0967, + "step": 33 + }, + { + "epoch": 0.150609080841639, + "grad_norm": 10.28000259399414, + "learning_rate": 8.823529411764706e-05, + "loss": 2.5619, + "step": 34 + }, + { + "epoch": 0.15503875968992248, + "grad_norm": 9.967302322387695, + "learning_rate": 9.11764705882353e-05, + "loss": 2.0039, + "step": 35 + }, + { + "epoch": 0.15946843853820597, + "grad_norm": 8.877910614013672, + "learning_rate": 9.411764705882353e-05, + "loss": 2.3817, + "step": 36 + }, + { + "epoch": 0.1638981173864895, + "grad_norm": 6.2244181632995605, + "learning_rate": 9.705882352941177e-05, + "loss": 1.9161, + "step": 37 + }, + { + "epoch": 0.16832779623477298, + "grad_norm": 6.652865409851074, + "learning_rate": 0.0001, + "loss": 2.2661, + "step": 38 + }, + { + "epoch": 0.17275747508305647, + "grad_norm": 5.198558330535889, + "learning_rate": 9.999979270446262e-05, + "loss": 2.2035, + "step": 39 + }, + { + "epoch": 0.17718715393134, + "grad_norm": 7.7754950523376465, + "learning_rate": 9.999917081956933e-05, + "loss": 1.8432, + "step": 40 + }, + { + "epoch": 0.18161683277962348, + "grad_norm": 7.527754306793213, + "learning_rate": 9.999813435047667e-05, + "loss": 2.2061, + "step": 41 + }, + { + "epoch": 0.18604651162790697, + "grad_norm": 9.0021390914917, + "learning_rate": 9.99966833057789e-05, + "loss": 2.3816, + "step": 42 + }, + { + "epoch": 0.19047619047619047, + "grad_norm": 6.2614827156066895, + "learning_rate": 9.999481769750779e-05, + "loss": 2.3224, + "step": 43 + }, + { + "epoch": 0.19490586932447398, + "grad_norm": 9.527298927307129, + "learning_rate": 9.999253754113262e-05, + "loss": 2.6791, + "step": 44 + }, + { + "epoch": 0.19933554817275748, + "grad_norm": 8.42602252960205, + "learning_rate": 9.998984285556007e-05, + "loss": 2.4413, + "step": 45 + }, + { + "epoch": 0.20376522702104097, + "grad_norm": 11.76278018951416, + "learning_rate": 9.998673366313398e-05, + "loss": 2.2867, + "step": 46 + }, + { + "epoch": 0.2081949058693245, + "grad_norm": 5.570592403411865, + "learning_rate": 9.998320998963521e-05, + "loss": 2.2111, + "step": 47 + }, + { + "epoch": 0.21262458471760798, + "grad_norm": 9.666138648986816, + "learning_rate": 9.997927186428145e-05, + "loss": 2.2846, + "step": 48 + }, + { + "epoch": 0.21705426356589147, + "grad_norm": 6.573619842529297, + "learning_rate": 9.997491931972694e-05, + "loss": 2.1636, + "step": 49 + }, + { + "epoch": 0.22148394241417496, + "grad_norm": 8.646806716918945, + "learning_rate": 9.997015239206215e-05, + "loss": 2.0793, + "step": 50 + }, + { + "epoch": 0.22591362126245848, + "grad_norm": 11.25146484375, + "learning_rate": 9.996497112081364e-05, + "loss": 2.523, + "step": 51 + }, + { + "epoch": 0.23034330011074197, + "grad_norm": 7.36045503616333, + "learning_rate": 9.99593755489436e-05, + "loss": 2.2237, + "step": 52 + }, + { + "epoch": 0.23477297895902546, + "grad_norm": 10.66434383392334, + "learning_rate": 9.995336572284946e-05, + "loss": 2.3307, + "step": 53 + }, + { + "epoch": 0.23920265780730898, + "grad_norm": 8.106691360473633, + "learning_rate": 9.994694169236365e-05, + "loss": 1.9581, + "step": 54 + }, + { + "epoch": 0.24363233665559247, + "grad_norm": 8.106691360473633, + "learning_rate": 9.994694169236365e-05, + "loss": 2.1817, + "step": 55 + }, + { + "epoch": 0.24806201550387597, + "grad_norm": 9.079061508178711, + "learning_rate": 9.994010351075309e-05, + "loss": 2.1612, + "step": 56 + }, + { + "epoch": 0.25249169435215946, + "grad_norm": 12.104523658752441, + "learning_rate": 9.993285123471877e-05, + "loss": 2.798, + "step": 57 + }, + { + "epoch": 0.25692137320044295, + "grad_norm": 10.687878608703613, + "learning_rate": 9.992518492439526e-05, + "loss": 1.9192, + "step": 58 + }, + { + "epoch": 0.26135105204872644, + "grad_norm": 13.860992431640625, + "learning_rate": 9.991710464335022e-05, + "loss": 2.0584, + "step": 59 + }, + { + "epoch": 0.26578073089701, + "grad_norm": 9.031172752380371, + "learning_rate": 9.990861045858392e-05, + "loss": 2.3866, + "step": 60 + }, + { + "epoch": 0.2702104097452935, + "grad_norm": 10.569039344787598, + "learning_rate": 9.989970244052861e-05, + "loss": 2.5264, + "step": 61 + }, + { + "epoch": 0.27464008859357697, + "grad_norm": 10.764253616333008, + "learning_rate": 9.989038066304799e-05, + "loss": 1.9073, + "step": 62 + }, + { + "epoch": 0.27906976744186046, + "grad_norm": 9.956701278686523, + "learning_rate": 9.988064520343659e-05, + "loss": 2.1363, + "step": 63 + }, + { + "epoch": 0.28349944629014395, + "grad_norm": 9.418704986572266, + "learning_rate": 9.987049614241905e-05, + "loss": 2.1378, + "step": 64 + }, + { + "epoch": 0.28792912513842744, + "grad_norm": 5.377594947814941, + "learning_rate": 9.985993356414966e-05, + "loss": 2.5805, + "step": 65 + }, + { + "epoch": 0.292358803986711, + "grad_norm": 5.151512622833252, + "learning_rate": 9.984895755621135e-05, + "loss": 2.4775, + "step": 66 + }, + { + "epoch": 0.2967884828349945, + "grad_norm": 7.1843743324279785, + "learning_rate": 9.983756820961528e-05, + "loss": 2.5418, + "step": 67 + }, + { + "epoch": 0.301218161683278, + "grad_norm": 6.158773899078369, + "learning_rate": 9.982576561879983e-05, + "loss": 2.2148, + "step": 68 + }, + { + "epoch": 0.30564784053156147, + "grad_norm": 6.614671230316162, + "learning_rate": 9.981354988163e-05, + "loss": 2.718, + "step": 69 + }, + { + "epoch": 0.31007751937984496, + "grad_norm": 4.830013751983643, + "learning_rate": 9.980092109939652e-05, + "loss": 2.151, + "step": 70 + }, + { + "epoch": 0.31450719822812845, + "grad_norm": 7.057981967926025, + "learning_rate": 9.978787937681496e-05, + "loss": 2.1143, + "step": 71 + }, + { + "epoch": 0.31893687707641194, + "grad_norm": 7.486720085144043, + "learning_rate": 9.977442482202498e-05, + "loss": 2.0879, + "step": 72 + }, + { + "epoch": 0.3233665559246955, + "grad_norm": 11.194393157958984, + "learning_rate": 9.976055754658935e-05, + "loss": 2.9186, + "step": 73 + }, + { + "epoch": 0.327796234772979, + "grad_norm": 4.733929634094238, + "learning_rate": 9.974627766549302e-05, + "loss": 1.9993, + "step": 74 + }, + { + "epoch": 0.33222591362126247, + "grad_norm": 9.492938995361328, + "learning_rate": 9.973158529714224e-05, + "loss": 2.3098, + "step": 75 + }, + { + "epoch": 0.33665559246954596, + "grad_norm": 6.460459232330322, + "learning_rate": 9.971648056336348e-05, + "loss": 2.0154, + "step": 76 + }, + { + "epoch": 0.34108527131782945, + "grad_norm": 6.839648246765137, + "learning_rate": 9.970096358940251e-05, + "loss": 2.8716, + "step": 77 + }, + { + "epoch": 0.34551495016611294, + "grad_norm": 11.949552536010742, + "learning_rate": 9.968503450392332e-05, + "loss": 3.2224, + "step": 78 + }, + { + "epoch": 0.34994462901439644, + "grad_norm": 11.671749114990234, + "learning_rate": 9.966869343900702e-05, + "loss": 2.4969, + "step": 79 + }, + { + "epoch": 0.35437430786268, + "grad_norm": 7.032209396362305, + "learning_rate": 9.965194053015083e-05, + "loss": 2.0895, + "step": 80 + }, + { + "epoch": 0.3588039867109635, + "grad_norm": 7.36803674697876, + "learning_rate": 9.963477591626687e-05, + "loss": 2.1803, + "step": 81 + }, + { + "epoch": 0.36323366555924697, + "grad_norm": 9.958213806152344, + "learning_rate": 9.961719973968102e-05, + "loss": 2.482, + "step": 82 + }, + { + "epoch": 0.36766334440753046, + "grad_norm": 8.602214813232422, + "learning_rate": 9.959921214613186e-05, + "loss": 1.9825, + "step": 83 + }, + { + "epoch": 0.37209302325581395, + "grad_norm": 8.817967414855957, + "learning_rate": 9.958081328476925e-05, + "loss": 2.2188, + "step": 84 + }, + { + "epoch": 0.37652270210409744, + "grad_norm": 6.849939823150635, + "learning_rate": 9.956200330815329e-05, + "loss": 2.2715, + "step": 85 + }, + { + "epoch": 0.38095238095238093, + "grad_norm": 7.644628524780273, + "learning_rate": 9.954278237225296e-05, + "loss": 2.1561, + "step": 86 + }, + { + "epoch": 0.3853820598006645, + "grad_norm": 10.270885467529297, + "learning_rate": 9.95231506364448e-05, + "loss": 2.3277, + "step": 87 + }, + { + "epoch": 0.38981173864894797, + "grad_norm": 4.501772403717041, + "learning_rate": 9.950310826351168e-05, + "loss": 2.0594, + "step": 88 + }, + { + "epoch": 0.39424141749723146, + "grad_norm": 6.094904899597168, + "learning_rate": 9.948265541964136e-05, + "loss": 2.5396, + "step": 89 + }, + { + "epoch": 0.39867109634551495, + "grad_norm": 4.792986869812012, + "learning_rate": 9.946179227442521e-05, + "loss": 2.0759, + "step": 90 + }, + { + "epoch": 0.40310077519379844, + "grad_norm": 11.445024490356445, + "learning_rate": 9.944051900085667e-05, + "loss": 1.8679, + "step": 91 + }, + { + "epoch": 0.40753045404208194, + "grad_norm": 4.9725422859191895, + "learning_rate": 9.941883577532993e-05, + "loss": 1.9542, + "step": 92 + }, + { + "epoch": 0.4119601328903654, + "grad_norm": 8.572590827941895, + "learning_rate": 9.939674277763844e-05, + "loss": 1.8557, + "step": 93 + }, + { + "epoch": 0.416389811738649, + "grad_norm": 8.919875144958496, + "learning_rate": 9.937424019097337e-05, + "loss": 1.9008, + "step": 94 + }, + { + "epoch": 0.42081949058693247, + "grad_norm": 6.352705955505371, + "learning_rate": 9.935132820192217e-05, + "loss": 2.1908, + "step": 95 + }, + { + "epoch": 0.42524916943521596, + "grad_norm": 6.315227031707764, + "learning_rate": 9.932800700046697e-05, + "loss": 2.2764, + "step": 96 + }, + { + "epoch": 0.42967884828349945, + "grad_norm": 8.06071949005127, + "learning_rate": 9.9304276779983e-05, + "loss": 2.0963, + "step": 97 + }, + { + "epoch": 0.43410852713178294, + "grad_norm": 7.619894027709961, + "learning_rate": 9.9280137737237e-05, + "loss": 2.4399, + "step": 98 + }, + { + "epoch": 0.43853820598006643, + "grad_norm": 13.512142181396484, + "learning_rate": 9.925559007238563e-05, + "loss": 2.3685, + "step": 99 + }, + { + "epoch": 0.4429678848283499, + "grad_norm": 5.173255920410156, + "learning_rate": 9.923063398897372e-05, + "loss": 2.2099, + "step": 100 + }, + { + "epoch": 0.44739756367663347, + "grad_norm": 10.244203567504883, + "learning_rate": 9.920526969393267e-05, + "loss": 2.2032, + "step": 101 + }, + { + "epoch": 0.45182724252491696, + "grad_norm": 9.113722801208496, + "learning_rate": 9.917949739757868e-05, + "loss": 2.5239, + "step": 102 + }, + { + "epoch": 0.45625692137320045, + "grad_norm": 6.485414505004883, + "learning_rate": 9.915331731361103e-05, + "loss": 2.3008, + "step": 103 + }, + { + "epoch": 0.46068660022148394, + "grad_norm": 8.831535339355469, + "learning_rate": 9.912672965911034e-05, + "loss": 2.2914, + "step": 104 + }, + { + "epoch": 0.46511627906976744, + "grad_norm": 8.91296100616455, + "learning_rate": 9.909973465453666e-05, + "loss": 2.4561, + "step": 105 + }, + { + "epoch": 0.4695459579180509, + "grad_norm": 7.34066915512085, + "learning_rate": 9.907233252372774e-05, + "loss": 2.4614, + "step": 106 + }, + { + "epoch": 0.4739756367663344, + "grad_norm": 5.878296375274658, + "learning_rate": 9.904452349389717e-05, + "loss": 2.0118, + "step": 107 + }, + { + "epoch": 0.47840531561461797, + "grad_norm": 8.123944282531738, + "learning_rate": 9.901630779563246e-05, + "loss": 2.3756, + "step": 108 + }, + { + "epoch": 0.48283499446290146, + "grad_norm": 11.739887237548828, + "learning_rate": 9.898768566289315e-05, + "loss": 2.365, + "step": 109 + }, + { + "epoch": 0.48726467331118495, + "grad_norm": 5.946976184844971, + "learning_rate": 9.895865733300886e-05, + "loss": 2.2305, + "step": 110 + }, + { + "epoch": 0.49169435215946844, + "grad_norm": 5.463192462921143, + "learning_rate": 9.89292230466773e-05, + "loss": 2.5732, + "step": 111 + }, + { + "epoch": 0.49612403100775193, + "grad_norm": 4.646864891052246, + "learning_rate": 9.889938304796236e-05, + "loss": 2.1547, + "step": 112 + }, + { + "epoch": 0.5005537098560354, + "grad_norm": 6.221797943115234, + "learning_rate": 9.886913758429193e-05, + "loss": 2.015, + "step": 113 + }, + { + "epoch": 0.5049833887043189, + "grad_norm": 6.859386444091797, + "learning_rate": 9.883848690645601e-05, + "loss": 2.2811, + "step": 114 + }, + { + "epoch": 0.5094130675526024, + "grad_norm": 10.671966552734375, + "learning_rate": 9.880743126860457e-05, + "loss": 1.9567, + "step": 115 + }, + { + "epoch": 0.5138427464008859, + "grad_norm": 5.32224702835083, + "learning_rate": 9.87759709282454e-05, + "loss": 2.5827, + "step": 116 + }, + { + "epoch": 0.5182724252491694, + "grad_norm": 8.69084644317627, + "learning_rate": 9.874410614624203e-05, + "loss": 1.81, + "step": 117 + }, + { + "epoch": 0.5227021040974529, + "grad_norm": 9.261399269104004, + "learning_rate": 9.871183718681153e-05, + "loss": 2.3125, + "step": 118 + }, + { + "epoch": 0.5271317829457365, + "grad_norm": 6.587903022766113, + "learning_rate": 9.867916431752238e-05, + "loss": 2.4587, + "step": 119 + }, + { + "epoch": 0.53156146179402, + "grad_norm": 7.4397430419921875, + "learning_rate": 9.864608780929216e-05, + "loss": 2.5349, + "step": 120 + }, + { + "epoch": 0.5359911406423035, + "grad_norm": 8.569662094116211, + "learning_rate": 9.861260793638538e-05, + "loss": 2.302, + "step": 121 + }, + { + "epoch": 0.540420819490587, + "grad_norm": 8.569662094116211, + "learning_rate": 9.861260793638538e-05, + "loss": 2.3263, + "step": 122 + }, + { + "epoch": 0.5448504983388704, + "grad_norm": 7.741957664489746, + "learning_rate": 9.857872497641117e-05, + "loss": 2.4619, + "step": 123 + }, + { + "epoch": 0.5492801771871539, + "grad_norm": 13.403289794921875, + "learning_rate": 9.854443921032097e-05, + "loss": 2.4491, + "step": 124 + }, + { + "epoch": 0.5537098560354374, + "grad_norm": 11.84448528289795, + "learning_rate": 9.850975092240625e-05, + "loss": 2.4692, + "step": 125 + }, + { + "epoch": 0.5581395348837209, + "grad_norm": 6.203007698059082, + "learning_rate": 9.847466040029609e-05, + "loss": 2.2906, + "step": 126 + }, + { + "epoch": 0.5625692137320044, + "grad_norm": 6.883812427520752, + "learning_rate": 9.843916793495487e-05, + "loss": 2.5539, + "step": 127 + }, + { + "epoch": 0.5669988925802879, + "grad_norm": 11.461130142211914, + "learning_rate": 9.840327382067973e-05, + "loss": 1.9333, + "step": 128 + }, + { + "epoch": 0.5714285714285714, + "grad_norm": 13.166450500488281, + "learning_rate": 9.836697835509827e-05, + "loss": 2.2345, + "step": 129 + }, + { + "epoch": 0.5758582502768549, + "grad_norm": 8.314001083374023, + "learning_rate": 9.833028183916602e-05, + "loss": 2.6494, + "step": 130 + }, + { + "epoch": 0.5802879291251384, + "grad_norm": 6.866865634918213, + "learning_rate": 9.829318457716394e-05, + "loss": 1.9577, + "step": 131 + }, + { + "epoch": 0.584717607973422, + "grad_norm": 5.3526082038879395, + "learning_rate": 9.82556868766959e-05, + "loss": 2.3536, + "step": 132 + }, + { + "epoch": 0.5891472868217055, + "grad_norm": 12.298832893371582, + "learning_rate": 9.821778904868616e-05, + "loss": 2.4454, + "step": 133 + }, + { + "epoch": 0.593576965669989, + "grad_norm": 5.376310348510742, + "learning_rate": 9.817949140737673e-05, + "loss": 2.763, + "step": 134 + }, + { + "epoch": 0.5980066445182725, + "grad_norm": 7.576231002807617, + "learning_rate": 9.814079427032478e-05, + "loss": 2.3395, + "step": 135 + }, + { + "epoch": 0.602436323366556, + "grad_norm": 6.101162433624268, + "learning_rate": 9.810169795840012e-05, + "loss": 1.8174, + "step": 136 + }, + { + "epoch": 0.6068660022148394, + "grad_norm": 9.156957626342773, + "learning_rate": 9.806220279578235e-05, + "loss": 2.5992, + "step": 137 + }, + { + "epoch": 0.6112956810631229, + "grad_norm": 9.112500190734863, + "learning_rate": 9.802230910995834e-05, + "loss": 2.1072, + "step": 138 + }, + { + "epoch": 0.6157253599114064, + "grad_norm": 8.444991111755371, + "learning_rate": 9.798201723171939e-05, + "loss": 2.2685, + "step": 139 + }, + { + "epoch": 0.6201550387596899, + "grad_norm": 4.799725532531738, + "learning_rate": 9.794132749515854e-05, + "loss": 2.136, + "step": 140 + }, + { + "epoch": 0.6245847176079734, + "grad_norm": 8.076310157775879, + "learning_rate": 9.790024023766789e-05, + "loss": 2.2673, + "step": 141 + }, + { + "epoch": 0.6290143964562569, + "grad_norm": 4.653082847595215, + "learning_rate": 9.785875579993557e-05, + "loss": 1.8902, + "step": 142 + }, + { + "epoch": 0.6334440753045404, + "grad_norm": 11.317476272583008, + "learning_rate": 9.781687452594318e-05, + "loss": 2.4088, + "step": 143 + }, + { + "epoch": 0.6378737541528239, + "grad_norm": 7.295245170593262, + "learning_rate": 9.777459676296275e-05, + "loss": 2.2239, + "step": 144 + }, + { + "epoch": 0.6423034330011074, + "grad_norm": 10.208337783813477, + "learning_rate": 9.773192286155394e-05, + "loss": 2.3007, + "step": 145 + }, + { + "epoch": 0.646733111849391, + "grad_norm": 9.138147354125977, + "learning_rate": 9.768885317556116e-05, + "loss": 2.1123, + "step": 146 + }, + { + "epoch": 0.6511627906976745, + "grad_norm": 8.081501960754395, + "learning_rate": 9.764538806211051e-05, + "loss": 2.1822, + "step": 147 + }, + { + "epoch": 0.655592469545958, + "grad_norm": 10.28136920928955, + "learning_rate": 9.760152788160697e-05, + "loss": 2.0472, + "step": 148 + }, + { + "epoch": 0.6600221483942414, + "grad_norm": 12.062602996826172, + "learning_rate": 9.755727299773134e-05, + "loss": 2.2228, + "step": 149 + }, + { + "epoch": 0.6644518272425249, + "grad_norm": 8.189140319824219, + "learning_rate": 9.75126237774372e-05, + "loss": 2.1186, + "step": 150 + }, + { + "epoch": 0.6688815060908084, + "grad_norm": 4.761727809906006, + "learning_rate": 9.746758059094791e-05, + "loss": 2.141, + "step": 151 + }, + { + "epoch": 0.6733111849390919, + "grad_norm": 7.920103549957275, + "learning_rate": 9.742214381175354e-05, + "loss": 2.2799, + "step": 152 + }, + { + "epoch": 0.6777408637873754, + "grad_norm": 7.213951587677002, + "learning_rate": 9.737631381660776e-05, + "loss": 2.1124, + "step": 153 + }, + { + "epoch": 0.6821705426356589, + "grad_norm": 9.37542724609375, + "learning_rate": 9.733009098552473e-05, + "loss": 2.1971, + "step": 154 + }, + { + "epoch": 0.6866002214839424, + "grad_norm": 8.1496000289917, + "learning_rate": 9.728347570177586e-05, + "loss": 2.0953, + "step": 155 + }, + { + "epoch": 0.6910299003322259, + "grad_norm": 5.148806095123291, + "learning_rate": 9.72364683518868e-05, + "loss": 2.1246, + "step": 156 + }, + { + "epoch": 0.6954595791805094, + "grad_norm": 5.193173885345459, + "learning_rate": 9.71890693256341e-05, + "loss": 1.9214, + "step": 157 + }, + { + "epoch": 0.6998892580287929, + "grad_norm": 7.136893272399902, + "learning_rate": 9.7141279016042e-05, + "loss": 2.2632, + "step": 158 + }, + { + "epoch": 0.7043189368770764, + "grad_norm": 7.360518455505371, + "learning_rate": 9.709309781937925e-05, + "loss": 2.4558, + "step": 159 + }, + { + "epoch": 0.70874861572536, + "grad_norm": 9.435393333435059, + "learning_rate": 9.704452613515571e-05, + "loss": 2.1096, + "step": 160 + }, + { + "epoch": 0.7131782945736435, + "grad_norm": 7.742511749267578, + "learning_rate": 9.699556436611912e-05, + "loss": 2.3128, + "step": 161 + }, + { + "epoch": 0.717607973421927, + "grad_norm": 9.153509140014648, + "learning_rate": 9.694621291825174e-05, + "loss": 2.2575, + "step": 162 + }, + { + "epoch": 0.7220376522702104, + "grad_norm": 8.447454452514648, + "learning_rate": 9.689647220076696e-05, + "loss": 1.9745, + "step": 163 + }, + { + "epoch": 0.7264673311184939, + "grad_norm": 5.4174909591674805, + "learning_rate": 9.684634262610593e-05, + "loss": 2.4049, + "step": 164 + }, + { + "epoch": 0.7308970099667774, + "grad_norm": 9.518255233764648, + "learning_rate": 9.679582460993413e-05, + "loss": 1.8485, + "step": 165 + }, + { + "epoch": 0.7353266888150609, + "grad_norm": 8.27046012878418, + "learning_rate": 9.674491857113792e-05, + "loss": 2.665, + "step": 166 + }, + { + "epoch": 0.7397563676633444, + "grad_norm": 6.9749836921691895, + "learning_rate": 9.669362493182111e-05, + "loss": 2.3234, + "step": 167 + }, + { + "epoch": 0.7441860465116279, + "grad_norm": 5.377409934997559, + "learning_rate": 9.664194411730141e-05, + "loss": 1.8631, + "step": 168 + }, + { + "epoch": 0.7486157253599114, + "grad_norm": 8.406085014343262, + "learning_rate": 9.658987655610687e-05, + "loss": 2.2899, + "step": 169 + }, + { + "epoch": 0.7530454042081949, + "grad_norm": 4.888492584228516, + "learning_rate": 9.653742267997246e-05, + "loss": 2.0495, + "step": 170 + }, + { + "epoch": 0.7574750830564784, + "grad_norm": 10.016488075256348, + "learning_rate": 9.648458292383631e-05, + "loss": 1.9254, + "step": 171 + }, + { + "epoch": 0.7619047619047619, + "grad_norm": 5.880348205566406, + "learning_rate": 9.643135772583627e-05, + "loss": 1.9746, + "step": 172 + }, + { + "epoch": 0.7663344407530454, + "grad_norm": 8.261670112609863, + "learning_rate": 9.63777475273062e-05, + "loss": 2.437, + "step": 173 + }, + { + "epoch": 0.770764119601329, + "grad_norm": 15.291732788085938, + "learning_rate": 9.632375277277226e-05, + "loss": 2.6626, + "step": 174 + }, + { + "epoch": 0.7751937984496124, + "grad_norm": 7.9028096199035645, + "learning_rate": 9.626937390994932e-05, + "loss": 2.2399, + "step": 175 + }, + { + "epoch": 0.7796234772978959, + "grad_norm": 4.564765930175781, + "learning_rate": 9.621461138973724e-05, + "loss": 2.083, + "step": 176 + }, + { + "epoch": 0.7840531561461794, + "grad_norm": 7.957129955291748, + "learning_rate": 9.615946566621702e-05, + "loss": 1.8967, + "step": 177 + }, + { + "epoch": 0.7884828349944629, + "grad_norm": 16.633319854736328, + "learning_rate": 9.610393719664719e-05, + "loss": 1.9869, + "step": 178 + }, + { + "epoch": 0.7929125138427464, + "grad_norm": 8.166046142578125, + "learning_rate": 9.604802644145989e-05, + "loss": 2.141, + "step": 179 + }, + { + "epoch": 0.7973421926910299, + "grad_norm": 10.27008056640625, + "learning_rate": 9.59917338642571e-05, + "loss": 2.077, + "step": 180 + }, + { + "epoch": 0.8017718715393134, + "grad_norm": 5.996699810028076, + "learning_rate": 9.593505993180687e-05, + "loss": 1.9777, + "step": 181 + }, + { + "epoch": 0.8062015503875969, + "grad_norm": 8.765548706054688, + "learning_rate": 9.58780051140393e-05, + "loss": 1.9158, + "step": 182 + }, + { + "epoch": 0.8106312292358804, + "grad_norm": 4.515958786010742, + "learning_rate": 9.582056988404275e-05, + "loss": 1.9419, + "step": 183 + }, + { + "epoch": 0.8150609080841639, + "grad_norm": 10.045695304870605, + "learning_rate": 9.576275471805993e-05, + "loss": 2.2376, + "step": 184 + }, + { + "epoch": 0.8194905869324474, + "grad_norm": 5.502155303955078, + "learning_rate": 9.570456009548383e-05, + "loss": 1.9097, + "step": 185 + }, + { + "epoch": 0.8239202657807309, + "grad_norm": 6.540369987487793, + "learning_rate": 9.56459864988539e-05, + "loss": 2.1779, + "step": 186 + }, + { + "epoch": 0.8283499446290143, + "grad_norm": 6.4399285316467285, + "learning_rate": 9.558703441385194e-05, + "loss": 2.0636, + "step": 187 + }, + { + "epoch": 0.832779623477298, + "grad_norm": 17.52277946472168, + "learning_rate": 9.552770432929811e-05, + "loss": 2.0766, + "step": 188 + }, + { + "epoch": 0.8372093023255814, + "grad_norm": 7.090964317321777, + "learning_rate": 9.54679967371469e-05, + "loss": 2.2198, + "step": 189 + }, + { + "epoch": 0.8416389811738649, + "grad_norm": 4.9996490478515625, + "learning_rate": 9.540791213248299e-05, + "loss": 2.1353, + "step": 190 + }, + { + "epoch": 0.8460686600221484, + "grad_norm": 10.12185001373291, + "learning_rate": 9.534745101351719e-05, + "loss": 2.4016, + "step": 191 + }, + { + "epoch": 0.8504983388704319, + "grad_norm": 7.592304706573486, + "learning_rate": 9.528661388158234e-05, + "loss": 2.1985, + "step": 192 + }, + { + "epoch": 0.8549280177187154, + "grad_norm": 6.609996318817139, + "learning_rate": 9.522540124112902e-05, + "loss": 2.1224, + "step": 193 + }, + { + "epoch": 0.8593576965669989, + "grad_norm": 4.852092742919922, + "learning_rate": 9.516381359972158e-05, + "loss": 2.2335, + "step": 194 + }, + { + "epoch": 0.8637873754152824, + "grad_norm": 6.790386199951172, + "learning_rate": 9.51018514680337e-05, + "loss": 1.7354, + "step": 195 + }, + { + "epoch": 0.8682170542635659, + "grad_norm": 4.828433990478516, + "learning_rate": 9.503951535984434e-05, + "loss": 1.9872, + "step": 196 + }, + { + "epoch": 0.8726467331118494, + "grad_norm": 4.812173843383789, + "learning_rate": 9.497680579203338e-05, + "loss": 2.0323, + "step": 197 + }, + { + "epoch": 0.8770764119601329, + "grad_norm": 3.76886248588562, + "learning_rate": 9.491372328457738e-05, + "loss": 2.308, + "step": 198 + }, + { + "epoch": 0.8815060908084164, + "grad_norm": 4.918368339538574, + "learning_rate": 9.485026836054519e-05, + "loss": 2.0527, + "step": 199 + }, + { + "epoch": 0.8859357696566998, + "grad_norm": 5.6282639503479, + "learning_rate": 9.478644154609372e-05, + "loss": 2.4852, + "step": 200 + }, + { + "epoch": 0.8903654485049833, + "grad_norm": 8.104965209960938, + "learning_rate": 9.472224337046357e-05, + "loss": 1.6854, + "step": 201 + }, + { + "epoch": 0.8947951273532669, + "grad_norm": 5.7505598068237305, + "learning_rate": 9.46576743659745e-05, + "loss": 2.1109, + "step": 202 + }, + { + "epoch": 0.8992248062015504, + "grad_norm": 5.078502178192139, + "learning_rate": 9.45927350680212e-05, + "loss": 2.2926, + "step": 203 + }, + { + "epoch": 0.9036544850498339, + "grad_norm": 6.620337009429932, + "learning_rate": 9.452742601506872e-05, + "loss": 2.1566, + "step": 204 + }, + { + "epoch": 0.9080841638981174, + "grad_norm": 6.228148937225342, + "learning_rate": 9.446174774864808e-05, + "loss": 2.141, + "step": 205 + }, + { + "epoch": 0.9125138427464009, + "grad_norm": 5.25639533996582, + "learning_rate": 9.439570081335173e-05, + "loss": 2.1721, + "step": 206 + }, + { + "epoch": 0.9169435215946844, + "grad_norm": 7.61407470703125, + "learning_rate": 9.432928575682907e-05, + "loss": 1.9253, + "step": 207 + }, + { + "epoch": 0.9213732004429679, + "grad_norm": 7.121610641479492, + "learning_rate": 9.42625031297819e-05, + "loss": 2.3093, + "step": 208 + }, + { + "epoch": 0.9258028792912514, + "grad_norm": 8.246429443359375, + "learning_rate": 9.419535348595985e-05, + "loss": 2.321, + "step": 209 + }, + { + "epoch": 0.9302325581395349, + "grad_norm": 8.831986427307129, + "learning_rate": 9.412783738215575e-05, + "loss": 2.3273, + "step": 210 + }, + { + "epoch": 0.9346622369878184, + "grad_norm": 7.420327186584473, + "learning_rate": 9.405995537820111e-05, + "loss": 2.2849, + "step": 211 + }, + { + "epoch": 0.9390919158361019, + "grad_norm": 7.654839992523193, + "learning_rate": 9.399170803696138e-05, + "loss": 1.8333, + "step": 212 + }, + { + "epoch": 0.9435215946843853, + "grad_norm": 7.557086944580078, + "learning_rate": 9.392309592433133e-05, + "loss": 2.1632, + "step": 213 + }, + { + "epoch": 0.9479512735326688, + "grad_norm": 7.721208572387695, + "learning_rate": 9.385411960923036e-05, + "loss": 1.9735, + "step": 214 + }, + { + "epoch": 0.9523809523809523, + "grad_norm": 7.448840618133545, + "learning_rate": 9.378477966359773e-05, + "loss": 2.4228, + "step": 215 + }, + { + "epoch": 0.9568106312292359, + "grad_norm": 5.7559733390808105, + "learning_rate": 9.371507666238794e-05, + "loss": 2.3595, + "step": 216 + }, + { + "epoch": 0.9612403100775194, + "grad_norm": 5.419656753540039, + "learning_rate": 9.36450111835658e-05, + "loss": 2.3156, + "step": 217 + }, + { + "epoch": 0.9656699889258029, + "grad_norm": 8.596833229064941, + "learning_rate": 9.357458380810175e-05, + "loss": 2.3778, + "step": 218 + }, + { + "epoch": 0.9700996677740864, + "grad_norm": 15.337946891784668, + "learning_rate": 9.350379511996705e-05, + "loss": 2.3743, + "step": 219 + }, + { + "epoch": 0.9745293466223699, + "grad_norm": 6.938776016235352, + "learning_rate": 9.343264570612883e-05, + "loss": 2.2887, + "step": 220 + }, + { + "epoch": 0.9789590254706534, + "grad_norm": 8.320263862609863, + "learning_rate": 9.336113615654534e-05, + "loss": 1.9961, + "step": 221 + }, + { + "epoch": 0.9833887043189369, + "grad_norm": 5.129944801330566, + "learning_rate": 9.328926706416101e-05, + "loss": 1.8434, + "step": 222 + }, + { + "epoch": 0.9878183831672204, + "grad_norm": 6.840932369232178, + "learning_rate": 9.321703902490151e-05, + "loss": 2.0304, + "step": 223 + }, + { + "epoch": 0.9922480620155039, + "grad_norm": 8.26124095916748, + "learning_rate": 9.314445263766888e-05, + "loss": 2.2058, + "step": 224 + }, + { + "epoch": 0.9966777408637874, + "grad_norm": 9.003861427307129, + "learning_rate": 9.307150850433643e-05, + "loss": 2.3681, + "step": 225 + }, + { + "epoch": 1.0022148394241417, + "grad_norm": 10.506808280944824, + "learning_rate": 9.299820722974395e-05, + "loss": 2.1082, + "step": 226 + }, + { + "epoch": 1.0066445182724253, + "grad_norm": 13.937885284423828, + "learning_rate": 9.29245494216925e-05, + "loss": 2.1381, + "step": 227 + }, + { + "epoch": 1.0110741971207087, + "grad_norm": 10.9959077835083, + "learning_rate": 9.285053569093946e-05, + "loss": 2.5619, + "step": 228 + }, + { + "epoch": 1.0155038759689923, + "grad_norm": 8.583209991455078, + "learning_rate": 9.27761666511935e-05, + "loss": 2.4306, + "step": 229 + }, + { + "epoch": 1.0199335548172757, + "grad_norm": 8.024839401245117, + "learning_rate": 9.270144291910941e-05, + "loss": 2.2235, + "step": 230 + }, + { + "epoch": 1.0243632336655593, + "grad_norm": 7.364706516265869, + "learning_rate": 9.262636511428304e-05, + "loss": 2.4616, + "step": 231 + }, + { + "epoch": 1.0287929125138426, + "grad_norm": 8.752544403076172, + "learning_rate": 9.255093385924614e-05, + "loss": 2.127, + "step": 232 + }, + { + "epoch": 1.0332225913621262, + "grad_norm": 6.984606742858887, + "learning_rate": 9.247514977946124e-05, + "loss": 2.1871, + "step": 233 + }, + { + "epoch": 1.0376522702104098, + "grad_norm": 8.737010955810547, + "learning_rate": 9.239901350331634e-05, + "loss": 2.3309, + "step": 234 + }, + { + "epoch": 1.0420819490586932, + "grad_norm": 6.438765525817871, + "learning_rate": 9.232252566211991e-05, + "loss": 2.1588, + "step": 235 + }, + { + "epoch": 1.0465116279069768, + "grad_norm": 9.31252384185791, + "learning_rate": 9.224568689009547e-05, + "loss": 2.0311, + "step": 236 + }, + { + "epoch": 1.0509413067552602, + "grad_norm": 11.200642585754395, + "learning_rate": 9.216849782437637e-05, + "loss": 1.8581, + "step": 237 + }, + { + "epoch": 1.0553709856035438, + "grad_norm": 6.259405136108398, + "learning_rate": 9.20909591050006e-05, + "loss": 2.0763, + "step": 238 + }, + { + "epoch": 1.0598006644518272, + "grad_norm": 7.228961944580078, + "learning_rate": 9.201307137490536e-05, + "loss": 2.5367, + "step": 239 + }, + { + "epoch": 1.0642303433001108, + "grad_norm": 10.137297630310059, + "learning_rate": 9.19348352799218e-05, + "loss": 1.8874, + "step": 240 + }, + { + "epoch": 1.0686600221483942, + "grad_norm": 6.983479976654053, + "learning_rate": 9.185625146876965e-05, + "loss": 1.8221, + "step": 241 + }, + { + "epoch": 1.0730897009966778, + "grad_norm": 5.6672444343566895, + "learning_rate": 9.177732059305186e-05, + "loss": 2.119, + "step": 242 + }, + { + "epoch": 1.0775193798449612, + "grad_norm": 6.37617826461792, + "learning_rate": 9.169804330724916e-05, + "loss": 2.1792, + "step": 243 + }, + { + "epoch": 1.0819490586932448, + "grad_norm": 7.426950931549072, + "learning_rate": 9.161842026871466e-05, + "loss": 1.8846, + "step": 244 + }, + { + "epoch": 1.0863787375415281, + "grad_norm": 5.454375743865967, + "learning_rate": 9.153845213766836e-05, + "loss": 2.2066, + "step": 245 + }, + { + "epoch": 1.0908084163898117, + "grad_norm": 8.249213218688965, + "learning_rate": 9.145813957719174e-05, + "loss": 2.2305, + "step": 246 + }, + { + "epoch": 1.0952380952380953, + "grad_norm": 3.8901290893554688, + "learning_rate": 9.137748325322222e-05, + "loss": 2.2013, + "step": 247 + }, + { + "epoch": 1.0996677740863787, + "grad_norm": 5.952533721923828, + "learning_rate": 9.129648383454764e-05, + "loss": 2.1404, + "step": 248 + }, + { + "epoch": 1.1040974529346623, + "grad_norm": 9.447667121887207, + "learning_rate": 9.121514199280071e-05, + "loss": 2.1473, + "step": 249 + }, + { + "epoch": 1.1085271317829457, + "grad_norm": 7.8357415199279785, + "learning_rate": 9.113345840245347e-05, + "loss": 2.1353, + "step": 250 + }, + { + "epoch": 1.1129568106312293, + "grad_norm": 5.7361860275268555, + "learning_rate": 9.105143374081168e-05, + "loss": 2.1992, + "step": 251 + }, + { + "epoch": 1.1173864894795127, + "grad_norm": 10.3837251663208, + "learning_rate": 9.096906868800917e-05, + "loss": 2.3124, + "step": 252 + }, + { + "epoch": 1.1218161683277963, + "grad_norm": 3.863896369934082, + "learning_rate": 9.088636392700227e-05, + "loss": 2.257, + "step": 253 + }, + { + "epoch": 1.1262458471760797, + "grad_norm": 9.265199661254883, + "learning_rate": 9.08033201435641e-05, + "loss": 2.0167, + "step": 254 + }, + { + "epoch": 1.1306755260243633, + "grad_norm": 4.3140177726745605, + "learning_rate": 9.071993802627887e-05, + "loss": 2.3914, + "step": 255 + }, + { + "epoch": 1.1351052048726467, + "grad_norm": 14.065458297729492, + "learning_rate": 9.063621826653624e-05, + "loss": 2.7486, + "step": 256 + }, + { + "epoch": 1.1395348837209303, + "grad_norm": 9.089619636535645, + "learning_rate": 9.055216155852548e-05, + "loss": 1.7132, + "step": 257 + }, + { + "epoch": 1.1439645625692136, + "grad_norm": 6.716838836669922, + "learning_rate": 9.046776859922983e-05, + "loss": 1.8824, + "step": 258 + }, + { + "epoch": 1.1483942414174972, + "grad_norm": 14.335271835327148, + "learning_rate": 9.038304008842064e-05, + "loss": 1.744, + "step": 259 + }, + { + "epoch": 1.1528239202657806, + "grad_norm": 7.002466678619385, + "learning_rate": 9.029797672865159e-05, + "loss": 2.4813, + "step": 260 + }, + { + "epoch": 1.1572535991140642, + "grad_norm": 6.661802291870117, + "learning_rate": 9.021257922525288e-05, + "loss": 1.8213, + "step": 261 + }, + { + "epoch": 1.1616832779623478, + "grad_norm": 7.363678455352783, + "learning_rate": 9.012684828632538e-05, + "loss": 1.7975, + "step": 262 + }, + { + "epoch": 1.1661129568106312, + "grad_norm": 6.833836555480957, + "learning_rate": 9.00407846227347e-05, + "loss": 2.1189, + "step": 263 + }, + { + "epoch": 1.1705426356589148, + "grad_norm": 6.624426364898682, + "learning_rate": 8.995438894810541e-05, + "loss": 1.9159, + "step": 264 + }, + { + "epoch": 1.1749723145071982, + "grad_norm": 7.053401470184326, + "learning_rate": 8.9867661978815e-05, + "loss": 1.8237, + "step": 265 + }, + { + "epoch": 1.1794019933554818, + "grad_norm": 6.560683250427246, + "learning_rate": 8.978060443398802e-05, + "loss": 1.9715, + "step": 266 + }, + { + "epoch": 1.1838316722037652, + "grad_norm": 6.258054733276367, + "learning_rate": 8.96932170354901e-05, + "loss": 1.8396, + "step": 267 + }, + { + "epoch": 1.1882613510520488, + "grad_norm": 9.02526569366455, + "learning_rate": 8.960550050792194e-05, + "loss": 2.1659, + "step": 268 + }, + { + "epoch": 1.1926910299003322, + "grad_norm": 6.251518726348877, + "learning_rate": 8.951745557861333e-05, + "loss": 1.8894, + "step": 269 + }, + { + "epoch": 1.1971207087486158, + "grad_norm": 8.956218719482422, + "learning_rate": 8.942908297761711e-05, + "loss": 2.5807, + "step": 270 + }, + { + "epoch": 1.2015503875968991, + "grad_norm": 6.6762237548828125, + "learning_rate": 8.934038343770312e-05, + "loss": 2.2264, + "step": 271 + }, + { + "epoch": 1.2059800664451827, + "grad_norm": 4.275567531585693, + "learning_rate": 8.925135769435211e-05, + "loss": 1.6638, + "step": 272 + }, + { + "epoch": 1.2104097452934663, + "grad_norm": 6.121718406677246, + "learning_rate": 8.916200648574964e-05, + "loss": 2.0015, + "step": 273 + }, + { + "epoch": 1.2148394241417497, + "grad_norm": 5.744752883911133, + "learning_rate": 8.907233055277999e-05, + "loss": 1.9907, + "step": 274 + }, + { + "epoch": 1.219269102990033, + "grad_norm": 6.730654239654541, + "learning_rate": 8.898233063902e-05, + "loss": 2.1793, + "step": 275 + }, + { + "epoch": 1.2236987818383167, + "grad_norm": 14.31635856628418, + "learning_rate": 8.889200749073285e-05, + "loss": 2.0648, + "step": 276 + }, + { + "epoch": 1.2281284606866003, + "grad_norm": 5.152917385101318, + "learning_rate": 8.880136185686201e-05, + "loss": 1.793, + "step": 277 + }, + { + "epoch": 1.2325581395348837, + "grad_norm": 10.570582389831543, + "learning_rate": 8.871039448902487e-05, + "loss": 1.7875, + "step": 278 + }, + { + "epoch": 1.2369878183831673, + "grad_norm": 9.247808456420898, + "learning_rate": 8.861910614150661e-05, + "loss": 1.7261, + "step": 279 + }, + { + "epoch": 1.2414174972314507, + "grad_norm": 13.006294250488281, + "learning_rate": 8.852749757125392e-05, + "loss": 1.6614, + "step": 280 + }, + { + "epoch": 1.2458471760797343, + "grad_norm": 6.007842063903809, + "learning_rate": 8.84355695378687e-05, + "loss": 1.9225, + "step": 281 + }, + { + "epoch": 1.2502768549280177, + "grad_norm": 7.5843658447265625, + "learning_rate": 8.83433228036018e-05, + "loss": 2.3472, + "step": 282 + }, + { + "epoch": 1.2547065337763013, + "grad_norm": 8.273632049560547, + "learning_rate": 8.825075813334669e-05, + "loss": 2.0889, + "step": 283 + }, + { + "epoch": 1.2591362126245846, + "grad_norm": 7.670727729797363, + "learning_rate": 8.815787629463305e-05, + "loss": 1.5007, + "step": 284 + }, + { + "epoch": 1.2635658914728682, + "grad_norm": 10.460335731506348, + "learning_rate": 8.806467805762055e-05, + "loss": 1.9641, + "step": 285 + }, + { + "epoch": 1.2679955703211516, + "grad_norm": 8.615412712097168, + "learning_rate": 8.79711641950923e-05, + "loss": 2.0096, + "step": 286 + }, + { + "epoch": 1.2724252491694352, + "grad_norm": 8.69423770904541, + "learning_rate": 8.787733548244858e-05, + "loss": 1.8392, + "step": 287 + }, + { + "epoch": 1.2768549280177188, + "grad_norm": 4.95367431640625, + "learning_rate": 8.778319269770034e-05, + "loss": 1.8004, + "step": 288 + }, + { + "epoch": 1.2812846068660022, + "grad_norm": 8.5960693359375, + "learning_rate": 8.768873662146271e-05, + "loss": 2.2377, + "step": 289 + }, + { + "epoch": 1.2857142857142856, + "grad_norm": 7.77866268157959, + "learning_rate": 8.759396803694861e-05, + "loss": 2.0425, + "step": 290 + }, + { + "epoch": 1.2901439645625692, + "grad_norm": 6.187042236328125, + "learning_rate": 8.749888772996226e-05, + "loss": 2.1688, + "step": 291 + }, + { + "epoch": 1.2945736434108528, + "grad_norm": 8.12923812866211, + "learning_rate": 8.740349648889259e-05, + "loss": 2.4057, + "step": 292 + }, + { + "epoch": 1.2990033222591362, + "grad_norm": 5.084444522857666, + "learning_rate": 8.730779510470671e-05, + "loss": 2.092, + "step": 293 + }, + { + "epoch": 1.3034330011074198, + "grad_norm": 8.99492359161377, + "learning_rate": 8.721178437094346e-05, + "loss": 1.9337, + "step": 294 + }, + { + "epoch": 1.3078626799557032, + "grad_norm": 4.3779706954956055, + "learning_rate": 8.711546508370666e-05, + "loss": 2.2644, + "step": 295 + }, + { + "epoch": 1.3122923588039868, + "grad_norm": 5.100649356842041, + "learning_rate": 8.701883804165866e-05, + "loss": 1.7897, + "step": 296 + }, + { + "epoch": 1.3167220376522701, + "grad_norm": 7.3021769523620605, + "learning_rate": 8.692190404601369e-05, + "loss": 1.8963, + "step": 297 + }, + { + "epoch": 1.3211517165005537, + "grad_norm": 8.583118438720703, + "learning_rate": 8.682466390053106e-05, + "loss": 2.3528, + "step": 298 + }, + { + "epoch": 1.3255813953488373, + "grad_norm": 9.092851638793945, + "learning_rate": 8.672711841150877e-05, + "loss": 2.2453, + "step": 299 + }, + { + "epoch": 1.3300110741971207, + "grad_norm": 4.446629524230957, + "learning_rate": 8.662926838777656e-05, + "loss": 1.8826, + "step": 300 + }, + { + "epoch": 1.334440753045404, + "grad_norm": 8.508023262023926, + "learning_rate": 8.653111464068937e-05, + "loss": 1.7581, + "step": 301 + }, + { + "epoch": 1.3388704318936877, + "grad_norm": 3.5442817211151123, + "learning_rate": 8.643265798412056e-05, + "loss": 2.124, + "step": 302 + }, + { + "epoch": 1.3433001107419713, + "grad_norm": 10.445211410522461, + "learning_rate": 8.633389923445514e-05, + "loss": 2.6855, + "step": 303 + }, + { + "epoch": 1.3477297895902547, + "grad_norm": 6.55064582824707, + "learning_rate": 8.623483921058303e-05, + "loss": 2.4239, + "step": 304 + }, + { + "epoch": 1.352159468438538, + "grad_norm": 7.995059013366699, + "learning_rate": 8.613547873389228e-05, + "loss": 1.828, + "step": 305 + }, + { + "epoch": 1.3565891472868217, + "grad_norm": 7.152670860290527, + "learning_rate": 8.603581862826222e-05, + "loss": 1.944, + "step": 306 + }, + { + "epoch": 1.3610188261351053, + "grad_norm": 6.228992938995361, + "learning_rate": 8.593585972005664e-05, + "loss": 2.0547, + "step": 307 + }, + { + "epoch": 1.3654485049833887, + "grad_norm": 6.894842624664307, + "learning_rate": 8.5835602838117e-05, + "loss": 1.7732, + "step": 308 + }, + { + "epoch": 1.3698781838316723, + "grad_norm": 7.37748384475708, + "learning_rate": 8.573504881375542e-05, + "loss": 1.8529, + "step": 309 + }, + { + "epoch": 1.3743078626799556, + "grad_norm": 7.970264911651611, + "learning_rate": 8.563419848074798e-05, + "loss": 1.9512, + "step": 310 + }, + { + "epoch": 1.3787375415282392, + "grad_norm": 7.616507530212402, + "learning_rate": 8.553305267532758e-05, + "loss": 1.9866, + "step": 311 + }, + { + "epoch": 1.3831672203765226, + "grad_norm": 7.430023670196533, + "learning_rate": 8.543161223617724e-05, + "loss": 2.0003, + "step": 312 + }, + { + "epoch": 1.3875968992248062, + "grad_norm": 7.171806812286377, + "learning_rate": 8.532987800442292e-05, + "loss": 1.9861, + "step": 313 + }, + { + "epoch": 1.3920265780730898, + "grad_norm": 7.678527355194092, + "learning_rate": 8.522785082362674e-05, + "loss": 2.0038, + "step": 314 + }, + { + "epoch": 1.3964562569213732, + "grad_norm": 7.096549034118652, + "learning_rate": 8.512553153977987e-05, + "loss": 2.0742, + "step": 315 + }, + { + "epoch": 1.4008859357696566, + "grad_norm": 7.600286960601807, + "learning_rate": 8.502292100129553e-05, + "loss": 1.9543, + "step": 316 + }, + { + "epoch": 1.4053156146179402, + "grad_norm": 8.969922065734863, + "learning_rate": 8.492002005900202e-05, + "loss": 1.6853, + "step": 317 + }, + { + "epoch": 1.4097452934662238, + "grad_norm": 5.5180745124816895, + "learning_rate": 8.481682956613555e-05, + "loss": 1.7093, + "step": 318 + }, + { + "epoch": 1.4141749723145072, + "grad_norm": 13.428471565246582, + "learning_rate": 8.471335037833328e-05, + "loss": 2.0241, + "step": 319 + }, + { + "epoch": 1.4186046511627908, + "grad_norm": 16.033464431762695, + "learning_rate": 8.460958335362616e-05, + "loss": 1.788, + "step": 320 + }, + { + "epoch": 1.4230343300110742, + "grad_norm": 5.455560684204102, + "learning_rate": 8.450552935243186e-05, + "loss": 2.0701, + "step": 321 + }, + { + "epoch": 1.4274640088593578, + "grad_norm": 7.2362189292907715, + "learning_rate": 8.440118923754757e-05, + "loss": 1.7665, + "step": 322 + }, + { + "epoch": 1.4318936877076411, + "grad_norm": 6.421346664428711, + "learning_rate": 8.429656387414289e-05, + "loss": 1.843, + "step": 323 + }, + { + "epoch": 1.4363233665559247, + "grad_norm": 16.439189910888672, + "learning_rate": 8.419165412975265e-05, + "loss": 2.2937, + "step": 324 + }, + { + "epoch": 1.4407530454042081, + "grad_norm": 9.863225936889648, + "learning_rate": 8.408646087426974e-05, + "loss": 2.0961, + "step": 325 + }, + { + "epoch": 1.4451827242524917, + "grad_norm": 5.041436195373535, + "learning_rate": 8.398098497993786e-05, + "loss": 2.1183, + "step": 326 + }, + { + "epoch": 1.449612403100775, + "grad_norm": 8.520334243774414, + "learning_rate": 8.387522732134428e-05, + "loss": 1.7906, + "step": 327 + }, + { + "epoch": 1.4540420819490587, + "grad_norm": 6.228928565979004, + "learning_rate": 8.376918877541264e-05, + "loss": 2.1264, + "step": 328 + }, + { + "epoch": 1.4584717607973423, + "grad_norm": 8.996322631835938, + "learning_rate": 8.366287022139561e-05, + "loss": 2.1557, + "step": 329 + }, + { + "epoch": 1.4629014396456257, + "grad_norm": 5.682254791259766, + "learning_rate": 8.35562725408677e-05, + "loss": 1.9131, + "step": 330 + }, + { + "epoch": 1.467331118493909, + "grad_norm": 8.119385719299316, + "learning_rate": 8.344939661771783e-05, + "loss": 2.4303, + "step": 331 + }, + { + "epoch": 1.4717607973421927, + "grad_norm": 8.717931747436523, + "learning_rate": 8.334224333814209e-05, + "loss": 1.9083, + "step": 332 + }, + { + "epoch": 1.4761904761904763, + "grad_norm": 7.539588928222656, + "learning_rate": 8.323481359063632e-05, + "loss": 1.7957, + "step": 333 + }, + { + "epoch": 1.4806201550387597, + "grad_norm": 6.015201568603516, + "learning_rate": 8.312710826598883e-05, + "loss": 1.9132, + "step": 334 + }, + { + "epoch": 1.4850498338870433, + "grad_norm": 8.548601150512695, + "learning_rate": 8.301912825727294e-05, + "loss": 1.9532, + "step": 335 + }, + { + "epoch": 1.4894795127353266, + "grad_norm": 12.072663307189941, + "learning_rate": 8.29108744598396e-05, + "loss": 2.2549, + "step": 336 + }, + { + "epoch": 1.4939091915836102, + "grad_norm": 4.701639175415039, + "learning_rate": 8.280234777131e-05, + "loss": 2.013, + "step": 337 + }, + { + "epoch": 1.4983388704318936, + "grad_norm": 6.194016456604004, + "learning_rate": 8.269354909156802e-05, + "loss": 1.677, + "step": 338 + }, + { + "epoch": 1.5027685492801772, + "grad_norm": 6.05522346496582, + "learning_rate": 8.258447932275295e-05, + "loss": 1.844, + "step": 339 + }, + { + "epoch": 1.5071982281284608, + "grad_norm": 5.3696980476379395, + "learning_rate": 8.24751393692518e-05, + "loss": 1.9202, + "step": 340 + }, + { + "epoch": 1.5116279069767442, + "grad_norm": 4.895171165466309, + "learning_rate": 8.236553013769197e-05, + "loss": 1.8623, + "step": 341 + }, + { + "epoch": 1.5160575858250276, + "grad_norm": 10.872806549072266, + "learning_rate": 8.225565253693364e-05, + "loss": 1.6916, + "step": 342 + }, + { + "epoch": 1.5204872646733112, + "grad_norm": 6.46589994430542, + "learning_rate": 8.214550747806227e-05, + "loss": 2.0726, + "step": 343 + }, + { + "epoch": 1.5249169435215948, + "grad_norm": 7.604471683502197, + "learning_rate": 8.203509587438098e-05, + "loss": 2.05, + "step": 344 + }, + { + "epoch": 1.5293466223698782, + "grad_norm": 16.37729263305664, + "learning_rate": 8.192441864140313e-05, + "loss": 2.1448, + "step": 345 + }, + { + "epoch": 1.5337763012181616, + "grad_norm": 11.779487609863281, + "learning_rate": 8.181347669684457e-05, + "loss": 1.8318, + "step": 346 + }, + { + "epoch": 1.5382059800664452, + "grad_norm": 11.811964988708496, + "learning_rate": 8.170227096061608e-05, + "loss": 2.3381, + "step": 347 + }, + { + "epoch": 1.5426356589147288, + "grad_norm": 6.4559645652771, + "learning_rate": 8.159080235481579e-05, + "loss": 2.2196, + "step": 348 + }, + { + "epoch": 1.5470653377630121, + "grad_norm": 8.946720123291016, + "learning_rate": 8.147907180372148e-05, + "loss": 2.0286, + "step": 349 + }, + { + "epoch": 1.5514950166112955, + "grad_norm": 13.929659843444824, + "learning_rate": 8.136708023378292e-05, + "loss": 2.3538, + "step": 350 + }, + { + "epoch": 1.5559246954595793, + "grad_norm": 5.848323345184326, + "learning_rate": 8.125482857361425e-05, + "loss": 1.943, + "step": 351 + }, + { + "epoch": 1.5603543743078627, + "grad_norm": 5.630266189575195, + "learning_rate": 8.114231775398617e-05, + "loss": 2.2482, + "step": 352 + }, + { + "epoch": 1.564784053156146, + "grad_norm": 10.885608673095703, + "learning_rate": 8.102954870781832e-05, + "loss": 1.8583, + "step": 353 + }, + { + "epoch": 1.5692137320044297, + "grad_norm": 6.998108863830566, + "learning_rate": 8.091652237017151e-05, + "loss": 1.9425, + "step": 354 + }, + { + "epoch": 1.5736434108527133, + "grad_norm": 8.657463073730469, + "learning_rate": 8.080323967823993e-05, + "loss": 1.9427, + "step": 355 + }, + { + "epoch": 1.5780730897009967, + "grad_norm": 11.812101364135742, + "learning_rate": 8.068970157134347e-05, + "loss": 2.1204, + "step": 356 + }, + { + "epoch": 1.58250276854928, + "grad_norm": 6.619724273681641, + "learning_rate": 8.057590899091984e-05, + "loss": 1.8593, + "step": 357 + }, + { + "epoch": 1.5869324473975637, + "grad_norm": 8.670092582702637, + "learning_rate": 8.046186288051681e-05, + "loss": 2.001, + "step": 358 + }, + { + "epoch": 1.5913621262458473, + "grad_norm": 8.840471267700195, + "learning_rate": 8.034756418578434e-05, + "loss": 1.852, + "step": 359 + }, + { + "epoch": 1.5957918050941307, + "grad_norm": 4.436490535736084, + "learning_rate": 8.023301385446682e-05, + "loss": 2.3419, + "step": 360 + }, + { + "epoch": 1.600221483942414, + "grad_norm": 9.356131553649902, + "learning_rate": 8.011821283639516e-05, + "loss": 1.8724, + "step": 361 + }, + { + "epoch": 1.6046511627906976, + "grad_norm": 9.546539306640625, + "learning_rate": 8.000316208347891e-05, + "loss": 1.8241, + "step": 362 + }, + { + "epoch": 1.6090808416389812, + "grad_norm": 12.444693565368652, + "learning_rate": 7.988786254969835e-05, + "loss": 2.2428, + "step": 363 + }, + { + "epoch": 1.6135105204872646, + "grad_norm": 8.764139175415039, + "learning_rate": 7.977231519109665e-05, + "loss": 1.7305, + "step": 364 + }, + { + "epoch": 1.6179401993355482, + "grad_norm": 8.302329063415527, + "learning_rate": 7.965652096577188e-05, + "loss": 2.3451, + "step": 365 + }, + { + "epoch": 1.6223698781838318, + "grad_norm": 13.382746696472168, + "learning_rate": 7.954048083386909e-05, + "loss": 1.8125, + "step": 366 + }, + { + "epoch": 1.6267995570321152, + "grad_norm": 6.948853015899658, + "learning_rate": 7.942419575757235e-05, + "loss": 1.9318, + "step": 367 + }, + { + "epoch": 1.6312292358803986, + "grad_norm": 6.62320613861084, + "learning_rate": 7.930766670109673e-05, + "loss": 1.9616, + "step": 368 + }, + { + "epoch": 1.6356589147286822, + "grad_norm": 6.08026123046875, + "learning_rate": 7.919089463068039e-05, + "loss": 1.9908, + "step": 369 + }, + { + "epoch": 1.6400885935769658, + "grad_norm": 9.736669540405273, + "learning_rate": 7.907388051457647e-05, + "loss": 1.9624, + "step": 370 + }, + { + "epoch": 1.6445182724252492, + "grad_norm": 7.714630126953125, + "learning_rate": 7.895662532304516e-05, + "loss": 1.8318, + "step": 371 + }, + { + "epoch": 1.6489479512735326, + "grad_norm": 9.872337341308594, + "learning_rate": 7.883913002834555e-05, + "loss": 2.0592, + "step": 372 + }, + { + "epoch": 1.6533776301218162, + "grad_norm": 10.687009811401367, + "learning_rate": 7.872139560472767e-05, + "loss": 1.9198, + "step": 373 + }, + { + "epoch": 1.6578073089700998, + "grad_norm": 7.267500400543213, + "learning_rate": 7.860342302842432e-05, + "loss": 1.5929, + "step": 374 + }, + { + "epoch": 1.6622369878183831, + "grad_norm": 9.017212867736816, + "learning_rate": 7.848521327764308e-05, + "loss": 1.6919, + "step": 375 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 7.3636860847473145, + "learning_rate": 7.836676733255809e-05, + "loss": 2.0157, + "step": 376 + }, + { + "epoch": 1.6710963455149501, + "grad_norm": 10.131976127624512, + "learning_rate": 7.824808617530197e-05, + "loss": 2.0387, + "step": 377 + }, + { + "epoch": 1.6755260243632337, + "grad_norm": 10.72778606414795, + "learning_rate": 7.81291707899577e-05, + "loss": 1.9898, + "step": 378 + }, + { + "epoch": 1.679955703211517, + "grad_norm": 8.190147399902344, + "learning_rate": 7.801002216255042e-05, + "loss": 1.9384, + "step": 379 + }, + { + "epoch": 1.6843853820598007, + "grad_norm": 8.989834785461426, + "learning_rate": 7.789064128103929e-05, + "loss": 1.9706, + "step": 380 + }, + { + "epoch": 1.6888150609080843, + "grad_norm": 6.380321025848389, + "learning_rate": 7.777102913530928e-05, + "loss": 2.0304, + "step": 381 + }, + { + "epoch": 1.6932447397563677, + "grad_norm": 6.1082258224487305, + "learning_rate": 7.76511867171629e-05, + "loss": 1.7976, + "step": 382 + }, + { + "epoch": 1.697674418604651, + "grad_norm": 8.573156356811523, + "learning_rate": 7.753111502031214e-05, + "loss": 1.59, + "step": 383 + }, + { + "epoch": 1.7021040974529347, + "grad_norm": 5.815869331359863, + "learning_rate": 7.741081504037008e-05, + "loss": 1.889, + "step": 384 + }, + { + "epoch": 1.7065337763012183, + "grad_norm": 8.750470161437988, + "learning_rate": 7.729028777484267e-05, + "loss": 1.7533, + "step": 385 + }, + { + "epoch": 1.7109634551495017, + "grad_norm": 12.4512300491333, + "learning_rate": 7.716953422312045e-05, + "loss": 2.0341, + "step": 386 + }, + { + "epoch": 1.715393133997785, + "grad_norm": 7.664438247680664, + "learning_rate": 7.704855538647033e-05, + "loss": 2.2063, + "step": 387 + }, + { + "epoch": 1.7198228128460686, + "grad_norm": 7.865803241729736, + "learning_rate": 7.692735226802728e-05, + "loss": 1.8626, + "step": 388 + }, + { + "epoch": 1.7242524916943522, + "grad_norm": 9.542865753173828, + "learning_rate": 7.680592587278585e-05, + "loss": 2.1088, + "step": 389 + }, + { + "epoch": 1.7286821705426356, + "grad_norm": 10.541374206542969, + "learning_rate": 7.668427720759207e-05, + "loss": 1.889, + "step": 390 + }, + { + "epoch": 1.733111849390919, + "grad_norm": 7.541441440582275, + "learning_rate": 7.656240728113493e-05, + "loss": 1.9824, + "step": 391 + }, + { + "epoch": 1.7375415282392026, + "grad_norm": 15.032777786254883, + "learning_rate": 7.644031710393815e-05, + "loss": 2.2166, + "step": 392 + }, + { + "epoch": 1.7419712070874862, + "grad_norm": 5.340550422668457, + "learning_rate": 7.631800768835166e-05, + "loss": 1.6734, + "step": 393 + }, + { + "epoch": 1.7464008859357696, + "grad_norm": 4.782181739807129, + "learning_rate": 7.619548004854333e-05, + "loss": 1.9972, + "step": 394 + }, + { + "epoch": 1.7508305647840532, + "grad_norm": 7.717755317687988, + "learning_rate": 7.607273520049041e-05, + "loss": 1.8174, + "step": 395 + }, + { + "epoch": 1.7552602436323368, + "grad_norm": 4.189438819885254, + "learning_rate": 7.594977416197133e-05, + "loss": 1.8464, + "step": 396 + }, + { + "epoch": 1.7596899224806202, + "grad_norm": 18.602073669433594, + "learning_rate": 7.582659795255706e-05, + "loss": 1.5871, + "step": 397 + }, + { + "epoch": 1.7641196013289036, + "grad_norm": 7.6945481300354, + "learning_rate": 7.570320759360273e-05, + "loss": 1.9943, + "step": 398 + }, + { + "epoch": 1.7685492801771872, + "grad_norm": 7.603806495666504, + "learning_rate": 7.557960410823917e-05, + "loss": 2.1524, + "step": 399 + }, + { + "epoch": 1.7729789590254708, + "grad_norm": 6.585241794586182, + "learning_rate": 7.545578852136443e-05, + "loss": 1.9966, + "step": 400 + }, + { + "epoch": 1.7774086378737541, + "grad_norm": 8.787303924560547, + "learning_rate": 7.533176185963523e-05, + "loss": 1.7846, + "step": 401 + }, + { + "epoch": 1.7818383167220375, + "grad_norm": 5.8807477951049805, + "learning_rate": 7.520752515145855e-05, + "loss": 1.7961, + "step": 402 + }, + { + "epoch": 1.7862679955703211, + "grad_norm": 11.430944442749023, + "learning_rate": 7.508307942698295e-05, + "loss": 1.5721, + "step": 403 + }, + { + "epoch": 1.7906976744186047, + "grad_norm": 9.947343826293945, + "learning_rate": 7.495842571809021e-05, + "loss": 1.8959, + "step": 404 + }, + { + "epoch": 1.795127353266888, + "grad_norm": 8.460419654846191, + "learning_rate": 7.483356505838659e-05, + "loss": 1.9603, + "step": 405 + }, + { + "epoch": 1.7995570321151715, + "grad_norm": 11.215185165405273, + "learning_rate": 7.470849848319444e-05, + "loss": 1.6616, + "step": 406 + }, + { + "epoch": 1.8039867109634553, + "grad_norm": 6.264514446258545, + "learning_rate": 7.458322702954341e-05, + "loss": 1.8412, + "step": 407 + }, + { + "epoch": 1.8084163898117387, + "grad_norm": 5.761272430419922, + "learning_rate": 7.44577517361621e-05, + "loss": 1.7221, + "step": 408 + }, + { + "epoch": 1.812846068660022, + "grad_norm": 5.823096752166748, + "learning_rate": 7.433207364346919e-05, + "loss": 1.8469, + "step": 409 + }, + { + "epoch": 1.8172757475083057, + "grad_norm": 7.28630256652832, + "learning_rate": 7.420619379356503e-05, + "loss": 1.8922, + "step": 410 + }, + { + "epoch": 1.8217054263565893, + "grad_norm": 5.539880275726318, + "learning_rate": 7.408011323022286e-05, + "loss": 1.9854, + "step": 411 + }, + { + "epoch": 1.8261351052048727, + "grad_norm": 6.604762077331543, + "learning_rate": 7.395383299888019e-05, + "loss": 1.927, + "step": 412 + }, + { + "epoch": 1.830564784053156, + "grad_norm": 6.813220977783203, + "learning_rate": 7.382735414663016e-05, + "loss": 1.9083, + "step": 413 + }, + { + "epoch": 1.8349944629014396, + "grad_norm": 4.795647621154785, + "learning_rate": 7.370067772221285e-05, + "loss": 1.9524, + "step": 414 + }, + { + "epoch": 1.8394241417497232, + "grad_norm": 7.515512466430664, + "learning_rate": 7.357380477600653e-05, + "loss": 2.0216, + "step": 415 + }, + { + "epoch": 1.8438538205980066, + "grad_norm": 10.164600372314453, + "learning_rate": 7.344673636001906e-05, + "loss": 2.2258, + "step": 416 + }, + { + "epoch": 1.84828349944629, + "grad_norm": 5.503859996795654, + "learning_rate": 7.331947352787905e-05, + "loss": 1.86, + "step": 417 + }, + { + "epoch": 1.8527131782945736, + "grad_norm": 6.191671371459961, + "learning_rate": 7.319201733482715e-05, + "loss": 1.9093, + "step": 418 + }, + { + "epoch": 1.8571428571428572, + "grad_norm": 7.223259925842285, + "learning_rate": 7.306436883770742e-05, + "loss": 2.1421, + "step": 419 + }, + { + "epoch": 1.8615725359911406, + "grad_norm": 7.5998992919921875, + "learning_rate": 7.293652909495836e-05, + "loss": 1.9365, + "step": 420 + }, + { + "epoch": 1.8660022148394242, + "grad_norm": 7.336147308349609, + "learning_rate": 7.280849916660434e-05, + "loss": 1.7375, + "step": 421 + }, + { + "epoch": 1.8704318936877078, + "grad_norm": 6.501189231872559, + "learning_rate": 7.268028011424663e-05, + "loss": 1.7906, + "step": 422 + }, + { + "epoch": 1.8748615725359912, + "grad_norm": 6.481664180755615, + "learning_rate": 7.255187300105476e-05, + "loss": 1.8908, + "step": 423 + }, + { + "epoch": 1.8792912513842746, + "grad_norm": 8.071866035461426, + "learning_rate": 7.24232788917576e-05, + "loss": 2.0077, + "step": 424 + }, + { + "epoch": 1.8837209302325582, + "grad_norm": 5.397129535675049, + "learning_rate": 7.22944988526345e-05, + "loss": 1.9436, + "step": 425 + }, + { + "epoch": 1.8881506090808418, + "grad_norm": 6.841665267944336, + "learning_rate": 7.21655339515066e-05, + "loss": 1.9098, + "step": 426 + }, + { + "epoch": 1.8925802879291251, + "grad_norm": 5.7756195068359375, + "learning_rate": 7.203638525772782e-05, + "loss": 1.8002, + "step": 427 + }, + { + "epoch": 1.8970099667774085, + "grad_norm": 5.458576679229736, + "learning_rate": 7.190705384217608e-05, + "loss": 1.6935, + "step": 428 + }, + { + "epoch": 1.9014396456256921, + "grad_norm": 5.86073112487793, + "learning_rate": 7.17775407772444e-05, + "loss": 2.2811, + "step": 429 + }, + { + "epoch": 1.9058693244739757, + "grad_norm": 6.207028388977051, + "learning_rate": 7.164784713683197e-05, + "loss": 1.7852, + "step": 430 + }, + { + "epoch": 1.910299003322259, + "grad_norm": 9.03230094909668, + "learning_rate": 7.151797399633533e-05, + "loss": 2.3004, + "step": 431 + }, + { + "epoch": 1.9147286821705425, + "grad_norm": 10.938101768493652, + "learning_rate": 7.138792243263936e-05, + "loss": 1.9295, + "step": 432 + }, + { + "epoch": 1.919158361018826, + "grad_norm": 4.8152995109558105, + "learning_rate": 7.125769352410845e-05, + "loss": 2.0202, + "step": 433 + }, + { + "epoch": 1.9235880398671097, + "grad_norm": 7.909144878387451, + "learning_rate": 7.112728835057741e-05, + "loss": 1.7597, + "step": 434 + }, + { + "epoch": 1.928017718715393, + "grad_norm": 7.012314319610596, + "learning_rate": 7.099670799334269e-05, + "loss": 2.0388, + "step": 435 + }, + { + "epoch": 1.9324473975636767, + "grad_norm": 8.393206596374512, + "learning_rate": 7.08659535351533e-05, + "loss": 1.9736, + "step": 436 + }, + { + "epoch": 1.9368770764119603, + "grad_norm": 10.293094635009766, + "learning_rate": 7.073502606020187e-05, + "loss": 2.1422, + "step": 437 + }, + { + "epoch": 1.9413067552602437, + "grad_norm": 9.445409774780273, + "learning_rate": 7.060392665411564e-05, + "loss": 1.6767, + "step": 438 + }, + { + "epoch": 1.945736434108527, + "grad_norm": 7.761300086975098, + "learning_rate": 7.04726564039475e-05, + "loss": 1.7759, + "step": 439 + }, + { + "epoch": 1.9501661129568106, + "grad_norm": 8.283308982849121, + "learning_rate": 7.03412163981669e-05, + "loss": 1.9943, + "step": 440 + }, + { + "epoch": 1.9545957918050942, + "grad_norm": 5.032713413238525, + "learning_rate": 7.020960772665096e-05, + "loss": 2.0963, + "step": 441 + }, + { + "epoch": 1.9590254706533776, + "grad_norm": 6.809930324554443, + "learning_rate": 7.007783148067523e-05, + "loss": 1.9811, + "step": 442 + }, + { + "epoch": 1.963455149501661, + "grad_norm": 8.002426147460938, + "learning_rate": 6.994588875290487e-05, + "loss": 2.2668, + "step": 443 + }, + { + "epoch": 1.9678848283499446, + "grad_norm": 11.3838529586792, + "learning_rate": 6.981378063738539e-05, + "loss": 2.0145, + "step": 444 + }, + { + "epoch": 1.9723145071982282, + "grad_norm": 14.399958610534668, + "learning_rate": 6.968150822953372e-05, + "loss": 2.1096, + "step": 445 + }, + { + "epoch": 1.9767441860465116, + "grad_norm": 4.975661277770996, + "learning_rate": 6.954907262612905e-05, + "loss": 1.6861, + "step": 446 + }, + { + "epoch": 1.981173864894795, + "grad_norm": 11.665627479553223, + "learning_rate": 6.941647492530377e-05, + "loss": 1.861, + "step": 447 + }, + { + "epoch": 1.9856035437430788, + "grad_norm": 5.3779706954956055, + "learning_rate": 6.928371622653433e-05, + "loss": 1.7886, + "step": 448 + }, + { + "epoch": 1.9900332225913622, + "grad_norm": 6.151978492736816, + "learning_rate": 6.91507976306322e-05, + "loss": 1.8549, + "step": 449 + }, + { + "epoch": 1.9944629014396456, + "grad_norm": 6.620397567749023, + "learning_rate": 6.901772023973459e-05, + "loss": 1.8999, + "step": 450 + }, + { + "epoch": 1.9988925802879292, + "grad_norm": 14.254983901977539, + "learning_rate": 6.888448515729552e-05, + "loss": 2.0419, + "step": 451 + }, + { + "epoch": 2.0044296788482834, + "grad_norm": 5.509189605712891, + "learning_rate": 6.875109348807649e-05, + "loss": 2.1954, + "step": 452 + }, + { + "epoch": 2.008859357696567, + "grad_norm": 5.627884864807129, + "learning_rate": 6.861754633813739e-05, + "loss": 1.8798, + "step": 453 + }, + { + "epoch": 2.0132890365448506, + "grad_norm": 9.019150733947754, + "learning_rate": 6.84838448148274e-05, + "loss": 2.1152, + "step": 454 + }, + { + "epoch": 2.017718715393134, + "grad_norm": 7.247636318206787, + "learning_rate": 6.834999002677565e-05, + "loss": 1.7317, + "step": 455 + }, + { + "epoch": 2.0221483942414173, + "grad_norm": 7.318861961364746, + "learning_rate": 6.821598308388216e-05, + "loss": 1.6497, + "step": 456 + }, + { + "epoch": 2.026578073089701, + "grad_norm": 7.973992347717285, + "learning_rate": 6.808182509730858e-05, + "loss": 2.0685, + "step": 457 + }, + { + "epoch": 2.0310077519379846, + "grad_norm": 6.912788391113281, + "learning_rate": 6.794751717946896e-05, + "loss": 1.8776, + "step": 458 + }, + { + "epoch": 2.035437430786268, + "grad_norm": 9.139986038208008, + "learning_rate": 6.781306044402063e-05, + "loss": 1.929, + "step": 459 + }, + { + "epoch": 2.0398671096345513, + "grad_norm": 10.722004890441895, + "learning_rate": 6.767845600585479e-05, + "loss": 1.9654, + "step": 460 + }, + { + "epoch": 2.044296788482835, + "grad_norm": 6.275202751159668, + "learning_rate": 6.754370498108746e-05, + "loss": 1.8465, + "step": 461 + }, + { + "epoch": 2.0487264673311185, + "grad_norm": 6.970125675201416, + "learning_rate": 6.740880848705004e-05, + "loss": 1.9397, + "step": 462 + }, + { + "epoch": 2.053156146179402, + "grad_norm": 6.303622245788574, + "learning_rate": 6.727376764228019e-05, + "loss": 1.7633, + "step": 463 + }, + { + "epoch": 2.0575858250276853, + "grad_norm": 6.273361682891846, + "learning_rate": 6.713858356651253e-05, + "loss": 1.748, + "step": 464 + }, + { + "epoch": 2.062015503875969, + "grad_norm": 10.793755531311035, + "learning_rate": 6.700325738066922e-05, + "loss": 1.7963, + "step": 465 + }, + { + "epoch": 2.0664451827242525, + "grad_norm": 6.142402172088623, + "learning_rate": 6.686779020685089e-05, + "loss": 1.7866, + "step": 466 + }, + { + "epoch": 2.070874861572536, + "grad_norm": 8.846867561340332, + "learning_rate": 6.673218316832714e-05, + "loss": 1.7278, + "step": 467 + }, + { + "epoch": 2.0753045404208197, + "grad_norm": 6.133031368255615, + "learning_rate": 6.659643738952732e-05, + "loss": 2.0824, + "step": 468 + }, + { + "epoch": 2.079734219269103, + "grad_norm": 6.360471725463867, + "learning_rate": 6.646055399603122e-05, + "loss": 1.8149, + "step": 469 + }, + { + "epoch": 2.0841638981173864, + "grad_norm": 10.420990943908691, + "learning_rate": 6.632453411455965e-05, + "loss": 1.7843, + "step": 470 + }, + { + "epoch": 2.08859357696567, + "grad_norm": 10.17164134979248, + "learning_rate": 6.618837887296522e-05, + "loss": 1.6479, + "step": 471 + }, + { + "epoch": 2.0930232558139537, + "grad_norm": 6.589805603027344, + "learning_rate": 6.60520894002229e-05, + "loss": 1.7869, + "step": 472 + }, + { + "epoch": 2.097452934662237, + "grad_norm": 15.679412841796875, + "learning_rate": 6.591566682642061e-05, + "loss": 1.9296, + "step": 473 + }, + { + "epoch": 2.1018826135105204, + "grad_norm": 5.550188064575195, + "learning_rate": 6.577911228275003e-05, + "loss": 1.5573, + "step": 474 + }, + { + "epoch": 2.106312292358804, + "grad_norm": 4.819809913635254, + "learning_rate": 6.564242690149704e-05, + "loss": 1.4723, + "step": 475 + }, + { + "epoch": 2.1107419712070876, + "grad_norm": 7.867583274841309, + "learning_rate": 6.550561181603244e-05, + "loss": 1.9197, + "step": 476 + }, + { + "epoch": 2.115171650055371, + "grad_norm": 7.505444526672363, + "learning_rate": 6.536866816080247e-05, + "loss": 1.96, + "step": 477 + }, + { + "epoch": 2.1196013289036544, + "grad_norm": 9.75400447845459, + "learning_rate": 6.523159707131951e-05, + "loss": 1.6, + "step": 478 + }, + { + "epoch": 2.124031007751938, + "grad_norm": 4.644130229949951, + "learning_rate": 6.509439968415253e-05, + "loss": 1.512, + "step": 479 + }, + { + "epoch": 2.1284606866002216, + "grad_norm": 5.214409828186035, + "learning_rate": 6.495707713691778e-05, + "loss": 1.7648, + "step": 480 + }, + { + "epoch": 2.132890365448505, + "grad_norm": 10.711379051208496, + "learning_rate": 6.481963056826932e-05, + "loss": 1.7958, + "step": 481 + }, + { + "epoch": 2.1373200442967883, + "grad_norm": 7.786211013793945, + "learning_rate": 6.468206111788957e-05, + "loss": 2.0541, + "step": 482 + }, + { + "epoch": 2.141749723145072, + "grad_norm": 9.271305084228516, + "learning_rate": 6.454436992647984e-05, + "loss": 1.9129, + "step": 483 + }, + { + "epoch": 2.1461794019933556, + "grad_norm": 9.571303367614746, + "learning_rate": 6.440655813575093e-05, + "loss": 1.6192, + "step": 484 + }, + { + "epoch": 2.150609080841639, + "grad_norm": 6.501114368438721, + "learning_rate": 6.426862688841359e-05, + "loss": 1.5886, + "step": 485 + }, + { + "epoch": 2.1550387596899223, + "grad_norm": 9.78244400024414, + "learning_rate": 6.41305773281691e-05, + "loss": 1.7534, + "step": 486 + }, + { + "epoch": 2.159468438538206, + "grad_norm": 6.028672218322754, + "learning_rate": 6.399241059969978e-05, + "loss": 1.7561, + "step": 487 + }, + { + "epoch": 2.1638981173864895, + "grad_norm": 6.525522708892822, + "learning_rate": 6.385412784865948e-05, + "loss": 1.9586, + "step": 488 + }, + { + "epoch": 2.168327796234773, + "grad_norm": 13.112713813781738, + "learning_rate": 6.371573022166408e-05, + "loss": 1.8899, + "step": 489 + }, + { + "epoch": 2.1727574750830563, + "grad_norm": 5.4037394523620605, + "learning_rate": 6.357721886628201e-05, + "loss": 1.7339, + "step": 490 + }, + { + "epoch": 2.17718715393134, + "grad_norm": 7.634363651275635, + "learning_rate": 6.34385949310247e-05, + "loss": 1.6131, + "step": 491 + }, + { + "epoch": 2.1816168327796235, + "grad_norm": 6.406160354614258, + "learning_rate": 6.329985956533707e-05, + "loss": 1.8587, + "step": 492 + }, + { + "epoch": 2.186046511627907, + "grad_norm": 6.028163909912109, + "learning_rate": 6.3161013919588e-05, + "loss": 1.6508, + "step": 493 + }, + { + "epoch": 2.1904761904761907, + "grad_norm": 7.243360996246338, + "learning_rate": 6.302205914506083e-05, + "loss": 1.8309, + "step": 494 + }, + { + "epoch": 2.194905869324474, + "grad_norm": 11.702513694763184, + "learning_rate": 6.288299639394371e-05, + "loss": 1.5447, + "step": 495 + }, + { + "epoch": 2.1993355481727574, + "grad_norm": 6.540017604827881, + "learning_rate": 6.274382681932018e-05, + "loss": 1.9629, + "step": 496 + }, + { + "epoch": 2.203765227021041, + "grad_norm": 7.083324432373047, + "learning_rate": 6.260455157515947e-05, + "loss": 1.8167, + "step": 497 + }, + { + "epoch": 2.2081949058693247, + "grad_norm": 4.912618160247803, + "learning_rate": 6.24651718163071e-05, + "loss": 1.88, + "step": 498 + }, + { + "epoch": 2.212624584717608, + "grad_norm": 7.98961067199707, + "learning_rate": 6.23256886984751e-05, + "loss": 1.9315, + "step": 499 + }, + { + "epoch": 2.2170542635658914, + "grad_norm": 8.828598976135254, + "learning_rate": 6.218610337823261e-05, + "loss": 2.0435, + "step": 500 + }, + { + "epoch": 2.221483942414175, + "grad_norm": 4.62863826751709, + "learning_rate": 6.204641701299619e-05, + "loss": 1.92, + "step": 501 + }, + { + "epoch": 2.2259136212624586, + "grad_norm": 6.152403354644775, + "learning_rate": 6.190663076102025e-05, + "loss": 1.8263, + "step": 502 + }, + { + "epoch": 2.230343300110742, + "grad_norm": 4.635473728179932, + "learning_rate": 6.17667457813874e-05, + "loss": 1.8494, + "step": 503 + }, + { + "epoch": 2.2347729789590254, + "grad_norm": 5.960703372955322, + "learning_rate": 6.162676323399897e-05, + "loss": 2.1242, + "step": 504 + }, + { + "epoch": 2.2392026578073088, + "grad_norm": 7.865222454071045, + "learning_rate": 6.148668427956524e-05, + "loss": 1.9256, + "step": 505 + }, + { + "epoch": 2.2436323366555926, + "grad_norm": 14.765120506286621, + "learning_rate": 6.134651007959586e-05, + "loss": 2.0467, + "step": 506 + }, + { + "epoch": 2.248062015503876, + "grad_norm": 9.482008934020996, + "learning_rate": 6.120624179639032e-05, + "loss": 1.7785, + "step": 507 + }, + { + "epoch": 2.2524916943521593, + "grad_norm": 5.620559215545654, + "learning_rate": 6.106588059302818e-05, + "loss": 1.9651, + "step": 508 + }, + { + "epoch": 2.256921373200443, + "grad_norm": 5.738358497619629, + "learning_rate": 6.0925427633359466e-05, + "loss": 1.5501, + "step": 509 + }, + { + "epoch": 2.2613510520487266, + "grad_norm": 8.628968238830566, + "learning_rate": 6.0784884081995064e-05, + "loss": 2.002, + "step": 510 + }, + { + "epoch": 2.26578073089701, + "grad_norm": 8.108263969421387, + "learning_rate": 6.064425110429699e-05, + "loss": 2.0662, + "step": 511 + }, + { + "epoch": 2.2702104097452933, + "grad_norm": 10.603029251098633, + "learning_rate": 6.0503529866368826e-05, + "loss": 1.6147, + "step": 512 + }, + { + "epoch": 2.274640088593577, + "grad_norm": 9.906684875488281, + "learning_rate": 6.036272153504592e-05, + "loss": 1.7281, + "step": 513 + }, + { + "epoch": 2.2790697674418605, + "grad_norm": 10.202539443969727, + "learning_rate": 6.0221827277885856e-05, + "loss": 1.7101, + "step": 514 + }, + { + "epoch": 2.283499446290144, + "grad_norm": 9.090194702148438, + "learning_rate": 6.008084826315863e-05, + "loss": 1.997, + "step": 515 + }, + { + "epoch": 2.2879291251384273, + "grad_norm": 9.38204574584961, + "learning_rate": 5.993978565983709e-05, + "loss": 1.7423, + "step": 516 + }, + { + "epoch": 2.292358803986711, + "grad_norm": 10.475700378417969, + "learning_rate": 5.979864063758717e-05, + "loss": 1.6178, + "step": 517 + }, + { + "epoch": 2.2967884828349945, + "grad_norm": 11.60288143157959, + "learning_rate": 5.965741436675816e-05, + "loss": 1.6708, + "step": 518 + }, + { + "epoch": 2.301218161683278, + "grad_norm": 4.961325168609619, + "learning_rate": 5.951610801837314e-05, + "loss": 1.585, + "step": 519 + }, + { + "epoch": 2.3056478405315612, + "grad_norm": 7.948910236358643, + "learning_rate": 5.937472276411909e-05, + "loss": 1.9334, + "step": 520 + }, + { + "epoch": 2.310077519379845, + "grad_norm": 12.114509582519531, + "learning_rate": 5.9233259776337316e-05, + "loss": 1.6873, + "step": 521 + }, + { + "epoch": 2.3145071982281284, + "grad_norm": 5.8998494148254395, + "learning_rate": 5.9091720228013636e-05, + "loss": 1.9225, + "step": 522 + }, + { + "epoch": 2.318936877076412, + "grad_norm": 5.098598480224609, + "learning_rate": 5.8950105292768756e-05, + "loss": 1.6197, + "step": 523 + }, + { + "epoch": 2.3233665559246957, + "grad_norm": 5.198586463928223, + "learning_rate": 5.880841614484841e-05, + "loss": 1.6607, + "step": 524 + }, + { + "epoch": 2.327796234772979, + "grad_norm": 10.038392066955566, + "learning_rate": 5.8666653959113746e-05, + "loss": 1.9545, + "step": 525 + }, + { + "epoch": 2.3322259136212624, + "grad_norm": 3.849628448486328, + "learning_rate": 5.8524819911031495e-05, + "loss": 1.6751, + "step": 526 + }, + { + "epoch": 2.336655592469546, + "grad_norm": 11.999637603759766, + "learning_rate": 5.838291517666427e-05, + "loss": 2.0971, + "step": 527 + }, + { + "epoch": 2.3410852713178296, + "grad_norm": 6.144494533538818, + "learning_rate": 5.824094093266077e-05, + "loss": 1.6829, + "step": 528 + }, + { + "epoch": 2.345514950166113, + "grad_norm": 7.545333385467529, + "learning_rate": 5.8098898356246115e-05, + "loss": 1.7714, + "step": 529 + }, + { + "epoch": 2.3499446290143964, + "grad_norm": 8.580190658569336, + "learning_rate": 5.795678862521197e-05, + "loss": 1.9958, + "step": 530 + }, + { + "epoch": 2.35437430786268, + "grad_norm": 7.868439197540283, + "learning_rate": 5.781461291790688e-05, + "loss": 2.0649, + "step": 531 + }, + { + "epoch": 2.3588039867109636, + "grad_norm": 4.296184539794922, + "learning_rate": 5.76723724132264e-05, + "loss": 1.7261, + "step": 532 + }, + { + "epoch": 2.363233665559247, + "grad_norm": 7.093977928161621, + "learning_rate": 5.753006829060344e-05, + "loss": 2.0423, + "step": 533 + }, + { + "epoch": 2.3676633444075303, + "grad_norm": 8.066122055053711, + "learning_rate": 5.7387701729998346e-05, + "loss": 1.8569, + "step": 534 + }, + { + "epoch": 2.3720930232558137, + "grad_norm": 9.265979766845703, + "learning_rate": 5.7245273911889276e-05, + "loss": 1.7729, + "step": 535 + }, + { + "epoch": 2.3765227021040976, + "grad_norm": 9.756134033203125, + "learning_rate": 5.710278601726222e-05, + "loss": 1.9864, + "step": 536 + }, + { + "epoch": 2.380952380952381, + "grad_norm": 9.198343276977539, + "learning_rate": 5.696023922760141e-05, + "loss": 1.7085, + "step": 537 + }, + { + "epoch": 2.3853820598006643, + "grad_norm": 7.285330295562744, + "learning_rate": 5.6817634724879333e-05, + "loss": 2.2659, + "step": 538 + }, + { + "epoch": 2.389811738648948, + "grad_norm": 11.868597030639648, + "learning_rate": 5.6674973691547115e-05, + "loss": 1.5843, + "step": 539 + }, + { + "epoch": 2.3942414174972315, + "grad_norm": 5.195461750030518, + "learning_rate": 5.653225731052456e-05, + "loss": 1.4188, + "step": 540 + }, + { + "epoch": 2.398671096345515, + "grad_norm": 6.345346450805664, + "learning_rate": 5.638948676519042e-05, + "loss": 2.1584, + "step": 541 + }, + { + "epoch": 2.4031007751937983, + "grad_norm": 10.57470703125, + "learning_rate": 5.6246663239372565e-05, + "loss": 1.6626, + "step": 542 + }, + { + "epoch": 2.407530454042082, + "grad_norm": 11.943028450012207, + "learning_rate": 5.6103787917338214e-05, + "loss": 1.959, + "step": 543 + }, + { + "epoch": 2.4119601328903655, + "grad_norm": 12.070340156555176, + "learning_rate": 5.5960861983783985e-05, + "loss": 2.0654, + "step": 544 + }, + { + "epoch": 2.416389811738649, + "grad_norm": 5.627765655517578, + "learning_rate": 5.581788662382624e-05, + "loss": 2.0118, + "step": 545 + }, + { + "epoch": 2.4208194905869327, + "grad_norm": 10.758605003356934, + "learning_rate": 5.567486302299112e-05, + "loss": 1.7541, + "step": 546 + }, + { + "epoch": 2.425249169435216, + "grad_norm": 6.83001708984375, + "learning_rate": 5.5531792367204816e-05, + "loss": 2.1466, + "step": 547 + }, + { + "epoch": 2.4296788482834994, + "grad_norm": 5.710721492767334, + "learning_rate": 5.5388675842783646e-05, + "loss": 1.6117, + "step": 548 + }, + { + "epoch": 2.434108527131783, + "grad_norm": 6.419375896453857, + "learning_rate": 5.5245514636424286e-05, + "loss": 2.098, + "step": 549 + }, + { + "epoch": 2.438538205980066, + "grad_norm": 6.893190860748291, + "learning_rate": 5.5102309935193906e-05, + "loss": 1.5716, + "step": 550 + }, + { + "epoch": 2.44296788482835, + "grad_norm": 10.09901237487793, + "learning_rate": 5.4959062926520345e-05, + "loss": 1.7729, + "step": 551 + }, + { + "epoch": 2.4473975636766334, + "grad_norm": 10.772427558898926, + "learning_rate": 5.481577479818219e-05, + "loss": 2.0269, + "step": 552 + }, + { + "epoch": 2.451827242524917, + "grad_norm": 10.259986877441406, + "learning_rate": 5.467244673829908e-05, + "loss": 2.1173, + "step": 553 + }, + { + "epoch": 2.4562569213732006, + "grad_norm": 4.576322078704834, + "learning_rate": 5.452907993532164e-05, + "loss": 1.9707, + "step": 554 + }, + { + "epoch": 2.460686600221484, + "grad_norm": 6.365279674530029, + "learning_rate": 5.438567557802186e-05, + "loss": 1.8728, + "step": 555 + }, + { + "epoch": 2.4651162790697674, + "grad_norm": 14.015559196472168, + "learning_rate": 5.424223485548303e-05, + "loss": 1.8777, + "step": 556 + }, + { + "epoch": 2.4695459579180508, + "grad_norm": 7.423595428466797, + "learning_rate": 5.4098758957090055e-05, + "loss": 1.7334, + "step": 557 + }, + { + "epoch": 2.4739756367663346, + "grad_norm": 6.120731830596924, + "learning_rate": 5.395524907251944e-05, + "loss": 1.8194, + "step": 558 + }, + { + "epoch": 2.478405315614618, + "grad_norm": 5.712774276733398, + "learning_rate": 5.3811706391729545e-05, + "loss": 1.8025, + "step": 559 + }, + { + "epoch": 2.4828349944629013, + "grad_norm": 5.074977397918701, + "learning_rate": 5.366813210495067e-05, + "loss": 1.6992, + "step": 560 + }, + { + "epoch": 2.487264673311185, + "grad_norm": 5.497018814086914, + "learning_rate": 5.352452740267515e-05, + "loss": 1.843, + "step": 561 + }, + { + "epoch": 2.4916943521594686, + "grad_norm": 6.377771854400635, + "learning_rate": 5.338089347564757e-05, + "loss": 1.7029, + "step": 562 + }, + { + "epoch": 2.496124031007752, + "grad_norm": 5.825841426849365, + "learning_rate": 5.323723151485477e-05, + "loss": 2.1683, + "step": 563 + }, + { + "epoch": 2.5005537098560353, + "grad_norm": 9.456023216247559, + "learning_rate": 5.309354271151613e-05, + "loss": 1.7453, + "step": 564 + }, + { + "epoch": 2.5049833887043187, + "grad_norm": 9.619424819946289, + "learning_rate": 5.294982825707352e-05, + "loss": 1.6813, + "step": 565 + }, + { + "epoch": 2.5094130675526025, + "grad_norm": 9.673927307128906, + "learning_rate": 5.2806089343181564e-05, + "loss": 1.5608, + "step": 566 + }, + { + "epoch": 2.513842746400886, + "grad_norm": 8.455759048461914, + "learning_rate": 5.266232716169769e-05, + "loss": 2.1658, + "step": 567 + }, + { + "epoch": 2.5182724252491693, + "grad_norm": 6.917920112609863, + "learning_rate": 5.251854290467221e-05, + "loss": 1.6381, + "step": 568 + }, + { + "epoch": 2.522702104097453, + "grad_norm": 17.744230270385742, + "learning_rate": 5.237473776433853e-05, + "loss": 2.1334, + "step": 569 + }, + { + "epoch": 2.5271317829457365, + "grad_norm": 8.317569732666016, + "learning_rate": 5.223091293310324e-05, + "loss": 1.6458, + "step": 570 + }, + { + "epoch": 2.53156146179402, + "grad_norm": 5.05155611038208, + "learning_rate": 5.208706960353611e-05, + "loss": 1.8317, + "step": 571 + }, + { + "epoch": 2.5359911406423032, + "grad_norm": 9.146858215332031, + "learning_rate": 5.194320896836039e-05, + "loss": 1.7776, + "step": 572 + }, + { + "epoch": 2.540420819490587, + "grad_norm": 8.478264808654785, + "learning_rate": 5.1799332220442776e-05, + "loss": 1.6552, + "step": 573 + }, + { + "epoch": 2.5448504983388704, + "grad_norm": 11.424778938293457, + "learning_rate": 5.1655440552783584e-05, + "loss": 1.6622, + "step": 574 + }, + { + "epoch": 2.549280177187154, + "grad_norm": 5.402507305145264, + "learning_rate": 5.151153515850682e-05, + "loss": 2.3321, + "step": 575 + }, + { + "epoch": 2.5537098560354377, + "grad_norm": 11.693721771240234, + "learning_rate": 5.136761723085035e-05, + "loss": 1.5945, + "step": 576 + }, + { + "epoch": 2.558139534883721, + "grad_norm": 7.056944847106934, + "learning_rate": 5.1223687963155906e-05, + "loss": 2.2514, + "step": 577 + }, + { + "epoch": 2.5625692137320044, + "grad_norm": 5.904301643371582, + "learning_rate": 5.107974854885933e-05, + "loss": 1.5878, + "step": 578 + }, + { + "epoch": 2.566998892580288, + "grad_norm": 10.586432456970215, + "learning_rate": 5.093580018148052e-05, + "loss": 1.5959, + "step": 579 + }, + { + "epoch": 2.571428571428571, + "grad_norm": 8.07880973815918, + "learning_rate": 5.0791844054613646e-05, + "loss": 1.874, + "step": 580 + }, + { + "epoch": 2.575858250276855, + "grad_norm": 8.128579139709473, + "learning_rate": 5.0647881361917224e-05, + "loss": 1.8369, + "step": 581 + }, + { + "epoch": 2.5802879291251384, + "grad_norm": 7.19783878326416, + "learning_rate": 5.05039132971042e-05, + "loss": 1.3947, + "step": 582 + }, + { + "epoch": 2.584717607973422, + "grad_norm": 13.905142784118652, + "learning_rate": 5.0359941053932056e-05, + "loss": 2.0522, + "step": 583 + }, + { + "epoch": 2.5891472868217056, + "grad_norm": 10.910123825073242, + "learning_rate": 5.0215965826192954e-05, + "loss": 2.1368, + "step": 584 + }, + { + "epoch": 2.593576965669989, + "grad_norm": 4.5783867835998535, + "learning_rate": 5.0071988807703776e-05, + "loss": 1.7897, + "step": 585 + }, + { + "epoch": 2.5980066445182723, + "grad_norm": 8.571553230285645, + "learning_rate": 4.9928011192296236e-05, + "loss": 2.1332, + "step": 586 + }, + { + "epoch": 2.6024363233665557, + "grad_norm": 8.85322380065918, + "learning_rate": 4.978403417380706e-05, + "loss": 1.4058, + "step": 587 + }, + { + "epoch": 2.6068660022148396, + "grad_norm": 8.1622896194458, + "learning_rate": 4.964005894606795e-05, + "loss": 1.6939, + "step": 588 + }, + { + "epoch": 2.611295681063123, + "grad_norm": 9.46420669555664, + "learning_rate": 4.949608670289582e-05, + "loss": 2.0323, + "step": 589 + }, + { + "epoch": 2.6157253599114063, + "grad_norm": 6.406886577606201, + "learning_rate": 4.935211863808279e-05, + "loss": 1.9075, + "step": 590 + }, + { + "epoch": 2.62015503875969, + "grad_norm": 5.835373878479004, + "learning_rate": 4.9208155945386345e-05, + "loss": 2.0062, + "step": 591 + }, + { + "epoch": 2.6245847176079735, + "grad_norm": 6.667347431182861, + "learning_rate": 4.9064199818519495e-05, + "loss": 1.8224, + "step": 592 + }, + { + "epoch": 2.629014396456257, + "grad_norm": 5.613856315612793, + "learning_rate": 4.892025145114067e-05, + "loss": 1.5333, + "step": 593 + }, + { + "epoch": 2.6334440753045403, + "grad_norm": 4.702713489532471, + "learning_rate": 4.87763120368441e-05, + "loss": 1.4557, + "step": 594 + }, + { + "epoch": 2.6378737541528237, + "grad_norm": 8.00310230255127, + "learning_rate": 4.863238276914966e-05, + "loss": 1.6452, + "step": 595 + }, + { + "epoch": 2.6423034330011075, + "grad_norm": 15.650694847106934, + "learning_rate": 4.84884648414932e-05, + "loss": 1.8716, + "step": 596 + }, + { + "epoch": 2.646733111849391, + "grad_norm": 9.945942878723145, + "learning_rate": 4.834455944721643e-05, + "loss": 1.9081, + "step": 597 + }, + { + "epoch": 2.6511627906976747, + "grad_norm": 11.016633987426758, + "learning_rate": 4.820066777955724e-05, + "loss": 1.7934, + "step": 598 + }, + { + "epoch": 2.655592469545958, + "grad_norm": 7.862950325012207, + "learning_rate": 4.8056791031639614e-05, + "loss": 1.9754, + "step": 599 + }, + { + "epoch": 2.6600221483942414, + "grad_norm": 5.507509708404541, + "learning_rate": 4.791293039646389e-05, + "loss": 1.9389, + "step": 600 + }, + { + "epoch": 2.664451827242525, + "grad_norm": 3.8648977279663086, + "learning_rate": 4.776908706689678e-05, + "loss": 1.4915, + "step": 601 + }, + { + "epoch": 2.668881506090808, + "grad_norm": 6.0287699699401855, + "learning_rate": 4.762526223566147e-05, + "loss": 1.8027, + "step": 602 + }, + { + "epoch": 2.673311184939092, + "grad_norm": 7.0374226570129395, + "learning_rate": 4.7481457095327804e-05, + "loss": 1.7526, + "step": 603 + }, + { + "epoch": 2.6777408637873754, + "grad_norm": 6.987399578094482, + "learning_rate": 4.733767283830233e-05, + "loss": 1.9017, + "step": 604 + }, + { + "epoch": 2.682170542635659, + "grad_norm": 10.102921485900879, + "learning_rate": 4.719391065681845e-05, + "loss": 1.7473, + "step": 605 + }, + { + "epoch": 2.6866002214839426, + "grad_norm": 5.184350967407227, + "learning_rate": 4.705017174292649e-05, + "loss": 1.5428, + "step": 606 + }, + { + "epoch": 2.691029900332226, + "grad_norm": 6.327332973480225, + "learning_rate": 4.6906457288483895e-05, + "loss": 2.0073, + "step": 607 + }, + { + "epoch": 2.6954595791805094, + "grad_norm": 4.9231672286987305, + "learning_rate": 4.676276848514524e-05, + "loss": 1.5135, + "step": 608 + }, + { + "epoch": 2.6998892580287928, + "grad_norm": 4.648147106170654, + "learning_rate": 4.6619106524352445e-05, + "loss": 1.7111, + "step": 609 + }, + { + "epoch": 2.704318936877076, + "grad_norm": 12.321041107177734, + "learning_rate": 4.647547259732486e-05, + "loss": 1.6586, + "step": 610 + }, + { + "epoch": 2.70874861572536, + "grad_norm": 10.489548683166504, + "learning_rate": 4.633186789504933e-05, + "loss": 1.9735, + "step": 611 + }, + { + "epoch": 2.7131782945736433, + "grad_norm": 7.285097599029541, + "learning_rate": 4.618829360827046e-05, + "loss": 1.9382, + "step": 612 + }, + { + "epoch": 2.717607973421927, + "grad_norm": 5.02518892288208, + "learning_rate": 4.604475092748057e-05, + "loss": 1.6625, + "step": 613 + }, + { + "epoch": 2.7220376522702106, + "grad_norm": 6.852099895477295, + "learning_rate": 4.590124104290997e-05, + "loss": 2.0571, + "step": 614 + }, + { + "epoch": 2.726467331118494, + "grad_norm": 15.236705780029297, + "learning_rate": 4.575776514451698e-05, + "loss": 1.7972, + "step": 615 + }, + { + "epoch": 2.7308970099667773, + "grad_norm": 7.635733127593994, + "learning_rate": 4.561432442197817e-05, + "loss": 1.7937, + "step": 616 + }, + { + "epoch": 2.7353266888150607, + "grad_norm": 7.20578670501709, + "learning_rate": 4.547092006467837e-05, + "loss": 1.8999, + "step": 617 + }, + { + "epoch": 2.7397563676633445, + "grad_norm": 6.789028644561768, + "learning_rate": 4.532755326170093e-05, + "loss": 1.816, + "step": 618 + }, + { + "epoch": 2.744186046511628, + "grad_norm": 5.359363555908203, + "learning_rate": 4.518422520181781e-05, + "loss": 1.9299, + "step": 619 + }, + { + "epoch": 2.7486157253599113, + "grad_norm": 14.938844680786133, + "learning_rate": 4.504093707347966e-05, + "loss": 1.7996, + "step": 620 + }, + { + "epoch": 2.753045404208195, + "grad_norm": 11.183375358581543, + "learning_rate": 4.4897690064806106e-05, + "loss": 1.7149, + "step": 621 + }, + { + "epoch": 2.7574750830564785, + "grad_norm": 6.135425567626953, + "learning_rate": 4.475448536357572e-05, + "loss": 1.8004, + "step": 622 + }, + { + "epoch": 2.761904761904762, + "grad_norm": 4.227088451385498, + "learning_rate": 4.461132415721637e-05, + "loss": 2.0701, + "step": 623 + }, + { + "epoch": 2.7663344407530452, + "grad_norm": 7.318582057952881, + "learning_rate": 4.44682076327952e-05, + "loss": 1.8482, + "step": 624 + }, + { + "epoch": 2.770764119601329, + "grad_norm": 5.547973155975342, + "learning_rate": 4.432513697700889e-05, + "loss": 2.215, + "step": 625 + }, + { + "epoch": 2.7751937984496124, + "grad_norm": 10.597505569458008, + "learning_rate": 4.418211337617377e-05, + "loss": 1.8472, + "step": 626 + }, + { + "epoch": 2.779623477297896, + "grad_norm": 7.636361598968506, + "learning_rate": 4.403913801621602e-05, + "loss": 2.1119, + "step": 627 + }, + { + "epoch": 2.7840531561461797, + "grad_norm": 5.91168737411499, + "learning_rate": 4.3896212082661804e-05, + "loss": 1.3616, + "step": 628 + }, + { + "epoch": 2.788482834994463, + "grad_norm": 6.863846778869629, + "learning_rate": 4.375333676062743e-05, + "loss": 1.5778, + "step": 629 + }, + { + "epoch": 2.7929125138427464, + "grad_norm": 5.835073471069336, + "learning_rate": 4.3610513234809596e-05, + "loss": 1.6205, + "step": 630 + }, + { + "epoch": 2.79734219269103, + "grad_norm": 7.397892475128174, + "learning_rate": 4.346774268947545e-05, + "loss": 1.8186, + "step": 631 + }, + { + "epoch": 2.801771871539313, + "grad_norm": 5.960362434387207, + "learning_rate": 4.3325026308452896e-05, + "loss": 1.5533, + "step": 632 + }, + { + "epoch": 2.806201550387597, + "grad_norm": 7.241276264190674, + "learning_rate": 4.318236527512067e-05, + "loss": 1.8679, + "step": 633 + }, + { + "epoch": 2.8106312292358804, + "grad_norm": 5.476749420166016, + "learning_rate": 4.3039760772398616e-05, + "loss": 1.7967, + "step": 634 + }, + { + "epoch": 2.8150609080841638, + "grad_norm": 7.25029993057251, + "learning_rate": 4.289721398273779e-05, + "loss": 1.8366, + "step": 635 + }, + { + "epoch": 2.8194905869324476, + "grad_norm": 6.801474571228027, + "learning_rate": 4.275472608811073e-05, + "loss": 1.9466, + "step": 636 + }, + { + "epoch": 2.823920265780731, + "grad_norm": 6.29557991027832, + "learning_rate": 4.261229827000166e-05, + "loss": 1.7434, + "step": 637 + }, + { + "epoch": 2.8283499446290143, + "grad_norm": 5.536020755767822, + "learning_rate": 4.246993170939657e-05, + "loss": 1.793, + "step": 638 + }, + { + "epoch": 2.8327796234772977, + "grad_norm": 8.520576477050781, + "learning_rate": 4.232762758677362e-05, + "loss": 1.6861, + "step": 639 + }, + { + "epoch": 2.8372093023255816, + "grad_norm": 7.413863182067871, + "learning_rate": 4.2185387082093134e-05, + "loss": 1.9494, + "step": 640 + }, + { + "epoch": 2.841638981173865, + "grad_norm": 5.839072227478027, + "learning_rate": 4.204321137478806e-05, + "loss": 1.7803, + "step": 641 + }, + { + "epoch": 2.8460686600221483, + "grad_norm": 4.603065490722656, + "learning_rate": 4.1901101643753904e-05, + "loss": 2.0012, + "step": 642 + }, + { + "epoch": 2.850498338870432, + "grad_norm": 7.950796127319336, + "learning_rate": 4.1759059067339254e-05, + "loss": 1.7754, + "step": 643 + }, + { + "epoch": 2.8549280177187155, + "grad_norm": 4.7786173820495605, + "learning_rate": 4.161708482333575e-05, + "loss": 1.8431, + "step": 644 + }, + { + "epoch": 2.859357696566999, + "grad_norm": 8.72305679321289, + "learning_rate": 4.147518008896851e-05, + "loss": 1.3848, + "step": 645 + }, + { + "epoch": 2.8637873754152823, + "grad_norm": 8.073031425476074, + "learning_rate": 4.133334604088626e-05, + "loss": 1.8088, + "step": 646 + }, + { + "epoch": 2.8682170542635657, + "grad_norm": 7.937218189239502, + "learning_rate": 4.119158385515159e-05, + "loss": 1.5952, + "step": 647 + }, + { + "epoch": 2.8726467331118495, + "grad_norm": 4.457007884979248, + "learning_rate": 4.104989470723126e-05, + "loss": 1.4987, + "step": 648 + }, + { + "epoch": 2.877076411960133, + "grad_norm": 8.611004829406738, + "learning_rate": 4.0908279771986376e-05, + "loss": 1.7993, + "step": 649 + }, + { + "epoch": 2.8815060908084162, + "grad_norm": 7.399827480316162, + "learning_rate": 4.0766740223662716e-05, + "loss": 1.6888, + "step": 650 + }, + { + "epoch": 2.8859357696567, + "grad_norm": 6.210876941680908, + "learning_rate": 4.062527723588093e-05, + "loss": 1.8918, + "step": 651 + }, + { + "epoch": 2.8903654485049834, + "grad_norm": 8.73986530303955, + "learning_rate": 4.0483891981626885e-05, + "loss": 1.9802, + "step": 652 + }, + { + "epoch": 2.894795127353267, + "grad_norm": 7.386151313781738, + "learning_rate": 4.0342585633241846e-05, + "loss": 1.791, + "step": 653 + }, + { + "epoch": 2.89922480620155, + "grad_norm": 6.179415225982666, + "learning_rate": 4.020135936241285e-05, + "loss": 1.862, + "step": 654 + }, + { + "epoch": 2.903654485049834, + "grad_norm": 5.072649002075195, + "learning_rate": 4.006021434016293e-05, + "loss": 1.7385, + "step": 655 + }, + { + "epoch": 2.9080841638981174, + "grad_norm": 7.220780849456787, + "learning_rate": 3.991915173684138e-05, + "loss": 1.9256, + "step": 656 + }, + { + "epoch": 2.912513842746401, + "grad_norm": 7.535258769989014, + "learning_rate": 3.977817272211417e-05, + "loss": 1.4881, + "step": 657 + }, + { + "epoch": 2.9169435215946846, + "grad_norm": 8.532382011413574, + "learning_rate": 3.963727846495409e-05, + "loss": 1.8854, + "step": 658 + }, + { + "epoch": 2.921373200442968, + "grad_norm": 4.557820796966553, + "learning_rate": 3.9496470133631206e-05, + "loss": 1.7547, + "step": 659 + }, + { + "epoch": 2.9258028792912514, + "grad_norm": 7.472973346710205, + "learning_rate": 3.935574889570302e-05, + "loss": 1.594, + "step": 660 + }, + { + "epoch": 2.9302325581395348, + "grad_norm": 11.085735321044922, + "learning_rate": 3.921511591800494e-05, + "loss": 1.7801, + "step": 661 + }, + { + "epoch": 2.934662236987818, + "grad_norm": 10.502659797668457, + "learning_rate": 3.907457236664055e-05, + "loss": 1.8752, + "step": 662 + }, + { + "epoch": 2.939091915836102, + "grad_norm": 6.8742523193359375, + "learning_rate": 3.893411940697182e-05, + "loss": 1.9468, + "step": 663 + }, + { + "epoch": 2.9435215946843853, + "grad_norm": 6.568576812744141, + "learning_rate": 3.8793758203609685e-05, + "loss": 1.9107, + "step": 664 + }, + { + "epoch": 2.9479512735326687, + "grad_norm": 7.630648612976074, + "learning_rate": 3.8653489920404136e-05, + "loss": 1.8425, + "step": 665 + }, + { + "epoch": 2.9523809523809526, + "grad_norm": 5.386789798736572, + "learning_rate": 3.851331572043478e-05, + "loss": 1.6793, + "step": 666 + }, + { + "epoch": 2.956810631229236, + "grad_norm": 7.836945056915283, + "learning_rate": 3.837323676600104e-05, + "loss": 1.9951, + "step": 667 + }, + { + "epoch": 2.9612403100775193, + "grad_norm": 6.634953498840332, + "learning_rate": 3.82332542186126e-05, + "loss": 1.6223, + "step": 668 + }, + { + "epoch": 2.9656699889258027, + "grad_norm": 4.823019504547119, + "learning_rate": 3.809336923897977e-05, + "loss": 1.8021, + "step": 669 + }, + { + "epoch": 2.9700996677740865, + "grad_norm": 5.209251880645752, + "learning_rate": 3.7953582987003806e-05, + "loss": 1.9003, + "step": 670 + }, + { + "epoch": 2.97452934662237, + "grad_norm": 6.102602481842041, + "learning_rate": 3.7813896621767384e-05, + "loss": 1.8635, + "step": 671 + }, + { + "epoch": 2.9789590254706533, + "grad_norm": 5.569439888000488, + "learning_rate": 3.76743113015249e-05, + "loss": 2.2471, + "step": 672 + }, + { + "epoch": 2.983388704318937, + "grad_norm": 7.8752827644348145, + "learning_rate": 3.753482818369291e-05, + "loss": 1.6478, + "step": 673 + }, + { + "epoch": 2.9878183831672205, + "grad_norm": 10.240179061889648, + "learning_rate": 3.739544842484053e-05, + "loss": 1.8428, + "step": 674 + }, + { + "epoch": 2.992248062015504, + "grad_norm": 6.250666618347168, + "learning_rate": 3.725617318067984e-05, + "loss": 1.5094, + "step": 675 + }, + { + "epoch": 2.9966777408637872, + "grad_norm": 5.396643161773682, + "learning_rate": 3.71170036060563e-05, + "loss": 1.8349, + "step": 676 + }, + { + "epoch": 3.0022148394241417, + "grad_norm": 6.2632269859313965, + "learning_rate": 3.697794085493919e-05, + "loss": 2.0916, + "step": 677 + }, + { + "epoch": 3.006644518272425, + "grad_norm": 4.901583194732666, + "learning_rate": 3.683898608041201e-05, + "loss": 1.4515, + "step": 678 + }, + { + "epoch": 3.011074197120709, + "grad_norm": 8.754711151123047, + "learning_rate": 3.670014043466293e-05, + "loss": 1.7908, + "step": 679 + }, + { + "epoch": 3.0155038759689923, + "grad_norm": 7.272785186767578, + "learning_rate": 3.656140506897532e-05, + "loss": 1.5918, + "step": 680 + }, + { + "epoch": 3.0199335548172757, + "grad_norm": 5.400629997253418, + "learning_rate": 3.642278113371799e-05, + "loss": 1.8925, + "step": 681 + }, + { + "epoch": 3.024363233665559, + "grad_norm": 5.312657833099365, + "learning_rate": 3.6284269778335935e-05, + "loss": 1.4915, + "step": 682 + }, + { + "epoch": 3.028792912513843, + "grad_norm": 11.454095840454102, + "learning_rate": 3.614587215134052e-05, + "loss": 1.6244, + "step": 683 + }, + { + "epoch": 3.0332225913621262, + "grad_norm": 9.896474838256836, + "learning_rate": 3.600758940030024e-05, + "loss": 1.7049, + "step": 684 + }, + { + "epoch": 3.0376522702104096, + "grad_norm": 9.982049942016602, + "learning_rate": 3.586942267183091e-05, + "loss": 1.4014, + "step": 685 + }, + { + "epoch": 3.0420819490586934, + "grad_norm": 4.595606327056885, + "learning_rate": 3.573137311158644e-05, + "loss": 1.7367, + "step": 686 + }, + { + "epoch": 3.046511627906977, + "grad_norm": 7.622084617614746, + "learning_rate": 3.559344186424908e-05, + "loss": 1.6358, + "step": 687 + }, + { + "epoch": 3.05094130675526, + "grad_norm": 4.9482645988464355, + "learning_rate": 3.545563007352016e-05, + "loss": 1.3935, + "step": 688 + }, + { + "epoch": 3.0553709856035436, + "grad_norm": 6.924054145812988, + "learning_rate": 3.531793888211044e-05, + "loss": 1.522, + "step": 689 + }, + { + "epoch": 3.0598006644518274, + "grad_norm": 6.1782636642456055, + "learning_rate": 3.518036943173069e-05, + "loss": 1.6191, + "step": 690 + }, + { + "epoch": 3.064230343300111, + "grad_norm": 5.369998455047607, + "learning_rate": 3.504292286308224e-05, + "loss": 1.7259, + "step": 691 + }, + { + "epoch": 3.068660022148394, + "grad_norm": 6.008235454559326, + "learning_rate": 3.490560031584749e-05, + "loss": 1.7566, + "step": 692 + }, + { + "epoch": 3.0730897009966776, + "grad_norm": 5.590559959411621, + "learning_rate": 3.476840292868052e-05, + "loss": 1.5846, + "step": 693 + }, + { + "epoch": 3.0775193798449614, + "grad_norm": 5.604452133178711, + "learning_rate": 3.463133183919754e-05, + "loss": 1.7103, + "step": 694 + }, + { + "epoch": 3.0819490586932448, + "grad_norm": 6.367408752441406, + "learning_rate": 3.449438818396759e-05, + "loss": 1.8149, + "step": 695 + }, + { + "epoch": 3.086378737541528, + "grad_norm": 7.243832588195801, + "learning_rate": 3.435757309850297e-05, + "loss": 1.4385, + "step": 696 + }, + { + "epoch": 3.0908084163898115, + "grad_norm": 11.65194034576416, + "learning_rate": 3.4220887717249984e-05, + "loss": 1.7527, + "step": 697 + }, + { + "epoch": 3.0952380952380953, + "grad_norm": 6.606711387634277, + "learning_rate": 3.4084333173579404e-05, + "loss": 1.516, + "step": 698 + }, + { + "epoch": 3.0996677740863787, + "grad_norm": 9.863051414489746, + "learning_rate": 3.394791059977712e-05, + "loss": 1.5221, + "step": 699 + }, + { + "epoch": 3.104097452934662, + "grad_norm": 6.656291961669922, + "learning_rate": 3.381162112703479e-05, + "loss": 2.0494, + "step": 700 + }, + { + "epoch": 3.108527131782946, + "grad_norm": 8.231369018554688, + "learning_rate": 3.367546588544035e-05, + "loss": 1.7698, + "step": 701 + }, + { + "epoch": 3.1129568106312293, + "grad_norm": 4.92387056350708, + "learning_rate": 3.35394460039688e-05, + "loss": 1.7323, + "step": 702 + }, + { + "epoch": 3.1173864894795127, + "grad_norm": 5.547698497772217, + "learning_rate": 3.340356261047269e-05, + "loss": 1.9306, + "step": 703 + }, + { + "epoch": 3.121816168327796, + "grad_norm": 12.515167236328125, + "learning_rate": 3.326781683167288e-05, + "loss": 1.6062, + "step": 704 + }, + { + "epoch": 3.12624584717608, + "grad_norm": 8.807191848754883, + "learning_rate": 3.313220979314913e-05, + "loss": 2.1803, + "step": 705 + }, + { + "epoch": 3.1306755260243633, + "grad_norm": 6.980557918548584, + "learning_rate": 3.2996742619330776e-05, + "loss": 1.6428, + "step": 706 + }, + { + "epoch": 3.1351052048726467, + "grad_norm": 11.408295631408691, + "learning_rate": 3.286141643348748e-05, + "loss": 1.7191, + "step": 707 + }, + { + "epoch": 3.13953488372093, + "grad_norm": 7.25669527053833, + "learning_rate": 3.27262323577198e-05, + "loss": 2.0474, + "step": 708 + }, + { + "epoch": 3.143964562569214, + "grad_norm": 7.197902679443359, + "learning_rate": 3.259119151294997e-05, + "loss": 1.4791, + "step": 709 + }, + { + "epoch": 3.1483942414174972, + "grad_norm": 7.758485317230225, + "learning_rate": 3.2456295018912554e-05, + "loss": 1.6882, + "step": 710 + }, + { + "epoch": 3.1528239202657806, + "grad_norm": 10.935338973999023, + "learning_rate": 3.232154399414521e-05, + "loss": 1.8307, + "step": 711 + }, + { + "epoch": 3.157253599114064, + "grad_norm": 9.585394859313965, + "learning_rate": 3.218693955597938e-05, + "loss": 1.6335, + "step": 712 + }, + { + "epoch": 3.161683277962348, + "grad_norm": 8.889345169067383, + "learning_rate": 3.205248282053104e-05, + "loss": 1.8018, + "step": 713 + }, + { + "epoch": 3.166112956810631, + "grad_norm": 10.92686653137207, + "learning_rate": 3.191817490269143e-05, + "loss": 1.5411, + "step": 714 + }, + { + "epoch": 3.1705426356589146, + "grad_norm": 5.509748935699463, + "learning_rate": 3.1784016916117845e-05, + "loss": 1.5074, + "step": 715 + }, + { + "epoch": 3.1749723145071984, + "grad_norm": 8.162956237792969, + "learning_rate": 3.165000997322435e-05, + "loss": 1.5617, + "step": 716 + }, + { + "epoch": 3.179401993355482, + "grad_norm": 7.924961566925049, + "learning_rate": 3.15161551851726e-05, + "loss": 1.8548, + "step": 717 + }, + { + "epoch": 3.183831672203765, + "grad_norm": 6.5894269943237305, + "learning_rate": 3.138245366186261e-05, + "loss": 1.3462, + "step": 718 + }, + { + "epoch": 3.1882613510520486, + "grad_norm": 5.109679698944092, + "learning_rate": 3.1248906511923526e-05, + "loss": 1.4762, + "step": 719 + }, + { + "epoch": 3.1926910299003324, + "grad_norm": 8.769123077392578, + "learning_rate": 3.11155148427045e-05, + "loss": 1.543, + "step": 720 + }, + { + "epoch": 3.1971207087486158, + "grad_norm": 9.410780906677246, + "learning_rate": 3.098227976026542e-05, + "loss": 1.6442, + "step": 721 + }, + { + "epoch": 3.201550387596899, + "grad_norm": 8.743534088134766, + "learning_rate": 3.084920236936783e-05, + "loss": 1.8122, + "step": 722 + }, + { + "epoch": 3.2059800664451825, + "grad_norm": 8.16320514678955, + "learning_rate": 3.071628377346567e-05, + "loss": 1.5269, + "step": 723 + }, + { + "epoch": 3.2104097452934663, + "grad_norm": 5.1464362144470215, + "learning_rate": 3.058352507469623e-05, + "loss": 1.8886, + "step": 724 + }, + { + "epoch": 3.2148394241417497, + "grad_norm": 4.970855712890625, + "learning_rate": 3.045092737387096e-05, + "loss": 1.6344, + "step": 725 + }, + { + "epoch": 3.219269102990033, + "grad_norm": 13.145638465881348, + "learning_rate": 3.0318491770466275e-05, + "loss": 2.1682, + "step": 726 + }, + { + "epoch": 3.223698781838317, + "grad_norm": 4.9592084884643555, + "learning_rate": 3.018621936261462e-05, + "loss": 1.5418, + "step": 727 + }, + { + "epoch": 3.2281284606866003, + "grad_norm": 6.089108467102051, + "learning_rate": 3.0054111247095134e-05, + "loss": 1.5348, + "step": 728 + }, + { + "epoch": 3.2325581395348837, + "grad_norm": 6.222757816314697, + "learning_rate": 2.992216851932478e-05, + "loss": 1.6028, + "step": 729 + }, + { + "epoch": 3.236987818383167, + "grad_norm": 9.714614868164062, + "learning_rate": 2.979039227334905e-05, + "loss": 1.5382, + "step": 730 + }, + { + "epoch": 3.241417497231451, + "grad_norm": 7.093183994293213, + "learning_rate": 2.9658783601833106e-05, + "loss": 1.9805, + "step": 731 + }, + { + "epoch": 3.2458471760797343, + "grad_norm": 5.3288397789001465, + "learning_rate": 2.9527343596052513e-05, + "loss": 1.61, + "step": 732 + }, + { + "epoch": 3.2502768549280177, + "grad_norm": 6.90171480178833, + "learning_rate": 2.9396073345884363e-05, + "loss": 1.5034, + "step": 733 + }, + { + "epoch": 3.254706533776301, + "grad_norm": 15.343708992004395, + "learning_rate": 2.926497393979814e-05, + "loss": 1.7569, + "step": 734 + }, + { + "epoch": 3.259136212624585, + "grad_norm": 6.383762836456299, + "learning_rate": 2.9134046464846703e-05, + "loss": 1.894, + "step": 735 + }, + { + "epoch": 3.2635658914728682, + "grad_norm": 7.399294376373291, + "learning_rate": 2.9003292006657335e-05, + "loss": 1.6227, + "step": 736 + }, + { + "epoch": 3.2679955703211516, + "grad_norm": 6.315852642059326, + "learning_rate": 2.887271164942259e-05, + "loss": 1.6429, + "step": 737 + }, + { + "epoch": 3.2724252491694354, + "grad_norm": 6.891193389892578, + "learning_rate": 2.8742306475891574e-05, + "loss": 1.3061, + "step": 738 + }, + { + "epoch": 3.276854928017719, + "grad_norm": 5.40971040725708, + "learning_rate": 2.861207756736064e-05, + "loss": 1.6944, + "step": 739 + }, + { + "epoch": 3.281284606866002, + "grad_norm": 6.299607276916504, + "learning_rate": 2.8482026003664695e-05, + "loss": 1.9616, + "step": 740 + }, + { + "epoch": 3.2857142857142856, + "grad_norm": 5.829094886779785, + "learning_rate": 2.8352152863168035e-05, + "loss": 2.0074, + "step": 741 + }, + { + "epoch": 3.290143964562569, + "grad_norm": 4.643031120300293, + "learning_rate": 2.8222459222755605e-05, + "loss": 1.2762, + "step": 742 + }, + { + "epoch": 3.294573643410853, + "grad_norm": 8.120580673217773, + "learning_rate": 2.8092946157823928e-05, + "loss": 1.4793, + "step": 743 + }, + { + "epoch": 3.299003322259136, + "grad_norm": 7.702285289764404, + "learning_rate": 2.7963614742272186e-05, + "loss": 1.7667, + "step": 744 + }, + { + "epoch": 3.3034330011074196, + "grad_norm": 7.583436012268066, + "learning_rate": 2.78344660484934e-05, + "loss": 1.6145, + "step": 745 + }, + { + "epoch": 3.3078626799557034, + "grad_norm": 8.428057670593262, + "learning_rate": 2.7705501147365498e-05, + "loss": 1.9334, + "step": 746 + }, + { + "epoch": 3.3122923588039868, + "grad_norm": 9.74374771118164, + "learning_rate": 2.757672110824242e-05, + "loss": 1.7522, + "step": 747 + }, + { + "epoch": 3.31672203765227, + "grad_norm": 5.473592281341553, + "learning_rate": 2.744812699894524e-05, + "loss": 1.8402, + "step": 748 + }, + { + "epoch": 3.3211517165005535, + "grad_norm": 5.208759784698486, + "learning_rate": 2.731971988575337e-05, + "loss": 1.5117, + "step": 749 + }, + { + "epoch": 3.3255813953488373, + "grad_norm": 5.262367248535156, + "learning_rate": 2.719150083339567e-05, + "loss": 1.7127, + "step": 750 + }, + { + "epoch": 3.3300110741971207, + "grad_norm": 5.584857940673828, + "learning_rate": 2.7063470905041643e-05, + "loss": 1.5979, + "step": 751 + }, + { + "epoch": 3.334440753045404, + "grad_norm": 5.59777307510376, + "learning_rate": 2.693563116229261e-05, + "loss": 1.9831, + "step": 752 + }, + { + "epoch": 3.338870431893688, + "grad_norm": 5.473875999450684, + "learning_rate": 2.6807982665172858e-05, + "loss": 1.6157, + "step": 753 + }, + { + "epoch": 3.3433001107419713, + "grad_norm": 7.791520595550537, + "learning_rate": 2.668052647212097e-05, + "loss": 1.9269, + "step": 754 + }, + { + "epoch": 3.3477297895902547, + "grad_norm": 8.027992248535156, + "learning_rate": 2.655326363998094e-05, + "loss": 1.5925, + "step": 755 + }, + { + "epoch": 3.352159468438538, + "grad_norm": 6.83018684387207, + "learning_rate": 2.6426195223993476e-05, + "loss": 1.6024, + "step": 756 + }, + { + "epoch": 3.356589147286822, + "grad_norm": 9.020490646362305, + "learning_rate": 2.6299322277787163e-05, + "loss": 1.8361, + "step": 757 + }, + { + "epoch": 3.3610188261351053, + "grad_norm": 5.8339524269104, + "learning_rate": 2.6172645853369842e-05, + "loss": 1.6429, + "step": 758 + }, + { + "epoch": 3.3654485049833887, + "grad_norm": 4.7356672286987305, + "learning_rate": 2.6046167001119816e-05, + "loss": 1.4711, + "step": 759 + }, + { + "epoch": 3.369878183831672, + "grad_norm": 10.283108711242676, + "learning_rate": 2.5919886769777136e-05, + "loss": 1.5224, + "step": 760 + }, + { + "epoch": 3.374307862679956, + "grad_norm": 6.590269088745117, + "learning_rate": 2.5793806206434977e-05, + "loss": 1.7833, + "step": 761 + }, + { + "epoch": 3.3787375415282392, + "grad_norm": 6.434593200683594, + "learning_rate": 2.5667926356530797e-05, + "loss": 1.8216, + "step": 762 + }, + { + "epoch": 3.3831672203765226, + "grad_norm": 5.936555862426758, + "learning_rate": 2.554224826383792e-05, + "loss": 1.5334, + "step": 763 + }, + { + "epoch": 3.387596899224806, + "grad_norm": 14.309653282165527, + "learning_rate": 2.5416772970456592e-05, + "loss": 1.4408, + "step": 764 + }, + { + "epoch": 3.39202657807309, + "grad_norm": 4.926191806793213, + "learning_rate": 2.5291501516805582e-05, + "loss": 1.553, + "step": 765 + }, + { + "epoch": 3.396456256921373, + "grad_norm": 6.072823524475098, + "learning_rate": 2.5166434941613408e-05, + "loss": 1.6255, + "step": 766 + }, + { + "epoch": 3.4008859357696566, + "grad_norm": 5.073078632354736, + "learning_rate": 2.5041574281909814e-05, + "loss": 1.4512, + "step": 767 + }, + { + "epoch": 3.4053156146179404, + "grad_norm": 5.974765300750732, + "learning_rate": 2.491692057301706e-05, + "loss": 1.5731, + "step": 768 + }, + { + "epoch": 3.409745293466224, + "grad_norm": 7.9622015953063965, + "learning_rate": 2.4792474848541468e-05, + "loss": 1.9739, + "step": 769 + }, + { + "epoch": 3.414174972314507, + "grad_norm": 6.868570327758789, + "learning_rate": 2.466823814036477e-05, + "loss": 1.8549, + "step": 770 + }, + { + "epoch": 3.4186046511627906, + "grad_norm": 5.171596527099609, + "learning_rate": 2.4544211478635582e-05, + "loss": 1.9368, + "step": 771 + }, + { + "epoch": 3.4230343300110744, + "grad_norm": 5.299581527709961, + "learning_rate": 2.442039589176085e-05, + "loss": 1.6896, + "step": 772 + }, + { + "epoch": 3.4274640088593578, + "grad_norm": 4.135156631469727, + "learning_rate": 2.429679240639729e-05, + "loss": 1.6126, + "step": 773 + }, + { + "epoch": 3.431893687707641, + "grad_norm": 10.067416191101074, + "learning_rate": 2.417340204744295e-05, + "loss": 1.5858, + "step": 774 + }, + { + "epoch": 3.4363233665559245, + "grad_norm": 6.024186611175537, + "learning_rate": 2.4050225838028673e-05, + "loss": 1.6173, + "step": 775 + }, + { + "epoch": 3.4407530454042083, + "grad_norm": 5.533918380737305, + "learning_rate": 2.3927264799509607e-05, + "loss": 1.6362, + "step": 776 + }, + { + "epoch": 3.4451827242524917, + "grad_norm": 11.559389114379883, + "learning_rate": 2.38045199514567e-05, + "loss": 1.7199, + "step": 777 + }, + { + "epoch": 3.449612403100775, + "grad_norm": 18.32662010192871, + "learning_rate": 2.368199231164832e-05, + "loss": 1.6391, + "step": 778 + }, + { + "epoch": 3.4540420819490585, + "grad_norm": 5.221166610717773, + "learning_rate": 2.3559682896061845e-05, + "loss": 1.9352, + "step": 779 + }, + { + "epoch": 3.4584717607973423, + "grad_norm": 10.909178733825684, + "learning_rate": 2.3437592718865064e-05, + "loss": 1.951, + "step": 780 + }, + { + "epoch": 3.4629014396456257, + "grad_norm": 11.137706756591797, + "learning_rate": 2.3315722792407963e-05, + "loss": 1.7804, + "step": 781 + }, + { + "epoch": 3.467331118493909, + "grad_norm": 5.704611301422119, + "learning_rate": 2.319407412721416e-05, + "loss": 1.7664, + "step": 782 + }, + { + "epoch": 3.471760797342193, + "grad_norm": 6.683074951171875, + "learning_rate": 2.3072647731972746e-05, + "loss": 1.7184, + "step": 783 + }, + { + "epoch": 3.4761904761904763, + "grad_norm": 5.341973781585693, + "learning_rate": 2.2951444613529673e-05, + "loss": 1.7536, + "step": 784 + }, + { + "epoch": 3.4806201550387597, + "grad_norm": 7.4396443367004395, + "learning_rate": 2.283046577687958e-05, + "loss": 1.8698, + "step": 785 + }, + { + "epoch": 3.485049833887043, + "grad_norm": 9.703819274902344, + "learning_rate": 2.270971222515735e-05, + "loss": 1.4894, + "step": 786 + }, + { + "epoch": 3.489479512735327, + "grad_norm": 7.606261730194092, + "learning_rate": 2.2589184959629918e-05, + "loss": 1.5671, + "step": 787 + }, + { + "epoch": 3.4939091915836102, + "grad_norm": 12.827896118164062, + "learning_rate": 2.2468884979687865e-05, + "loss": 1.8357, + "step": 788 + }, + { + "epoch": 3.4983388704318936, + "grad_norm": 5.452163219451904, + "learning_rate": 2.2348813282837112e-05, + "loss": 1.6886, + "step": 789 + }, + { + "epoch": 3.5027685492801774, + "grad_norm": 7.322004318237305, + "learning_rate": 2.2228970864690746e-05, + "loss": 1.4266, + "step": 790 + }, + { + "epoch": 3.507198228128461, + "grad_norm": 10.763130187988281, + "learning_rate": 2.2109358718960723e-05, + "loss": 1.861, + "step": 791 + }, + { + "epoch": 3.511627906976744, + "grad_norm": 5.822150707244873, + "learning_rate": 2.1989977837449605e-05, + "loss": 2.3509, + "step": 792 + }, + { + "epoch": 3.5160575858250276, + "grad_norm": 5.79690408706665, + "learning_rate": 2.187082921004232e-05, + "loss": 1.6493, + "step": 793 + }, + { + "epoch": 3.520487264673311, + "grad_norm": 8.250429153442383, + "learning_rate": 2.1751913824698043e-05, + "loss": 1.7764, + "step": 794 + }, + { + "epoch": 3.524916943521595, + "grad_norm": 8.250429153442383, + "learning_rate": 2.1751913824698043e-05, + "loss": 1.7552, + "step": 795 + }, + { + "epoch": 3.529346622369878, + "grad_norm": 11.57303237915039, + "learning_rate": 2.163323266744192e-05, + "loss": 1.6983, + "step": 796 + }, + { + "epoch": 3.5337763012181616, + "grad_norm": 11.181726455688477, + "learning_rate": 2.1514786722356915e-05, + "loss": 1.6178, + "step": 797 + }, + { + "epoch": 3.5382059800664454, + "grad_norm": 8.835942268371582, + "learning_rate": 2.1396576971575687e-05, + "loss": 1.9932, + "step": 798 + }, + { + "epoch": 3.5426356589147288, + "grad_norm": 9.03427505493164, + "learning_rate": 2.1278604395272344e-05, + "loss": 1.736, + "step": 799 + }, + { + "epoch": 3.547065337763012, + "grad_norm": 8.337313652038574, + "learning_rate": 2.1160869971654463e-05, + "loss": 1.7021, + "step": 800 + }, + { + "epoch": 3.5514950166112955, + "grad_norm": 7.044549942016602, + "learning_rate": 2.1043374676954852e-05, + "loss": 1.277, + "step": 801 + }, + { + "epoch": 3.5559246954595793, + "grad_norm": 7.318129062652588, + "learning_rate": 2.0926119485423545e-05, + "loss": 1.6063, + "step": 802 + }, + { + "epoch": 3.5603543743078627, + "grad_norm": 5.060410022735596, + "learning_rate": 2.0809105369319627e-05, + "loss": 1.5807, + "step": 803 + }, + { + "epoch": 3.564784053156146, + "grad_norm": 6.682000637054443, + "learning_rate": 2.0692333298903262e-05, + "loss": 1.5304, + "step": 804 + }, + { + "epoch": 3.56921373200443, + "grad_norm": 6.671335220336914, + "learning_rate": 2.057580424242766e-05, + "loss": 1.4664, + "step": 805 + }, + { + "epoch": 3.5736434108527133, + "grad_norm": 6.632824420928955, + "learning_rate": 2.0459519166130908e-05, + "loss": 1.6361, + "step": 806 + }, + { + "epoch": 3.5780730897009967, + "grad_norm": 8.885310173034668, + "learning_rate": 2.034347903422812e-05, + "loss": 1.7954, + "step": 807 + }, + { + "epoch": 3.58250276854928, + "grad_norm": 10.805225372314453, + "learning_rate": 2.022768480890335e-05, + "loss": 1.6574, + "step": 808 + }, + { + "epoch": 3.5869324473975635, + "grad_norm": 9.142356872558594, + "learning_rate": 2.0112137450301673e-05, + "loss": 2.0787, + "step": 809 + }, + { + "epoch": 3.5913621262458473, + "grad_norm": 6.482480525970459, + "learning_rate": 1.999683791652111e-05, + "loss": 1.6388, + "step": 810 + }, + { + "epoch": 3.5957918050941307, + "grad_norm": 5.05092191696167, + "learning_rate": 1.9881787163604842e-05, + "loss": 1.7928, + "step": 811 + }, + { + "epoch": 3.600221483942414, + "grad_norm": 11.357190132141113, + "learning_rate": 1.976698614553318e-05, + "loss": 1.3903, + "step": 812 + }, + { + "epoch": 3.604651162790698, + "grad_norm": 5.155529022216797, + "learning_rate": 1.965243581421566e-05, + "loss": 1.56, + "step": 813 + }, + { + "epoch": 3.6090808416389812, + "grad_norm": 8.153807640075684, + "learning_rate": 1.9538137119483206e-05, + "loss": 1.5867, + "step": 814 + }, + { + "epoch": 3.6135105204872646, + "grad_norm": 5.8331708908081055, + "learning_rate": 1.9424091009080157e-05, + "loss": 1.4878, + "step": 815 + }, + { + "epoch": 3.617940199335548, + "grad_norm": 6.762333869934082, + "learning_rate": 1.9310298428656522e-05, + "loss": 1.6243, + "step": 816 + }, + { + "epoch": 3.622369878183832, + "grad_norm": 5.154025077819824, + "learning_rate": 1.919676032176006e-05, + "loss": 1.7056, + "step": 817 + }, + { + "epoch": 3.626799557032115, + "grad_norm": 8.215972900390625, + "learning_rate": 1.908347762982851e-05, + "loss": 1.6746, + "step": 818 + }, + { + "epoch": 3.6312292358803986, + "grad_norm": 6.393702507019043, + "learning_rate": 1.897045129218169e-05, + "loss": 1.4001, + "step": 819 + }, + { + "epoch": 3.6356589147286824, + "grad_norm": 6.989351272583008, + "learning_rate": 1.885768224601384e-05, + "loss": 1.7093, + "step": 820 + }, + { + "epoch": 3.640088593576966, + "grad_norm": 7.712499618530273, + "learning_rate": 1.874517142638576e-05, + "loss": 1.9538, + "step": 821 + }, + { + "epoch": 3.644518272425249, + "grad_norm": 8.841022491455078, + "learning_rate": 1.8632919766217077e-05, + "loss": 1.9689, + "step": 822 + }, + { + "epoch": 3.6489479512735326, + "grad_norm": 7.386549949645996, + "learning_rate": 1.852092819627854e-05, + "loss": 1.6141, + "step": 823 + }, + { + "epoch": 3.653377630121816, + "grad_norm": 5.60568904876709, + "learning_rate": 1.8409197645184205e-05, + "loss": 1.4079, + "step": 824 + }, + { + "epoch": 3.6578073089700998, + "grad_norm": 10.555825233459473, + "learning_rate": 1.829772903938393e-05, + "loss": 1.7793, + "step": 825 + }, + { + "epoch": 3.662236987818383, + "grad_norm": 7.500188827514648, + "learning_rate": 1.818652330315544e-05, + "loss": 1.836, + "step": 826 + }, + { + "epoch": 3.6666666666666665, + "grad_norm": 4.860312461853027, + "learning_rate": 1.807558135859689e-05, + "loss": 1.5209, + "step": 827 + }, + { + "epoch": 3.6710963455149503, + "grad_norm": 5.605367183685303, + "learning_rate": 1.7964904125619014e-05, + "loss": 1.4998, + "step": 828 + }, + { + "epoch": 3.6755260243632337, + "grad_norm": 4.448379993438721, + "learning_rate": 1.7854492521937753e-05, + "loss": 1.3537, + "step": 829 + }, + { + "epoch": 3.679955703211517, + "grad_norm": 8.981001853942871, + "learning_rate": 1.7744347463066364e-05, + "loss": 1.7037, + "step": 830 + }, + { + "epoch": 3.6843853820598005, + "grad_norm": 5.554767608642578, + "learning_rate": 1.7634469862308028e-05, + "loss": 1.5133, + "step": 831 + }, + { + "epoch": 3.6888150609080843, + "grad_norm": 6.952533721923828, + "learning_rate": 1.7524860630748196e-05, + "loss": 1.9494, + "step": 832 + }, + { + "epoch": 3.6932447397563677, + "grad_norm": 5.131599426269531, + "learning_rate": 1.741552067724705e-05, + "loss": 1.7359, + "step": 833 + }, + { + "epoch": 3.697674418604651, + "grad_norm": 5.388726711273193, + "learning_rate": 1.7306450908431988e-05, + "loss": 1.6177, + "step": 834 + }, + { + "epoch": 3.702104097452935, + "grad_norm": 4.925945281982422, + "learning_rate": 1.719765222869002e-05, + "loss": 1.6502, + "step": 835 + }, + { + "epoch": 3.7065337763012183, + "grad_norm": 6.9254679679870605, + "learning_rate": 1.7089125540160405e-05, + "loss": 1.4616, + "step": 836 + }, + { + "epoch": 3.7109634551495017, + "grad_norm": 5.429500579833984, + "learning_rate": 1.6980871742727067e-05, + "loss": 1.9383, + "step": 837 + }, + { + "epoch": 3.715393133997785, + "grad_norm": 6.535275936126709, + "learning_rate": 1.6872891734011194e-05, + "loss": 1.7912, + "step": 838 + }, + { + "epoch": 3.7198228128460684, + "grad_norm": 10.143779754638672, + "learning_rate": 1.6765186409363697e-05, + "loss": 1.6256, + "step": 839 + }, + { + "epoch": 3.7242524916943522, + "grad_norm": 5.600742340087891, + "learning_rate": 1.6657756661857926e-05, + "loss": 1.3974, + "step": 840 + }, + { + "epoch": 3.7286821705426356, + "grad_norm": 4.404951095581055, + "learning_rate": 1.655060338228217e-05, + "loss": 1.7132, + "step": 841 + }, + { + "epoch": 3.733111849390919, + "grad_norm": 5.553696632385254, + "learning_rate": 1.6443727459132295e-05, + "loss": 1.6708, + "step": 842 + }, + { + "epoch": 3.737541528239203, + "grad_norm": 5.281304359436035, + "learning_rate": 1.6337129778604404e-05, + "loss": 1.4539, + "step": 843 + }, + { + "epoch": 3.741971207087486, + "grad_norm": 5.668034076690674, + "learning_rate": 1.6230811224587388e-05, + "loss": 1.7885, + "step": 844 + }, + { + "epoch": 3.7464008859357696, + "grad_norm": 10.235897064208984, + "learning_rate": 1.6124772678655742e-05, + "loss": 1.4588, + "step": 845 + }, + { + "epoch": 3.750830564784053, + "grad_norm": 5.400420188903809, + "learning_rate": 1.6019015020062163e-05, + "loss": 1.62, + "step": 846 + }, + { + "epoch": 3.755260243632337, + "grad_norm": 6.550075054168701, + "learning_rate": 1.591353912573028e-05, + "loss": 1.975, + "step": 847 + }, + { + "epoch": 3.75968992248062, + "grad_norm": 7.559651851654053, + "learning_rate": 1.580834587024737e-05, + "loss": 1.8536, + "step": 848 + }, + { + "epoch": 3.7641196013289036, + "grad_norm": 7.971467971801758, + "learning_rate": 1.5703436125857117e-05, + "loss": 1.5991, + "step": 849 + }, + { + "epoch": 3.7685492801771874, + "grad_norm": 5.560451507568359, + "learning_rate": 1.5598810762452443e-05, + "loss": 1.7564, + "step": 850 + }, + { + "epoch": 3.7729789590254708, + "grad_norm": 4.291793346405029, + "learning_rate": 1.5494470647568144e-05, + "loss": 1.5643, + "step": 851 + }, + { + "epoch": 3.777408637873754, + "grad_norm": 4.999351978302002, + "learning_rate": 1.5390416646373835e-05, + "loss": 1.5744, + "step": 852 + }, + { + "epoch": 3.7818383167220375, + "grad_norm": 9.698875427246094, + "learning_rate": 1.5286649621666727e-05, + "loss": 1.5056, + "step": 853 + }, + { + "epoch": 3.786267995570321, + "grad_norm": 10.186306953430176, + "learning_rate": 1.5183170433864474e-05, + "loss": 1.8846, + "step": 854 + }, + { + "epoch": 3.7906976744186047, + "grad_norm": 4.416024208068848, + "learning_rate": 1.5079979940998e-05, + "loss": 1.5333, + "step": 855 + }, + { + "epoch": 3.795127353266888, + "grad_norm": 5.287939071655273, + "learning_rate": 1.497707899870448e-05, + "loss": 1.7457, + "step": 856 + }, + { + "epoch": 3.7995570321151715, + "grad_norm": 7.013801574707031, + "learning_rate": 1.4874468460220143e-05, + "loss": 1.7936, + "step": 857 + }, + { + "epoch": 3.8039867109634553, + "grad_norm": 5.881109714508057, + "learning_rate": 1.4772149176373262e-05, + "loss": 1.6287, + "step": 858 + }, + { + "epoch": 3.8084163898117387, + "grad_norm": 6.294797420501709, + "learning_rate": 1.4670121995577096e-05, + "loss": 1.8849, + "step": 859 + }, + { + "epoch": 3.812846068660022, + "grad_norm": 4.177390098571777, + "learning_rate": 1.4568387763822777e-05, + "loss": 1.5562, + "step": 860 + }, + { + "epoch": 3.8172757475083055, + "grad_norm": 6.2013678550720215, + "learning_rate": 1.446694732467242e-05, + "loss": 1.5161, + "step": 861 + }, + { + "epoch": 3.8217054263565893, + "grad_norm": 11.345988273620605, + "learning_rate": 1.4365801519252025e-05, + "loss": 1.9987, + "step": 862 + }, + { + "epoch": 3.8261351052048727, + "grad_norm": 7.59104061126709, + "learning_rate": 1.4264951186244585e-05, + "loss": 1.6589, + "step": 863 + }, + { + "epoch": 3.830564784053156, + "grad_norm": 6.380128860473633, + "learning_rate": 1.4164397161883019e-05, + "loss": 1.4321, + "step": 864 + }, + { + "epoch": 3.83499446290144, + "grad_norm": 7.260960578918457, + "learning_rate": 1.4064140279943367e-05, + "loss": 1.9097, + "step": 865 + }, + { + "epoch": 3.8394241417497232, + "grad_norm": 4.328764915466309, + "learning_rate": 1.3964181371737794e-05, + "loss": 1.6674, + "step": 866 + }, + { + "epoch": 3.8438538205980066, + "grad_norm": 10.491178512573242, + "learning_rate": 1.3864521266107728e-05, + "loss": 1.6915, + "step": 867 + }, + { + "epoch": 3.84828349944629, + "grad_norm": 5.77324914932251, + "learning_rate": 1.3765160789416986e-05, + "loss": 2.1619, + "step": 868 + }, + { + "epoch": 3.8527131782945734, + "grad_norm": 11.36082935333252, + "learning_rate": 1.3666100765544865e-05, + "loss": 1.9967, + "step": 869 + }, + { + "epoch": 3.857142857142857, + "grad_norm": 4.28300666809082, + "learning_rate": 1.3567342015879453e-05, + "loss": 1.6779, + "step": 870 + }, + { + "epoch": 3.8615725359911406, + "grad_norm": 6.903509616851807, + "learning_rate": 1.346888535931064e-05, + "loss": 1.6562, + "step": 871 + }, + { + "epoch": 3.8660022148394244, + "grad_norm": 6.6412672996521, + "learning_rate": 1.3370731612223463e-05, + "loss": 1.834, + "step": 872 + }, + { + "epoch": 3.870431893687708, + "grad_norm": 5.490871429443359, + "learning_rate": 1.3272881588491237e-05, + "loss": 1.5606, + "step": 873 + }, + { + "epoch": 3.874861572535991, + "grad_norm": 7.209272384643555, + "learning_rate": 1.3175336099468943e-05, + "loss": 1.4649, + "step": 874 + }, + { + "epoch": 3.8792912513842746, + "grad_norm": 4.529653549194336, + "learning_rate": 1.307809595398633e-05, + "loss": 1.8369, + "step": 875 + }, + { + "epoch": 3.883720930232558, + "grad_norm": 6.793308734893799, + "learning_rate": 1.2981161958341337e-05, + "loss": 1.6708, + "step": 876 + }, + { + "epoch": 3.8881506090808418, + "grad_norm": 5.094226360321045, + "learning_rate": 1.2884534916293345e-05, + "loss": 1.8261, + "step": 877 + }, + { + "epoch": 3.892580287929125, + "grad_norm": 5.952579021453857, + "learning_rate": 1.2788215629056548e-05, + "loss": 1.8971, + "step": 878 + }, + { + "epoch": 3.8970099667774085, + "grad_norm": 4.1918745040893555, + "learning_rate": 1.2692204895293298e-05, + "loss": 1.5702, + "step": 879 + }, + { + "epoch": 3.9014396456256923, + "grad_norm": 6.7971391677856445, + "learning_rate": 1.2596503511107428e-05, + "loss": 1.2105, + "step": 880 + }, + { + "epoch": 3.9058693244739757, + "grad_norm": 3.9530787467956543, + "learning_rate": 1.250111227003774e-05, + "loss": 1.9222, + "step": 881 + }, + { + "epoch": 3.910299003322259, + "grad_norm": 5.808813095092773, + "learning_rate": 1.240603196305139e-05, + "loss": 1.5273, + "step": 882 + }, + { + "epoch": 3.9147286821705425, + "grad_norm": 4.9905242919921875, + "learning_rate": 1.2311263378537314e-05, + "loss": 1.9195, + "step": 883 + }, + { + "epoch": 3.919158361018826, + "grad_norm": 5.769536972045898, + "learning_rate": 1.2216807302299683e-05, + "loss": 1.8533, + "step": 884 + }, + { + "epoch": 3.9235880398671097, + "grad_norm": 6.07841682434082, + "learning_rate": 1.2122664517551418e-05, + "loss": 1.4366, + "step": 885 + }, + { + "epoch": 3.928017718715393, + "grad_norm": 5.433627128601074, + "learning_rate": 1.2028835804907701e-05, + "loss": 1.6119, + "step": 886 + }, + { + "epoch": 3.932447397563677, + "grad_norm": 5.320455551147461, + "learning_rate": 1.1935321942379456e-05, + "loss": 1.555, + "step": 887 + }, + { + "epoch": 3.9368770764119603, + "grad_norm": 5.33215856552124, + "learning_rate": 1.1842123705366959e-05, + "loss": 1.3726, + "step": 888 + }, + { + "epoch": 3.9413067552602437, + "grad_norm": 5.91862678527832, + "learning_rate": 1.1749241866653327e-05, + "loss": 1.6878, + "step": 889 + }, + { + "epoch": 3.945736434108527, + "grad_norm": 5.898477554321289, + "learning_rate": 1.1656677196398197e-05, + "loss": 2.099, + "step": 890 + }, + { + "epoch": 3.9501661129568104, + "grad_norm": 6.435544967651367, + "learning_rate": 1.1564430462131299e-05, + "loss": 1.7103, + "step": 891 + }, + { + "epoch": 3.9545957918050942, + "grad_norm": 5.444723129272461, + "learning_rate": 1.147250242874609e-05, + "loss": 1.9926, + "step": 892 + }, + { + "epoch": 3.9590254706533776, + "grad_norm": 7.465640068054199, + "learning_rate": 1.138089385849338e-05, + "loss": 1.473, + "step": 893 + }, + { + "epoch": 3.963455149501661, + "grad_norm": 5.691059112548828, + "learning_rate": 1.128960551097512e-05, + "loss": 1.7822, + "step": 894 + }, + { + "epoch": 3.967884828349945, + "grad_norm": 4.731316566467285, + "learning_rate": 1.1198638143137996e-05, + "loss": 1.4145, + "step": 895 + }, + { + "epoch": 3.972314507198228, + "grad_norm": 7.516561031341553, + "learning_rate": 1.110799250926715e-05, + "loss": 1.5675, + "step": 896 + }, + { + "epoch": 3.9767441860465116, + "grad_norm": 10.761414527893066, + "learning_rate": 1.1017669360980016e-05, + "loss": 1.6808, + "step": 897 + }, + { + "epoch": 3.981173864894795, + "grad_norm": 8.457561492919922, + "learning_rate": 1.0927669447220012e-05, + "loss": 1.6556, + "step": 898 + }, + { + "epoch": 3.985603543743079, + "grad_norm": 5.558786869049072, + "learning_rate": 1.0837993514250371e-05, + "loss": 1.5594, + "step": 899 + }, + { + "epoch": 3.990033222591362, + "grad_norm": 7.166381359100342, + "learning_rate": 1.07486423056479e-05, + "loss": 1.279, + "step": 900 + }, + { + "epoch": 3.9944629014396456, + "grad_norm": 4.120462417602539, + "learning_rate": 1.0659616562296882e-05, + "loss": 1.6244, + "step": 901 + }, + { + "epoch": 3.9988925802879294, + "grad_norm": 7.515431880950928, + "learning_rate": 1.0570917022382887e-05, + "loss": 1.4077, + "step": 902 + }, + { + "epoch": 4.004429678848283, + "grad_norm": 3.734769821166992, + "learning_rate": 1.0482544421386669e-05, + "loss": 1.4449, + "step": 903 + }, + { + "epoch": 4.008859357696567, + "grad_norm": 6.960366249084473, + "learning_rate": 1.039449949207808e-05, + "loss": 1.5066, + "step": 904 + }, + { + "epoch": 4.01328903654485, + "grad_norm": 6.179312705993652, + "learning_rate": 1.030678296450992e-05, + "loss": 1.6181, + "step": 905 + }, + { + "epoch": 4.017718715393134, + "grad_norm": 4.362559795379639, + "learning_rate": 1.0219395566011992e-05, + "loss": 1.6309, + "step": 906 + }, + { + "epoch": 4.022148394241418, + "grad_norm": 5.700815200805664, + "learning_rate": 1.0132338021185012e-05, + "loss": 2.0716, + "step": 907 + }, + { + "epoch": 4.026578073089701, + "grad_norm": 6.670535087585449, + "learning_rate": 1.004561105189461e-05, + "loss": 1.4746, + "step": 908 + }, + { + "epoch": 4.0310077519379846, + "grad_norm": 5.524834632873535, + "learning_rate": 9.959215377265312e-06, + "loss": 1.324, + "step": 909 + }, + { + "epoch": 4.035437430786268, + "grad_norm": 5.835739612579346, + "learning_rate": 9.873151713674638e-06, + "loss": 1.7052, + "step": 910 + }, + { + "epoch": 4.039867109634551, + "grad_norm": 5.3091044425964355, + "learning_rate": 9.78742077474713e-06, + "loss": 1.6702, + "step": 911 + }, + { + "epoch": 4.044296788482835, + "grad_norm": 6.9273762702941895, + "learning_rate": 9.70202327134842e-06, + "loss": 1.9085, + "step": 912 + }, + { + "epoch": 4.048726467331118, + "grad_norm": 4.531795978546143, + "learning_rate": 9.616959911579388e-06, + "loss": 1.3284, + "step": 913 + }, + { + "epoch": 4.053156146179402, + "grad_norm": 8.328673362731934, + "learning_rate": 9.532231400770181e-06, + "loss": 1.4044, + "step": 914 + }, + { + "epoch": 4.057585825027686, + "grad_norm": 5.203913688659668, + "learning_rate": 9.447838441474538e-06, + "loss": 1.6471, + "step": 915 + }, + { + "epoch": 4.062015503875969, + "grad_norm": 4.853526592254639, + "learning_rate": 9.363781733463784e-06, + "loss": 1.707, + "step": 916 + }, + { + "epoch": 4.0664451827242525, + "grad_norm": 5.6800737380981445, + "learning_rate": 9.280061973721149e-06, + "loss": 1.7479, + "step": 917 + }, + { + "epoch": 4.070874861572536, + "grad_norm": 4.587559700012207, + "learning_rate": 9.196679856435908e-06, + "loss": 1.589, + "step": 918 + }, + { + "epoch": 4.075304540420819, + "grad_norm": 4.8105788230896, + "learning_rate": 9.113636072997744e-06, + "loss": 1.5716, + "step": 919 + }, + { + "epoch": 4.079734219269103, + "grad_norm": 5.17236328125, + "learning_rate": 9.030931311990842e-06, + "loss": 1.499, + "step": 920 + }, + { + "epoch": 4.084163898117387, + "grad_norm": 6.545801162719727, + "learning_rate": 8.948566259188334e-06, + "loss": 1.7694, + "step": 921 + }, + { + "epoch": 4.08859357696567, + "grad_norm": 5.068369388580322, + "learning_rate": 8.866541597546529e-06, + "loss": 1.8775, + "step": 922 + }, + { + "epoch": 4.093023255813954, + "grad_norm": 6.716525554656982, + "learning_rate": 8.784858007199292e-06, + "loss": 1.6109, + "step": 923 + }, + { + "epoch": 4.097452934662237, + "grad_norm": 6.743626594543457, + "learning_rate": 8.703516165452375e-06, + "loss": 1.4904, + "step": 924 + }, + { + "epoch": 4.10188261351052, + "grad_norm": 7.2740983963012695, + "learning_rate": 8.622516746777787e-06, + "loss": 1.7105, + "step": 925 + }, + { + "epoch": 4.106312292358804, + "grad_norm": 4.581962585449219, + "learning_rate": 8.541860422808267e-06, + "loss": 1.4977, + "step": 926 + }, + { + "epoch": 4.110741971207087, + "grad_norm": 5.813825607299805, + "learning_rate": 8.461547862331649e-06, + "loss": 1.3004, + "step": 927 + }, + { + "epoch": 4.1151716500553706, + "grad_norm": 4.484111785888672, + "learning_rate": 8.38157973128535e-06, + "loss": 1.3782, + "step": 928 + }, + { + "epoch": 4.119601328903655, + "grad_norm": 4.56226110458374, + "learning_rate": 8.30195669275085e-06, + "loss": 1.6109, + "step": 929 + }, + { + "epoch": 4.124031007751938, + "grad_norm": 7.081721305847168, + "learning_rate": 8.222679406948148e-06, + "loss": 1.7734, + "step": 930 + }, + { + "epoch": 4.128460686600222, + "grad_norm": 7.574245452880859, + "learning_rate": 8.143748531230361e-06, + "loss": 1.6593, + "step": 931 + }, + { + "epoch": 4.132890365448505, + "grad_norm": 9.047913551330566, + "learning_rate": 8.065164720078216e-06, + "loss": 1.5733, + "step": 932 + }, + { + "epoch": 4.137320044296788, + "grad_norm": 4.532729625701904, + "learning_rate": 7.986928625094663e-06, + "loss": 1.6855, + "step": 933 + }, + { + "epoch": 4.141749723145072, + "grad_norm": 6.263223648071289, + "learning_rate": 7.909040894999403e-06, + "loss": 1.7791, + "step": 934 + }, + { + "epoch": 4.146179401993355, + "grad_norm": 6.938450813293457, + "learning_rate": 7.831502175623628e-06, + "loss": 1.6734, + "step": 935 + }, + { + "epoch": 4.150609080841639, + "grad_norm": 5.863696098327637, + "learning_rate": 7.754313109904533e-06, + "loss": 1.8766, + "step": 936 + }, + { + "epoch": 4.155038759689923, + "grad_norm": 7.904068470001221, + "learning_rate": 7.67747433788008e-06, + "loss": 1.5805, + "step": 937 + }, + { + "epoch": 4.159468438538206, + "grad_norm": 6.500148773193359, + "learning_rate": 7.6009864966836546e-06, + "loss": 1.5432, + "step": 938 + }, + { + "epoch": 4.1638981173864895, + "grad_norm": 5.123464584350586, + "learning_rate": 7.5248502205387705e-06, + "loss": 1.2146, + "step": 939 + }, + { + "epoch": 4.168327796234773, + "grad_norm": 4.51666259765625, + "learning_rate": 7.4490661407538635e-06, + "loss": 1.4389, + "step": 940 + }, + { + "epoch": 4.172757475083056, + "grad_norm": 7.513207912445068, + "learning_rate": 7.373634885716968e-06, + "loss": 1.4889, + "step": 941 + }, + { + "epoch": 4.17718715393134, + "grad_norm": 8.832489013671875, + "learning_rate": 7.298557080890594e-06, + "loss": 1.8399, + "step": 942 + }, + { + "epoch": 4.181616832779623, + "grad_norm": 4.61073637008667, + "learning_rate": 7.223833348806502e-06, + "loss": 1.5412, + "step": 943 + }, + { + "epoch": 4.186046511627907, + "grad_norm": 10.34188461303711, + "learning_rate": 7.149464309060549e-06, + "loss": 1.4855, + "step": 944 + }, + { + "epoch": 4.190476190476191, + "grad_norm": 7.247960090637207, + "learning_rate": 7.075450578307513e-06, + "loss": 1.5399, + "step": 945 + }, + { + "epoch": 4.194905869324474, + "grad_norm": 6.24155855178833, + "learning_rate": 7.001792770256055e-06, + "loss": 1.7671, + "step": 946 + }, + { + "epoch": 4.1993355481727574, + "grad_norm": 4.38225793838501, + "learning_rate": 6.928491495663564e-06, + "loss": 1.8296, + "step": 947 + }, + { + "epoch": 4.203765227021041, + "grad_norm": 6.093568801879883, + "learning_rate": 6.8555473623311275e-06, + "loss": 1.4753, + "step": 948 + }, + { + "epoch": 4.208194905869324, + "grad_norm": 5.5349626541137695, + "learning_rate": 6.78296097509849e-06, + "loss": 1.5375, + "step": 949 + }, + { + "epoch": 4.212624584717608, + "grad_norm": 6.367927074432373, + "learning_rate": 6.710732935838998e-06, + "loss": 1.7022, + "step": 950 + }, + { + "epoch": 4.217054263565892, + "grad_norm": 6.881635665893555, + "learning_rate": 6.638863843454662e-06, + "loss": 1.6797, + "step": 951 + }, + { + "epoch": 4.221483942414175, + "grad_norm": 14.884693145751953, + "learning_rate": 6.567354293871181e-06, + "loss": 1.432, + "step": 952 + }, + { + "epoch": 4.225913621262459, + "grad_norm": 8.686039924621582, + "learning_rate": 6.496204880032969e-06, + "loss": 1.6911, + "step": 953 + }, + { + "epoch": 4.230343300110742, + "grad_norm": 5.975069999694824, + "learning_rate": 6.425416191898259e-06, + "loss": 1.5197, + "step": 954 + }, + { + "epoch": 4.234772978959025, + "grad_norm": 7.568436622619629, + "learning_rate": 6.3549888164342046e-06, + "loss": 1.5966, + "step": 955 + }, + { + "epoch": 4.239202657807309, + "grad_norm": 6.951284885406494, + "learning_rate": 6.284923337612069e-06, + "loss": 1.7411, + "step": 956 + }, + { + "epoch": 4.243632336655592, + "grad_norm": 6.59275484085083, + "learning_rate": 6.215220336402272e-06, + "loss": 1.6294, + "step": 957 + }, + { + "epoch": 4.248062015503876, + "grad_norm": 6.854479789733887, + "learning_rate": 6.145880390769665e-06, + "loss": 1.1773, + "step": 958 + }, + { + "epoch": 4.25249169435216, + "grad_norm": 6.827851295471191, + "learning_rate": 6.076904075668671e-06, + "loss": 1.6421, + "step": 959 + }, + { + "epoch": 4.256921373200443, + "grad_norm": 4.86449670791626, + "learning_rate": 6.008291963038632e-06, + "loss": 1.6358, + "step": 960 + }, + { + "epoch": 4.2613510520487266, + "grad_norm": 4.6412835121154785, + "learning_rate": 5.940044621798896e-06, + "loss": 1.5333, + "step": 961 + }, + { + "epoch": 4.26578073089701, + "grad_norm": 4.715771198272705, + "learning_rate": 5.872162617844268e-06, + "loss": 1.7593, + "step": 962 + }, + { + "epoch": 4.270210409745293, + "grad_norm": 4.89564323425293, + "learning_rate": 5.804646514040163e-06, + "loss": 1.7706, + "step": 963 + }, + { + "epoch": 4.274640088593577, + "grad_norm": 6.569485187530518, + "learning_rate": 5.737496870218101e-06, + "loss": 1.4384, + "step": 964 + }, + { + "epoch": 4.27906976744186, + "grad_norm": 7.028257846832275, + "learning_rate": 5.670714243170938e-06, + "loss": 1.6896, + "step": 965 + }, + { + "epoch": 4.283499446290144, + "grad_norm": 5.102827072143555, + "learning_rate": 5.604299186648282e-06, + "loss": 1.8287, + "step": 966 + }, + { + "epoch": 4.287929125138428, + "grad_norm": 8.853278160095215, + "learning_rate": 5.538252251351933e-06, + "loss": 2.0511, + "step": 967 + }, + { + "epoch": 4.292358803986711, + "grad_norm": 5.434422016143799, + "learning_rate": 5.472573984931284e-06, + "loss": 1.3771, + "step": 968 + }, + { + "epoch": 4.2967884828349945, + "grad_norm": 7.120089054107666, + "learning_rate": 5.407264931978812e-06, + "loss": 1.7419, + "step": 969 + }, + { + "epoch": 4.301218161683278, + "grad_norm": 5.602823734283447, + "learning_rate": 5.3423256340255026e-06, + "loss": 1.6085, + "step": 970 + }, + { + "epoch": 4.305647840531561, + "grad_norm": 8.626879692077637, + "learning_rate": 5.277756629536434e-06, + "loss": 1.549, + "step": 971 + }, + { + "epoch": 4.310077519379845, + "grad_norm": 5.62658166885376, + "learning_rate": 5.21355845390627e-06, + "loss": 1.4692, + "step": 972 + }, + { + "epoch": 4.314507198228128, + "grad_norm": 4.512428283691406, + "learning_rate": 5.149731639454819e-06, + "loss": 1.5082, + "step": 973 + }, + { + "epoch": 4.318936877076412, + "grad_norm": 8.768449783325195, + "learning_rate": 5.086276715422644e-06, + "loss": 1.5685, + "step": 974 + }, + { + "epoch": 4.323366555924696, + "grad_norm": 6.369165897369385, + "learning_rate": 5.023194207966614e-06, + "loss": 1.6781, + "step": 975 + }, + { + "epoch": 4.327796234772979, + "grad_norm": 9.242612838745117, + "learning_rate": 4.9604846401556656e-06, + "loss": 1.7372, + "step": 976 + }, + { + "epoch": 4.332225913621262, + "grad_norm": 7.073103427886963, + "learning_rate": 4.898148531966307e-06, + "loss": 1.7124, + "step": 977 + }, + { + "epoch": 4.336655592469546, + "grad_norm": 7.333794116973877, + "learning_rate": 4.836186400278442e-06, + "loss": 1.742, + "step": 978 + }, + { + "epoch": 4.341085271317829, + "grad_norm": 6.159682750701904, + "learning_rate": 4.774598758870979e-06, + "loss": 1.2728, + "step": 979 + }, + { + "epoch": 4.3455149501661126, + "grad_norm": 6.649118423461914, + "learning_rate": 4.713386118417684e-06, + "loss": 1.5894, + "step": 980 + }, + { + "epoch": 4.349944629014397, + "grad_norm": 7.892576694488525, + "learning_rate": 4.652548986482813e-06, + "loss": 1.8148, + "step": 981 + }, + { + "epoch": 4.35437430786268, + "grad_norm": 4.917545795440674, + "learning_rate": 4.59208786751702e-06, + "loss": 1.5938, + "step": 982 + }, + { + "epoch": 4.358803986710964, + "grad_norm": 5.065957546234131, + "learning_rate": 4.532003262853107e-06, + "loss": 1.4148, + "step": 983 + }, + { + "epoch": 4.363233665559247, + "grad_norm": 5.68073034286499, + "learning_rate": 4.472295670701887e-06, + "loss": 1.4353, + "step": 984 + }, + { + "epoch": 4.36766334440753, + "grad_norm": 6.360957145690918, + "learning_rate": 4.412965586148071e-06, + "loss": 1.4035, + "step": 985 + }, + { + "epoch": 4.372093023255814, + "grad_norm": 4.0210981369018555, + "learning_rate": 4.354013501146109e-06, + "loss": 1.3468, + "step": 986 + }, + { + "epoch": 4.376522702104097, + "grad_norm": 6.830559730529785, + "learning_rate": 4.2954399045161746e-06, + "loss": 1.5327, + "step": 987 + }, + { + "epoch": 4.380952380952381, + "grad_norm": 4.727903842926025, + "learning_rate": 4.237245281940078e-06, + "loss": 1.4903, + "step": 988 + }, + { + "epoch": 4.385382059800665, + "grad_norm": 7.276794910430908, + "learning_rate": 4.17943011595725e-06, + "loss": 1.9205, + "step": 989 + }, + { + "epoch": 4.389811738648948, + "grad_norm": 4.359097003936768, + "learning_rate": 4.121994885960706e-06, + "loss": 1.6313, + "step": 990 + }, + { + "epoch": 4.3942414174972315, + "grad_norm": 4.636504173278809, + "learning_rate": 4.064940068193129e-06, + "loss": 1.3116, + "step": 991 + }, + { + "epoch": 4.398671096345515, + "grad_norm": 5.2092604637146, + "learning_rate": 4.008266135742894e-06, + "loss": 1.6403, + "step": 992 + }, + { + "epoch": 4.403100775193798, + "grad_norm": 5.949525356292725, + "learning_rate": 3.951973558540118e-06, + "loss": 1.3996, + "step": 993 + }, + { + "epoch": 4.407530454042082, + "grad_norm": 4.8397626876831055, + "learning_rate": 3.896062803352818e-06, + "loss": 1.8168, + "step": 994 + }, + { + "epoch": 4.411960132890365, + "grad_norm": 5.553168773651123, + "learning_rate": 3.84053433378298e-06, + "loss": 1.2202, + "step": 995 + }, + { + "epoch": 4.416389811738649, + "grad_norm": 5.277307987213135, + "learning_rate": 3.785388610262769e-06, + "loss": 1.3852, + "step": 996 + }, + { + "epoch": 4.420819490586933, + "grad_norm": 5.157022953033447, + "learning_rate": 3.730626090050676e-06, + "loss": 1.2935, + "step": 997 + }, + { + "epoch": 4.425249169435216, + "grad_norm": 6.345918655395508, + "learning_rate": 3.6762472272277582e-06, + "loss": 1.8446, + "step": 998 + }, + { + "epoch": 4.4296788482834994, + "grad_norm": 6.063523769378662, + "learning_rate": 3.6222524726938157e-06, + "loss": 1.5922, + "step": 999 + }, + { + "epoch": 4.434108527131783, + "grad_norm": 6.320123672485352, + "learning_rate": 3.568642274163725e-06, + "loss": 1.6638, + "step": 1000 + }, + { + "epoch": 4.438538205980066, + "grad_norm": 4.40836238861084, + "learning_rate": 3.5154170761636963e-06, + "loss": 1.703, + "step": 1001 + }, + { + "epoch": 4.44296788482835, + "grad_norm": 5.5650715827941895, + "learning_rate": 3.462577320027555e-06, + "loss": 1.6481, + "step": 1002 + }, + { + "epoch": 4.447397563676634, + "grad_norm": 6.705456733703613, + "learning_rate": 3.410123443893137e-06, + "loss": 1.4962, + "step": 1003 + }, + { + "epoch": 4.451827242524917, + "grad_norm": 6.810803413391113, + "learning_rate": 3.3580558826985963e-06, + "loss": 1.4737, + "step": 1004 + }, + { + "epoch": 4.456256921373201, + "grad_norm": 5.0352091789245605, + "learning_rate": 3.306375068178891e-06, + "loss": 1.6533, + "step": 1005 + }, + { + "epoch": 4.460686600221484, + "grad_norm": 5.479842185974121, + "learning_rate": 3.2550814288620834e-06, + "loss": 1.3461, + "step": 1006 + }, + { + "epoch": 4.465116279069767, + "grad_norm": 4.528697967529297, + "learning_rate": 3.204175390065889e-06, + "loss": 1.6675, + "step": 1007 + }, + { + "epoch": 4.469545957918051, + "grad_norm": 5.891976833343506, + "learning_rate": 3.1536573738940812e-06, + "loss": 1.6043, + "step": 1008 + }, + { + "epoch": 4.473975636766334, + "grad_norm": 5.913291931152344, + "learning_rate": 3.103527799233047e-06, + "loss": 1.4799, + "step": 1009 + }, + { + "epoch": 4.4784053156146175, + "grad_norm": 4.8424458503723145, + "learning_rate": 3.0537870817482663e-06, + "loss": 1.3448, + "step": 1010 + }, + { + "epoch": 4.482834994462902, + "grad_norm": 4.321615695953369, + "learning_rate": 3.004435633880881e-06, + "loss": 1.8712, + "step": 1011 + }, + { + "epoch": 4.487264673311185, + "grad_norm": 7.836488723754883, + "learning_rate": 2.955473864844299e-06, + "loss": 1.6831, + "step": 1012 + }, + { + "epoch": 4.4916943521594686, + "grad_norm": 7.034804344177246, + "learning_rate": 2.9069021806207573e-06, + "loss": 1.7364, + "step": 1013 + }, + { + "epoch": 4.496124031007752, + "grad_norm": 8.025514602661133, + "learning_rate": 2.858720983958013e-06, + "loss": 1.5019, + "step": 1014 + }, + { + "epoch": 4.500553709856035, + "grad_norm": 3.540686845779419, + "learning_rate": 2.8109306743659192e-06, + "loss": 1.3985, + "step": 1015 + }, + { + "epoch": 4.504983388704319, + "grad_norm": 10.38813591003418, + "learning_rate": 2.76353164811321e-06, + "loss": 1.8905, + "step": 1016 + }, + { + "epoch": 4.509413067552602, + "grad_norm": 4.8591628074646, + "learning_rate": 2.7165242982241436e-06, + "loss": 1.7566, + "step": 1017 + }, + { + "epoch": 4.513842746400886, + "grad_norm": 6.675249099731445, + "learning_rate": 2.6699090144752803e-06, + "loss": 1.5333, + "step": 1018 + }, + { + "epoch": 4.51827242524917, + "grad_norm": 4.969045162200928, + "learning_rate": 2.6236861833922376e-06, + "loss": 1.1801, + "step": 1019 + }, + { + "epoch": 4.522702104097453, + "grad_norm": 4.894344329833984, + "learning_rate": 2.5778561882464525e-06, + "loss": 1.629, + "step": 1020 + }, + { + "epoch": 4.5271317829457365, + "grad_norm": 6.36023473739624, + "learning_rate": 2.5324194090521002e-06, + "loss": 1.59, + "step": 1021 + }, + { + "epoch": 4.53156146179402, + "grad_norm": 8.083828926086426, + "learning_rate": 2.487376222562815e-06, + "loss": 1.3789, + "step": 1022 + }, + { + "epoch": 4.535991140642303, + "grad_norm": 6.636145114898682, + "learning_rate": 2.44272700226868e-06, + "loss": 1.9007, + "step": 1023 + }, + { + "epoch": 4.540420819490587, + "grad_norm": 6.143908500671387, + "learning_rate": 2.3984721183930303e-06, + "loss": 1.7806, + "step": 1024 + }, + { + "epoch": 4.544850498338871, + "grad_norm": 5.793661594390869, + "learning_rate": 2.3546119378895005e-06, + "loss": 1.874, + "step": 1025 + }, + { + "epoch": 4.549280177187154, + "grad_norm": 4.639335632324219, + "learning_rate": 2.311146824438848e-06, + "loss": 1.8634, + "step": 1026 + }, + { + "epoch": 4.553709856035438, + "grad_norm": 12.957326889038086, + "learning_rate": 2.2680771384460507e-06, + "loss": 1.5008, + "step": 1027 + }, + { + "epoch": 4.558139534883721, + "grad_norm": 5.100373268127441, + "learning_rate": 2.2254032370372492e-06, + "loss": 1.8479, + "step": 1028 + }, + { + "epoch": 4.562569213732004, + "grad_norm": 4.1848015785217285, + "learning_rate": 2.1831254740568242e-06, + "loss": 1.4017, + "step": 1029 + }, + { + "epoch": 4.566998892580288, + "grad_norm": 7.201414585113525, + "learning_rate": 2.1412442000644352e-06, + "loss": 1.756, + "step": 1030 + }, + { + "epoch": 4.571428571428571, + "grad_norm": 7.138917446136475, + "learning_rate": 2.0997597623321264e-06, + "loss": 1.7556, + "step": 1031 + }, + { + "epoch": 4.5758582502768546, + "grad_norm": 8.494391441345215, + "learning_rate": 2.058672504841458e-06, + "loss": 2.0976, + "step": 1032 + }, + { + "epoch": 4.580287929125138, + "grad_norm": 9.561163902282715, + "learning_rate": 2.017982768280624e-06, + "loss": 1.1794, + "step": 1033 + }, + { + "epoch": 4.584717607973422, + "grad_norm": 6.571170806884766, + "learning_rate": 1.9776908900416777e-06, + "loss": 1.4607, + "step": 1034 + }, + { + "epoch": 4.589147286821706, + "grad_norm": 4.655641555786133, + "learning_rate": 1.937797204217656e-06, + "loss": 1.8607, + "step": 1035 + }, + { + "epoch": 4.593576965669989, + "grad_norm": 4.481752872467041, + "learning_rate": 1.898302041599892e-06, + "loss": 1.6295, + "step": 1036 + }, + { + "epoch": 4.598006644518272, + "grad_norm": 6.839178562164307, + "learning_rate": 1.859205729675223e-06, + "loss": 1.5424, + "step": 1037 + }, + { + "epoch": 4.602436323366556, + "grad_norm": 5.645241737365723, + "learning_rate": 1.820508592623288e-06, + "loss": 1.2222, + "step": 1038 + }, + { + "epoch": 4.606866002214839, + "grad_norm": 6.006126880645752, + "learning_rate": 1.7822109513138453e-06, + "loss": 1.6035, + "step": 1039 + }, + { + "epoch": 4.6112956810631225, + "grad_norm": 7.534673690795898, + "learning_rate": 1.744313123304092e-06, + "loss": 1.4323, + "step": 1040 + }, + { + "epoch": 4.615725359911407, + "grad_norm": 6.148129940032959, + "learning_rate": 1.706815422836061e-06, + "loss": 1.7501, + "step": 1041 + }, + { + "epoch": 4.62015503875969, + "grad_norm": 4.745789051055908, + "learning_rate": 1.6697181608339828e-06, + "loss": 1.4226, + "step": 1042 + }, + { + "epoch": 4.6245847176079735, + "grad_norm": 4.307113170623779, + "learning_rate": 1.6330216449017443e-06, + "loss": 1.6688, + "step": 1043 + }, + { + "epoch": 4.629014396456257, + "grad_norm": 5.552404403686523, + "learning_rate": 1.5967261793202904e-06, + "loss": 1.5026, + "step": 1044 + }, + { + "epoch": 4.63344407530454, + "grad_norm": 9.312074661254883, + "learning_rate": 1.5608320650451425e-06, + "loss": 1.3092, + "step": 1045 + }, + { + "epoch": 4.637873754152824, + "grad_norm": 4.408802032470703, + "learning_rate": 1.5253395997039066e-06, + "loss": 1.4818, + "step": 1046 + }, + { + "epoch": 4.642303433001107, + "grad_norm": 8.572196960449219, + "learning_rate": 1.4902490775937527e-06, + "loss": 1.4766, + "step": 1047 + }, + { + "epoch": 4.646733111849391, + "grad_norm": 4.4483208656311035, + "learning_rate": 1.4555607896790446e-06, + "loss": 1.7217, + "step": 1048 + }, + { + "epoch": 4.651162790697675, + "grad_norm": 5.438201904296875, + "learning_rate": 1.4212750235888416e-06, + "loss": 1.7277, + "step": 1049 + }, + { + "epoch": 4.655592469545958, + "grad_norm": 4.61346960067749, + "learning_rate": 1.3873920636146342e-06, + "loss": 1.2524, + "step": 1050 + }, + { + "epoch": 4.6600221483942414, + "grad_norm": 12.7802095413208, + "learning_rate": 1.353912190707851e-06, + "loss": 1.3947, + "step": 1051 + }, + { + "epoch": 4.664451827242525, + "grad_norm": 9.374958992004395, + "learning_rate": 1.320835682477628e-06, + "loss": 1.8787, + "step": 1052 + }, + { + "epoch": 4.668881506090808, + "grad_norm": 7.309841632843018, + "learning_rate": 1.288162813188476e-06, + "loss": 1.5597, + "step": 1053 + }, + { + "epoch": 4.673311184939092, + "grad_norm": 7.758358478546143, + "learning_rate": 1.2558938537579835e-06, + "loss": 1.5036, + "step": 1054 + }, + { + "epoch": 4.677740863787376, + "grad_norm": 5.403035640716553, + "learning_rate": 1.2240290717546176e-06, + "loss": 1.6294, + "step": 1055 + }, + { + "epoch": 4.682170542635659, + "grad_norm": 4.8274102210998535, + "learning_rate": 1.1925687313954437e-06, + "loss": 1.5611, + "step": 1056 + }, + { + "epoch": 4.686600221483943, + "grad_norm": 7.348108768463135, + "learning_rate": 1.1615130935439978e-06, + "loss": 1.504, + "step": 1057 + }, + { + "epoch": 4.691029900332226, + "grad_norm": 7.85156774520874, + "learning_rate": 1.1308624157080837e-06, + "loss": 1.5926, + "step": 1058 + }, + { + "epoch": 4.695459579180509, + "grad_norm": 7.822666168212891, + "learning_rate": 1.1006169520376586e-06, + "loss": 1.6488, + "step": 1059 + }, + { + "epoch": 4.699889258028793, + "grad_norm": 13.061169624328613, + "learning_rate": 1.070776953322694e-06, + "loss": 1.2314, + "step": 1060 + }, + { + "epoch": 4.704318936877076, + "grad_norm": 7.329505443572998, + "learning_rate": 1.0413426669911408e-06, + "loss": 1.7047, + "step": 1061 + }, + { + "epoch": 4.70874861572536, + "grad_norm": 7.315627574920654, + "learning_rate": 1.0123143371068456e-06, + "loss": 1.4365, + "step": 1062 + }, + { + "epoch": 4.713178294573644, + "grad_norm": 5.628126621246338, + "learning_rate": 9.836922043675368e-07, + "loss": 1.6823, + "step": 1063 + }, + { + "epoch": 4.717607973421927, + "grad_norm": 6.389928340911865, + "learning_rate": 9.554765061028371e-07, + "loss": 1.4238, + "step": 1064 + }, + { + "epoch": 4.7220376522702106, + "grad_norm": 5.262420177459717, + "learning_rate": 9.276674762722704e-07, + "loss": 1.486, + "step": 1065 + }, + { + "epoch": 4.726467331118494, + "grad_norm": 4.576107025146484, + "learning_rate": 9.002653454633581e-07, + "loss": 1.5454, + "step": 1066 + }, + { + "epoch": 4.730897009966777, + "grad_norm": 5.287619113922119, + "learning_rate": 8.732703408896648e-07, + "loss": 1.8704, + "step": 1067 + }, + { + "epoch": 4.735326688815061, + "grad_norm": 4.985103607177734, + "learning_rate": 8.46682686388961e-07, + "loss": 1.3233, + "step": 1068 + }, + { + "epoch": 4.739756367663344, + "grad_norm": 4.814030647277832, + "learning_rate": 8.205026024213192e-07, + "loss": 1.5228, + "step": 1069 + }, + { + "epoch": 4.7441860465116275, + "grad_norm": 6.874385356903076, + "learning_rate": 7.947303060673372e-07, + "loss": 1.8679, + "step": 1070 + }, + { + "epoch": 4.748615725359912, + "grad_norm": 5.1823225021362305, + "learning_rate": 7.693660110262902e-07, + "loss": 1.513, + "step": 1071 + }, + { + "epoch": 4.753045404208195, + "grad_norm": 6.403878211975098, + "learning_rate": 7.444099276143812e-07, + "loss": 1.4745, + "step": 1072 + }, + { + "epoch": 4.7574750830564785, + "grad_norm": 5.249782562255859, + "learning_rate": 7.198622627630047e-07, + "loss": 1.5437, + "step": 1073 + }, + { + "epoch": 4.761904761904762, + "grad_norm": 11.559475898742676, + "learning_rate": 6.957232200170083e-07, + "loss": 1.7486, + "step": 1074 + }, + { + "epoch": 4.766334440753045, + "grad_norm": 5.234459400177002, + "learning_rate": 6.719929995330388e-07, + "loss": 1.4821, + "step": 1075 + }, + { + "epoch": 4.770764119601329, + "grad_norm": 6.172218322753906, + "learning_rate": 6.486717980778323e-07, + "loss": 1.6629, + "step": 1076 + }, + { + "epoch": 4.775193798449612, + "grad_norm": 4.527020454406738, + "learning_rate": 6.257598090266325e-07, + "loss": 1.444, + "step": 1077 + }, + { + "epoch": 4.779623477297896, + "grad_norm": 7.31265926361084, + "learning_rate": 6.032572223615695e-07, + "loss": 1.4468, + "step": 1078 + }, + { + "epoch": 4.78405315614618, + "grad_norm": 5.2357988357543945, + "learning_rate": 5.811642246700722e-07, + "loss": 1.6563, + "step": 1079 + }, + { + "epoch": 4.788482834994463, + "grad_norm": 4.705022811889648, + "learning_rate": 5.594809991433414e-07, + "loss": 1.6064, + "step": 1080 + }, + { + "epoch": 4.792912513842746, + "grad_norm": 6.180661678314209, + "learning_rate": 5.382077255747964e-07, + "loss": 1.2207, + "step": 1081 + }, + { + "epoch": 4.79734219269103, + "grad_norm": 6.87699031829834, + "learning_rate": 5.173445803586307e-07, + "loss": 1.7086, + "step": 1082 + }, + { + "epoch": 4.801771871539313, + "grad_norm": 4.476713180541992, + "learning_rate": 4.968917364883196e-07, + "loss": 1.579, + "step": 1083 + }, + { + "epoch": 4.8062015503875966, + "grad_norm": 7.013663291931152, + "learning_rate": 4.76849363555204e-07, + "loss": 1.3113, + "step": 1084 + }, + { + "epoch": 4.810631229235881, + "grad_norm": 5.766423225402832, + "learning_rate": 4.5721762774704747e-07, + "loss": 1.646, + "step": 1085 + }, + { + "epoch": 4.815060908084164, + "grad_norm": 5.878317832946777, + "learning_rate": 4.379966918467093e-07, + "loss": 1.4083, + "step": 1086 + }, + { + "epoch": 4.819490586932448, + "grad_norm": 4.808867454528809, + "learning_rate": 4.1918671523075716e-07, + "loss": 1.5285, + "step": 1087 + }, + { + "epoch": 4.823920265780731, + "grad_norm": 6.050039768218994, + "learning_rate": 4.0078785386815644e-07, + "loss": 1.8131, + "step": 1088 + }, + { + "epoch": 4.828349944629014, + "grad_norm": 6.054554462432861, + "learning_rate": 3.8280026031898287e-07, + "loss": 1.2736, + "step": 1089 + }, + { + "epoch": 4.832779623477298, + "grad_norm": 11.464051246643066, + "learning_rate": 3.6522408373314553e-07, + "loss": 1.8098, + "step": 1090 + }, + { + "epoch": 4.837209302325581, + "grad_norm": 5.715487957000732, + "learning_rate": 3.4805946984917683e-07, + "loss": 1.7507, + "step": 1091 + }, + { + "epoch": 4.841638981173865, + "grad_norm": 5.533278465270996, + "learning_rate": 3.3130656099297777e-07, + "loss": 1.6937, + "step": 1092 + }, + { + "epoch": 4.846068660022149, + "grad_norm": 4.647676944732666, + "learning_rate": 3.149654960766857e-07, + "loss": 1.8874, + "step": 1093 + }, + { + "epoch": 4.850498338870432, + "grad_norm": 6.739248752593994, + "learning_rate": 2.9903641059749186e-07, + "loss": 1.5665, + "step": 1094 + }, + { + "epoch": 4.8549280177187155, + "grad_norm": 5.865329742431641, + "learning_rate": 2.8351943663653124e-07, + "loss": 1.4985, + "step": 1095 + }, + { + "epoch": 4.859357696566999, + "grad_norm": 5.861200332641602, + "learning_rate": 2.684147028577777e-07, + "loss": 1.6865, + "step": 1096 + }, + { + "epoch": 4.863787375415282, + "grad_norm": 5.523236274719238, + "learning_rate": 2.537223345069895e-07, + "loss": 1.5327, + "step": 1097 + }, + { + "epoch": 4.868217054263566, + "grad_norm": 5.882645130157471, + "learning_rate": 2.3944245341065987e-07, + "loss": 1.6286, + "step": 1098 + }, + { + "epoch": 4.872646733111849, + "grad_norm": 8.576865196228027, + "learning_rate": 2.255751779750237e-07, + "loss": 1.7435, + "step": 1099 + }, + { + "epoch": 4.877076411960132, + "grad_norm": 5.604006767272949, + "learning_rate": 2.12120623185047e-07, + "loss": 1.4963, + "step": 1100 + }, + { + "epoch": 4.881506090808417, + "grad_norm": 10.904509544372559, + "learning_rate": 1.9907890060348878e-07, + "loss": 1.636, + "step": 1101 + }, + { + "epoch": 4.8859357696567, + "grad_norm": 5.717619895935059, + "learning_rate": 1.8645011836999626e-07, + "loss": 1.5698, + "step": 1102 + }, + { + "epoch": 4.8903654485049834, + "grad_norm": 5.974473476409912, + "learning_rate": 1.7423438120017234e-07, + "loss": 1.7177, + "step": 1103 + }, + { + "epoch": 4.894795127353267, + "grad_norm": 5.090288162231445, + "learning_rate": 1.624317903847372e-07, + "loss": 1.4201, + "step": 1104 + }, + { + "epoch": 4.89922480620155, + "grad_norm": 4.930177688598633, + "learning_rate": 1.5104244378865152e-07, + "loss": 1.246, + "step": 1105 + }, + { + "epoch": 4.903654485049834, + "grad_norm": 9.135576248168945, + "learning_rate": 1.4006643585035006e-07, + "loss": 1.6311, + "step": 1106 + }, + { + "epoch": 4.908084163898117, + "grad_norm": 5.666680812835693, + "learning_rate": 1.2950385758094263e-07, + "loss": 1.513, + "step": 1107 + }, + { + "epoch": 4.912513842746401, + "grad_norm": 6.164681434631348, + "learning_rate": 1.1935479656342562e-07, + "loss": 1.8122, + "step": 1108 + }, + { + "epoch": 4.916943521594685, + "grad_norm": 8.394211769104004, + "learning_rate": 1.0961933695201598e-07, + "loss": 1.5239, + "step": 1109 + }, + { + "epoch": 4.921373200442968, + "grad_norm": 6.864090919494629, + "learning_rate": 1.0029755947139618e-07, + "loss": 1.0922, + "step": 1110 + }, + { + "epoch": 4.925802879291251, + "grad_norm": 5.1737751960754395, + "learning_rate": 9.138954141608702e-08, + "loss": 1.5132, + "step": 1111 + }, + { + "epoch": 4.930232558139535, + "grad_norm": 4.374934673309326, + "learning_rate": 8.289535664978698e-08, + "loss": 1.495, + "step": 1112 + }, + { + "epoch": 4.934662236987818, + "grad_norm": 4.535995960235596, + "learning_rate": 7.481507560475053e-08, + "loss": 1.787, + "step": 1113 + }, + { + "epoch": 4.9390919158361015, + "grad_norm": 5.156623840332031, + "learning_rate": 6.714876528123304e-08, + "loss": 1.4265, + "step": 1114 + }, + { + "epoch": 4.943521594684386, + "grad_norm": 5.1794328689575195, + "learning_rate": 5.989648924690782e-08, + "loss": 1.6533, + "step": 1115 + }, + { + "epoch": 4.947951273532669, + "grad_norm": 6.805692672729492, + "learning_rate": 5.305830763635555e-08, + "loss": 1.4241, + "step": 1116 + }, + { + "epoch": 4.9523809523809526, + "grad_norm": 4.86265230178833, + "learning_rate": 4.663427715055346e-08, + "loss": 1.4307, + "step": 1117 + }, + { + "epoch": 4.956810631229236, + "grad_norm": 5.848089218139648, + "learning_rate": 4.062445105642021e-08, + "loss": 1.7906, + "step": 1118 + }, + { + "epoch": 4.961240310077519, + "grad_norm": 6.41409969329834, + "learning_rate": 3.502887918636066e-08, + "loss": 1.5248, + "step": 1119 + }, + { + "epoch": 4.965669988925803, + "grad_norm": 4.486702919006348, + "learning_rate": 2.984760793785513e-08, + "loss": 1.8782, + "step": 1120 + }, + { + "epoch": 4.970099667774086, + "grad_norm": 10.109087944030762, + "learning_rate": 2.508068027308186e-08, + "loss": 1.5929, + "step": 1121 + }, + { + "epoch": 4.97452934662237, + "grad_norm": 5.497599124908447, + "learning_rate": 2.07281357185507e-08, + "loss": 1.2086, + "step": 1122 + }, + { + "epoch": 4.978959025470654, + "grad_norm": 5.0335001945495605, + "learning_rate": 1.6790010364786667e-08, + "loss": 1.5011, + "step": 1123 + }, + { + "epoch": 4.983388704318937, + "grad_norm": 6.264035224914551, + "learning_rate": 1.3266336866024631e-08, + "loss": 1.5586, + "step": 1124 + }, + { + "epoch": 4.9878183831672205, + "grad_norm": 4.8113274574279785, + "learning_rate": 1.0157144439937317e-08, + "loss": 1.4083, + "step": 1125 + }, + { + "epoch": 4.9878183831672205, + "step": 1125, + "total_flos": 1.8752222545942938e+18, + "train_loss": 1.867327324125502, + "train_runtime": 15337.9497, + "train_samples_per_second": 2.353, + "train_steps_per_second": 0.073 + } + ], + "logging_steps": 1.0, + "max_steps": 1125, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.8752222545942938e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}