{ "best_metric": 0.7813047170639038, "best_model_checkpoint": "model/checkpoints/run1-python-codegen/checkpoint-10000", "epoch": 4.999356582164458, "eval_steps": 1000, "global_step": 19425, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025736713421696047, "grad_norm": 1.7038387060165405, "learning_rate": 4.997425997425998e-05, "loss": 0.9335, "step": 10 }, { "epoch": 0.0051473426843392095, "grad_norm": 1.1648929119110107, "learning_rate": 4.994851994851995e-05, "loss": 0.9433, "step": 20 }, { "epoch": 0.007721014026508815, "grad_norm": 3.9555516242980957, "learning_rate": 4.992277992277993e-05, "loss": 0.9713, "step": 30 }, { "epoch": 0.010294685368678419, "grad_norm": 0.9799448847770691, "learning_rate": 4.9897039897039896e-05, "loss": 0.9518, "step": 40 }, { "epoch": 0.012868356710848025, "grad_norm": 1.26755952835083, "learning_rate": 4.9871299871299874e-05, "loss": 1.0754, "step": 50 }, { "epoch": 0.01544202805301763, "grad_norm": 4.542232513427734, "learning_rate": 4.9845559845559846e-05, "loss": 0.9471, "step": 60 }, { "epoch": 0.018015699395187233, "grad_norm": 1.2068519592285156, "learning_rate": 4.9819819819819824e-05, "loss": 0.9576, "step": 70 }, { "epoch": 0.020589370737356838, "grad_norm": 1.019250750541687, "learning_rate": 4.9794079794079796e-05, "loss": 0.9516, "step": 80 }, { "epoch": 0.023163042079526446, "grad_norm": 1.2534639835357666, "learning_rate": 4.9768339768339774e-05, "loss": 1.0046, "step": 90 }, { "epoch": 0.02573671342169605, "grad_norm": 0.8234790563583374, "learning_rate": 4.9742599742599746e-05, "loss": 0.8847, "step": 100 }, { "epoch": 0.028310384763865656, "grad_norm": 0.9790653586387634, "learning_rate": 4.971685971685972e-05, "loss": 0.8495, "step": 110 }, { "epoch": 0.03088405610603526, "grad_norm": 4.180076599121094, "learning_rate": 4.969111969111969e-05, "loss": 0.8801, "step": 120 }, { "epoch": 0.03345772744820486, "grad_norm": 0.9933150410652161, "learning_rate": 4.966537966537967e-05, "loss": 0.9608, "step": 130 }, { "epoch": 0.036031398790374466, "grad_norm": 1.1884509325027466, "learning_rate": 4.963963963963964e-05, "loss": 0.9133, "step": 140 }, { "epoch": 0.03860507013254407, "grad_norm": 4.013946056365967, "learning_rate": 4.961389961389962e-05, "loss": 0.9142, "step": 150 }, { "epoch": 0.041178741474713676, "grad_norm": 2.0833263397216797, "learning_rate": 4.958815958815959e-05, "loss": 0.9864, "step": 160 }, { "epoch": 0.04375241281688328, "grad_norm": 3.602749824523926, "learning_rate": 4.956241956241957e-05, "loss": 0.876, "step": 170 }, { "epoch": 0.04632608415905289, "grad_norm": 1.2331393957138062, "learning_rate": 4.953667953667954e-05, "loss": 0.9233, "step": 180 }, { "epoch": 0.0488997555012225, "grad_norm": 1.0604982376098633, "learning_rate": 4.951093951093951e-05, "loss": 0.9497, "step": 190 }, { "epoch": 0.0514734268433921, "grad_norm": 1.0822978019714355, "learning_rate": 4.948519948519948e-05, "loss": 0.8719, "step": 200 }, { "epoch": 0.05404709818556171, "grad_norm": 0.9496659636497498, "learning_rate": 4.945945945945946e-05, "loss": 0.9186, "step": 210 }, { "epoch": 0.05662076952773131, "grad_norm": 1.0661075115203857, "learning_rate": 4.943371943371944e-05, "loss": 0.8779, "step": 220 }, { "epoch": 0.059194440869900916, "grad_norm": 1.1287846565246582, "learning_rate": 4.940797940797941e-05, "loss": 0.8501, "step": 230 }, { "epoch": 0.06176811221207052, "grad_norm": 1.16519296169281, "learning_rate": 4.938223938223939e-05, "loss": 0.9317, "step": 240 }, { "epoch": 0.06434178355424013, "grad_norm": 1.5323387384414673, "learning_rate": 4.935649935649936e-05, "loss": 0.9511, "step": 250 }, { "epoch": 0.06691545489640972, "grad_norm": 1.1095296144485474, "learning_rate": 4.933075933075933e-05, "loss": 0.9372, "step": 260 }, { "epoch": 0.06948912623857934, "grad_norm": 1.1270581483840942, "learning_rate": 4.9305019305019304e-05, "loss": 0.9549, "step": 270 }, { "epoch": 0.07206279758074893, "grad_norm": 1.0347909927368164, "learning_rate": 4.927927927927928e-05, "loss": 0.882, "step": 280 }, { "epoch": 0.07463646892291854, "grad_norm": 0.9273091554641724, "learning_rate": 4.9253539253539254e-05, "loss": 0.9191, "step": 290 }, { "epoch": 0.07721014026508814, "grad_norm": 0.9906871318817139, "learning_rate": 4.922779922779923e-05, "loss": 0.9259, "step": 300 }, { "epoch": 0.07978381160725775, "grad_norm": 0.8945180773735046, "learning_rate": 4.9202059202059204e-05, "loss": 0.8202, "step": 310 }, { "epoch": 0.08235748294942735, "grad_norm": 1.2765549421310425, "learning_rate": 4.917631917631918e-05, "loss": 0.9122, "step": 320 }, { "epoch": 0.08493115429159696, "grad_norm": 0.8597398400306702, "learning_rate": 4.9150579150579154e-05, "loss": 0.9114, "step": 330 }, { "epoch": 0.08750482563376656, "grad_norm": 3.5975146293640137, "learning_rate": 4.9124839124839125e-05, "loss": 0.8598, "step": 340 }, { "epoch": 0.09007849697593617, "grad_norm": 4.273248672485352, "learning_rate": 4.90990990990991e-05, "loss": 0.9325, "step": 350 }, { "epoch": 0.09265216831810578, "grad_norm": 0.9573097229003906, "learning_rate": 4.9073359073359075e-05, "loss": 0.992, "step": 360 }, { "epoch": 0.09522583966027538, "grad_norm": 1.2998064756393433, "learning_rate": 4.904761904761905e-05, "loss": 0.8984, "step": 370 }, { "epoch": 0.097799511002445, "grad_norm": 1.3857780694961548, "learning_rate": 4.9021879021879025e-05, "loss": 1.0102, "step": 380 }, { "epoch": 0.10037318234461459, "grad_norm": 1.2418707609176636, "learning_rate": 4.8996138996139e-05, "loss": 0.9023, "step": 390 }, { "epoch": 0.1029468536867842, "grad_norm": 1.2056580781936646, "learning_rate": 4.8970398970398975e-05, "loss": 0.8691, "step": 400 }, { "epoch": 0.1055205250289538, "grad_norm": 0.9754079580307007, "learning_rate": 4.894465894465895e-05, "loss": 0.9484, "step": 410 }, { "epoch": 0.10809419637112341, "grad_norm": 0.9066883325576782, "learning_rate": 4.891891891891892e-05, "loss": 0.9009, "step": 420 }, { "epoch": 0.11066786771329301, "grad_norm": 1.480106234550476, "learning_rate": 4.88931788931789e-05, "loss": 0.8764, "step": 430 }, { "epoch": 0.11324153905546262, "grad_norm": 1.1571600437164307, "learning_rate": 4.886743886743887e-05, "loss": 0.8711, "step": 440 }, { "epoch": 0.11581521039763222, "grad_norm": 0.9223948121070862, "learning_rate": 4.884169884169885e-05, "loss": 0.8466, "step": 450 }, { "epoch": 0.11838888173980183, "grad_norm": 1.7378956079483032, "learning_rate": 4.881595881595882e-05, "loss": 0.98, "step": 460 }, { "epoch": 0.12096255308197143, "grad_norm": 1.114259958267212, "learning_rate": 4.87902187902188e-05, "loss": 0.9181, "step": 470 }, { "epoch": 0.12353622442414104, "grad_norm": 1.0394957065582275, "learning_rate": 4.876447876447877e-05, "loss": 0.9139, "step": 480 }, { "epoch": 0.12610989576631065, "grad_norm": 1.100290298461914, "learning_rate": 4.873873873873874e-05, "loss": 0.8707, "step": 490 }, { "epoch": 0.12868356710848025, "grad_norm": 0.9463397264480591, "learning_rate": 4.871299871299871e-05, "loss": 0.8608, "step": 500 }, { "epoch": 0.13125723845064985, "grad_norm": 0.9843112230300903, "learning_rate": 4.868725868725869e-05, "loss": 0.931, "step": 510 }, { "epoch": 0.13383090979281945, "grad_norm": 1.050361156463623, "learning_rate": 4.866151866151866e-05, "loss": 0.9192, "step": 520 }, { "epoch": 0.13640458113498907, "grad_norm": 1.1644479036331177, "learning_rate": 4.863577863577864e-05, "loss": 0.8797, "step": 530 }, { "epoch": 0.13897825247715867, "grad_norm": 1.2110307216644287, "learning_rate": 4.861003861003861e-05, "loss": 0.9662, "step": 540 }, { "epoch": 0.14155192381932827, "grad_norm": 1.7485480308532715, "learning_rate": 4.858429858429859e-05, "loss": 0.9391, "step": 550 }, { "epoch": 0.14412559516149787, "grad_norm": 4.363419532775879, "learning_rate": 4.855855855855856e-05, "loss": 0.9487, "step": 560 }, { "epoch": 0.1466992665036675, "grad_norm": 1.028351068496704, "learning_rate": 4.853281853281853e-05, "loss": 0.7843, "step": 570 }, { "epoch": 0.1492729378458371, "grad_norm": 1.3750572204589844, "learning_rate": 4.8507078507078505e-05, "loss": 0.9074, "step": 580 }, { "epoch": 0.1518466091880067, "grad_norm": 4.166394233703613, "learning_rate": 4.848133848133848e-05, "loss": 0.9494, "step": 590 }, { "epoch": 0.15442028053017628, "grad_norm": 1.1192243099212646, "learning_rate": 4.8455598455598455e-05, "loss": 0.9937, "step": 600 }, { "epoch": 0.1569939518723459, "grad_norm": 1.243043065071106, "learning_rate": 4.842985842985843e-05, "loss": 0.8259, "step": 610 }, { "epoch": 0.1595676232145155, "grad_norm": 1.0789657831192017, "learning_rate": 4.8404118404118405e-05, "loss": 0.8466, "step": 620 }, { "epoch": 0.1621412945566851, "grad_norm": 1.2137137651443481, "learning_rate": 4.837837837837838e-05, "loss": 0.8358, "step": 630 }, { "epoch": 0.1647149658988547, "grad_norm": 5.4415974617004395, "learning_rate": 4.8352638352638355e-05, "loss": 0.9067, "step": 640 }, { "epoch": 0.16728863724102433, "grad_norm": 1.1608994007110596, "learning_rate": 4.8326898326898327e-05, "loss": 0.8828, "step": 650 }, { "epoch": 0.16986230858319393, "grad_norm": 1.287034034729004, "learning_rate": 4.8301158301158305e-05, "loss": 0.8905, "step": 660 }, { "epoch": 0.17243597992536353, "grad_norm": 1.0879331827163696, "learning_rate": 4.8275418275418277e-05, "loss": 0.8821, "step": 670 }, { "epoch": 0.17500965126753312, "grad_norm": 1.1660597324371338, "learning_rate": 4.8249678249678255e-05, "loss": 0.8955, "step": 680 }, { "epoch": 0.17758332260970275, "grad_norm": 0.8420181274414062, "learning_rate": 4.8223938223938227e-05, "loss": 0.8927, "step": 690 }, { "epoch": 0.18015699395187235, "grad_norm": 0.7901207208633423, "learning_rate": 4.8198198198198205e-05, "loss": 0.8926, "step": 700 }, { "epoch": 0.18273066529404194, "grad_norm": 1.2035242319107056, "learning_rate": 4.8172458172458177e-05, "loss": 0.8742, "step": 710 }, { "epoch": 0.18530433663621157, "grad_norm": 0.8758509159088135, "learning_rate": 4.814671814671815e-05, "loss": 0.9302, "step": 720 }, { "epoch": 0.18787800797838117, "grad_norm": 0.9391271471977234, "learning_rate": 4.812097812097812e-05, "loss": 0.8675, "step": 730 }, { "epoch": 0.19045167932055077, "grad_norm": 4.989656448364258, "learning_rate": 4.80952380952381e-05, "loss": 0.9314, "step": 740 }, { "epoch": 0.19302535066272036, "grad_norm": 2.094479560852051, "learning_rate": 4.806949806949807e-05, "loss": 0.8705, "step": 750 }, { "epoch": 0.19559902200489, "grad_norm": 1.1212397813796997, "learning_rate": 4.804375804375805e-05, "loss": 0.9013, "step": 760 }, { "epoch": 0.1981726933470596, "grad_norm": 3.019252300262451, "learning_rate": 4.801801801801802e-05, "loss": 0.8546, "step": 770 }, { "epoch": 0.20074636468922918, "grad_norm": 2.4471728801727295, "learning_rate": 4.7992277992278e-05, "loss": 0.8738, "step": 780 }, { "epoch": 0.20332003603139878, "grad_norm": 1.226980209350586, "learning_rate": 4.796653796653797e-05, "loss": 0.921, "step": 790 }, { "epoch": 0.2058937073735684, "grad_norm": 0.9777143001556396, "learning_rate": 4.794079794079794e-05, "loss": 0.893, "step": 800 }, { "epoch": 0.208467378715738, "grad_norm": 1.05388343334198, "learning_rate": 4.791505791505791e-05, "loss": 0.8911, "step": 810 }, { "epoch": 0.2110410500579076, "grad_norm": 3.421910285949707, "learning_rate": 4.788931788931789e-05, "loss": 0.9105, "step": 820 }, { "epoch": 0.2136147214000772, "grad_norm": 1.2801098823547363, "learning_rate": 4.786357786357786e-05, "loss": 0.8692, "step": 830 }, { "epoch": 0.21618839274224683, "grad_norm": 0.8514150381088257, "learning_rate": 4.783783783783784e-05, "loss": 0.8643, "step": 840 }, { "epoch": 0.21876206408441642, "grad_norm": 3.815889596939087, "learning_rate": 4.781209781209782e-05, "loss": 0.8395, "step": 850 }, { "epoch": 0.22133573542658602, "grad_norm": 1.0590102672576904, "learning_rate": 4.778635778635779e-05, "loss": 0.9168, "step": 860 }, { "epoch": 0.22390940676875562, "grad_norm": 0.9641560316085815, "learning_rate": 4.776061776061776e-05, "loss": 0.8777, "step": 870 }, { "epoch": 0.22648307811092525, "grad_norm": 1.1582986116409302, "learning_rate": 4.7734877734877735e-05, "loss": 0.9626, "step": 880 }, { "epoch": 0.22905674945309484, "grad_norm": 1.3446133136749268, "learning_rate": 4.770913770913771e-05, "loss": 0.8254, "step": 890 }, { "epoch": 0.23163042079526444, "grad_norm": 0.9514161944389343, "learning_rate": 4.7683397683397685e-05, "loss": 0.9177, "step": 900 }, { "epoch": 0.23420409213743404, "grad_norm": 1.3230582475662231, "learning_rate": 4.765765765765766e-05, "loss": 0.9428, "step": 910 }, { "epoch": 0.23677776347960366, "grad_norm": 3.5208494663238525, "learning_rate": 4.7631917631917635e-05, "loss": 0.9657, "step": 920 }, { "epoch": 0.23935143482177326, "grad_norm": 1.2036529779434204, "learning_rate": 4.760617760617761e-05, "loss": 0.9157, "step": 930 }, { "epoch": 0.24192510616394286, "grad_norm": 0.9241206645965576, "learning_rate": 4.7580437580437585e-05, "loss": 0.931, "step": 940 }, { "epoch": 0.24449877750611246, "grad_norm": 0.9766173958778381, "learning_rate": 4.7554697554697556e-05, "loss": 0.8657, "step": 950 }, { "epoch": 0.24707244884828208, "grad_norm": 1.0403867959976196, "learning_rate": 4.752895752895753e-05, "loss": 1.002, "step": 960 }, { "epoch": 0.24964612019045168, "grad_norm": 1.6326649188995361, "learning_rate": 4.7503217503217506e-05, "loss": 0.8585, "step": 970 }, { "epoch": 0.2522197915326213, "grad_norm": 0.8411473035812378, "learning_rate": 4.747747747747748e-05, "loss": 0.8956, "step": 980 }, { "epoch": 0.2547934628747909, "grad_norm": 1.342337727546692, "learning_rate": 4.7451737451737456e-05, "loss": 0.9417, "step": 990 }, { "epoch": 0.2573671342169605, "grad_norm": 1.0992860794067383, "learning_rate": 4.742599742599743e-05, "loss": 0.8915, "step": 1000 }, { "epoch": 0.2573671342169605, "eval_loss": 0.8595815300941467, "eval_runtime": 395.6739, "eval_samples_per_second": 49.096, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.0003721938368800249, "step": 1000 }, { "epoch": 0.2599408055591301, "grad_norm": 0.776602566242218, "learning_rate": 4.7400257400257406e-05, "loss": 0.8822, "step": 1010 }, { "epoch": 0.2625144769012997, "grad_norm": 1.099664568901062, "learning_rate": 4.737451737451737e-05, "loss": 0.8394, "step": 1020 }, { "epoch": 0.2650881482434693, "grad_norm": 0.80992192029953, "learning_rate": 4.734877734877735e-05, "loss": 0.8998, "step": 1030 }, { "epoch": 0.2676618195856389, "grad_norm": 0.8013533353805542, "learning_rate": 4.732303732303733e-05, "loss": 0.8744, "step": 1040 }, { "epoch": 0.2702354909278085, "grad_norm": 0.8478692770004272, "learning_rate": 4.72972972972973e-05, "loss": 0.8574, "step": 1050 }, { "epoch": 0.27280916226997814, "grad_norm": 1.3109067678451538, "learning_rate": 4.727155727155728e-05, "loss": 0.914, "step": 1060 }, { "epoch": 0.2753828336121477, "grad_norm": 1.043544888496399, "learning_rate": 4.724581724581725e-05, "loss": 0.8326, "step": 1070 }, { "epoch": 0.27795650495431734, "grad_norm": 0.9514428973197937, "learning_rate": 4.722007722007723e-05, "loss": 0.9688, "step": 1080 }, { "epoch": 0.28053017629648697, "grad_norm": 2.1324360370635986, "learning_rate": 4.719433719433719e-05, "loss": 0.9619, "step": 1090 }, { "epoch": 0.28310384763865654, "grad_norm": 1.0498496294021606, "learning_rate": 4.716859716859717e-05, "loss": 0.8134, "step": 1100 }, { "epoch": 0.28567751898082616, "grad_norm": 1.7257237434387207, "learning_rate": 4.714285714285714e-05, "loss": 0.8972, "step": 1110 }, { "epoch": 0.28825119032299573, "grad_norm": 3.86918568611145, "learning_rate": 4.711711711711712e-05, "loss": 0.9396, "step": 1120 }, { "epoch": 0.29082486166516536, "grad_norm": 0.8664824962615967, "learning_rate": 4.709137709137709e-05, "loss": 0.8619, "step": 1130 }, { "epoch": 0.293398533007335, "grad_norm": 1.012129783630371, "learning_rate": 4.706563706563707e-05, "loss": 0.858, "step": 1140 }, { "epoch": 0.29597220434950455, "grad_norm": 1.0971994400024414, "learning_rate": 4.703989703989704e-05, "loss": 0.8932, "step": 1150 }, { "epoch": 0.2985458756916742, "grad_norm": 1.346057415008545, "learning_rate": 4.7014157014157014e-05, "loss": 0.9042, "step": 1160 }, { "epoch": 0.3011195470338438, "grad_norm": 1.1237205266952515, "learning_rate": 4.6988416988416986e-05, "loss": 0.7794, "step": 1170 }, { "epoch": 0.3036932183760134, "grad_norm": 0.975764274597168, "learning_rate": 4.6962676962676964e-05, "loss": 0.8831, "step": 1180 }, { "epoch": 0.306266889718183, "grad_norm": 1.2752641439437866, "learning_rate": 4.6936936936936936e-05, "loss": 0.8896, "step": 1190 }, { "epoch": 0.30884056106035257, "grad_norm": 4.031650543212891, "learning_rate": 4.6911196911196914e-05, "loss": 0.9143, "step": 1200 }, { "epoch": 0.3114142324025222, "grad_norm": 1.1821297407150269, "learning_rate": 4.6885456885456886e-05, "loss": 0.8443, "step": 1210 }, { "epoch": 0.3139879037446918, "grad_norm": 1.1812553405761719, "learning_rate": 4.6859716859716864e-05, "loss": 0.9018, "step": 1220 }, { "epoch": 0.3165615750868614, "grad_norm": 0.8742527365684509, "learning_rate": 4.6833976833976836e-05, "loss": 0.7639, "step": 1230 }, { "epoch": 0.319135246429031, "grad_norm": 1.2843698263168335, "learning_rate": 4.680823680823681e-05, "loss": 0.8889, "step": 1240 }, { "epoch": 0.32170891777120064, "grad_norm": 0.6974703073501587, "learning_rate": 4.6782496782496786e-05, "loss": 0.9007, "step": 1250 }, { "epoch": 0.3242825891133702, "grad_norm": 3.426180839538574, "learning_rate": 4.675675675675676e-05, "loss": 0.8939, "step": 1260 }, { "epoch": 0.32685626045553984, "grad_norm": 1.3513213396072388, "learning_rate": 4.6731016731016736e-05, "loss": 0.8406, "step": 1270 }, { "epoch": 0.3294299317977094, "grad_norm": 1.0384771823883057, "learning_rate": 4.670527670527671e-05, "loss": 0.8806, "step": 1280 }, { "epoch": 0.33200360313987903, "grad_norm": 1.1321779489517212, "learning_rate": 4.6679536679536686e-05, "loss": 0.7905, "step": 1290 }, { "epoch": 0.33457727448204866, "grad_norm": 0.8733064532279968, "learning_rate": 4.665379665379666e-05, "loss": 0.8215, "step": 1300 }, { "epoch": 0.33715094582421823, "grad_norm": 0.7575940489768982, "learning_rate": 4.662805662805663e-05, "loss": 0.8211, "step": 1310 }, { "epoch": 0.33972461716638785, "grad_norm": 0.6808631420135498, "learning_rate": 4.66023166023166e-05, "loss": 0.8255, "step": 1320 }, { "epoch": 0.3422982885085575, "grad_norm": 1.1921888589859009, "learning_rate": 4.657657657657658e-05, "loss": 0.8856, "step": 1330 }, { "epoch": 0.34487195985072705, "grad_norm": 1.2014636993408203, "learning_rate": 4.655083655083655e-05, "loss": 0.8523, "step": 1340 }, { "epoch": 0.3474456311928967, "grad_norm": 1.0800600051879883, "learning_rate": 4.652509652509653e-05, "loss": 0.8697, "step": 1350 }, { "epoch": 0.35001930253506625, "grad_norm": 1.1528334617614746, "learning_rate": 4.64993564993565e-05, "loss": 0.8628, "step": 1360 }, { "epoch": 0.35259297387723587, "grad_norm": 1.2347609996795654, "learning_rate": 4.647361647361648e-05, "loss": 0.8431, "step": 1370 }, { "epoch": 0.3551666452194055, "grad_norm": 1.086850881576538, "learning_rate": 4.644787644787645e-05, "loss": 0.8429, "step": 1380 }, { "epoch": 0.35774031656157507, "grad_norm": 1.0318641662597656, "learning_rate": 4.642213642213642e-05, "loss": 0.9399, "step": 1390 }, { "epoch": 0.3603139879037447, "grad_norm": 0.9806049466133118, "learning_rate": 4.6396396396396394e-05, "loss": 0.8605, "step": 1400 }, { "epoch": 0.3628876592459143, "grad_norm": 1.040371298789978, "learning_rate": 4.637065637065637e-05, "loss": 0.8756, "step": 1410 }, { "epoch": 0.3654613305880839, "grad_norm": 3.93825626373291, "learning_rate": 4.6344916344916344e-05, "loss": 0.8493, "step": 1420 }, { "epoch": 0.3680350019302535, "grad_norm": 1.530876874923706, "learning_rate": 4.631917631917632e-05, "loss": 0.8755, "step": 1430 }, { "epoch": 0.37060867327242314, "grad_norm": 1.0033499002456665, "learning_rate": 4.6293436293436294e-05, "loss": 0.8141, "step": 1440 }, { "epoch": 0.3731823446145927, "grad_norm": 1.1045700311660767, "learning_rate": 4.626769626769627e-05, "loss": 0.8565, "step": 1450 }, { "epoch": 0.37575601595676233, "grad_norm": 0.9272418022155762, "learning_rate": 4.6241956241956244e-05, "loss": 0.8335, "step": 1460 }, { "epoch": 0.3783296872989319, "grad_norm": 2.702624559402466, "learning_rate": 4.6216216216216215e-05, "loss": 0.8326, "step": 1470 }, { "epoch": 0.38090335864110153, "grad_norm": 1.0378378629684448, "learning_rate": 4.6190476190476194e-05, "loss": 0.8696, "step": 1480 }, { "epoch": 0.38347702998327116, "grad_norm": 0.9696075320243835, "learning_rate": 4.6164736164736165e-05, "loss": 0.8559, "step": 1490 }, { "epoch": 0.3860507013254407, "grad_norm": 0.9262058734893799, "learning_rate": 4.6138996138996144e-05, "loss": 0.9104, "step": 1500 }, { "epoch": 0.38862437266761035, "grad_norm": 3.3203704357147217, "learning_rate": 4.6113256113256115e-05, "loss": 0.803, "step": 1510 }, { "epoch": 0.39119804400978, "grad_norm": 1.3920564651489258, "learning_rate": 4.6087516087516094e-05, "loss": 0.9708, "step": 1520 }, { "epoch": 0.39377171535194955, "grad_norm": 0.9011679291725159, "learning_rate": 4.6061776061776065e-05, "loss": 0.872, "step": 1530 }, { "epoch": 0.3963453866941192, "grad_norm": 1.1053688526153564, "learning_rate": 4.603603603603604e-05, "loss": 0.882, "step": 1540 }, { "epoch": 0.39891905803628874, "grad_norm": 0.8040546774864197, "learning_rate": 4.601029601029601e-05, "loss": 0.8539, "step": 1550 }, { "epoch": 0.40149272937845837, "grad_norm": 3.191498041152954, "learning_rate": 4.598455598455599e-05, "loss": 0.9703, "step": 1560 }, { "epoch": 0.404066400720628, "grad_norm": 1.2510138750076294, "learning_rate": 4.595881595881596e-05, "loss": 0.8436, "step": 1570 }, { "epoch": 0.40664007206279756, "grad_norm": 0.8919886350631714, "learning_rate": 4.593307593307594e-05, "loss": 0.8437, "step": 1580 }, { "epoch": 0.4092137434049672, "grad_norm": 0.8489089608192444, "learning_rate": 4.590733590733591e-05, "loss": 0.9111, "step": 1590 }, { "epoch": 0.4117874147471368, "grad_norm": 1.2358871698379517, "learning_rate": 4.588159588159589e-05, "loss": 0.8806, "step": 1600 }, { "epoch": 0.4143610860893064, "grad_norm": 1.3173274993896484, "learning_rate": 4.585585585585586e-05, "loss": 0.833, "step": 1610 }, { "epoch": 0.416934757431476, "grad_norm": 1.3292251825332642, "learning_rate": 4.583011583011583e-05, "loss": 0.8926, "step": 1620 }, { "epoch": 0.4195084287736456, "grad_norm": 1.155118465423584, "learning_rate": 4.58043758043758e-05, "loss": 0.8289, "step": 1630 }, { "epoch": 0.4220821001158152, "grad_norm": 0.9190046787261963, "learning_rate": 4.577863577863578e-05, "loss": 0.8332, "step": 1640 }, { "epoch": 0.42465577145798483, "grad_norm": 0.9945672750473022, "learning_rate": 4.575289575289575e-05, "loss": 0.8378, "step": 1650 }, { "epoch": 0.4272294428001544, "grad_norm": 1.0910547971725464, "learning_rate": 4.572715572715573e-05, "loss": 0.8843, "step": 1660 }, { "epoch": 0.429803114142324, "grad_norm": 1.24397611618042, "learning_rate": 4.570141570141571e-05, "loss": 0.8926, "step": 1670 }, { "epoch": 0.43237678548449365, "grad_norm": 1.0651038885116577, "learning_rate": 4.567567567567568e-05, "loss": 0.8542, "step": 1680 }, { "epoch": 0.4349504568266632, "grad_norm": 1.0350909233093262, "learning_rate": 4.564993564993565e-05, "loss": 0.8258, "step": 1690 }, { "epoch": 0.43752412816883285, "grad_norm": 1.0753473043441772, "learning_rate": 4.5624195624195624e-05, "loss": 0.8557, "step": 1700 }, { "epoch": 0.4400977995110024, "grad_norm": 1.2361218929290771, "learning_rate": 4.55984555984556e-05, "loss": 0.9459, "step": 1710 }, { "epoch": 0.44267147085317204, "grad_norm": 1.2962043285369873, "learning_rate": 4.5572715572715574e-05, "loss": 0.8822, "step": 1720 }, { "epoch": 0.44524514219534167, "grad_norm": 2.557123899459839, "learning_rate": 4.554697554697555e-05, "loss": 0.8018, "step": 1730 }, { "epoch": 0.44781881353751124, "grad_norm": 0.8217394351959229, "learning_rate": 4.5521235521235524e-05, "loss": 0.8012, "step": 1740 }, { "epoch": 0.45039248487968087, "grad_norm": 0.9458732008934021, "learning_rate": 4.54954954954955e-05, "loss": 0.835, "step": 1750 }, { "epoch": 0.4529661562218505, "grad_norm": 0.9185703992843628, "learning_rate": 4.5469755469755473e-05, "loss": 0.8571, "step": 1760 }, { "epoch": 0.45553982756402006, "grad_norm": 1.190473198890686, "learning_rate": 4.5444015444015445e-05, "loss": 0.81, "step": 1770 }, { "epoch": 0.4581134989061897, "grad_norm": 1.1734472513198853, "learning_rate": 4.541827541827542e-05, "loss": 0.8926, "step": 1780 }, { "epoch": 0.46068717024835926, "grad_norm": 1.0630743503570557, "learning_rate": 4.5392535392535395e-05, "loss": 0.8848, "step": 1790 }, { "epoch": 0.4632608415905289, "grad_norm": 3.2704575061798096, "learning_rate": 4.536679536679537e-05, "loss": 0.923, "step": 1800 }, { "epoch": 0.4658345129326985, "grad_norm": 0.8586927056312561, "learning_rate": 4.5341055341055345e-05, "loss": 0.8056, "step": 1810 }, { "epoch": 0.4684081842748681, "grad_norm": 1.2519656419754028, "learning_rate": 4.531531531531532e-05, "loss": 0.8505, "step": 1820 }, { "epoch": 0.4709818556170377, "grad_norm": 0.8403356075286865, "learning_rate": 4.5289575289575295e-05, "loss": 0.8093, "step": 1830 }, { "epoch": 0.47355552695920733, "grad_norm": 1.1714375019073486, "learning_rate": 4.526383526383527e-05, "loss": 0.9739, "step": 1840 }, { "epoch": 0.4761291983013769, "grad_norm": 0.9783056378364563, "learning_rate": 4.523809523809524e-05, "loss": 0.8602, "step": 1850 }, { "epoch": 0.4787028696435465, "grad_norm": 2.869663715362549, "learning_rate": 4.521235521235521e-05, "loss": 0.8444, "step": 1860 }, { "epoch": 0.48127654098571615, "grad_norm": 0.8355293273925781, "learning_rate": 4.518661518661519e-05, "loss": 0.9036, "step": 1870 }, { "epoch": 0.4838502123278857, "grad_norm": 1.140135407447815, "learning_rate": 4.516087516087517e-05, "loss": 0.8022, "step": 1880 }, { "epoch": 0.48642388367005535, "grad_norm": 0.9020267724990845, "learning_rate": 4.513513513513514e-05, "loss": 0.8493, "step": 1890 }, { "epoch": 0.4889975550122249, "grad_norm": 3.613842248916626, "learning_rate": 4.510939510939512e-05, "loss": 0.8602, "step": 1900 }, { "epoch": 0.49157122635439454, "grad_norm": 3.0518836975097656, "learning_rate": 4.508365508365509e-05, "loss": 0.8379, "step": 1910 }, { "epoch": 0.49414489769656417, "grad_norm": 0.9278186559677124, "learning_rate": 4.505791505791506e-05, "loss": 0.9156, "step": 1920 }, { "epoch": 0.49671856903873374, "grad_norm": 4.282588481903076, "learning_rate": 4.503217503217503e-05, "loss": 0.9301, "step": 1930 }, { "epoch": 0.49929224038090336, "grad_norm": 3.971865653991699, "learning_rate": 4.500643500643501e-05, "loss": 0.8885, "step": 1940 }, { "epoch": 0.5018659117230729, "grad_norm": 1.102675199508667, "learning_rate": 4.498069498069498e-05, "loss": 0.8936, "step": 1950 }, { "epoch": 0.5044395830652426, "grad_norm": 0.9979203939437866, "learning_rate": 4.495495495495496e-05, "loss": 0.8014, "step": 1960 }, { "epoch": 0.5070132544074122, "grad_norm": 4.093165397644043, "learning_rate": 4.492921492921493e-05, "loss": 0.8699, "step": 1970 }, { "epoch": 0.5095869257495818, "grad_norm": 1.107493281364441, "learning_rate": 4.490347490347491e-05, "loss": 0.8319, "step": 1980 }, { "epoch": 0.5121605970917514, "grad_norm": 0.8577767014503479, "learning_rate": 4.487773487773488e-05, "loss": 0.8874, "step": 1990 }, { "epoch": 0.514734268433921, "grad_norm": 1.2690497636795044, "learning_rate": 4.485199485199485e-05, "loss": 0.8586, "step": 2000 }, { "epoch": 0.514734268433921, "eval_loss": 0.8306575417518616, "eval_runtime": 395.638, "eval_samples_per_second": 49.1, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00039326141255247916, "step": 2000 }, { "epoch": 0.5173079397760906, "grad_norm": 1.1434204578399658, "learning_rate": 4.4826254826254825e-05, "loss": 0.9541, "step": 2010 }, { "epoch": 0.5198816111182601, "grad_norm": 3.3350868225097656, "learning_rate": 4.48005148005148e-05, "loss": 0.8295, "step": 2020 }, { "epoch": 0.5224552824604298, "grad_norm": 1.2531917095184326, "learning_rate": 4.4774774774774775e-05, "loss": 0.8997, "step": 2030 }, { "epoch": 0.5250289538025994, "grad_norm": 0.9255910515785217, "learning_rate": 4.474903474903475e-05, "loss": 0.8375, "step": 2040 }, { "epoch": 0.527602625144769, "grad_norm": 1.0736703872680664, "learning_rate": 4.4723294723294725e-05, "loss": 0.8293, "step": 2050 }, { "epoch": 0.5301762964869386, "grad_norm": 1.5314749479293823, "learning_rate": 4.46975546975547e-05, "loss": 0.8036, "step": 2060 }, { "epoch": 0.5327499678291082, "grad_norm": 1.0924575328826904, "learning_rate": 4.467181467181467e-05, "loss": 0.7548, "step": 2070 }, { "epoch": 0.5353236391712778, "grad_norm": 0.9415216445922852, "learning_rate": 4.4646074646074646e-05, "loss": 0.8181, "step": 2080 }, { "epoch": 0.5378973105134475, "grad_norm": 1.0533065795898438, "learning_rate": 4.4620334620334625e-05, "loss": 0.8045, "step": 2090 }, { "epoch": 0.540470981855617, "grad_norm": 1.1450499296188354, "learning_rate": 4.4594594594594596e-05, "loss": 0.8571, "step": 2100 }, { "epoch": 0.5430446531977866, "grad_norm": 3.1443707942962646, "learning_rate": 4.4568854568854575e-05, "loss": 0.8687, "step": 2110 }, { "epoch": 0.5456183245399563, "grad_norm": 0.8034738302230835, "learning_rate": 4.4543114543114546e-05, "loss": 0.8746, "step": 2120 }, { "epoch": 0.5481919958821259, "grad_norm": 0.8792498707771301, "learning_rate": 4.4517374517374525e-05, "loss": 0.8886, "step": 2130 }, { "epoch": 0.5507656672242954, "grad_norm": 3.1286137104034424, "learning_rate": 4.449163449163449e-05, "loss": 0.8574, "step": 2140 }, { "epoch": 0.5533393385664651, "grad_norm": 0.9840642213821411, "learning_rate": 4.446589446589447e-05, "loss": 0.815, "step": 2150 }, { "epoch": 0.5559130099086347, "grad_norm": 1.1499487161636353, "learning_rate": 4.444015444015444e-05, "loss": 0.8691, "step": 2160 }, { "epoch": 0.5584866812508043, "grad_norm": 0.8926144242286682, "learning_rate": 4.441441441441442e-05, "loss": 0.8082, "step": 2170 }, { "epoch": 0.5610603525929739, "grad_norm": 0.7535387873649597, "learning_rate": 4.438867438867439e-05, "loss": 0.9033, "step": 2180 }, { "epoch": 0.5636340239351435, "grad_norm": 1.196241021156311, "learning_rate": 4.436293436293437e-05, "loss": 0.9039, "step": 2190 }, { "epoch": 0.5662076952773131, "grad_norm": 2.5870370864868164, "learning_rate": 4.433719433719434e-05, "loss": 0.8692, "step": 2200 }, { "epoch": 0.5687813666194826, "grad_norm": 1.209121823310852, "learning_rate": 4.431145431145432e-05, "loss": 0.8467, "step": 2210 }, { "epoch": 0.5713550379616523, "grad_norm": 0.988710343837738, "learning_rate": 4.428571428571428e-05, "loss": 0.7887, "step": 2220 }, { "epoch": 0.5739287093038219, "grad_norm": 0.9768078327178955, "learning_rate": 4.425997425997426e-05, "loss": 0.8597, "step": 2230 }, { "epoch": 0.5765023806459915, "grad_norm": 1.3332276344299316, "learning_rate": 4.423423423423423e-05, "loss": 0.9351, "step": 2240 }, { "epoch": 0.5790760519881611, "grad_norm": 2.8126509189605713, "learning_rate": 4.420849420849421e-05, "loss": 0.8706, "step": 2250 }, { "epoch": 0.5816497233303307, "grad_norm": 1.1589819192886353, "learning_rate": 4.418275418275418e-05, "loss": 0.874, "step": 2260 }, { "epoch": 0.5842233946725003, "grad_norm": 1.085998296737671, "learning_rate": 4.415701415701416e-05, "loss": 0.8321, "step": 2270 }, { "epoch": 0.58679706601467, "grad_norm": 0.8814729452133179, "learning_rate": 4.413127413127413e-05, "loss": 0.8407, "step": 2280 }, { "epoch": 0.5893707373568395, "grad_norm": 3.6216423511505127, "learning_rate": 4.4105534105534104e-05, "loss": 0.7986, "step": 2290 }, { "epoch": 0.5919444086990091, "grad_norm": 3.6593005657196045, "learning_rate": 4.407979407979408e-05, "loss": 0.8252, "step": 2300 }, { "epoch": 0.5945180800411788, "grad_norm": 0.9884054660797119, "learning_rate": 4.4054054054054054e-05, "loss": 0.8606, "step": 2310 }, { "epoch": 0.5970917513833484, "grad_norm": 1.0792969465255737, "learning_rate": 4.402831402831403e-05, "loss": 0.7866, "step": 2320 }, { "epoch": 0.5996654227255179, "grad_norm": 1.0255166292190552, "learning_rate": 4.4002574002574004e-05, "loss": 0.7659, "step": 2330 }, { "epoch": 0.6022390940676876, "grad_norm": 0.7235389947891235, "learning_rate": 4.397683397683398e-05, "loss": 0.8475, "step": 2340 }, { "epoch": 0.6048127654098572, "grad_norm": 0.8555417060852051, "learning_rate": 4.3951093951093954e-05, "loss": 0.8234, "step": 2350 }, { "epoch": 0.6073864367520267, "grad_norm": 1.1346522569656372, "learning_rate": 4.3925353925353926e-05, "loss": 0.8572, "step": 2360 }, { "epoch": 0.6099601080941963, "grad_norm": 3.7354023456573486, "learning_rate": 4.38996138996139e-05, "loss": 0.842, "step": 2370 }, { "epoch": 0.612533779436366, "grad_norm": 1.0269885063171387, "learning_rate": 4.3873873873873876e-05, "loss": 0.8483, "step": 2380 }, { "epoch": 0.6151074507785356, "grad_norm": 4.06434965133667, "learning_rate": 4.384813384813385e-05, "loss": 0.9082, "step": 2390 }, { "epoch": 0.6176811221207051, "grad_norm": 1.0140035152435303, "learning_rate": 4.3822393822393826e-05, "loss": 0.8993, "step": 2400 }, { "epoch": 0.6202547934628748, "grad_norm": 0.9860438108444214, "learning_rate": 4.37966537966538e-05, "loss": 0.8143, "step": 2410 }, { "epoch": 0.6228284648050444, "grad_norm": 0.9462414979934692, "learning_rate": 4.3770913770913776e-05, "loss": 0.8752, "step": 2420 }, { "epoch": 0.625402136147214, "grad_norm": 0.889310896396637, "learning_rate": 4.374517374517375e-05, "loss": 0.8576, "step": 2430 }, { "epoch": 0.6279758074893836, "grad_norm": 1.0256364345550537, "learning_rate": 4.371943371943372e-05, "loss": 0.8331, "step": 2440 }, { "epoch": 0.6305494788315532, "grad_norm": 1.242521047592163, "learning_rate": 4.369369369369369e-05, "loss": 0.917, "step": 2450 }, { "epoch": 0.6331231501737228, "grad_norm": 0.8450289368629456, "learning_rate": 4.366795366795367e-05, "loss": 0.8463, "step": 2460 }, { "epoch": 0.6356968215158925, "grad_norm": 2.8590588569641113, "learning_rate": 4.364221364221364e-05, "loss": 0.8509, "step": 2470 }, { "epoch": 0.638270492858062, "grad_norm": 0.9202858805656433, "learning_rate": 4.361647361647362e-05, "loss": 0.8259, "step": 2480 }, { "epoch": 0.6408441642002316, "grad_norm": 1.0481261014938354, "learning_rate": 4.359073359073359e-05, "loss": 0.8764, "step": 2490 }, { "epoch": 0.6434178355424013, "grad_norm": 1.0081082582473755, "learning_rate": 4.356499356499357e-05, "loss": 0.8097, "step": 2500 }, { "epoch": 0.6459915068845709, "grad_norm": 0.9049886465072632, "learning_rate": 4.353925353925354e-05, "loss": 0.8617, "step": 2510 }, { "epoch": 0.6485651782267404, "grad_norm": 1.0273019075393677, "learning_rate": 4.351351351351351e-05, "loss": 0.8457, "step": 2520 }, { "epoch": 0.6511388495689101, "grad_norm": 3.1183676719665527, "learning_rate": 4.348777348777349e-05, "loss": 0.8202, "step": 2530 }, { "epoch": 0.6537125209110797, "grad_norm": 0.9991865158081055, "learning_rate": 4.346203346203346e-05, "loss": 0.8147, "step": 2540 }, { "epoch": 0.6562861922532492, "grad_norm": 1.0085402727127075, "learning_rate": 4.343629343629344e-05, "loss": 0.8073, "step": 2550 }, { "epoch": 0.6588598635954188, "grad_norm": 0.8995606899261475, "learning_rate": 4.341055341055341e-05, "loss": 0.8394, "step": 2560 }, { "epoch": 0.6614335349375885, "grad_norm": 0.9390326142311096, "learning_rate": 4.338481338481339e-05, "loss": 0.8784, "step": 2570 }, { "epoch": 0.6640072062797581, "grad_norm": 0.7828226089477539, "learning_rate": 4.335907335907336e-05, "loss": 0.861, "step": 2580 }, { "epoch": 0.6665808776219276, "grad_norm": 0.752167284488678, "learning_rate": 4.3333333333333334e-05, "loss": 0.8099, "step": 2590 }, { "epoch": 0.6691545489640973, "grad_norm": 0.9056263566017151, "learning_rate": 4.3307593307593306e-05, "loss": 0.9575, "step": 2600 }, { "epoch": 0.6717282203062669, "grad_norm": 1.1399377584457397, "learning_rate": 4.3281853281853284e-05, "loss": 0.8491, "step": 2610 }, { "epoch": 0.6743018916484365, "grad_norm": 1.1799904108047485, "learning_rate": 4.3256113256113256e-05, "loss": 0.8375, "step": 2620 }, { "epoch": 0.6768755629906061, "grad_norm": 3.550177812576294, "learning_rate": 4.3230373230373234e-05, "loss": 0.8797, "step": 2630 }, { "epoch": 0.6794492343327757, "grad_norm": 0.8482165336608887, "learning_rate": 4.3204633204633206e-05, "loss": 0.8424, "step": 2640 }, { "epoch": 0.6820229056749453, "grad_norm": 0.8729701042175293, "learning_rate": 4.3178893178893184e-05, "loss": 0.7896, "step": 2650 }, { "epoch": 0.684596577017115, "grad_norm": 1.0203555822372437, "learning_rate": 4.3153153153153156e-05, "loss": 0.8836, "step": 2660 }, { "epoch": 0.6871702483592845, "grad_norm": 1.1605058908462524, "learning_rate": 4.312741312741313e-05, "loss": 0.7561, "step": 2670 }, { "epoch": 0.6897439197014541, "grad_norm": 1.3112170696258545, "learning_rate": 4.31016731016731e-05, "loss": 0.8504, "step": 2680 }, { "epoch": 0.6923175910436238, "grad_norm": 2.7593069076538086, "learning_rate": 4.307593307593308e-05, "loss": 0.9337, "step": 2690 }, { "epoch": 0.6948912623857934, "grad_norm": 0.7685155272483826, "learning_rate": 4.305019305019305e-05, "loss": 0.9312, "step": 2700 }, { "epoch": 0.6974649337279629, "grad_norm": 0.9489961862564087, "learning_rate": 4.302445302445303e-05, "loss": 0.839, "step": 2710 }, { "epoch": 0.7000386050701325, "grad_norm": 2.002380132675171, "learning_rate": 4.2998712998713006e-05, "loss": 0.7451, "step": 2720 }, { "epoch": 0.7026122764123022, "grad_norm": 1.3577697277069092, "learning_rate": 4.297297297297298e-05, "loss": 0.823, "step": 2730 }, { "epoch": 0.7051859477544717, "grad_norm": 0.991814911365509, "learning_rate": 4.294723294723295e-05, "loss": 0.7832, "step": 2740 }, { "epoch": 0.7077596190966413, "grad_norm": 1.302040696144104, "learning_rate": 4.292149292149292e-05, "loss": 0.8151, "step": 2750 }, { "epoch": 0.710333290438811, "grad_norm": 1.1003100872039795, "learning_rate": 4.28957528957529e-05, "loss": 0.8792, "step": 2760 }, { "epoch": 0.7129069617809806, "grad_norm": 1.0496855974197388, "learning_rate": 4.287001287001287e-05, "loss": 0.8354, "step": 2770 }, { "epoch": 0.7154806331231501, "grad_norm": 1.375705599784851, "learning_rate": 4.284427284427285e-05, "loss": 0.7915, "step": 2780 }, { "epoch": 0.7180543044653198, "grad_norm": 1.030092716217041, "learning_rate": 4.281853281853282e-05, "loss": 0.828, "step": 2790 }, { "epoch": 0.7206279758074894, "grad_norm": 2.520185708999634, "learning_rate": 4.27927927927928e-05, "loss": 0.8099, "step": 2800 }, { "epoch": 0.723201647149659, "grad_norm": 0.7944628596305847, "learning_rate": 4.276705276705277e-05, "loss": 0.8974, "step": 2810 }, { "epoch": 0.7257753184918286, "grad_norm": 1.2783660888671875, "learning_rate": 4.274131274131274e-05, "loss": 0.7787, "step": 2820 }, { "epoch": 0.7283489898339982, "grad_norm": 1.3568511009216309, "learning_rate": 4.2715572715572714e-05, "loss": 0.8208, "step": 2830 }, { "epoch": 0.7309226611761678, "grad_norm": 2.9468472003936768, "learning_rate": 4.268983268983269e-05, "loss": 0.8063, "step": 2840 }, { "epoch": 0.7334963325183375, "grad_norm": 0.871306300163269, "learning_rate": 4.2664092664092664e-05, "loss": 0.7988, "step": 2850 }, { "epoch": 0.736070003860507, "grad_norm": 1.0941967964172363, "learning_rate": 4.263835263835264e-05, "loss": 0.8288, "step": 2860 }, { "epoch": 0.7386436752026766, "grad_norm": 0.992978036403656, "learning_rate": 4.2612612612612614e-05, "loss": 0.8742, "step": 2870 }, { "epoch": 0.7412173465448463, "grad_norm": 0.8720837831497192, "learning_rate": 4.258687258687259e-05, "loss": 0.8032, "step": 2880 }, { "epoch": 0.7437910178870158, "grad_norm": 3.578805446624756, "learning_rate": 4.2561132561132564e-05, "loss": 0.8462, "step": 2890 }, { "epoch": 0.7463646892291854, "grad_norm": 1.0471611022949219, "learning_rate": 4.2535392535392535e-05, "loss": 0.8461, "step": 2900 }, { "epoch": 0.748938360571355, "grad_norm": 0.9668796062469482, "learning_rate": 4.2509652509652514e-05, "loss": 0.8373, "step": 2910 }, { "epoch": 0.7515120319135247, "grad_norm": 0.8879560232162476, "learning_rate": 4.2483912483912485e-05, "loss": 0.8707, "step": 2920 }, { "epoch": 0.7540857032556942, "grad_norm": 0.8577817678451538, "learning_rate": 4.2458172458172464e-05, "loss": 0.8934, "step": 2930 }, { "epoch": 0.7566593745978638, "grad_norm": 2.37251877784729, "learning_rate": 4.2432432432432435e-05, "loss": 0.8708, "step": 2940 }, { "epoch": 0.7592330459400335, "grad_norm": 0.7796174883842468, "learning_rate": 4.2406692406692414e-05, "loss": 0.8832, "step": 2950 }, { "epoch": 0.7618067172822031, "grad_norm": 1.032285213470459, "learning_rate": 4.2380952380952385e-05, "loss": 0.8615, "step": 2960 }, { "epoch": 0.7643803886243726, "grad_norm": 4.347442150115967, "learning_rate": 4.235521235521236e-05, "loss": 0.8413, "step": 2970 }, { "epoch": 0.7669540599665423, "grad_norm": 1.1776065826416016, "learning_rate": 4.232947232947233e-05, "loss": 0.8533, "step": 2980 }, { "epoch": 0.7695277313087119, "grad_norm": 0.956089973449707, "learning_rate": 4.230373230373231e-05, "loss": 0.8051, "step": 2990 }, { "epoch": 0.7721014026508815, "grad_norm": 2.812290906906128, "learning_rate": 4.227799227799228e-05, "loss": 0.816, "step": 3000 }, { "epoch": 0.7721014026508815, "eval_loss": 0.8144821524620056, "eval_runtime": 395.614, "eval_samples_per_second": 49.103, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00037428161464936723, "step": 3000 }, { "epoch": 0.7746750739930511, "grad_norm": 0.971863865852356, "learning_rate": 4.225225225225226e-05, "loss": 0.8316, "step": 3010 }, { "epoch": 0.7772487453352207, "grad_norm": 0.9860455393791199, "learning_rate": 4.222651222651223e-05, "loss": 0.8641, "step": 3020 }, { "epoch": 0.7798224166773903, "grad_norm": 1.0967594385147095, "learning_rate": 4.220077220077221e-05, "loss": 0.8423, "step": 3030 }, { "epoch": 0.78239608801956, "grad_norm": 0.7346376776695251, "learning_rate": 4.217503217503218e-05, "loss": 0.8773, "step": 3040 }, { "epoch": 0.7849697593617295, "grad_norm": 0.7431369423866272, "learning_rate": 4.214929214929215e-05, "loss": 0.8448, "step": 3050 }, { "epoch": 0.7875434307038991, "grad_norm": 1.0294193029403687, "learning_rate": 4.212355212355212e-05, "loss": 0.7999, "step": 3060 }, { "epoch": 0.7901171020460687, "grad_norm": 3.059105157852173, "learning_rate": 4.20978120978121e-05, "loss": 0.8155, "step": 3070 }, { "epoch": 0.7926907733882383, "grad_norm": 1.1290513277053833, "learning_rate": 4.207207207207207e-05, "loss": 0.9172, "step": 3080 }, { "epoch": 0.7952644447304079, "grad_norm": 0.7203499674797058, "learning_rate": 4.204633204633205e-05, "loss": 0.8059, "step": 3090 }, { "epoch": 0.7978381160725775, "grad_norm": 3.0688018798828125, "learning_rate": 4.202059202059202e-05, "loss": 0.9509, "step": 3100 }, { "epoch": 0.8004117874147472, "grad_norm": 4.001659393310547, "learning_rate": 4.1994851994852e-05, "loss": 0.7508, "step": 3110 }, { "epoch": 0.8029854587569167, "grad_norm": 1.1222361326217651, "learning_rate": 4.196911196911197e-05, "loss": 0.9271, "step": 3120 }, { "epoch": 0.8055591300990863, "grad_norm": 0.8683832287788391, "learning_rate": 4.194337194337194e-05, "loss": 0.8042, "step": 3130 }, { "epoch": 0.808132801441256, "grad_norm": 0.89637291431427, "learning_rate": 4.191763191763192e-05, "loss": 0.7162, "step": 3140 }, { "epoch": 0.8107064727834256, "grad_norm": 0.8857401609420776, "learning_rate": 4.189189189189189e-05, "loss": 0.7789, "step": 3150 }, { "epoch": 0.8132801441255951, "grad_norm": 1.2026958465576172, "learning_rate": 4.186615186615187e-05, "loss": 0.8782, "step": 3160 }, { "epoch": 0.8158538154677648, "grad_norm": 0.8439406156539917, "learning_rate": 4.184041184041184e-05, "loss": 0.8504, "step": 3170 }, { "epoch": 0.8184274868099344, "grad_norm": 0.9551829099655151, "learning_rate": 4.181467181467182e-05, "loss": 0.8152, "step": 3180 }, { "epoch": 0.821001158152104, "grad_norm": 1.0437098741531372, "learning_rate": 4.1788931788931787e-05, "loss": 0.8451, "step": 3190 }, { "epoch": 0.8235748294942736, "grad_norm": 0.9579282402992249, "learning_rate": 4.1763191763191765e-05, "loss": 0.8714, "step": 3200 }, { "epoch": 0.8261485008364432, "grad_norm": 3.2837677001953125, "learning_rate": 4.1737451737451737e-05, "loss": 0.9142, "step": 3210 }, { "epoch": 0.8287221721786128, "grad_norm": 0.8237940073013306, "learning_rate": 4.1711711711711715e-05, "loss": 0.8592, "step": 3220 }, { "epoch": 0.8312958435207825, "grad_norm": 0.9125580787658691, "learning_rate": 4.1685971685971687e-05, "loss": 0.8598, "step": 3230 }, { "epoch": 0.833869514862952, "grad_norm": 3.1465914249420166, "learning_rate": 4.1660231660231665e-05, "loss": 0.756, "step": 3240 }, { "epoch": 0.8364431862051216, "grad_norm": 0.9401727318763733, "learning_rate": 4.1634491634491637e-05, "loss": 0.765, "step": 3250 }, { "epoch": 0.8390168575472912, "grad_norm": 1.0945508480072021, "learning_rate": 4.1608751608751615e-05, "loss": 0.8335, "step": 3260 }, { "epoch": 0.8415905288894608, "grad_norm": 1.5504162311553955, "learning_rate": 4.158301158301158e-05, "loss": 0.8269, "step": 3270 }, { "epoch": 0.8441642002316304, "grad_norm": 0.7810274362564087, "learning_rate": 4.155727155727156e-05, "loss": 0.8326, "step": 3280 }, { "epoch": 0.8467378715738, "grad_norm": 1.2022150754928589, "learning_rate": 4.153153153153153e-05, "loss": 0.7965, "step": 3290 }, { "epoch": 0.8493115429159697, "grad_norm": 2.462810754776001, "learning_rate": 4.150579150579151e-05, "loss": 0.8781, "step": 3300 }, { "epoch": 0.8518852142581392, "grad_norm": 1.1842944622039795, "learning_rate": 4.148005148005148e-05, "loss": 0.7906, "step": 3310 }, { "epoch": 0.8544588856003088, "grad_norm": 0.8692466020584106, "learning_rate": 4.145431145431146e-05, "loss": 0.8167, "step": 3320 }, { "epoch": 0.8570325569424785, "grad_norm": 1.1452484130859375, "learning_rate": 4.1428571428571437e-05, "loss": 0.795, "step": 3330 }, { "epoch": 0.859606228284648, "grad_norm": 1.0664327144622803, "learning_rate": 4.14028314028314e-05, "loss": 0.8176, "step": 3340 }, { "epoch": 0.8621798996268176, "grad_norm": 1.0237557888031006, "learning_rate": 4.137709137709138e-05, "loss": 0.8243, "step": 3350 }, { "epoch": 0.8647535709689873, "grad_norm": 1.301072597503662, "learning_rate": 4.135135135135135e-05, "loss": 0.7853, "step": 3360 }, { "epoch": 0.8673272423111569, "grad_norm": 1.061145544052124, "learning_rate": 4.132561132561133e-05, "loss": 0.8669, "step": 3370 }, { "epoch": 0.8699009136533264, "grad_norm": 4.076564788818359, "learning_rate": 4.12998712998713e-05, "loss": 0.8657, "step": 3380 }, { "epoch": 0.8724745849954961, "grad_norm": 0.8901606798171997, "learning_rate": 4.127413127413128e-05, "loss": 0.8867, "step": 3390 }, { "epoch": 0.8750482563376657, "grad_norm": 1.0653889179229736, "learning_rate": 4.124839124839125e-05, "loss": 0.8954, "step": 3400 }, { "epoch": 0.8776219276798353, "grad_norm": 0.9028094410896301, "learning_rate": 4.122265122265122e-05, "loss": 0.8455, "step": 3410 }, { "epoch": 0.8801955990220048, "grad_norm": 0.7900987267494202, "learning_rate": 4.1196911196911195e-05, "loss": 0.8002, "step": 3420 }, { "epoch": 0.8827692703641745, "grad_norm": 3.244790554046631, "learning_rate": 4.117117117117117e-05, "loss": 0.7287, "step": 3430 }, { "epoch": 0.8853429417063441, "grad_norm": 1.0797080993652344, "learning_rate": 4.1145431145431145e-05, "loss": 0.8955, "step": 3440 }, { "epoch": 0.8879166130485137, "grad_norm": 1.0693559646606445, "learning_rate": 4.111969111969112e-05, "loss": 0.8617, "step": 3450 }, { "epoch": 0.8904902843906833, "grad_norm": 0.8851631283760071, "learning_rate": 4.1093951093951095e-05, "loss": 0.8405, "step": 3460 }, { "epoch": 0.8930639557328529, "grad_norm": 2.6843698024749756, "learning_rate": 4.106821106821107e-05, "loss": 0.8352, "step": 3470 }, { "epoch": 0.8956376270750225, "grad_norm": 3.100241184234619, "learning_rate": 4.1042471042471045e-05, "loss": 0.7919, "step": 3480 }, { "epoch": 0.8982112984171922, "grad_norm": 0.8943722248077393, "learning_rate": 4.1016731016731016e-05, "loss": 0.8122, "step": 3490 }, { "epoch": 0.9007849697593617, "grad_norm": 0.827441394329071, "learning_rate": 4.099099099099099e-05, "loss": 0.8883, "step": 3500 }, { "epoch": 0.9033586411015313, "grad_norm": 1.1157348155975342, "learning_rate": 4.0965250965250966e-05, "loss": 0.8779, "step": 3510 }, { "epoch": 0.905932312443701, "grad_norm": 1.1023356914520264, "learning_rate": 4.093951093951094e-05, "loss": 0.8149, "step": 3520 }, { "epoch": 0.9085059837858706, "grad_norm": 0.661834180355072, "learning_rate": 4.0913770913770916e-05, "loss": 0.7732, "step": 3530 }, { "epoch": 0.9110796551280401, "grad_norm": 1.1258097887039185, "learning_rate": 4.0888030888030895e-05, "loss": 0.7748, "step": 3540 }, { "epoch": 0.9136533264702098, "grad_norm": 1.0704106092453003, "learning_rate": 4.0862290862290866e-05, "loss": 0.8873, "step": 3550 }, { "epoch": 0.9162269978123794, "grad_norm": 0.9617460370063782, "learning_rate": 4.083655083655084e-05, "loss": 0.7, "step": 3560 }, { "epoch": 0.9188006691545489, "grad_norm": 0.8261712789535522, "learning_rate": 4.081081081081081e-05, "loss": 0.7724, "step": 3570 }, { "epoch": 0.9213743404967185, "grad_norm": 0.8264958262443542, "learning_rate": 4.078507078507079e-05, "loss": 0.8189, "step": 3580 }, { "epoch": 0.9239480118388882, "grad_norm": 0.9358187913894653, "learning_rate": 4.075933075933076e-05, "loss": 0.823, "step": 3590 }, { "epoch": 0.9265216831810578, "grad_norm": 0.9269980788230896, "learning_rate": 4.073359073359074e-05, "loss": 0.7901, "step": 3600 }, { "epoch": 0.9290953545232273, "grad_norm": 1.0300779342651367, "learning_rate": 4.070785070785071e-05, "loss": 0.8129, "step": 3610 }, { "epoch": 0.931669025865397, "grad_norm": 2.253840446472168, "learning_rate": 4.068211068211069e-05, "loss": 0.785, "step": 3620 }, { "epoch": 0.9342426972075666, "grad_norm": 0.7381809949874878, "learning_rate": 4.065637065637066e-05, "loss": 0.8316, "step": 3630 }, { "epoch": 0.9368163685497362, "grad_norm": 2.6727185249328613, "learning_rate": 4.063063063063063e-05, "loss": 0.88, "step": 3640 }, { "epoch": 0.9393900398919058, "grad_norm": 0.7248988747596741, "learning_rate": 4.06048906048906e-05, "loss": 0.8239, "step": 3650 }, { "epoch": 0.9419637112340754, "grad_norm": 0.8684216141700745, "learning_rate": 4.057915057915058e-05, "loss": 0.8899, "step": 3660 }, { "epoch": 0.944537382576245, "grad_norm": 2.2277228832244873, "learning_rate": 4.055341055341055e-05, "loss": 0.8437, "step": 3670 }, { "epoch": 0.9471110539184147, "grad_norm": 0.946140706539154, "learning_rate": 4.052767052767053e-05, "loss": 0.7992, "step": 3680 }, { "epoch": 0.9496847252605842, "grad_norm": 0.7557298541069031, "learning_rate": 4.05019305019305e-05, "loss": 0.849, "step": 3690 }, { "epoch": 0.9522583966027538, "grad_norm": 1.0458942651748657, "learning_rate": 4.047619047619048e-05, "loss": 0.8492, "step": 3700 }, { "epoch": 0.9548320679449235, "grad_norm": 1.1995244026184082, "learning_rate": 4.045045045045045e-05, "loss": 0.7823, "step": 3710 }, { "epoch": 0.957405739287093, "grad_norm": 3.7634544372558594, "learning_rate": 4.0424710424710424e-05, "loss": 0.7751, "step": 3720 }, { "epoch": 0.9599794106292626, "grad_norm": 1.2692762613296509, "learning_rate": 4.0398970398970396e-05, "loss": 0.8069, "step": 3730 }, { "epoch": 0.9625530819714323, "grad_norm": 1.2810945510864258, "learning_rate": 4.0373230373230374e-05, "loss": 0.9275, "step": 3740 }, { "epoch": 0.9651267533136019, "grad_norm": 1.7654497623443604, "learning_rate": 4.034749034749035e-05, "loss": 0.8789, "step": 3750 }, { "epoch": 0.9677004246557714, "grad_norm": 3.197882652282715, "learning_rate": 4.0321750321750324e-05, "loss": 0.8377, "step": 3760 }, { "epoch": 0.970274095997941, "grad_norm": 0.8221775889396667, "learning_rate": 4.02960102960103e-05, "loss": 0.7834, "step": 3770 }, { "epoch": 0.9728477673401107, "grad_norm": 0.9172367453575134, "learning_rate": 4.0270270270270274e-05, "loss": 0.8551, "step": 3780 }, { "epoch": 0.9754214386822803, "grad_norm": 0.8287565112113953, "learning_rate": 4.0244530244530246e-05, "loss": 0.8427, "step": 3790 }, { "epoch": 0.9779951100244498, "grad_norm": 0.9844699501991272, "learning_rate": 4.021879021879022e-05, "loss": 0.8414, "step": 3800 }, { "epoch": 0.9805687813666195, "grad_norm": 0.8020986318588257, "learning_rate": 4.0193050193050196e-05, "loss": 0.9036, "step": 3810 }, { "epoch": 0.9831424527087891, "grad_norm": 0.9322710633277893, "learning_rate": 4.016731016731017e-05, "loss": 0.773, "step": 3820 }, { "epoch": 0.9857161240509587, "grad_norm": 1.5379607677459717, "learning_rate": 4.0141570141570146e-05, "loss": 0.8547, "step": 3830 }, { "epoch": 0.9882897953931283, "grad_norm": 3.0354602336883545, "learning_rate": 4.011583011583012e-05, "loss": 0.8487, "step": 3840 }, { "epoch": 0.9908634667352979, "grad_norm": 0.9367321133613586, "learning_rate": 4.0090090090090096e-05, "loss": 0.8564, "step": 3850 }, { "epoch": 0.9934371380774675, "grad_norm": 0.9189753532409668, "learning_rate": 4.006435006435007e-05, "loss": 0.8674, "step": 3860 }, { "epoch": 0.9960108094196372, "grad_norm": 1.0535331964492798, "learning_rate": 4.003861003861004e-05, "loss": 0.8038, "step": 3870 }, { "epoch": 0.9985844807618067, "grad_norm": 0.8121357560157776, "learning_rate": 4.001287001287001e-05, "loss": 0.8706, "step": 3880 }, { "epoch": 1.0011581521039763, "grad_norm": 1.0434505939483643, "learning_rate": 3.998712998712999e-05, "loss": 0.7571, "step": 3890 }, { "epoch": 1.0037318234461459, "grad_norm": 0.9326075315475464, "learning_rate": 3.996138996138996e-05, "loss": 0.736, "step": 3900 }, { "epoch": 1.0063054947883154, "grad_norm": 0.7963811755180359, "learning_rate": 3.993564993564994e-05, "loss": 0.6516, "step": 3910 }, { "epoch": 1.0088791661304852, "grad_norm": 3.7952184677124023, "learning_rate": 3.990990990990991e-05, "loss": 0.7163, "step": 3920 }, { "epoch": 1.0114528374726548, "grad_norm": 3.2275214195251465, "learning_rate": 3.988416988416989e-05, "loss": 0.7229, "step": 3930 }, { "epoch": 1.0140265088148244, "grad_norm": 0.9668965339660645, "learning_rate": 3.985842985842986e-05, "loss": 0.6849, "step": 3940 }, { "epoch": 1.016600180156994, "grad_norm": 1.117182731628418, "learning_rate": 3.983268983268983e-05, "loss": 0.7061, "step": 3950 }, { "epoch": 1.0191738514991635, "grad_norm": 2.518076181411743, "learning_rate": 3.980694980694981e-05, "loss": 0.7505, "step": 3960 }, { "epoch": 1.021747522841333, "grad_norm": 1.0619583129882812, "learning_rate": 3.978120978120978e-05, "loss": 0.7439, "step": 3970 }, { "epoch": 1.0243211941835029, "grad_norm": 0.724909245967865, "learning_rate": 3.975546975546976e-05, "loss": 0.6356, "step": 3980 }, { "epoch": 1.0268948655256724, "grad_norm": 0.8852310180664062, "learning_rate": 3.972972972972973e-05, "loss": 0.6923, "step": 3990 }, { "epoch": 1.029468536867842, "grad_norm": 1.0579869747161865, "learning_rate": 3.970398970398971e-05, "loss": 0.6591, "step": 4000 }, { "epoch": 1.029468536867842, "eval_loss": 0.8084650039672852, "eval_runtime": 395.6791, "eval_samples_per_second": 49.095, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00037788777625095846, "step": 4000 }, { "epoch": 1.0320422082100116, "grad_norm": 1.5254931449890137, "learning_rate": 3.967824967824968e-05, "loss": 0.7753, "step": 4010 }, { "epoch": 1.0346158795521812, "grad_norm": 0.9235827922821045, "learning_rate": 3.9652509652509654e-05, "loss": 0.6961, "step": 4020 }, { "epoch": 1.0371895508943507, "grad_norm": 1.1861381530761719, "learning_rate": 3.9626769626769626e-05, "loss": 0.6979, "step": 4030 }, { "epoch": 1.0397632222365203, "grad_norm": 1.0676695108413696, "learning_rate": 3.9601029601029604e-05, "loss": 0.7373, "step": 4040 }, { "epoch": 1.04233689357869, "grad_norm": 0.9689359664916992, "learning_rate": 3.9575289575289576e-05, "loss": 0.7406, "step": 4050 }, { "epoch": 1.0449105649208597, "grad_norm": 1.1008409261703491, "learning_rate": 3.9549549549549554e-05, "loss": 0.6522, "step": 4060 }, { "epoch": 1.0474842362630292, "grad_norm": 1.293125033378601, "learning_rate": 3.9523809523809526e-05, "loss": 0.6762, "step": 4070 }, { "epoch": 1.0500579076051988, "grad_norm": 0.9956673979759216, "learning_rate": 3.9498069498069504e-05, "loss": 0.6625, "step": 4080 }, { "epoch": 1.0526315789473684, "grad_norm": 0.8510293960571289, "learning_rate": 3.9472329472329476e-05, "loss": 0.7396, "step": 4090 }, { "epoch": 1.055205250289538, "grad_norm": 1.1811062097549438, "learning_rate": 3.944658944658945e-05, "loss": 0.6562, "step": 4100 }, { "epoch": 1.0577789216317077, "grad_norm": 1.1089203357696533, "learning_rate": 3.942084942084942e-05, "loss": 0.6791, "step": 4110 }, { "epoch": 1.0603525929738773, "grad_norm": 1.135820984840393, "learning_rate": 3.93951093951094e-05, "loss": 0.7083, "step": 4120 }, { "epoch": 1.0629262643160469, "grad_norm": 0.8153538107872009, "learning_rate": 3.936936936936937e-05, "loss": 0.6713, "step": 4130 }, { "epoch": 1.0654999356582164, "grad_norm": 1.1000064611434937, "learning_rate": 3.934362934362935e-05, "loss": 0.7594, "step": 4140 }, { "epoch": 1.068073607000386, "grad_norm": 0.9205760955810547, "learning_rate": 3.931788931788932e-05, "loss": 0.7066, "step": 4150 }, { "epoch": 1.0706472783425556, "grad_norm": 1.1012518405914307, "learning_rate": 3.92921492921493e-05, "loss": 0.691, "step": 4160 }, { "epoch": 1.0732209496847251, "grad_norm": 0.9490509629249573, "learning_rate": 3.926640926640927e-05, "loss": 0.6361, "step": 4170 }, { "epoch": 1.075794621026895, "grad_norm": 0.9934425950050354, "learning_rate": 3.924066924066924e-05, "loss": 0.6546, "step": 4180 }, { "epoch": 1.0783682923690645, "grad_norm": 0.9373430609703064, "learning_rate": 3.921492921492922e-05, "loss": 0.7585, "step": 4190 }, { "epoch": 1.080941963711234, "grad_norm": 0.7465323209762573, "learning_rate": 3.918918918918919e-05, "loss": 0.7026, "step": 4200 }, { "epoch": 1.0835156350534036, "grad_norm": 0.8605757355690002, "learning_rate": 3.916344916344917e-05, "loss": 0.7215, "step": 4210 }, { "epoch": 1.0860893063955732, "grad_norm": 1.27156400680542, "learning_rate": 3.913770913770914e-05, "loss": 0.7393, "step": 4220 }, { "epoch": 1.0886629777377428, "grad_norm": 3.140281915664673, "learning_rate": 3.911196911196912e-05, "loss": 0.6781, "step": 4230 }, { "epoch": 1.0912366490799126, "grad_norm": 1.438792109489441, "learning_rate": 3.908622908622909e-05, "loss": 0.7344, "step": 4240 }, { "epoch": 1.0938103204220821, "grad_norm": 1.1474467515945435, "learning_rate": 3.906048906048906e-05, "loss": 0.6722, "step": 4250 }, { "epoch": 1.0963839917642517, "grad_norm": 0.7898831963539124, "learning_rate": 3.9034749034749034e-05, "loss": 0.709, "step": 4260 }, { "epoch": 1.0989576631064213, "grad_norm": 4.05173397064209, "learning_rate": 3.900900900900901e-05, "loss": 0.682, "step": 4270 }, { "epoch": 1.1015313344485909, "grad_norm": 1.2369790077209473, "learning_rate": 3.8983268983268984e-05, "loss": 0.6313, "step": 4280 }, { "epoch": 1.1041050057907604, "grad_norm": 1.0742212533950806, "learning_rate": 3.895752895752896e-05, "loss": 0.6878, "step": 4290 }, { "epoch": 1.1066786771329302, "grad_norm": 4.394085884094238, "learning_rate": 3.8931788931788934e-05, "loss": 0.6974, "step": 4300 }, { "epoch": 1.1092523484750998, "grad_norm": 1.285081148147583, "learning_rate": 3.890604890604891e-05, "loss": 0.6822, "step": 4310 }, { "epoch": 1.1118260198172694, "grad_norm": 1.0450791120529175, "learning_rate": 3.888030888030888e-05, "loss": 0.6848, "step": 4320 }, { "epoch": 1.114399691159439, "grad_norm": 0.808032214641571, "learning_rate": 3.8854568854568855e-05, "loss": 0.6127, "step": 4330 }, { "epoch": 1.1169733625016085, "grad_norm": 0.6634959578514099, "learning_rate": 3.882882882882883e-05, "loss": 0.6332, "step": 4340 }, { "epoch": 1.119547033843778, "grad_norm": 0.7773045301437378, "learning_rate": 3.8803088803088805e-05, "loss": 0.6558, "step": 4350 }, { "epoch": 1.1221207051859476, "grad_norm": 0.9749086499214172, "learning_rate": 3.877734877734878e-05, "loss": 0.7365, "step": 4360 }, { "epoch": 1.1246943765281174, "grad_norm": 1.0104619264602661, "learning_rate": 3.8751608751608755e-05, "loss": 0.731, "step": 4370 }, { "epoch": 1.127268047870287, "grad_norm": 1.246193766593933, "learning_rate": 3.8725868725868734e-05, "loss": 0.6772, "step": 4380 }, { "epoch": 1.1298417192124566, "grad_norm": 0.9654109477996826, "learning_rate": 3.87001287001287e-05, "loss": 0.7359, "step": 4390 }, { "epoch": 1.1324153905546261, "grad_norm": 0.8749091029167175, "learning_rate": 3.867438867438868e-05, "loss": 0.6908, "step": 4400 }, { "epoch": 1.1349890618967957, "grad_norm": 2.6671011447906494, "learning_rate": 3.864864864864865e-05, "loss": 0.6935, "step": 4410 }, { "epoch": 1.1375627332389655, "grad_norm": 0.834434986114502, "learning_rate": 3.862290862290863e-05, "loss": 0.6978, "step": 4420 }, { "epoch": 1.140136404581135, "grad_norm": 1.0454275608062744, "learning_rate": 3.85971685971686e-05, "loss": 0.7106, "step": 4430 }, { "epoch": 1.1427100759233046, "grad_norm": 1.1424705982208252, "learning_rate": 3.857142857142858e-05, "loss": 0.6781, "step": 4440 }, { "epoch": 1.1452837472654742, "grad_norm": 0.8291309475898743, "learning_rate": 3.854568854568855e-05, "loss": 0.6827, "step": 4450 }, { "epoch": 1.1478574186076438, "grad_norm": 1.2139686346054077, "learning_rate": 3.851994851994852e-05, "loss": 0.6321, "step": 4460 }, { "epoch": 1.1504310899498134, "grad_norm": 0.935051441192627, "learning_rate": 3.849420849420849e-05, "loss": 0.7246, "step": 4470 }, { "epoch": 1.153004761291983, "grad_norm": 2.4891180992126465, "learning_rate": 3.846846846846847e-05, "loss": 0.6742, "step": 4480 }, { "epoch": 1.1555784326341527, "grad_norm": 1.248016357421875, "learning_rate": 3.844272844272844e-05, "loss": 0.7204, "step": 4490 }, { "epoch": 1.1581521039763223, "grad_norm": 3.4102649688720703, "learning_rate": 3.841698841698842e-05, "loss": 0.6985, "step": 4500 }, { "epoch": 1.1607257753184919, "grad_norm": 1.1724501848220825, "learning_rate": 3.839124839124839e-05, "loss": 0.6828, "step": 4510 }, { "epoch": 1.1632994466606614, "grad_norm": 0.8048924207687378, "learning_rate": 3.836550836550837e-05, "loss": 0.6608, "step": 4520 }, { "epoch": 1.165873118002831, "grad_norm": 0.8188682794570923, "learning_rate": 3.833976833976834e-05, "loss": 0.671, "step": 4530 }, { "epoch": 1.1684467893450006, "grad_norm": 0.879068911075592, "learning_rate": 3.831402831402831e-05, "loss": 0.7447, "step": 4540 }, { "epoch": 1.1710204606871701, "grad_norm": 0.9891297817230225, "learning_rate": 3.8288288288288285e-05, "loss": 0.7087, "step": 4550 }, { "epoch": 1.17359413202934, "grad_norm": 1.1570535898208618, "learning_rate": 3.826254826254826e-05, "loss": 0.6537, "step": 4560 }, { "epoch": 1.1761678033715095, "grad_norm": 1.1008327007293701, "learning_rate": 3.8236808236808235e-05, "loss": 0.7138, "step": 4570 }, { "epoch": 1.178741474713679, "grad_norm": 0.8416620492935181, "learning_rate": 3.821106821106821e-05, "loss": 0.6743, "step": 4580 }, { "epoch": 1.1813151460558486, "grad_norm": 0.8778625726699829, "learning_rate": 3.818532818532819e-05, "loss": 0.6572, "step": 4590 }, { "epoch": 1.1838888173980182, "grad_norm": 2.828808307647705, "learning_rate": 3.815958815958816e-05, "loss": 0.7036, "step": 4600 }, { "epoch": 1.1864624887401878, "grad_norm": 1.3439350128173828, "learning_rate": 3.8133848133848135e-05, "loss": 0.759, "step": 4610 }, { "epoch": 1.1890361600823576, "grad_norm": 0.7776147723197937, "learning_rate": 3.8108108108108106e-05, "loss": 0.7424, "step": 4620 }, { "epoch": 1.1916098314245271, "grad_norm": 0.7925708293914795, "learning_rate": 3.8082368082368085e-05, "loss": 0.7157, "step": 4630 }, { "epoch": 1.1941835027666967, "grad_norm": 0.9618962407112122, "learning_rate": 3.8056628056628056e-05, "loss": 0.6598, "step": 4640 }, { "epoch": 1.1967571741088663, "grad_norm": 1.325016975402832, "learning_rate": 3.8030888030888035e-05, "loss": 0.7199, "step": 4650 }, { "epoch": 1.1993308454510359, "grad_norm": 1.2265050411224365, "learning_rate": 3.8005148005148006e-05, "loss": 0.6643, "step": 4660 }, { "epoch": 1.2019045167932054, "grad_norm": 3.7649102210998535, "learning_rate": 3.7979407979407985e-05, "loss": 0.6875, "step": 4670 }, { "epoch": 1.2044781881353752, "grad_norm": 0.7999979257583618, "learning_rate": 3.7953667953667956e-05, "loss": 0.5883, "step": 4680 }, { "epoch": 1.2070518594775448, "grad_norm": 1.012054681777954, "learning_rate": 3.792792792792793e-05, "loss": 0.707, "step": 4690 }, { "epoch": 1.2096255308197144, "grad_norm": 3.4517581462860107, "learning_rate": 3.79021879021879e-05, "loss": 0.6276, "step": 4700 }, { "epoch": 1.212199202161884, "grad_norm": 1.0213309526443481, "learning_rate": 3.787644787644788e-05, "loss": 0.6809, "step": 4710 }, { "epoch": 1.2147728735040535, "grad_norm": 1.3979082107543945, "learning_rate": 3.785070785070785e-05, "loss": 0.7044, "step": 4720 }, { "epoch": 1.217346544846223, "grad_norm": 0.9082636833190918, "learning_rate": 3.782496782496783e-05, "loss": 0.6827, "step": 4730 }, { "epoch": 1.2199202161883926, "grad_norm": 0.9929207563400269, "learning_rate": 3.77992277992278e-05, "loss": 0.6552, "step": 4740 }, { "epoch": 1.2224938875305624, "grad_norm": 0.873182475566864, "learning_rate": 3.777348777348778e-05, "loss": 0.5661, "step": 4750 }, { "epoch": 1.225067558872732, "grad_norm": 1.0508198738098145, "learning_rate": 3.774774774774775e-05, "loss": 0.7024, "step": 4760 }, { "epoch": 1.2276412302149016, "grad_norm": 1.1543952226638794, "learning_rate": 3.772200772200772e-05, "loss": 0.6994, "step": 4770 }, { "epoch": 1.2302149015570711, "grad_norm": 3.8903653621673584, "learning_rate": 3.76962676962677e-05, "loss": 0.7234, "step": 4780 }, { "epoch": 1.2327885728992407, "grad_norm": 1.0679028034210205, "learning_rate": 3.767052767052767e-05, "loss": 0.6768, "step": 4790 }, { "epoch": 1.2353622442414103, "grad_norm": 2.860356330871582, "learning_rate": 3.764478764478765e-05, "loss": 0.7189, "step": 4800 }, { "epoch": 1.23793591558358, "grad_norm": 1.192459225654602, "learning_rate": 3.761904761904762e-05, "loss": 0.7076, "step": 4810 }, { "epoch": 1.2405095869257496, "grad_norm": 0.8065189719200134, "learning_rate": 3.75933075933076e-05, "loss": 0.6556, "step": 4820 }, { "epoch": 1.2430832582679192, "grad_norm": 1.1386466026306152, "learning_rate": 3.756756756756757e-05, "loss": 0.7085, "step": 4830 }, { "epoch": 1.2456569296100888, "grad_norm": 1.2543619871139526, "learning_rate": 3.754182754182754e-05, "loss": 0.7253, "step": 4840 }, { "epoch": 1.2482306009522584, "grad_norm": 1.2314350605010986, "learning_rate": 3.7516087516087514e-05, "loss": 0.6282, "step": 4850 }, { "epoch": 1.250804272294428, "grad_norm": 1.4268747568130493, "learning_rate": 3.749034749034749e-05, "loss": 0.705, "step": 4860 }, { "epoch": 1.2533779436365977, "grad_norm": 0.9429617524147034, "learning_rate": 3.7464607464607464e-05, "loss": 0.702, "step": 4870 }, { "epoch": 1.2559516149787673, "grad_norm": 1.0535614490509033, "learning_rate": 3.743886743886744e-05, "loss": 0.7376, "step": 4880 }, { "epoch": 1.2585252863209369, "grad_norm": 4.010643005371094, "learning_rate": 3.7413127413127414e-05, "loss": 0.6941, "step": 4890 }, { "epoch": 1.2610989576631064, "grad_norm": 0.8559678196907043, "learning_rate": 3.738738738738739e-05, "loss": 0.711, "step": 4900 }, { "epoch": 1.263672629005276, "grad_norm": 0.8388733863830566, "learning_rate": 3.7361647361647364e-05, "loss": 0.7521, "step": 4910 }, { "epoch": 1.2662463003474456, "grad_norm": 3.9231414794921875, "learning_rate": 3.7335907335907336e-05, "loss": 0.6587, "step": 4920 }, { "epoch": 1.2688199716896151, "grad_norm": 1.1947028636932373, "learning_rate": 3.731016731016731e-05, "loss": 0.7029, "step": 4930 }, { "epoch": 1.271393643031785, "grad_norm": 1.23485267162323, "learning_rate": 3.7284427284427286e-05, "loss": 0.7298, "step": 4940 }, { "epoch": 1.2739673143739545, "grad_norm": 1.133660912513733, "learning_rate": 3.725868725868726e-05, "loss": 0.6935, "step": 4950 }, { "epoch": 1.276540985716124, "grad_norm": 3.402822256088257, "learning_rate": 3.7232947232947236e-05, "loss": 0.649, "step": 4960 }, { "epoch": 1.2791146570582936, "grad_norm": 1.2218265533447266, "learning_rate": 3.720720720720721e-05, "loss": 0.734, "step": 4970 }, { "epoch": 1.2816883284004632, "grad_norm": 2.865407705307007, "learning_rate": 3.7181467181467186e-05, "loss": 0.6461, "step": 4980 }, { "epoch": 1.284261999742633, "grad_norm": 1.4591176509857178, "learning_rate": 3.715572715572716e-05, "loss": 0.7582, "step": 4990 }, { "epoch": 1.2868356710848023, "grad_norm": 3.178297519683838, "learning_rate": 3.712998712998713e-05, "loss": 0.7112, "step": 5000 }, { "epoch": 1.2868356710848023, "eval_loss": 0.7959006428718567, "eval_runtime": 395.735, "eval_samples_per_second": 49.088, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.0003792163621041763, "step": 5000 }, { "epoch": 1.2894093424269721, "grad_norm": 0.7849406003952026, "learning_rate": 3.710424710424711e-05, "loss": 0.6349, "step": 5010 }, { "epoch": 1.2919830137691417, "grad_norm": 1.164996862411499, "learning_rate": 3.707850707850708e-05, "loss": 0.6836, "step": 5020 }, { "epoch": 1.2945566851113113, "grad_norm": 4.410824298858643, "learning_rate": 3.705276705276706e-05, "loss": 0.7274, "step": 5030 }, { "epoch": 1.2971303564534808, "grad_norm": 1.3747788667678833, "learning_rate": 3.702702702702703e-05, "loss": 0.713, "step": 5040 }, { "epoch": 1.2997040277956504, "grad_norm": 3.5286660194396973, "learning_rate": 3.700128700128701e-05, "loss": 0.691, "step": 5050 }, { "epoch": 1.3022776991378202, "grad_norm": 1.0159038305282593, "learning_rate": 3.697554697554698e-05, "loss": 0.7158, "step": 5060 }, { "epoch": 1.3048513704799898, "grad_norm": 1.5588384866714478, "learning_rate": 3.694980694980695e-05, "loss": 0.7042, "step": 5070 }, { "epoch": 1.3074250418221594, "grad_norm": 1.0354678630828857, "learning_rate": 3.692406692406692e-05, "loss": 0.6915, "step": 5080 }, { "epoch": 1.309998713164329, "grad_norm": 3.5660855770111084, "learning_rate": 3.68983268983269e-05, "loss": 0.6111, "step": 5090 }, { "epoch": 1.3125723845064985, "grad_norm": 1.010709285736084, "learning_rate": 3.687258687258687e-05, "loss": 0.6168, "step": 5100 }, { "epoch": 1.315146055848668, "grad_norm": 2.2821505069732666, "learning_rate": 3.684684684684685e-05, "loss": 0.7227, "step": 5110 }, { "epoch": 1.3177197271908376, "grad_norm": 0.8663123250007629, "learning_rate": 3.682110682110682e-05, "loss": 0.645, "step": 5120 }, { "epoch": 1.3202933985330074, "grad_norm": 0.9988008141517639, "learning_rate": 3.67953667953668e-05, "loss": 0.7149, "step": 5130 }, { "epoch": 1.322867069875177, "grad_norm": 1.1684341430664062, "learning_rate": 3.676962676962677e-05, "loss": 0.6919, "step": 5140 }, { "epoch": 1.3254407412173466, "grad_norm": 0.9359066486358643, "learning_rate": 3.6743886743886744e-05, "loss": 0.6732, "step": 5150 }, { "epoch": 1.3280144125595161, "grad_norm": 0.785476565361023, "learning_rate": 3.6718146718146716e-05, "loss": 0.7868, "step": 5160 }, { "epoch": 1.3305880839016857, "grad_norm": 3.040825366973877, "learning_rate": 3.6692406692406694e-05, "loss": 0.704, "step": 5170 }, { "epoch": 1.3331617552438555, "grad_norm": 1.1442739963531494, "learning_rate": 3.6666666666666666e-05, "loss": 0.7099, "step": 5180 }, { "epoch": 1.3357354265860248, "grad_norm": 1.0111069679260254, "learning_rate": 3.6640926640926644e-05, "loss": 0.6733, "step": 5190 }, { "epoch": 1.3383090979281946, "grad_norm": 0.922022819519043, "learning_rate": 3.661518661518662e-05, "loss": 0.7364, "step": 5200 }, { "epoch": 1.3408827692703642, "grad_norm": 1.0119050741195679, "learning_rate": 3.6589446589446594e-05, "loss": 0.7414, "step": 5210 }, { "epoch": 1.3434564406125338, "grad_norm": 1.0058611631393433, "learning_rate": 3.6563706563706566e-05, "loss": 0.7387, "step": 5220 }, { "epoch": 1.3460301119547033, "grad_norm": 2.5699520111083984, "learning_rate": 3.653796653796654e-05, "loss": 0.6831, "step": 5230 }, { "epoch": 1.348603783296873, "grad_norm": 0.723125159740448, "learning_rate": 3.6512226512226516e-05, "loss": 0.6642, "step": 5240 }, { "epoch": 1.3511774546390427, "grad_norm": 0.9540688991546631, "learning_rate": 3.648648648648649e-05, "loss": 0.7312, "step": 5250 }, { "epoch": 1.3537511259812123, "grad_norm": 1.160875916481018, "learning_rate": 3.6460746460746466e-05, "loss": 0.7023, "step": 5260 }, { "epoch": 1.3563247973233818, "grad_norm": 1.0321364402770996, "learning_rate": 3.643500643500644e-05, "loss": 0.6955, "step": 5270 }, { "epoch": 1.3588984686655514, "grad_norm": 0.8098282814025879, "learning_rate": 3.6409266409266416e-05, "loss": 0.6461, "step": 5280 }, { "epoch": 1.361472140007721, "grad_norm": 1.080040454864502, "learning_rate": 3.638352638352639e-05, "loss": 0.7128, "step": 5290 }, { "epoch": 1.3640458113498906, "grad_norm": 0.9302416443824768, "learning_rate": 3.635778635778636e-05, "loss": 0.7007, "step": 5300 }, { "epoch": 1.3666194826920601, "grad_norm": 1.0521005392074585, "learning_rate": 3.633204633204633e-05, "loss": 0.6293, "step": 5310 }, { "epoch": 1.36919315403423, "grad_norm": 1.0561240911483765, "learning_rate": 3.630630630630631e-05, "loss": 0.702, "step": 5320 }, { "epoch": 1.3717668253763995, "grad_norm": 3.3522236347198486, "learning_rate": 3.628056628056628e-05, "loss": 0.6907, "step": 5330 }, { "epoch": 1.374340496718569, "grad_norm": 1.2843527793884277, "learning_rate": 3.625482625482626e-05, "loss": 0.7478, "step": 5340 }, { "epoch": 1.3769141680607386, "grad_norm": 0.7206987142562866, "learning_rate": 3.622908622908623e-05, "loss": 0.7065, "step": 5350 }, { "epoch": 1.3794878394029082, "grad_norm": 2.9926466941833496, "learning_rate": 3.620334620334621e-05, "loss": 0.7363, "step": 5360 }, { "epoch": 1.3820615107450778, "grad_norm": 3.599290609359741, "learning_rate": 3.6177606177606174e-05, "loss": 0.7904, "step": 5370 }, { "epoch": 1.3846351820872473, "grad_norm": 1.1965876817703247, "learning_rate": 3.615186615186615e-05, "loss": 0.6879, "step": 5380 }, { "epoch": 1.3872088534294171, "grad_norm": 0.8417690396308899, "learning_rate": 3.6126126126126124e-05, "loss": 0.7154, "step": 5390 }, { "epoch": 1.3897825247715867, "grad_norm": 2.286024570465088, "learning_rate": 3.61003861003861e-05, "loss": 0.7073, "step": 5400 }, { "epoch": 1.3923561961137563, "grad_norm": 0.7775086164474487, "learning_rate": 3.607464607464608e-05, "loss": 0.6656, "step": 5410 }, { "epoch": 1.3949298674559258, "grad_norm": 0.9922925233840942, "learning_rate": 3.604890604890605e-05, "loss": 0.7027, "step": 5420 }, { "epoch": 1.3975035387980954, "grad_norm": 3.689131021499634, "learning_rate": 3.602316602316603e-05, "loss": 0.6914, "step": 5430 }, { "epoch": 1.4000772101402652, "grad_norm": 1.1228481531143188, "learning_rate": 3.5997425997425995e-05, "loss": 0.6495, "step": 5440 }, { "epoch": 1.4026508814824348, "grad_norm": 1.0339312553405762, "learning_rate": 3.5971685971685974e-05, "loss": 0.6677, "step": 5450 }, { "epoch": 1.4052245528246043, "grad_norm": 1.3286689519882202, "learning_rate": 3.5945945945945945e-05, "loss": 0.7425, "step": 5460 }, { "epoch": 1.407798224166774, "grad_norm": 1.022578477859497, "learning_rate": 3.5920205920205924e-05, "loss": 0.6938, "step": 5470 }, { "epoch": 1.4103718955089435, "grad_norm": 0.8292086124420166, "learning_rate": 3.5894465894465895e-05, "loss": 0.6519, "step": 5480 }, { "epoch": 1.412945566851113, "grad_norm": 0.8953461647033691, "learning_rate": 3.5868725868725874e-05, "loss": 0.6347, "step": 5490 }, { "epoch": 1.4155192381932826, "grad_norm": 1.0440467596054077, "learning_rate": 3.5842985842985845e-05, "loss": 0.6647, "step": 5500 }, { "epoch": 1.4180929095354524, "grad_norm": 1.0310571193695068, "learning_rate": 3.581724581724582e-05, "loss": 0.7084, "step": 5510 }, { "epoch": 1.420666580877622, "grad_norm": 1.1693068742752075, "learning_rate": 3.579150579150579e-05, "loss": 0.6517, "step": 5520 }, { "epoch": 1.4232402522197916, "grad_norm": 0.7255993485450745, "learning_rate": 3.576576576576577e-05, "loss": 0.7003, "step": 5530 }, { "epoch": 1.4258139235619611, "grad_norm": 0.8931599259376526, "learning_rate": 3.574002574002574e-05, "loss": 0.6894, "step": 5540 }, { "epoch": 1.4283875949041307, "grad_norm": 1.551103115081787, "learning_rate": 3.571428571428572e-05, "loss": 0.7459, "step": 5550 }, { "epoch": 1.4309612662463003, "grad_norm": 0.9363420605659485, "learning_rate": 3.568854568854569e-05, "loss": 0.7155, "step": 5560 }, { "epoch": 1.4335349375884698, "grad_norm": 0.7762523889541626, "learning_rate": 3.566280566280567e-05, "loss": 0.6855, "step": 5570 }, { "epoch": 1.4361086089306396, "grad_norm": 3.491201400756836, "learning_rate": 3.563706563706564e-05, "loss": 0.7322, "step": 5580 }, { "epoch": 1.4386822802728092, "grad_norm": 1.4872835874557495, "learning_rate": 3.561132561132561e-05, "loss": 0.7379, "step": 5590 }, { "epoch": 1.4412559516149788, "grad_norm": 3.8220291137695312, "learning_rate": 3.558558558558558e-05, "loss": 0.7969, "step": 5600 }, { "epoch": 1.4438296229571483, "grad_norm": 1.0501987934112549, "learning_rate": 3.555984555984556e-05, "loss": 0.7007, "step": 5610 }, { "epoch": 1.446403294299318, "grad_norm": 0.7753943204879761, "learning_rate": 3.553410553410554e-05, "loss": 0.6857, "step": 5620 }, { "epoch": 1.4489769656414877, "grad_norm": 0.8087158799171448, "learning_rate": 3.550836550836551e-05, "loss": 0.6824, "step": 5630 }, { "epoch": 1.451550636983657, "grad_norm": 1.105362892150879, "learning_rate": 3.548262548262549e-05, "loss": 0.6828, "step": 5640 }, { "epoch": 1.4541243083258268, "grad_norm": 4.095690727233887, "learning_rate": 3.545688545688546e-05, "loss": 0.6885, "step": 5650 }, { "epoch": 1.4566979796679964, "grad_norm": 1.0908024311065674, "learning_rate": 3.543114543114543e-05, "loss": 0.7319, "step": 5660 }, { "epoch": 1.459271651010166, "grad_norm": 1.1036632061004639, "learning_rate": 3.5405405405405403e-05, "loss": 0.713, "step": 5670 }, { "epoch": 1.4618453223523356, "grad_norm": 0.9031351208686829, "learning_rate": 3.537966537966538e-05, "loss": 0.6322, "step": 5680 }, { "epoch": 1.4644189936945051, "grad_norm": 1.1566755771636963, "learning_rate": 3.5353925353925353e-05, "loss": 0.6899, "step": 5690 }, { "epoch": 1.466992665036675, "grad_norm": 0.8092819452285767, "learning_rate": 3.532818532818533e-05, "loss": 0.6701, "step": 5700 }, { "epoch": 1.4695663363788445, "grad_norm": 0.9954015612602234, "learning_rate": 3.5302445302445303e-05, "loss": 0.7205, "step": 5710 }, { "epoch": 1.472140007721014, "grad_norm": 0.8541660308837891, "learning_rate": 3.527670527670528e-05, "loss": 0.6193, "step": 5720 }, { "epoch": 1.4747136790631836, "grad_norm": 3.6179423332214355, "learning_rate": 3.5250965250965253e-05, "loss": 0.7084, "step": 5730 }, { "epoch": 1.4772873504053532, "grad_norm": 1.010477900505066, "learning_rate": 3.5225225225225225e-05, "loss": 0.7267, "step": 5740 }, { "epoch": 1.4798610217475228, "grad_norm": 0.8246733546257019, "learning_rate": 3.51994851994852e-05, "loss": 0.6874, "step": 5750 }, { "epoch": 1.4824346930896923, "grad_norm": 0.7221471071243286, "learning_rate": 3.5173745173745175e-05, "loss": 0.7386, "step": 5760 }, { "epoch": 1.4850083644318621, "grad_norm": 1.1807109117507935, "learning_rate": 3.514800514800515e-05, "loss": 0.7059, "step": 5770 }, { "epoch": 1.4875820357740317, "grad_norm": 1.179556131362915, "learning_rate": 3.5122265122265125e-05, "loss": 0.675, "step": 5780 }, { "epoch": 1.4901557071162013, "grad_norm": 3.564592123031616, "learning_rate": 3.50965250965251e-05, "loss": 0.628, "step": 5790 }, { "epoch": 1.4927293784583708, "grad_norm": 0.8945732712745667, "learning_rate": 3.5070785070785075e-05, "loss": 0.7223, "step": 5800 }, { "epoch": 1.4953030498005404, "grad_norm": 1.0145196914672852, "learning_rate": 3.504504504504505e-05, "loss": 0.6822, "step": 5810 }, { "epoch": 1.4978767211427102, "grad_norm": 1.1235549449920654, "learning_rate": 3.501930501930502e-05, "loss": 0.6738, "step": 5820 }, { "epoch": 1.5004503924848795, "grad_norm": 1.2400606870651245, "learning_rate": 3.4993564993565e-05, "loss": 0.6637, "step": 5830 }, { "epoch": 1.5030240638270493, "grad_norm": 4.3333821296691895, "learning_rate": 3.496782496782497e-05, "loss": 0.6351, "step": 5840 }, { "epoch": 1.505597735169219, "grad_norm": 1.126704216003418, "learning_rate": 3.4942084942084947e-05, "loss": 0.654, "step": 5850 }, { "epoch": 1.5081714065113885, "grad_norm": 1.162214994430542, "learning_rate": 3.491634491634492e-05, "loss": 0.6906, "step": 5860 }, { "epoch": 1.510745077853558, "grad_norm": 1.1057708263397217, "learning_rate": 3.4890604890604897e-05, "loss": 0.6567, "step": 5870 }, { "epoch": 1.5133187491957276, "grad_norm": 2.668668270111084, "learning_rate": 3.486486486486487e-05, "loss": 0.7647, "step": 5880 }, { "epoch": 1.5158924205378974, "grad_norm": 4.099859714508057, "learning_rate": 3.483912483912484e-05, "loss": 0.75, "step": 5890 }, { "epoch": 1.5184660918800668, "grad_norm": 0.9541329145431519, "learning_rate": 3.481338481338481e-05, "loss": 0.7209, "step": 5900 }, { "epoch": 1.5210397632222366, "grad_norm": 0.9891238212585449, "learning_rate": 3.478764478764479e-05, "loss": 0.6793, "step": 5910 }, { "epoch": 1.5236134345644061, "grad_norm": 1.2427529096603394, "learning_rate": 3.476190476190476e-05, "loss": 0.7109, "step": 5920 }, { "epoch": 1.5261871059065757, "grad_norm": 0.8798409104347229, "learning_rate": 3.473616473616474e-05, "loss": 0.7161, "step": 5930 }, { "epoch": 1.5287607772487455, "grad_norm": 0.9166936278343201, "learning_rate": 3.471042471042471e-05, "loss": 0.6958, "step": 5940 }, { "epoch": 1.5313344485909148, "grad_norm": 1.1786912679672241, "learning_rate": 3.468468468468469e-05, "loss": 0.6598, "step": 5950 }, { "epoch": 1.5339081199330846, "grad_norm": 0.8544349670410156, "learning_rate": 3.465894465894466e-05, "loss": 0.6384, "step": 5960 }, { "epoch": 1.5364817912752542, "grad_norm": 1.3337987661361694, "learning_rate": 3.463320463320463e-05, "loss": 0.72, "step": 5970 }, { "epoch": 1.5390554626174238, "grad_norm": 1.6757917404174805, "learning_rate": 3.4607464607464605e-05, "loss": 0.6767, "step": 5980 }, { "epoch": 1.5416291339595933, "grad_norm": 1.042203426361084, "learning_rate": 3.458172458172458e-05, "loss": 0.6199, "step": 5990 }, { "epoch": 1.544202805301763, "grad_norm": 0.9380660057067871, "learning_rate": 3.4555984555984555e-05, "loss": 0.635, "step": 6000 }, { "epoch": 1.544202805301763, "eval_loss": 0.7921908497810364, "eval_runtime": 395.6711, "eval_samples_per_second": 49.096, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.0003712448469848693, "step": 6000 }, { "epoch": 1.5467764766439327, "grad_norm": 1.0668463706970215, "learning_rate": 3.453024453024453e-05, "loss": 0.7029, "step": 6010 }, { "epoch": 1.549350147986102, "grad_norm": 0.9166046380996704, "learning_rate": 3.4504504504504505e-05, "loss": 0.7513, "step": 6020 }, { "epoch": 1.5519238193282718, "grad_norm": 1.067365050315857, "learning_rate": 3.447876447876448e-05, "loss": 0.6937, "step": 6030 }, { "epoch": 1.5544974906704414, "grad_norm": 1.0622658729553223, "learning_rate": 3.4453024453024455e-05, "loss": 0.7011, "step": 6040 }, { "epoch": 1.557071162012611, "grad_norm": 0.8257264494895935, "learning_rate": 3.4427284427284426e-05, "loss": 0.6929, "step": 6050 }, { "epoch": 1.5596448333547805, "grad_norm": 1.0909701585769653, "learning_rate": 3.440411840411841e-05, "loss": 0.6587, "step": 6060 }, { "epoch": 1.5622185046969501, "grad_norm": 1.0562666654586792, "learning_rate": 3.437837837837838e-05, "loss": 0.7101, "step": 6070 }, { "epoch": 1.56479217603912, "grad_norm": 0.9798877239227295, "learning_rate": 3.4352638352638353e-05, "loss": 0.7049, "step": 6080 }, { "epoch": 1.5673658473812893, "grad_norm": 1.297472357749939, "learning_rate": 3.4326898326898325e-05, "loss": 0.7213, "step": 6090 }, { "epoch": 1.569939518723459, "grad_norm": 0.9715744256973267, "learning_rate": 3.4301158301158303e-05, "loss": 0.667, "step": 6100 }, { "epoch": 1.5725131900656286, "grad_norm": 4.2070817947387695, "learning_rate": 3.4275418275418275e-05, "loss": 0.6737, "step": 6110 }, { "epoch": 1.5750868614077982, "grad_norm": 0.8832525014877319, "learning_rate": 3.4249678249678253e-05, "loss": 0.8282, "step": 6120 }, { "epoch": 1.577660532749968, "grad_norm": 1.0799152851104736, "learning_rate": 3.4223938223938225e-05, "loss": 0.6991, "step": 6130 }, { "epoch": 1.5802342040921373, "grad_norm": 0.986268162727356, "learning_rate": 3.4198198198198203e-05, "loss": 0.6998, "step": 6140 }, { "epoch": 1.5828078754343071, "grad_norm": 1.0623130798339844, "learning_rate": 3.417245817245817e-05, "loss": 0.7329, "step": 6150 }, { "epoch": 1.5853815467764767, "grad_norm": 0.9408344626426697, "learning_rate": 3.414671814671815e-05, "loss": 0.6361, "step": 6160 }, { "epoch": 1.5879552181186463, "grad_norm": 1.0099743604660034, "learning_rate": 3.412097812097812e-05, "loss": 0.6459, "step": 6170 }, { "epoch": 1.5905288894608158, "grad_norm": 1.1268500089645386, "learning_rate": 3.40952380952381e-05, "loss": 0.663, "step": 6180 }, { "epoch": 1.5931025608029854, "grad_norm": 1.2413302659988403, "learning_rate": 3.4069498069498075e-05, "loss": 0.6333, "step": 6190 }, { "epoch": 1.5956762321451552, "grad_norm": 1.3237611055374146, "learning_rate": 3.404375804375805e-05, "loss": 0.6517, "step": 6200 }, { "epoch": 1.5982499034873245, "grad_norm": 0.8348039388656616, "learning_rate": 3.4018018018018025e-05, "loss": 0.6988, "step": 6210 }, { "epoch": 1.6008235748294943, "grad_norm": 1.1034482717514038, "learning_rate": 3.3992277992278e-05, "loss": 0.6542, "step": 6220 }, { "epoch": 1.603397246171664, "grad_norm": 0.817833662033081, "learning_rate": 3.396653796653797e-05, "loss": 0.6705, "step": 6230 }, { "epoch": 1.6059709175138335, "grad_norm": 0.8758569955825806, "learning_rate": 3.394079794079794e-05, "loss": 0.6661, "step": 6240 }, { "epoch": 1.608544588856003, "grad_norm": 1.172839641571045, "learning_rate": 3.391505791505792e-05, "loss": 0.6228, "step": 6250 }, { "epoch": 1.6111182601981726, "grad_norm": 0.9346746802330017, "learning_rate": 3.388931788931789e-05, "loss": 0.6691, "step": 6260 }, { "epoch": 1.6136919315403424, "grad_norm": 0.942145586013794, "learning_rate": 3.386357786357787e-05, "loss": 0.6838, "step": 6270 }, { "epoch": 1.6162656028825118, "grad_norm": 4.026924133300781, "learning_rate": 3.383783783783784e-05, "loss": 0.6771, "step": 6280 }, { "epoch": 1.6188392742246815, "grad_norm": 0.9714291095733643, "learning_rate": 3.381209781209782e-05, "loss": 0.6274, "step": 6290 }, { "epoch": 1.6214129455668511, "grad_norm": 3.0900719165802, "learning_rate": 3.378635778635778e-05, "loss": 0.6962, "step": 6300 }, { "epoch": 1.6239866169090207, "grad_norm": 1.203955888748169, "learning_rate": 3.376061776061776e-05, "loss": 0.7344, "step": 6310 }, { "epoch": 1.6265602882511905, "grad_norm": 1.0462515354156494, "learning_rate": 3.373487773487773e-05, "loss": 0.698, "step": 6320 }, { "epoch": 1.6291339595933598, "grad_norm": 0.8859325647354126, "learning_rate": 3.370913770913771e-05, "loss": 0.6949, "step": 6330 }, { "epoch": 1.6317076309355296, "grad_norm": 1.0180076360702515, "learning_rate": 3.368339768339768e-05, "loss": 0.6681, "step": 6340 }, { "epoch": 1.6342813022776992, "grad_norm": 3.5328500270843506, "learning_rate": 3.365765765765766e-05, "loss": 0.6147, "step": 6350 }, { "epoch": 1.6368549736198688, "grad_norm": 0.9282681345939636, "learning_rate": 3.363191763191763e-05, "loss": 0.7505, "step": 6360 }, { "epoch": 1.6394286449620383, "grad_norm": 0.9516205191612244, "learning_rate": 3.3606177606177605e-05, "loss": 0.7205, "step": 6370 }, { "epoch": 1.642002316304208, "grad_norm": 1.07736337184906, "learning_rate": 3.358043758043758e-05, "loss": 0.6584, "step": 6380 }, { "epoch": 1.6445759876463777, "grad_norm": 0.8120790123939514, "learning_rate": 3.3554697554697555e-05, "loss": 0.6473, "step": 6390 }, { "epoch": 1.647149658988547, "grad_norm": 0.781129777431488, "learning_rate": 3.352895752895753e-05, "loss": 0.673, "step": 6400 }, { "epoch": 1.6497233303307168, "grad_norm": 1.0196880102157593, "learning_rate": 3.3503217503217505e-05, "loss": 0.6674, "step": 6410 }, { "epoch": 1.6522970016728864, "grad_norm": 1.0872300863265991, "learning_rate": 3.347747747747748e-05, "loss": 0.674, "step": 6420 }, { "epoch": 1.654870673015056, "grad_norm": 0.8442071676254272, "learning_rate": 3.3451737451737455e-05, "loss": 0.6944, "step": 6430 }, { "epoch": 1.6574443443572255, "grad_norm": 0.9583492279052734, "learning_rate": 3.3425997425997426e-05, "loss": 0.7546, "step": 6440 }, { "epoch": 1.660018015699395, "grad_norm": 1.018364667892456, "learning_rate": 3.34002574002574e-05, "loss": 0.6987, "step": 6450 }, { "epoch": 1.662591687041565, "grad_norm": 0.833738386631012, "learning_rate": 3.3374517374517376e-05, "loss": 0.7055, "step": 6460 }, { "epoch": 1.6651653583837343, "grad_norm": 1.014647126197815, "learning_rate": 3.334877734877735e-05, "loss": 0.7545, "step": 6470 }, { "epoch": 1.667739029725904, "grad_norm": 1.2637064456939697, "learning_rate": 3.3323037323037326e-05, "loss": 0.728, "step": 6480 }, { "epoch": 1.6703127010680736, "grad_norm": 0.9318623542785645, "learning_rate": 3.32972972972973e-05, "loss": 0.663, "step": 6490 }, { "epoch": 1.6728863724102432, "grad_norm": 1.4659100770950317, "learning_rate": 3.3271557271557276e-05, "loss": 0.7003, "step": 6500 }, { "epoch": 1.675460043752413, "grad_norm": 0.7381783127784729, "learning_rate": 3.324581724581725e-05, "loss": 0.7492, "step": 6510 }, { "epoch": 1.6780337150945823, "grad_norm": 0.9546041488647461, "learning_rate": 3.322007722007722e-05, "loss": 0.6817, "step": 6520 }, { "epoch": 1.6806073864367521, "grad_norm": 0.8316722512245178, "learning_rate": 3.319433719433719e-05, "loss": 0.7269, "step": 6530 }, { "epoch": 1.6831810577789217, "grad_norm": 0.8137519955635071, "learning_rate": 3.316859716859717e-05, "loss": 0.7524, "step": 6540 }, { "epoch": 1.6857547291210913, "grad_norm": 0.772885262966156, "learning_rate": 3.314285714285714e-05, "loss": 0.661, "step": 6550 }, { "epoch": 1.6883284004632608, "grad_norm": 0.9286116361618042, "learning_rate": 3.311711711711712e-05, "loss": 0.7038, "step": 6560 }, { "epoch": 1.6909020718054304, "grad_norm": 1.0171113014221191, "learning_rate": 3.309137709137709e-05, "loss": 0.715, "step": 6570 }, { "epoch": 1.6934757431476002, "grad_norm": 1.098702311515808, "learning_rate": 3.306563706563707e-05, "loss": 0.673, "step": 6580 }, { "epoch": 1.6960494144897695, "grad_norm": 0.9981555342674255, "learning_rate": 3.303989703989704e-05, "loss": 0.6945, "step": 6590 }, { "epoch": 1.6986230858319393, "grad_norm": 1.3243064880371094, "learning_rate": 3.301415701415701e-05, "loss": 0.7054, "step": 6600 }, { "epoch": 1.701196757174109, "grad_norm": 2.013225555419922, "learning_rate": 3.298841698841699e-05, "loss": 0.6962, "step": 6610 }, { "epoch": 1.7037704285162785, "grad_norm": 0.7733390927314758, "learning_rate": 3.296267696267696e-05, "loss": 0.6986, "step": 6620 }, { "epoch": 1.706344099858448, "grad_norm": 0.9102844595909119, "learning_rate": 3.293693693693694e-05, "loss": 0.7269, "step": 6630 }, { "epoch": 1.7089177712006176, "grad_norm": 0.9313980340957642, "learning_rate": 3.291119691119691e-05, "loss": 0.6498, "step": 6640 }, { "epoch": 1.7114914425427874, "grad_norm": 0.9392831325531006, "learning_rate": 3.288545688545689e-05, "loss": 0.7894, "step": 6650 }, { "epoch": 1.7140651138849567, "grad_norm": 1.059606909751892, "learning_rate": 3.285971685971686e-05, "loss": 0.6854, "step": 6660 }, { "epoch": 1.7166387852271265, "grad_norm": 0.7358580827713013, "learning_rate": 3.2833976833976834e-05, "loss": 0.6555, "step": 6670 }, { "epoch": 1.719212456569296, "grad_norm": 0.9238846302032471, "learning_rate": 3.2808236808236806e-05, "loss": 0.7385, "step": 6680 }, { "epoch": 1.7217861279114657, "grad_norm": 2.5383057594299316, "learning_rate": 3.2782496782496784e-05, "loss": 0.6411, "step": 6690 }, { "epoch": 1.7243597992536355, "grad_norm": 1.0750601291656494, "learning_rate": 3.2756756756756756e-05, "loss": 0.6677, "step": 6700 }, { "epoch": 1.7269334705958048, "grad_norm": 3.5669894218444824, "learning_rate": 3.2731016731016734e-05, "loss": 0.7254, "step": 6710 }, { "epoch": 1.7295071419379746, "grad_norm": 1.3590166568756104, "learning_rate": 3.2705276705276706e-05, "loss": 0.5707, "step": 6720 }, { "epoch": 1.732080813280144, "grad_norm": 3.761894702911377, "learning_rate": 3.2679536679536684e-05, "loss": 0.654, "step": 6730 }, { "epoch": 1.7346544846223138, "grad_norm": 1.5816675424575806, "learning_rate": 3.2653796653796656e-05, "loss": 0.7373, "step": 6740 }, { "epoch": 1.7372281559644833, "grad_norm": 1.1728434562683105, "learning_rate": 3.262805662805663e-05, "loss": 0.6679, "step": 6750 }, { "epoch": 1.739801827306653, "grad_norm": 1.1865681409835815, "learning_rate": 3.26023166023166e-05, "loss": 0.6444, "step": 6760 }, { "epoch": 1.7423754986488227, "grad_norm": 3.5320911407470703, "learning_rate": 3.257657657657658e-05, "loss": 0.7412, "step": 6770 }, { "epoch": 1.744949169990992, "grad_norm": 0.9932678937911987, "learning_rate": 3.255083655083655e-05, "loss": 0.6619, "step": 6780 }, { "epoch": 1.7475228413331618, "grad_norm": 0.9329503774642944, "learning_rate": 3.252509652509653e-05, "loss": 0.668, "step": 6790 }, { "epoch": 1.7500965126753314, "grad_norm": 0.91359943151474, "learning_rate": 3.2499356499356506e-05, "loss": 0.7856, "step": 6800 }, { "epoch": 1.752670184017501, "grad_norm": 2.197328567504883, "learning_rate": 3.247361647361648e-05, "loss": 0.7182, "step": 6810 }, { "epoch": 1.7552438553596705, "grad_norm": 1.2823790311813354, "learning_rate": 3.244787644787645e-05, "loss": 0.7202, "step": 6820 }, { "epoch": 1.75781752670184, "grad_norm": 1.1787232160568237, "learning_rate": 3.242213642213642e-05, "loss": 0.6817, "step": 6830 }, { "epoch": 1.76039119804401, "grad_norm": 0.9441166520118713, "learning_rate": 3.23963963963964e-05, "loss": 0.629, "step": 6840 }, { "epoch": 1.7629648693861792, "grad_norm": 2.959947109222412, "learning_rate": 3.237065637065637e-05, "loss": 0.6758, "step": 6850 }, { "epoch": 1.765538540728349, "grad_norm": 0.9195284247398376, "learning_rate": 3.234491634491635e-05, "loss": 0.6811, "step": 6860 }, { "epoch": 1.7681122120705186, "grad_norm": 1.753348708152771, "learning_rate": 3.231917631917632e-05, "loss": 0.6048, "step": 6870 }, { "epoch": 1.7706858834126882, "grad_norm": 0.8950250744819641, "learning_rate": 3.22934362934363e-05, "loss": 0.7056, "step": 6880 }, { "epoch": 1.773259554754858, "grad_norm": 0.9784029722213745, "learning_rate": 3.226769626769627e-05, "loss": 0.664, "step": 6890 }, { "epoch": 1.7758332260970273, "grad_norm": 3.3027687072753906, "learning_rate": 3.224195624195624e-05, "loss": 0.6592, "step": 6900 }, { "epoch": 1.778406897439197, "grad_norm": 2.5597102642059326, "learning_rate": 3.2216216216216214e-05, "loss": 0.6791, "step": 6910 }, { "epoch": 1.7809805687813665, "grad_norm": 1.0470184087753296, "learning_rate": 3.219047619047619e-05, "loss": 0.6964, "step": 6920 }, { "epoch": 1.7835542401235362, "grad_norm": 2.2413344383239746, "learning_rate": 3.2164736164736164e-05, "loss": 0.6795, "step": 6930 }, { "epoch": 1.7861279114657058, "grad_norm": 1.297616958618164, "learning_rate": 3.213899613899614e-05, "loss": 0.6991, "step": 6940 }, { "epoch": 1.7887015828078754, "grad_norm": 0.9208124279975891, "learning_rate": 3.2113256113256114e-05, "loss": 0.706, "step": 6950 }, { "epoch": 1.7912752541500452, "grad_norm": 1.298047423362732, "learning_rate": 3.208751608751609e-05, "loss": 0.6466, "step": 6960 }, { "epoch": 1.7938489254922145, "grad_norm": 0.9276260733604431, "learning_rate": 3.2061776061776064e-05, "loss": 0.6572, "step": 6970 }, { "epoch": 1.7964225968343843, "grad_norm": 0.9074581861495972, "learning_rate": 3.2036036036036036e-05, "loss": 0.743, "step": 6980 }, { "epoch": 1.798996268176554, "grad_norm": 0.9442815184593201, "learning_rate": 3.201029601029601e-05, "loss": 0.7033, "step": 6990 }, { "epoch": 1.8015699395187235, "grad_norm": 1.1410044431686401, "learning_rate": 3.1984555984555986e-05, "loss": 0.7442, "step": 7000 }, { "epoch": 1.8015699395187235, "eval_loss": 0.7747899889945984, "eval_runtime": 395.6047, "eval_samples_per_second": 49.105, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00037655919039774065, "step": 7000 }, { "epoch": 1.804143610860893, "grad_norm": 0.8645033836364746, "learning_rate": 3.1958815958815964e-05, "loss": 0.6423, "step": 7010 }, { "epoch": 1.8067172822030626, "grad_norm": 1.2555421590805054, "learning_rate": 3.1933075933075936e-05, "loss": 0.5997, "step": 7020 }, { "epoch": 1.8092909535452324, "grad_norm": 2.7066619396209717, "learning_rate": 3.1907335907335914e-05, "loss": 0.659, "step": 7030 }, { "epoch": 1.8118646248874017, "grad_norm": 0.806398868560791, "learning_rate": 3.1881595881595886e-05, "loss": 0.6708, "step": 7040 }, { "epoch": 1.8144382962295715, "grad_norm": 3.4134740829467773, "learning_rate": 3.185585585585586e-05, "loss": 0.6872, "step": 7050 }, { "epoch": 1.817011967571741, "grad_norm": 1.0480446815490723, "learning_rate": 3.183011583011583e-05, "loss": 0.6778, "step": 7060 }, { "epoch": 1.8195856389139107, "grad_norm": 0.9471696019172668, "learning_rate": 3.180437580437581e-05, "loss": 0.6911, "step": 7070 }, { "epoch": 1.8221593102560802, "grad_norm": 4.224599361419678, "learning_rate": 3.177863577863578e-05, "loss": 0.7447, "step": 7080 }, { "epoch": 1.8247329815982498, "grad_norm": 1.0805798768997192, "learning_rate": 3.175289575289576e-05, "loss": 0.704, "step": 7090 }, { "epoch": 1.8273066529404196, "grad_norm": 1.3605204820632935, "learning_rate": 3.172715572715573e-05, "loss": 0.7009, "step": 7100 }, { "epoch": 1.829880324282589, "grad_norm": 1.358485460281372, "learning_rate": 3.170141570141571e-05, "loss": 0.6511, "step": 7110 }, { "epoch": 1.8324539956247587, "grad_norm": 1.068803071975708, "learning_rate": 3.167567567567568e-05, "loss": 0.662, "step": 7120 }, { "epoch": 1.8350276669669283, "grad_norm": 1.6438531875610352, "learning_rate": 3.164993564993565e-05, "loss": 0.7509, "step": 7130 }, { "epoch": 1.8376013383090979, "grad_norm": 0.9765356183052063, "learning_rate": 3.162419562419562e-05, "loss": 0.6796, "step": 7140 }, { "epoch": 1.8401750096512677, "grad_norm": 0.8123814463615417, "learning_rate": 3.15984555984556e-05, "loss": 0.6197, "step": 7150 }, { "epoch": 1.842748680993437, "grad_norm": 0.9948647618293762, "learning_rate": 3.157271557271557e-05, "loss": 0.7236, "step": 7160 }, { "epoch": 1.8453223523356068, "grad_norm": 0.8912683725357056, "learning_rate": 3.154697554697555e-05, "loss": 0.666, "step": 7170 }, { "epoch": 1.8478960236777764, "grad_norm": 1.1189854145050049, "learning_rate": 3.152123552123552e-05, "loss": 0.7255, "step": 7180 }, { "epoch": 1.850469695019946, "grad_norm": 1.395694375038147, "learning_rate": 3.14954954954955e-05, "loss": 0.6849, "step": 7190 }, { "epoch": 1.8530433663621155, "grad_norm": 0.9775906205177307, "learning_rate": 3.1469755469755465e-05, "loss": 0.7228, "step": 7200 }, { "epoch": 1.855617037704285, "grad_norm": 1.1532434225082397, "learning_rate": 3.1444015444015444e-05, "loss": 0.7402, "step": 7210 }, { "epoch": 1.858190709046455, "grad_norm": 0.9169228076934814, "learning_rate": 3.141827541827542e-05, "loss": 0.6879, "step": 7220 }, { "epoch": 1.8607643803886242, "grad_norm": 1.0046635866165161, "learning_rate": 3.1392535392535394e-05, "loss": 0.6767, "step": 7230 }, { "epoch": 1.863338051730794, "grad_norm": 0.9030658602714539, "learning_rate": 3.136679536679537e-05, "loss": 0.6526, "step": 7240 }, { "epoch": 1.8659117230729636, "grad_norm": 0.8485135436058044, "learning_rate": 3.1341055341055344e-05, "loss": 0.6785, "step": 7250 }, { "epoch": 1.8684853944151332, "grad_norm": 1.1969527006149292, "learning_rate": 3.131531531531532e-05, "loss": 0.7079, "step": 7260 }, { "epoch": 1.8710590657573027, "grad_norm": 1.0521395206451416, "learning_rate": 3.1289575289575294e-05, "loss": 0.6408, "step": 7270 }, { "epoch": 1.8736327370994723, "grad_norm": 1.203535556793213, "learning_rate": 3.1263835263835265e-05, "loss": 0.6577, "step": 7280 }, { "epoch": 1.876206408441642, "grad_norm": 1.180917501449585, "learning_rate": 3.123809523809524e-05, "loss": 0.7297, "step": 7290 }, { "epoch": 1.8787800797838115, "grad_norm": 1.5537686347961426, "learning_rate": 3.1212355212355215e-05, "loss": 0.7638, "step": 7300 }, { "epoch": 1.8813537511259812, "grad_norm": 1.0814871788024902, "learning_rate": 3.118661518661519e-05, "loss": 0.6378, "step": 7310 }, { "epoch": 1.8839274224681508, "grad_norm": 7.449766159057617, "learning_rate": 3.1160875160875165e-05, "loss": 0.7747, "step": 7320 }, { "epoch": 1.8865010938103204, "grad_norm": 1.0792584419250488, "learning_rate": 3.113513513513514e-05, "loss": 0.7314, "step": 7330 }, { "epoch": 1.8890747651524902, "grad_norm": 0.8132847547531128, "learning_rate": 3.1109395109395115e-05, "loss": 0.7088, "step": 7340 }, { "epoch": 1.8916484364946595, "grad_norm": 0.9090826511383057, "learning_rate": 3.108365508365508e-05, "loss": 0.6871, "step": 7350 }, { "epoch": 1.8942221078368293, "grad_norm": 0.9906513094902039, "learning_rate": 3.105791505791506e-05, "loss": 0.6722, "step": 7360 }, { "epoch": 1.8967957791789989, "grad_norm": 1.0088332891464233, "learning_rate": 3.103217503217503e-05, "loss": 0.6669, "step": 7370 }, { "epoch": 1.8993694505211685, "grad_norm": 3.6802520751953125, "learning_rate": 3.100643500643501e-05, "loss": 0.6908, "step": 7380 }, { "epoch": 1.901943121863338, "grad_norm": 1.407848596572876, "learning_rate": 3.098069498069498e-05, "loss": 0.6203, "step": 7390 }, { "epoch": 1.9045167932055076, "grad_norm": 1.0382202863693237, "learning_rate": 3.095495495495496e-05, "loss": 0.6618, "step": 7400 }, { "epoch": 1.9070904645476774, "grad_norm": 1.39858877658844, "learning_rate": 3.092921492921493e-05, "loss": 0.6262, "step": 7410 }, { "epoch": 1.9096641358898467, "grad_norm": 0.9530224204063416, "learning_rate": 3.09034749034749e-05, "loss": 0.7161, "step": 7420 }, { "epoch": 1.9122378072320165, "grad_norm": 1.1237123012542725, "learning_rate": 3.087773487773488e-05, "loss": 0.7333, "step": 7430 }, { "epoch": 1.914811478574186, "grad_norm": 2.556382417678833, "learning_rate": 3.085199485199485e-05, "loss": 0.6714, "step": 7440 }, { "epoch": 1.9173851499163557, "grad_norm": 0.9731000065803528, "learning_rate": 3.082625482625483e-05, "loss": 0.6965, "step": 7450 }, { "epoch": 1.9199588212585252, "grad_norm": 3.8856775760650635, "learning_rate": 3.08005148005148e-05, "loss": 0.6913, "step": 7460 }, { "epoch": 1.9225324926006948, "grad_norm": 0.8493612408638, "learning_rate": 3.077477477477478e-05, "loss": 0.6683, "step": 7470 }, { "epoch": 1.9251061639428646, "grad_norm": 3.1507325172424316, "learning_rate": 3.074903474903475e-05, "loss": 0.6804, "step": 7480 }, { "epoch": 1.927679835285034, "grad_norm": 1.0596544742584229, "learning_rate": 3.072329472329472e-05, "loss": 0.6675, "step": 7490 }, { "epoch": 1.9302535066272037, "grad_norm": 0.8844677209854126, "learning_rate": 3.0697554697554695e-05, "loss": 0.7052, "step": 7500 }, { "epoch": 1.9328271779693733, "grad_norm": 3.017733335494995, "learning_rate": 3.067181467181467e-05, "loss": 0.6967, "step": 7510 }, { "epoch": 1.9354008493115429, "grad_norm": 2.120694875717163, "learning_rate": 3.0646074646074645e-05, "loss": 0.7005, "step": 7520 }, { "epoch": 1.9379745206537127, "grad_norm": 0.8085142970085144, "learning_rate": 3.062033462033462e-05, "loss": 0.7189, "step": 7530 }, { "epoch": 1.940548191995882, "grad_norm": 0.9313369393348694, "learning_rate": 3.0594594594594595e-05, "loss": 0.6743, "step": 7540 }, { "epoch": 1.9431218633380518, "grad_norm": 1.2053954601287842, "learning_rate": 3.056885456885457e-05, "loss": 0.6897, "step": 7550 }, { "epoch": 1.9456955346802214, "grad_norm": 0.8504372239112854, "learning_rate": 3.0543114543114545e-05, "loss": 0.6518, "step": 7560 }, { "epoch": 1.948269206022391, "grad_norm": 3.263662576675415, "learning_rate": 3.051737451737452e-05, "loss": 0.6482, "step": 7570 }, { "epoch": 1.9508428773645605, "grad_norm": 0.7058959007263184, "learning_rate": 3.049163449163449e-05, "loss": 0.6701, "step": 7580 }, { "epoch": 1.95341654870673, "grad_norm": 0.9414685964584351, "learning_rate": 3.0465894465894466e-05, "loss": 0.6468, "step": 7590 }, { "epoch": 1.9559902200488999, "grad_norm": 0.9003808498382568, "learning_rate": 3.0440154440154438e-05, "loss": 0.7726, "step": 7600 }, { "epoch": 1.9585638913910692, "grad_norm": 0.8774452209472656, "learning_rate": 3.0414414414414416e-05, "loss": 0.6717, "step": 7610 }, { "epoch": 1.961137562733239, "grad_norm": 1.403334379196167, "learning_rate": 3.0388674388674388e-05, "loss": 0.7249, "step": 7620 }, { "epoch": 1.9637112340754086, "grad_norm": 0.7909294962882996, "learning_rate": 3.0362934362934363e-05, "loss": 0.6854, "step": 7630 }, { "epoch": 1.9662849054175782, "grad_norm": 0.72877037525177, "learning_rate": 3.033719433719434e-05, "loss": 0.6771, "step": 7640 }, { "epoch": 1.9688585767597477, "grad_norm": 1.0888760089874268, "learning_rate": 3.0311454311454313e-05, "loss": 0.723, "step": 7650 }, { "epoch": 1.9714322481019173, "grad_norm": 1.091776728630066, "learning_rate": 3.0285714285714288e-05, "loss": 0.7481, "step": 7660 }, { "epoch": 1.974005919444087, "grad_norm": 1.1127774715423584, "learning_rate": 3.025997425997426e-05, "loss": 0.6342, "step": 7670 }, { "epoch": 1.9765795907862564, "grad_norm": 1.1457735300064087, "learning_rate": 3.0234234234234238e-05, "loss": 0.6916, "step": 7680 }, { "epoch": 1.9791532621284262, "grad_norm": 1.0062847137451172, "learning_rate": 3.020849420849421e-05, "loss": 0.6477, "step": 7690 }, { "epoch": 1.9817269334705958, "grad_norm": 0.9393907785415649, "learning_rate": 3.0182754182754185e-05, "loss": 0.6596, "step": 7700 }, { "epoch": 1.9843006048127654, "grad_norm": 0.8132648468017578, "learning_rate": 3.0157014157014156e-05, "loss": 0.7459, "step": 7710 }, { "epoch": 1.9868742761549352, "grad_norm": 0.8783884644508362, "learning_rate": 3.0131274131274135e-05, "loss": 0.6531, "step": 7720 }, { "epoch": 1.9894479474971045, "grad_norm": 0.982463538646698, "learning_rate": 3.0105534105534106e-05, "loss": 0.6378, "step": 7730 }, { "epoch": 1.9920216188392743, "grad_norm": 2.940614700317383, "learning_rate": 3.007979407979408e-05, "loss": 0.6651, "step": 7740 }, { "epoch": 1.9945952901814439, "grad_norm": 1.3412425518035889, "learning_rate": 3.0054054054054053e-05, "loss": 0.7025, "step": 7750 }, { "epoch": 1.9971689615236135, "grad_norm": 4.413862228393555, "learning_rate": 3.002831402831403e-05, "loss": 0.7143, "step": 7760 }, { "epoch": 1.999742632865783, "grad_norm": 1.016687035560608, "learning_rate": 3.0002574002574003e-05, "loss": 0.6285, "step": 7770 }, { "epoch": 2.0023163042079526, "grad_norm": 2.836697816848755, "learning_rate": 2.9976833976833978e-05, "loss": 0.584, "step": 7780 }, { "epoch": 2.0048899755501224, "grad_norm": 1.2601174116134644, "learning_rate": 2.995109395109395e-05, "loss": 0.509, "step": 7790 }, { "epoch": 2.0074636468922917, "grad_norm": 0.9443026781082153, "learning_rate": 2.9925353925353928e-05, "loss": 0.5627, "step": 7800 }, { "epoch": 2.0100373182344615, "grad_norm": 1.8750293254852295, "learning_rate": 2.98996138996139e-05, "loss": 0.544, "step": 7810 }, { "epoch": 2.012610989576631, "grad_norm": 1.145896553993225, "learning_rate": 2.9873873873873875e-05, "loss": 0.5443, "step": 7820 }, { "epoch": 2.0151846609188007, "grad_norm": 1.1090469360351562, "learning_rate": 2.9848133848133846e-05, "loss": 0.5265, "step": 7830 }, { "epoch": 2.0177583322609705, "grad_norm": 0.9929465651512146, "learning_rate": 2.9822393822393825e-05, "loss": 0.5974, "step": 7840 }, { "epoch": 2.02033200360314, "grad_norm": 0.9841685891151428, "learning_rate": 2.97966537966538e-05, "loss": 0.5093, "step": 7850 }, { "epoch": 2.0229056749453096, "grad_norm": 0.946467399597168, "learning_rate": 2.977091377091377e-05, "loss": 0.4556, "step": 7860 }, { "epoch": 2.025479346287479, "grad_norm": 3.8299388885498047, "learning_rate": 2.974517374517375e-05, "loss": 0.5295, "step": 7870 }, { "epoch": 2.0280530176296487, "grad_norm": 0.9343544840812683, "learning_rate": 2.971943371943372e-05, "loss": 0.4939, "step": 7880 }, { "epoch": 2.030626688971818, "grad_norm": 1.0280673503875732, "learning_rate": 2.9693693693693696e-05, "loss": 0.55, "step": 7890 }, { "epoch": 2.033200360313988, "grad_norm": 1.1726477146148682, "learning_rate": 2.9667953667953668e-05, "loss": 0.5473, "step": 7900 }, { "epoch": 2.0357740316561577, "grad_norm": 1.583446741104126, "learning_rate": 2.9642213642213646e-05, "loss": 0.5337, "step": 7910 }, { "epoch": 2.038347702998327, "grad_norm": 1.0653187036514282, "learning_rate": 2.9616473616473618e-05, "loss": 0.5612, "step": 7920 }, { "epoch": 2.040921374340497, "grad_norm": 1.5087721347808838, "learning_rate": 2.9590733590733593e-05, "loss": 0.5698, "step": 7930 }, { "epoch": 2.043495045682666, "grad_norm": 2.7015230655670166, "learning_rate": 2.9564993564993564e-05, "loss": 0.5721, "step": 7940 }, { "epoch": 2.046068717024836, "grad_norm": 2.5739264488220215, "learning_rate": 2.9539253539253543e-05, "loss": 0.5572, "step": 7950 }, { "epoch": 2.0486423883670057, "grad_norm": 1.3727151155471802, "learning_rate": 2.9513513513513514e-05, "loss": 0.5949, "step": 7960 }, { "epoch": 2.051216059709175, "grad_norm": 1.1039259433746338, "learning_rate": 2.948777348777349e-05, "loss": 0.5954, "step": 7970 }, { "epoch": 2.053789731051345, "grad_norm": 3.637061834335327, "learning_rate": 2.946203346203346e-05, "loss": 0.6065, "step": 7980 }, { "epoch": 2.0563634023935142, "grad_norm": 1.178566336631775, "learning_rate": 2.943629343629344e-05, "loss": 0.5253, "step": 7990 }, { "epoch": 2.058937073735684, "grad_norm": 0.8954353332519531, "learning_rate": 2.941055341055341e-05, "loss": 0.5421, "step": 8000 }, { "epoch": 2.058937073735684, "eval_loss": 0.7958057522773743, "eval_runtime": 395.6991, "eval_samples_per_second": 49.093, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.0003653611096349046, "step": 8000 }, { "epoch": 2.0615107450778534, "grad_norm": 1.133931040763855, "learning_rate": 2.9384813384813386e-05, "loss": 0.5644, "step": 8010 }, { "epoch": 2.064084416420023, "grad_norm": 1.2688968181610107, "learning_rate": 2.9359073359073358e-05, "loss": 0.5866, "step": 8020 }, { "epoch": 2.066658087762193, "grad_norm": 0.861214816570282, "learning_rate": 2.9333333333333336e-05, "loss": 0.5449, "step": 8030 }, { "epoch": 2.0692317591043623, "grad_norm": 0.9040514230728149, "learning_rate": 2.9307593307593308e-05, "loss": 0.5398, "step": 8040 }, { "epoch": 2.071805430446532, "grad_norm": 1.010221004486084, "learning_rate": 2.9281853281853283e-05, "loss": 0.5111, "step": 8050 }, { "epoch": 2.0743791017887014, "grad_norm": 0.7724061012268066, "learning_rate": 2.925611325611326e-05, "loss": 0.5378, "step": 8060 }, { "epoch": 2.0769527731308712, "grad_norm": 1.2511149644851685, "learning_rate": 2.9230373230373233e-05, "loss": 0.5746, "step": 8070 }, { "epoch": 2.0795264444730406, "grad_norm": 1.4768840074539185, "learning_rate": 2.9204633204633208e-05, "loss": 0.5502, "step": 8080 }, { "epoch": 2.0821001158152104, "grad_norm": 1.175214171409607, "learning_rate": 2.917889317889318e-05, "loss": 0.5351, "step": 8090 }, { "epoch": 2.08467378715738, "grad_norm": 0.9883387088775635, "learning_rate": 2.9153153153153158e-05, "loss": 0.5447, "step": 8100 }, { "epoch": 2.0872474584995495, "grad_norm": 1.3396533727645874, "learning_rate": 2.912998712998713e-05, "loss": 0.641, "step": 8110 }, { "epoch": 2.0898211298417193, "grad_norm": 2.0846426486968994, "learning_rate": 2.910682110682111e-05, "loss": 0.5135, "step": 8120 }, { "epoch": 2.0923948011838887, "grad_norm": 2.5317373275756836, "learning_rate": 2.9081081081081087e-05, "loss": 0.5579, "step": 8130 }, { "epoch": 2.0949684725260584, "grad_norm": 3.6536707878112793, "learning_rate": 2.9055341055341055e-05, "loss": 0.5584, "step": 8140 }, { "epoch": 2.0975421438682282, "grad_norm": 0.9044039249420166, "learning_rate": 2.9029601029601033e-05, "loss": 0.546, "step": 8150 }, { "epoch": 2.1001158152103976, "grad_norm": 1.1351265907287598, "learning_rate": 2.9003861003861005e-05, "loss": 0.6126, "step": 8160 }, { "epoch": 2.1026894865525674, "grad_norm": 1.3066338300704956, "learning_rate": 2.897812097812098e-05, "loss": 0.5059, "step": 8170 }, { "epoch": 2.1052631578947367, "grad_norm": 0.9377551674842834, "learning_rate": 2.8952380952380952e-05, "loss": 0.5245, "step": 8180 }, { "epoch": 2.1078368292369065, "grad_norm": 1.2370498180389404, "learning_rate": 2.892664092664093e-05, "loss": 0.5056, "step": 8190 }, { "epoch": 2.110410500579076, "grad_norm": 2.860896348953247, "learning_rate": 2.8900900900900902e-05, "loss": 0.575, "step": 8200 }, { "epoch": 2.1129841719212457, "grad_norm": 1.3423675298690796, "learning_rate": 2.8875160875160877e-05, "loss": 0.5974, "step": 8210 }, { "epoch": 2.1155578432634154, "grad_norm": 1.3402422666549683, "learning_rate": 2.884942084942085e-05, "loss": 0.582, "step": 8220 }, { "epoch": 2.118131514605585, "grad_norm": 1.225522756576538, "learning_rate": 2.8823680823680827e-05, "loss": 0.5424, "step": 8230 }, { "epoch": 2.1207051859477546, "grad_norm": 2.7710585594177246, "learning_rate": 2.87979407979408e-05, "loss": 0.5898, "step": 8240 }, { "epoch": 2.123278857289924, "grad_norm": 1.2279800176620483, "learning_rate": 2.8772200772200773e-05, "loss": 0.5657, "step": 8250 }, { "epoch": 2.1258525286320937, "grad_norm": 3.662959575653076, "learning_rate": 2.8746460746460745e-05, "loss": 0.4772, "step": 8260 }, { "epoch": 2.128426199974263, "grad_norm": 1.2062429189682007, "learning_rate": 2.8720720720720723e-05, "loss": 0.5685, "step": 8270 }, { "epoch": 2.130999871316433, "grad_norm": 3.6382572650909424, "learning_rate": 2.8694980694980695e-05, "loss": 0.5208, "step": 8280 }, { "epoch": 2.1335735426586027, "grad_norm": 1.2735857963562012, "learning_rate": 2.866924066924067e-05, "loss": 0.6405, "step": 8290 }, { "epoch": 2.136147214000772, "grad_norm": 1.1604303121566772, "learning_rate": 2.864350064350064e-05, "loss": 0.5217, "step": 8300 }, { "epoch": 2.138720885342942, "grad_norm": 5.992644786834717, "learning_rate": 2.861776061776062e-05, "loss": 0.5425, "step": 8310 }, { "epoch": 2.141294556685111, "grad_norm": 1.644361972808838, "learning_rate": 2.859202059202059e-05, "loss": 0.601, "step": 8320 }, { "epoch": 2.143868228027281, "grad_norm": 1.2567894458770752, "learning_rate": 2.8566280566280567e-05, "loss": 0.614, "step": 8330 }, { "epoch": 2.1464418993694503, "grad_norm": 1.1754887104034424, "learning_rate": 2.8540540540540545e-05, "loss": 0.5085, "step": 8340 }, { "epoch": 2.14901557071162, "grad_norm": 0.9711121320724487, "learning_rate": 2.8514800514800517e-05, "loss": 0.6065, "step": 8350 }, { "epoch": 2.15158924205379, "grad_norm": 0.790381669998169, "learning_rate": 2.848906048906049e-05, "loss": 0.5743, "step": 8360 }, { "epoch": 2.1541629133959592, "grad_norm": 1.308838129043579, "learning_rate": 2.8463320463320463e-05, "loss": 0.5701, "step": 8370 }, { "epoch": 2.156736584738129, "grad_norm": 1.1394035816192627, "learning_rate": 2.843758043758044e-05, "loss": 0.5446, "step": 8380 }, { "epoch": 2.1593102560802984, "grad_norm": 0.9981347918510437, "learning_rate": 2.8411840411840413e-05, "loss": 0.5563, "step": 8390 }, { "epoch": 2.161883927422468, "grad_norm": 1.1559290885925293, "learning_rate": 2.8386100386100388e-05, "loss": 0.554, "step": 8400 }, { "epoch": 2.164457598764638, "grad_norm": 1.4176164865493774, "learning_rate": 2.836036036036036e-05, "loss": 0.5751, "step": 8410 }, { "epoch": 2.1670312701068073, "grad_norm": 1.236528992652893, "learning_rate": 2.8334620334620338e-05, "loss": 0.6102, "step": 8420 }, { "epoch": 2.169604941448977, "grad_norm": 1.4529129266738892, "learning_rate": 2.830888030888031e-05, "loss": 0.5281, "step": 8430 }, { "epoch": 2.1721786127911464, "grad_norm": 2.4904539585113525, "learning_rate": 2.8283140283140285e-05, "loss": 0.5257, "step": 8440 }, { "epoch": 2.1747522841333162, "grad_norm": 2.638392448425293, "learning_rate": 2.8257400257400256e-05, "loss": 0.5261, "step": 8450 }, { "epoch": 2.1773259554754856, "grad_norm": 0.9878402352333069, "learning_rate": 2.8231660231660235e-05, "loss": 0.6199, "step": 8460 }, { "epoch": 2.1798996268176554, "grad_norm": 1.2372921705245972, "learning_rate": 2.8205920205920206e-05, "loss": 0.5189, "step": 8470 }, { "epoch": 2.182473298159825, "grad_norm": 1.3234797716140747, "learning_rate": 2.818018018018018e-05, "loss": 0.5475, "step": 8480 }, { "epoch": 2.1850469695019945, "grad_norm": 1.4524154663085938, "learning_rate": 2.8154440154440153e-05, "loss": 0.5446, "step": 8490 }, { "epoch": 2.1876206408441643, "grad_norm": 2.357632637023926, "learning_rate": 2.812870012870013e-05, "loss": 0.5848, "step": 8500 }, { "epoch": 2.1901943121863336, "grad_norm": 0.8181639909744263, "learning_rate": 2.8102960102960103e-05, "loss": 0.4966, "step": 8510 }, { "epoch": 2.1927679835285034, "grad_norm": 1.1121617555618286, "learning_rate": 2.8077220077220078e-05, "loss": 0.5481, "step": 8520 }, { "epoch": 2.1953416548706732, "grad_norm": 1.105263113975525, "learning_rate": 2.805148005148005e-05, "loss": 0.5626, "step": 8530 }, { "epoch": 2.1979153262128426, "grad_norm": 3.460721731185913, "learning_rate": 2.8025740025740028e-05, "loss": 0.55, "step": 8540 }, { "epoch": 2.2004889975550124, "grad_norm": 1.0655298233032227, "learning_rate": 2.8000000000000003e-05, "loss": 0.555, "step": 8550 }, { "epoch": 2.2030626688971817, "grad_norm": 1.5254201889038086, "learning_rate": 2.7974259974259975e-05, "loss": 0.5404, "step": 8560 }, { "epoch": 2.2056363402393515, "grad_norm": 0.9580132961273193, "learning_rate": 2.7948519948519953e-05, "loss": 0.574, "step": 8570 }, { "epoch": 2.208210011581521, "grad_norm": 0.87156081199646, "learning_rate": 2.7922779922779925e-05, "loss": 0.5618, "step": 8580 }, { "epoch": 2.2107836829236907, "grad_norm": 1.130436897277832, "learning_rate": 2.78970398970399e-05, "loss": 0.555, "step": 8590 }, { "epoch": 2.2133573542658604, "grad_norm": 1.059656023979187, "learning_rate": 2.787129987129987e-05, "loss": 0.6065, "step": 8600 }, { "epoch": 2.21593102560803, "grad_norm": 2.916320562362671, "learning_rate": 2.784555984555985e-05, "loss": 0.4926, "step": 8610 }, { "epoch": 2.2185046969501996, "grad_norm": 1.392271876335144, "learning_rate": 2.781981981981982e-05, "loss": 0.5869, "step": 8620 }, { "epoch": 2.221078368292369, "grad_norm": 0.8909618258476257, "learning_rate": 2.7794079794079796e-05, "loss": 0.522, "step": 8630 }, { "epoch": 2.2236520396345387, "grad_norm": 1.329585075378418, "learning_rate": 2.7768339768339768e-05, "loss": 0.6073, "step": 8640 }, { "epoch": 2.226225710976708, "grad_norm": 0.8332410454750061, "learning_rate": 2.7742599742599746e-05, "loss": 0.5942, "step": 8650 }, { "epoch": 2.228799382318878, "grad_norm": 1.180031180381775, "learning_rate": 2.7716859716859718e-05, "loss": 0.5056, "step": 8660 }, { "epoch": 2.2313730536610477, "grad_norm": 0.7509261965751648, "learning_rate": 2.7691119691119693e-05, "loss": 0.5434, "step": 8670 }, { "epoch": 2.233946725003217, "grad_norm": 1.0426994562149048, "learning_rate": 2.7665379665379664e-05, "loss": 0.5641, "step": 8680 }, { "epoch": 2.236520396345387, "grad_norm": 0.997373640537262, "learning_rate": 2.7639639639639643e-05, "loss": 0.5676, "step": 8690 }, { "epoch": 2.239094067687556, "grad_norm": 1.2469213008880615, "learning_rate": 2.7613899613899614e-05, "loss": 0.5751, "step": 8700 }, { "epoch": 2.241667739029726, "grad_norm": 1.0776273012161255, "learning_rate": 2.758815958815959e-05, "loss": 0.529, "step": 8710 }, { "epoch": 2.2442414103718953, "grad_norm": 1.02994966506958, "learning_rate": 2.756241956241956e-05, "loss": 0.5762, "step": 8720 }, { "epoch": 2.246815081714065, "grad_norm": 3.1107001304626465, "learning_rate": 2.753667953667954e-05, "loss": 0.5354, "step": 8730 }, { "epoch": 2.249388753056235, "grad_norm": 5.489597797393799, "learning_rate": 2.751093951093951e-05, "loss": 0.554, "step": 8740 }, { "epoch": 2.251962424398404, "grad_norm": 1.1937365531921387, "learning_rate": 2.7485199485199486e-05, "loss": 0.5603, "step": 8750 }, { "epoch": 2.254536095740574, "grad_norm": 0.7883516550064087, "learning_rate": 2.7459459459459464e-05, "loss": 0.5309, "step": 8760 }, { "epoch": 2.2571097670827434, "grad_norm": 0.9160059094429016, "learning_rate": 2.7433719433719436e-05, "loss": 0.5477, "step": 8770 }, { "epoch": 2.259683438424913, "grad_norm": 1.0978742837905884, "learning_rate": 2.740797940797941e-05, "loss": 0.5444, "step": 8780 }, { "epoch": 2.2622571097670825, "grad_norm": 5.506754398345947, "learning_rate": 2.7382239382239383e-05, "loss": 0.5292, "step": 8790 }, { "epoch": 2.2648307811092523, "grad_norm": 1.0431208610534668, "learning_rate": 2.735649935649936e-05, "loss": 0.522, "step": 8800 }, { "epoch": 2.267404452451422, "grad_norm": 1.238174319267273, "learning_rate": 2.7330759330759333e-05, "loss": 0.5596, "step": 8810 }, { "epoch": 2.2699781237935914, "grad_norm": 3.145578384399414, "learning_rate": 2.7305019305019308e-05, "loss": 0.5143, "step": 8820 }, { "epoch": 2.272551795135761, "grad_norm": 0.794948399066925, "learning_rate": 2.727927927927928e-05, "loss": 0.5643, "step": 8830 }, { "epoch": 2.275125466477931, "grad_norm": 0.9516599774360657, "learning_rate": 2.7253539253539258e-05, "loss": 0.55, "step": 8840 }, { "epoch": 2.2776991378201004, "grad_norm": 1.01409912109375, "learning_rate": 2.722779922779923e-05, "loss": 0.4978, "step": 8850 }, { "epoch": 2.28027280916227, "grad_norm": 2.1791932582855225, "learning_rate": 2.7202059202059204e-05, "loss": 0.568, "step": 8860 }, { "epoch": 2.2828464805044395, "grad_norm": 0.9807432889938354, "learning_rate": 2.7176319176319176e-05, "loss": 0.6153, "step": 8870 }, { "epoch": 2.2854201518466093, "grad_norm": 0.947486400604248, "learning_rate": 2.7150579150579154e-05, "loss": 0.5095, "step": 8880 }, { "epoch": 2.2879938231887786, "grad_norm": 1.0243611335754395, "learning_rate": 2.7124839124839126e-05, "loss": 0.5296, "step": 8890 }, { "epoch": 2.2905674945309484, "grad_norm": 0.9641122817993164, "learning_rate": 2.70990990990991e-05, "loss": 0.5597, "step": 8900 }, { "epoch": 2.2931411658731182, "grad_norm": 0.9149156212806702, "learning_rate": 2.7073359073359072e-05, "loss": 0.5361, "step": 8910 }, { "epoch": 2.2957148372152876, "grad_norm": 1.1571760177612305, "learning_rate": 2.704761904761905e-05, "loss": 0.5116, "step": 8920 }, { "epoch": 2.2982885085574574, "grad_norm": 1.1743327379226685, "learning_rate": 2.7021879021879022e-05, "loss": 0.5913, "step": 8930 }, { "epoch": 2.3008621798996267, "grad_norm": 0.9054014682769775, "learning_rate": 2.6996138996138997e-05, "loss": 0.5819, "step": 8940 }, { "epoch": 2.3034358512417965, "grad_norm": 1.0355494022369385, "learning_rate": 2.697039897039897e-05, "loss": 0.5633, "step": 8950 }, { "epoch": 2.306009522583966, "grad_norm": 2.785370111465454, "learning_rate": 2.6944658944658947e-05, "loss": 0.5018, "step": 8960 }, { "epoch": 2.3085831939261356, "grad_norm": 1.2618205547332764, "learning_rate": 2.6918918918918922e-05, "loss": 0.5516, "step": 8970 }, { "epoch": 2.3111568652683054, "grad_norm": 1.069515347480774, "learning_rate": 2.6893178893178894e-05, "loss": 0.5578, "step": 8980 }, { "epoch": 2.313730536610475, "grad_norm": 2.962970018386841, "learning_rate": 2.6867438867438872e-05, "loss": 0.565, "step": 8990 }, { "epoch": 2.3163042079526446, "grad_norm": 1.353009581565857, "learning_rate": 2.6841698841698844e-05, "loss": 0.507, "step": 9000 }, { "epoch": 2.3163042079526446, "eval_loss": 0.7905660271644592, "eval_runtime": 395.6161, "eval_samples_per_second": 49.103, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00036479171569781123, "step": 9000 }, { "epoch": 2.318877879294814, "grad_norm": 1.2040934562683105, "learning_rate": 2.681595881595882e-05, "loss": 0.5755, "step": 9010 }, { "epoch": 2.3214515506369837, "grad_norm": 0.8578842282295227, "learning_rate": 2.679021879021879e-05, "loss": 0.5335, "step": 9020 }, { "epoch": 2.324025221979153, "grad_norm": 0.989274799823761, "learning_rate": 2.676447876447877e-05, "loss": 0.5647, "step": 9030 }, { "epoch": 2.326598893321323, "grad_norm": 1.0703078508377075, "learning_rate": 2.6738738738738737e-05, "loss": 0.5929, "step": 9040 }, { "epoch": 2.3291725646634927, "grad_norm": 0.8797844052314758, "learning_rate": 2.6712998712998716e-05, "loss": 0.548, "step": 9050 }, { "epoch": 2.331746236005662, "grad_norm": 1.1639105081558228, "learning_rate": 2.6687258687258687e-05, "loss": 0.5716, "step": 9060 }, { "epoch": 2.334319907347832, "grad_norm": 1.3019888401031494, "learning_rate": 2.6661518661518666e-05, "loss": 0.5555, "step": 9070 }, { "epoch": 2.336893578690001, "grad_norm": 4.5147271156311035, "learning_rate": 2.6635778635778634e-05, "loss": 0.6358, "step": 9080 }, { "epoch": 2.339467250032171, "grad_norm": 1.306016206741333, "learning_rate": 2.6610038610038612e-05, "loss": 0.5318, "step": 9090 }, { "epoch": 2.3420409213743403, "grad_norm": 1.211195468902588, "learning_rate": 2.6584298584298584e-05, "loss": 0.5253, "step": 9100 }, { "epoch": 2.34461459271651, "grad_norm": 1.1488116979599, "learning_rate": 2.655855855855856e-05, "loss": 0.6381, "step": 9110 }, { "epoch": 2.34718826405868, "grad_norm": 1.1285436153411865, "learning_rate": 2.653281853281853e-05, "loss": 0.5281, "step": 9120 }, { "epoch": 2.349761935400849, "grad_norm": 1.651699423789978, "learning_rate": 2.650707850707851e-05, "loss": 0.5664, "step": 9130 }, { "epoch": 2.352335606743019, "grad_norm": 1.1540205478668213, "learning_rate": 2.648133848133848e-05, "loss": 0.5808, "step": 9140 }, { "epoch": 2.3549092780851884, "grad_norm": 1.0008643865585327, "learning_rate": 2.6455598455598455e-05, "loss": 0.5613, "step": 9150 }, { "epoch": 2.357482949427358, "grad_norm": 0.9590166211128235, "learning_rate": 2.6429858429858427e-05, "loss": 0.6062, "step": 9160 }, { "epoch": 2.3600566207695275, "grad_norm": 0.9757869243621826, "learning_rate": 2.6404118404118405e-05, "loss": 0.5632, "step": 9170 }, { "epoch": 2.3626302921116973, "grad_norm": 0.9153413772583008, "learning_rate": 2.6378378378378384e-05, "loss": 0.5501, "step": 9180 }, { "epoch": 2.365203963453867, "grad_norm": 1.4172649383544922, "learning_rate": 2.6352638352638352e-05, "loss": 0.5803, "step": 9190 }, { "epoch": 2.3677776347960364, "grad_norm": 1.612728238105774, "learning_rate": 2.632689832689833e-05, "loss": 0.6375, "step": 9200 }, { "epoch": 2.370351306138206, "grad_norm": 1.0975571870803833, "learning_rate": 2.6301158301158302e-05, "loss": 0.5791, "step": 9210 }, { "epoch": 2.3729249774803756, "grad_norm": 3.388500452041626, "learning_rate": 2.6275418275418277e-05, "loss": 0.5247, "step": 9220 }, { "epoch": 2.3754986488225454, "grad_norm": 0.9185710549354553, "learning_rate": 2.624967824967825e-05, "loss": 0.508, "step": 9230 }, { "epoch": 2.378072320164715, "grad_norm": 1.2001655101776123, "learning_rate": 2.6223938223938227e-05, "loss": 0.5771, "step": 9240 }, { "epoch": 2.3806459915068845, "grad_norm": 1.817859411239624, "learning_rate": 2.61981981981982e-05, "loss": 0.5443, "step": 9250 }, { "epoch": 2.3832196628490543, "grad_norm": 1.0994833707809448, "learning_rate": 2.6172458172458174e-05, "loss": 0.5538, "step": 9260 }, { "epoch": 2.3857933341912236, "grad_norm": 1.3039990663528442, "learning_rate": 2.6146718146718145e-05, "loss": 0.5196, "step": 9270 }, { "epoch": 2.3883670055333934, "grad_norm": 2.651137113571167, "learning_rate": 2.6120978120978124e-05, "loss": 0.5538, "step": 9280 }, { "epoch": 2.390940676875563, "grad_norm": 1.00489342212677, "learning_rate": 2.6095238095238095e-05, "loss": 0.5978, "step": 9290 }, { "epoch": 2.3935143482177326, "grad_norm": 2.647886276245117, "learning_rate": 2.606949806949807e-05, "loss": 0.5674, "step": 9300 }, { "epoch": 2.3960880195599024, "grad_norm": 1.094959020614624, "learning_rate": 2.6043758043758042e-05, "loss": 0.5768, "step": 9310 }, { "epoch": 2.3986616909020717, "grad_norm": 3.2877941131591797, "learning_rate": 2.601801801801802e-05, "loss": 0.5772, "step": 9320 }, { "epoch": 2.4012353622442415, "grad_norm": 0.7123093008995056, "learning_rate": 2.5992277992277992e-05, "loss": 0.5598, "step": 9330 }, { "epoch": 2.403809033586411, "grad_norm": 0.902870237827301, "learning_rate": 2.5966537966537967e-05, "loss": 0.5089, "step": 9340 }, { "epoch": 2.4063827049285806, "grad_norm": 1.3696403503417969, "learning_rate": 2.594079794079794e-05, "loss": 0.5726, "step": 9350 }, { "epoch": 2.4089563762707504, "grad_norm": 3.460519790649414, "learning_rate": 2.5915057915057917e-05, "loss": 0.5474, "step": 9360 }, { "epoch": 2.41153004761292, "grad_norm": 1.307590365409851, "learning_rate": 2.588931788931789e-05, "loss": 0.6438, "step": 9370 }, { "epoch": 2.4141037189550896, "grad_norm": 3.9630610942840576, "learning_rate": 2.5863577863577864e-05, "loss": 0.5163, "step": 9380 }, { "epoch": 2.416677390297259, "grad_norm": 1.028784990310669, "learning_rate": 2.5837837837837842e-05, "loss": 0.5345, "step": 9390 }, { "epoch": 2.4192510616394287, "grad_norm": 0.9117390513420105, "learning_rate": 2.5812097812097814e-05, "loss": 0.5966, "step": 9400 }, { "epoch": 2.421824732981598, "grad_norm": 3.028024911880493, "learning_rate": 2.578635778635779e-05, "loss": 0.5833, "step": 9410 }, { "epoch": 2.424398404323768, "grad_norm": 1.1407766342163086, "learning_rate": 2.576061776061776e-05, "loss": 0.5501, "step": 9420 }, { "epoch": 2.4269720756659376, "grad_norm": 1.0818943977355957, "learning_rate": 2.573487773487774e-05, "loss": 0.5323, "step": 9430 }, { "epoch": 2.429545747008107, "grad_norm": 1.1713372468948364, "learning_rate": 2.570913770913771e-05, "loss": 0.6247, "step": 9440 }, { "epoch": 2.432119418350277, "grad_norm": 1.3347368240356445, "learning_rate": 2.5683397683397685e-05, "loss": 0.6161, "step": 9450 }, { "epoch": 2.434693089692446, "grad_norm": 1.3403719663619995, "learning_rate": 2.5657657657657657e-05, "loss": 0.543, "step": 9460 }, { "epoch": 2.437266761034616, "grad_norm": 1.3469754457473755, "learning_rate": 2.5631917631917635e-05, "loss": 0.5616, "step": 9470 }, { "epoch": 2.4398404323767853, "grad_norm": 1.171751856803894, "learning_rate": 2.5606177606177607e-05, "loss": 0.5982, "step": 9480 }, { "epoch": 2.442414103718955, "grad_norm": 2.609652042388916, "learning_rate": 2.5580437580437582e-05, "loss": 0.5782, "step": 9490 }, { "epoch": 2.444987775061125, "grad_norm": 1.0340471267700195, "learning_rate": 2.5554697554697553e-05, "loss": 0.5698, "step": 9500 }, { "epoch": 2.447561446403294, "grad_norm": 0.9706631302833557, "learning_rate": 2.5528957528957532e-05, "loss": 0.5886, "step": 9510 }, { "epoch": 2.450135117745464, "grad_norm": 1.287369966506958, "learning_rate": 2.5503217503217503e-05, "loss": 0.5721, "step": 9520 }, { "epoch": 2.4527087890876333, "grad_norm": 0.9612070918083191, "learning_rate": 2.547747747747748e-05, "loss": 0.6088, "step": 9530 }, { "epoch": 2.455282460429803, "grad_norm": 2.4824914932250977, "learning_rate": 2.545173745173745e-05, "loss": 0.5057, "step": 9540 }, { "epoch": 2.4578561317719725, "grad_norm": 2.4112958908081055, "learning_rate": 2.542599742599743e-05, "loss": 0.5689, "step": 9550 }, { "epoch": 2.4604298031141423, "grad_norm": 0.7653863430023193, "learning_rate": 2.54002574002574e-05, "loss": 0.5119, "step": 9560 }, { "epoch": 2.463003474456312, "grad_norm": 1.1979918479919434, "learning_rate": 2.5374517374517375e-05, "loss": 0.5162, "step": 9570 }, { "epoch": 2.4655771457984814, "grad_norm": 1.1979789733886719, "learning_rate": 2.5348777348777347e-05, "loss": 0.5627, "step": 9580 }, { "epoch": 2.468150817140651, "grad_norm": 2.0756313800811768, "learning_rate": 2.5323037323037325e-05, "loss": 0.5502, "step": 9590 }, { "epoch": 2.4707244884828206, "grad_norm": 3.4498813152313232, "learning_rate": 2.52972972972973e-05, "loss": 0.5584, "step": 9600 }, { "epoch": 2.4732981598249903, "grad_norm": 1.145731806755066, "learning_rate": 2.527155727155727e-05, "loss": 0.5082, "step": 9610 }, { "epoch": 2.47587183116716, "grad_norm": 1.515875220298767, "learning_rate": 2.524581724581725e-05, "loss": 0.5051, "step": 9620 }, { "epoch": 2.4784455025093295, "grad_norm": 0.8977373838424683, "learning_rate": 2.522007722007722e-05, "loss": 0.6196, "step": 9630 }, { "epoch": 2.4810191738514993, "grad_norm": 1.2553268671035767, "learning_rate": 2.5194337194337197e-05, "loss": 0.5503, "step": 9640 }, { "epoch": 2.4835928451936686, "grad_norm": 0.8410692811012268, "learning_rate": 2.5168597168597168e-05, "loss": 0.5605, "step": 9650 }, { "epoch": 2.4861665165358384, "grad_norm": 1.2535247802734375, "learning_rate": 2.5142857142857147e-05, "loss": 0.5164, "step": 9660 }, { "epoch": 2.488740187878008, "grad_norm": 1.1014875173568726, "learning_rate": 2.5117117117117118e-05, "loss": 0.5934, "step": 9670 }, { "epoch": 2.4913138592201776, "grad_norm": 1.0416326522827148, "learning_rate": 2.5091377091377093e-05, "loss": 0.6187, "step": 9680 }, { "epoch": 2.4938875305623474, "grad_norm": 1.3807120323181152, "learning_rate": 2.5065637065637065e-05, "loss": 0.5462, "step": 9690 }, { "epoch": 2.4964612019045167, "grad_norm": 1.4746110439300537, "learning_rate": 2.5039897039897043e-05, "loss": 0.5576, "step": 9700 }, { "epoch": 2.4990348732466865, "grad_norm": 1.372180461883545, "learning_rate": 2.5014157014157015e-05, "loss": 0.5218, "step": 9710 }, { "epoch": 2.501608544588856, "grad_norm": 0.8345748782157898, "learning_rate": 2.498841698841699e-05, "loss": 0.5393, "step": 9720 }, { "epoch": 2.5041822159310256, "grad_norm": 1.1809195280075073, "learning_rate": 2.4962676962676965e-05, "loss": 0.5577, "step": 9730 }, { "epoch": 2.5067558872731954, "grad_norm": 1.1558030843734741, "learning_rate": 2.493693693693694e-05, "loss": 0.6091, "step": 9740 }, { "epoch": 2.5093295586153648, "grad_norm": 0.8379091620445251, "learning_rate": 2.491119691119691e-05, "loss": 0.5787, "step": 9750 }, { "epoch": 2.5119032299575346, "grad_norm": 1.2100491523742676, "learning_rate": 2.4885456885456886e-05, "loss": 0.5179, "step": 9760 }, { "epoch": 2.514476901299704, "grad_norm": 2.15959095954895, "learning_rate": 2.485971685971686e-05, "loss": 0.5448, "step": 9770 }, { "epoch": 2.5170505726418737, "grad_norm": 1.1132477521896362, "learning_rate": 2.4833976833976836e-05, "loss": 0.5354, "step": 9780 }, { "epoch": 2.519624243984043, "grad_norm": 1.1005936861038208, "learning_rate": 2.4808236808236808e-05, "loss": 0.6376, "step": 9790 }, { "epoch": 2.522197915326213, "grad_norm": 1.0520378351211548, "learning_rate": 2.4782496782496783e-05, "loss": 0.5468, "step": 9800 }, { "epoch": 2.5247715866683826, "grad_norm": 3.0992591381073, "learning_rate": 2.4756756756756758e-05, "loss": 0.6128, "step": 9810 }, { "epoch": 2.527345258010552, "grad_norm": 0.9129619002342224, "learning_rate": 2.4731016731016733e-05, "loss": 0.5671, "step": 9820 }, { "epoch": 2.529918929352722, "grad_norm": 4.400023460388184, "learning_rate": 2.4705276705276705e-05, "loss": 0.5863, "step": 9830 }, { "epoch": 2.532492600694891, "grad_norm": 1.1237571239471436, "learning_rate": 2.467953667953668e-05, "loss": 0.5115, "step": 9840 }, { "epoch": 2.535066272037061, "grad_norm": 1.2072199583053589, "learning_rate": 2.4653796653796655e-05, "loss": 0.5748, "step": 9850 }, { "epoch": 2.5376399433792303, "grad_norm": 0.8158368468284607, "learning_rate": 2.462805662805663e-05, "loss": 0.6466, "step": 9860 }, { "epoch": 2.5402136147214, "grad_norm": 1.3030297756195068, "learning_rate": 2.46023166023166e-05, "loss": 0.5114, "step": 9870 }, { "epoch": 2.54278728606357, "grad_norm": 1.1183769702911377, "learning_rate": 2.4576576576576576e-05, "loss": 0.5518, "step": 9880 }, { "epoch": 2.545360957405739, "grad_norm": 0.9065691828727722, "learning_rate": 2.455083655083655e-05, "loss": 0.522, "step": 9890 }, { "epoch": 2.547934628747909, "grad_norm": 1.4766273498535156, "learning_rate": 2.4525096525096526e-05, "loss": 0.587, "step": 9900 }, { "epoch": 2.5505083000900783, "grad_norm": 4.340095043182373, "learning_rate": 2.44993564993565e-05, "loss": 0.6021, "step": 9910 }, { "epoch": 2.553081971432248, "grad_norm": 0.8867147564888, "learning_rate": 2.4473616473616476e-05, "loss": 0.5076, "step": 9920 }, { "epoch": 2.5556556427744175, "grad_norm": 1.2805358171463013, "learning_rate": 2.444787644787645e-05, "loss": 0.5817, "step": 9930 }, { "epoch": 2.5582293141165873, "grad_norm": 1.0419846773147583, "learning_rate": 2.4422136422136423e-05, "loss": 0.5846, "step": 9940 }, { "epoch": 2.560802985458757, "grad_norm": 1.3988200426101685, "learning_rate": 2.4396396396396398e-05, "loss": 0.5483, "step": 9950 }, { "epoch": 2.5633766568009264, "grad_norm": 1.3084689378738403, "learning_rate": 2.4370656370656373e-05, "loss": 0.5685, "step": 9960 }, { "epoch": 2.565950328143096, "grad_norm": 1.160992980003357, "learning_rate": 2.4344916344916348e-05, "loss": 0.5852, "step": 9970 }, { "epoch": 2.568523999485266, "grad_norm": 1.859298825263977, "learning_rate": 2.431917631917632e-05, "loss": 0.5561, "step": 9980 }, { "epoch": 2.5710976708274353, "grad_norm": 0.9575828313827515, "learning_rate": 2.4293436293436294e-05, "loss": 0.5501, "step": 9990 }, { "epoch": 2.5736713421696047, "grad_norm": 1.3349659442901611, "learning_rate": 2.426769626769627e-05, "loss": 0.5458, "step": 10000 }, { "epoch": 2.5736713421696047, "eval_loss": 0.7813047170639038, "eval_runtime": 395.5545, "eval_samples_per_second": 49.111, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.0003636529278236245, "step": 10000 }, { "epoch": 2.5762450135117745, "grad_norm": 0.8265610933303833, "learning_rate": 2.4241956241956244e-05, "loss": 0.5199, "step": 10010 }, { "epoch": 2.5788186848539443, "grad_norm": 1.1226060390472412, "learning_rate": 2.4216216216216216e-05, "loss": 0.5416, "step": 10020 }, { "epoch": 2.5813923561961136, "grad_norm": 1.006882905960083, "learning_rate": 2.419047619047619e-05, "loss": 0.6422, "step": 10030 }, { "epoch": 2.5839660275382834, "grad_norm": 1.1571662425994873, "learning_rate": 2.4164736164736166e-05, "loss": 0.5548, "step": 10040 }, { "epoch": 2.586539698880453, "grad_norm": 0.9538158774375916, "learning_rate": 2.413899613899614e-05, "loss": 0.5416, "step": 10050 }, { "epoch": 2.5891133702226226, "grad_norm": 3.1586387157440186, "learning_rate": 2.4113256113256113e-05, "loss": 0.5445, "step": 10060 }, { "epoch": 2.591687041564792, "grad_norm": 0.9174941778182983, "learning_rate": 2.4087516087516088e-05, "loss": 0.5539, "step": 10070 }, { "epoch": 2.5942607129069617, "grad_norm": 0.8916310667991638, "learning_rate": 2.4061776061776063e-05, "loss": 0.5487, "step": 10080 }, { "epoch": 2.5968343842491315, "grad_norm": 3.3344202041625977, "learning_rate": 2.4036036036036034e-05, "loss": 0.5935, "step": 10090 }, { "epoch": 2.599408055591301, "grad_norm": 1.8310283422470093, "learning_rate": 2.4010296010296013e-05, "loss": 0.5507, "step": 10100 }, { "epoch": 2.6019817269334706, "grad_norm": 0.95711749792099, "learning_rate": 2.3984555984555988e-05, "loss": 0.5899, "step": 10110 }, { "epoch": 2.6045553982756404, "grad_norm": 1.1293368339538574, "learning_rate": 2.3958815958815963e-05, "loss": 0.5468, "step": 10120 }, { "epoch": 2.6071290696178098, "grad_norm": 0.8633250594139099, "learning_rate": 2.3933075933075934e-05, "loss": 0.5763, "step": 10130 }, { "epoch": 2.6097027409599796, "grad_norm": 1.343990445137024, "learning_rate": 2.390733590733591e-05, "loss": 0.5839, "step": 10140 }, { "epoch": 2.612276412302149, "grad_norm": 1.5311845541000366, "learning_rate": 2.3881595881595884e-05, "loss": 0.531, "step": 10150 }, { "epoch": 2.6148500836443187, "grad_norm": 1.0829716920852661, "learning_rate": 2.385585585585586e-05, "loss": 0.5402, "step": 10160 }, { "epoch": 2.617423754986488, "grad_norm": 1.5562556982040405, "learning_rate": 2.383011583011583e-05, "loss": 0.5652, "step": 10170 }, { "epoch": 2.619997426328658, "grad_norm": 1.0857690572738647, "learning_rate": 2.3804375804375806e-05, "loss": 0.6161, "step": 10180 }, { "epoch": 2.6225710976708276, "grad_norm": 1.1553452014923096, "learning_rate": 2.377863577863578e-05, "loss": 0.5975, "step": 10190 }, { "epoch": 2.625144769012997, "grad_norm": 0.9992988109588623, "learning_rate": 2.3752895752895752e-05, "loss": 0.5837, "step": 10200 }, { "epoch": 2.6277184403551668, "grad_norm": 1.0627033710479736, "learning_rate": 2.3727155727155727e-05, "loss": 0.5527, "step": 10210 }, { "epoch": 2.630292111697336, "grad_norm": 0.8239421248435974, "learning_rate": 2.3701415701415702e-05, "loss": 0.5822, "step": 10220 }, { "epoch": 2.632865783039506, "grad_norm": 2.0370047092437744, "learning_rate": 2.3675675675675677e-05, "loss": 0.55, "step": 10230 }, { "epoch": 2.6354394543816753, "grad_norm": 1.0203136205673218, "learning_rate": 2.364993564993565e-05, "loss": 0.5289, "step": 10240 }, { "epoch": 2.638013125723845, "grad_norm": 1.1402500867843628, "learning_rate": 2.3624195624195624e-05, "loss": 0.5492, "step": 10250 }, { "epoch": 2.640586797066015, "grad_norm": 1.1327241659164429, "learning_rate": 2.35984555984556e-05, "loss": 0.5416, "step": 10260 }, { "epoch": 2.643160468408184, "grad_norm": 0.9209975004196167, "learning_rate": 2.3572715572715574e-05, "loss": 0.5442, "step": 10270 }, { "epoch": 2.645734139750354, "grad_norm": 1.3790748119354248, "learning_rate": 2.3546975546975546e-05, "loss": 0.6103, "step": 10280 }, { "epoch": 2.6483078110925233, "grad_norm": 1.2023276090621948, "learning_rate": 2.352123552123552e-05, "loss": 0.5886, "step": 10290 }, { "epoch": 2.650881482434693, "grad_norm": 1.0166372060775757, "learning_rate": 2.3495495495495496e-05, "loss": 0.5741, "step": 10300 }, { "epoch": 2.6534551537768625, "grad_norm": 1.1085914373397827, "learning_rate": 2.346975546975547e-05, "loss": 0.5302, "step": 10310 }, { "epoch": 2.6560288251190323, "grad_norm": 1.1305118799209595, "learning_rate": 2.3444015444015446e-05, "loss": 0.5851, "step": 10320 }, { "epoch": 2.658602496461202, "grad_norm": 4.382242202758789, "learning_rate": 2.341827541827542e-05, "loss": 0.6207, "step": 10330 }, { "epoch": 2.6611761678033714, "grad_norm": 1.3495676517486572, "learning_rate": 2.3392535392535396e-05, "loss": 0.531, "step": 10340 }, { "epoch": 2.663749839145541, "grad_norm": 1.2372089624404907, "learning_rate": 2.3366795366795367e-05, "loss": 0.5543, "step": 10350 }, { "epoch": 2.666323510487711, "grad_norm": 1.2781168222427368, "learning_rate": 2.3341055341055342e-05, "loss": 0.5379, "step": 10360 }, { "epoch": 2.6688971818298803, "grad_norm": 0.9602215886116028, "learning_rate": 2.3315315315315317e-05, "loss": 0.5677, "step": 10370 }, { "epoch": 2.6714708531720497, "grad_norm": 1.4778882265090942, "learning_rate": 2.3289575289575292e-05, "loss": 0.5196, "step": 10380 }, { "epoch": 2.6740445245142195, "grad_norm": 1.1406737565994263, "learning_rate": 2.3263835263835264e-05, "loss": 0.5886, "step": 10390 }, { "epoch": 2.6766181958563893, "grad_norm": 1.3045039176940918, "learning_rate": 2.323809523809524e-05, "loss": 0.5468, "step": 10400 }, { "epoch": 2.6791918671985586, "grad_norm": 1.1120688915252686, "learning_rate": 2.3212355212355214e-05, "loss": 0.6246, "step": 10410 }, { "epoch": 2.6817655385407284, "grad_norm": 3.4695723056793213, "learning_rate": 2.318661518661519e-05, "loss": 0.5557, "step": 10420 }, { "epoch": 2.684339209882898, "grad_norm": 1.4263606071472168, "learning_rate": 2.316087516087516e-05, "loss": 0.5656, "step": 10430 }, { "epoch": 2.6869128812250676, "grad_norm": 0.7746553421020508, "learning_rate": 2.3135135135135136e-05, "loss": 0.5458, "step": 10440 }, { "epoch": 2.689486552567237, "grad_norm": 1.4084376096725464, "learning_rate": 2.310939510939511e-05, "loss": 0.5506, "step": 10450 }, { "epoch": 2.6920602239094067, "grad_norm": 0.9711858034133911, "learning_rate": 2.3083655083655086e-05, "loss": 0.5219, "step": 10460 }, { "epoch": 2.6946338952515765, "grad_norm": 1.1122053861618042, "learning_rate": 2.3057915057915057e-05, "loss": 0.5287, "step": 10470 }, { "epoch": 2.697207566593746, "grad_norm": 1.0087629556655884, "learning_rate": 2.3032175032175032e-05, "loss": 0.6167, "step": 10480 }, { "epoch": 2.6997812379359156, "grad_norm": 1.0684638023376465, "learning_rate": 2.3009009009009013e-05, "loss": 0.5191, "step": 10490 }, { "epoch": 2.7023549092780854, "grad_norm": 1.1910923719406128, "learning_rate": 2.2983268983268984e-05, "loss": 0.551, "step": 10500 }, { "epoch": 2.7049285806202548, "grad_norm": 1.287802815437317, "learning_rate": 2.295752895752896e-05, "loss": 0.5438, "step": 10510 }, { "epoch": 2.7075022519624246, "grad_norm": 0.9688395857810974, "learning_rate": 2.2931788931788934e-05, "loss": 0.5567, "step": 10520 }, { "epoch": 2.710075923304594, "grad_norm": 1.0128886699676514, "learning_rate": 2.2906048906048906e-05, "loss": 0.6127, "step": 10530 }, { "epoch": 2.7126495946467637, "grad_norm": 1.974613070487976, "learning_rate": 2.288030888030888e-05, "loss": 0.5642, "step": 10540 }, { "epoch": 2.715223265988933, "grad_norm": 2.323258876800537, "learning_rate": 2.2854568854568856e-05, "loss": 0.5192, "step": 10550 }, { "epoch": 2.717796937331103, "grad_norm": 2.6239235401153564, "learning_rate": 2.282882882882883e-05, "loss": 0.5786, "step": 10560 }, { "epoch": 2.7203706086732726, "grad_norm": 2.535557508468628, "learning_rate": 2.2803088803088802e-05, "loss": 0.5732, "step": 10570 }, { "epoch": 2.722944280015442, "grad_norm": 1.4049146175384521, "learning_rate": 2.2777348777348777e-05, "loss": 0.5613, "step": 10580 }, { "epoch": 2.7255179513576118, "grad_norm": 0.9087219834327698, "learning_rate": 2.2751608751608752e-05, "loss": 0.5429, "step": 10590 }, { "epoch": 2.728091622699781, "grad_norm": 1.2348556518554688, "learning_rate": 2.2725868725868727e-05, "loss": 0.6106, "step": 10600 }, { "epoch": 2.730665294041951, "grad_norm": 0.954634964466095, "learning_rate": 2.27001287001287e-05, "loss": 0.5648, "step": 10610 }, { "epoch": 2.7332389653841203, "grad_norm": 0.9417296051979065, "learning_rate": 2.2674388674388674e-05, "loss": 0.5641, "step": 10620 }, { "epoch": 2.73581263672629, "grad_norm": 3.07540225982666, "learning_rate": 2.264864864864865e-05, "loss": 0.6197, "step": 10630 }, { "epoch": 2.73838630806846, "grad_norm": 1.1706743240356445, "learning_rate": 2.2622908622908624e-05, "loss": 0.5237, "step": 10640 }, { "epoch": 2.740959979410629, "grad_norm": 1.0390594005584717, "learning_rate": 2.2597168597168596e-05, "loss": 0.5377, "step": 10650 }, { "epoch": 2.743533650752799, "grad_norm": 4.094557762145996, "learning_rate": 2.257142857142857e-05, "loss": 0.5717, "step": 10660 }, { "epoch": 2.7461073220949683, "grad_norm": 3.71659255027771, "learning_rate": 2.254568854568855e-05, "loss": 0.5725, "step": 10670 }, { "epoch": 2.748680993437138, "grad_norm": 1.1988261938095093, "learning_rate": 2.251994851994852e-05, "loss": 0.565, "step": 10680 }, { "epoch": 2.7512546647793075, "grad_norm": 1.5038524866104126, "learning_rate": 2.2494208494208496e-05, "loss": 0.5142, "step": 10690 }, { "epoch": 2.7538283361214773, "grad_norm": 1.0578770637512207, "learning_rate": 2.246846846846847e-05, "loss": 0.5922, "step": 10700 }, { "epoch": 2.756402007463647, "grad_norm": 1.151419997215271, "learning_rate": 2.2442728442728446e-05, "loss": 0.6072, "step": 10710 }, { "epoch": 2.7589756788058164, "grad_norm": 1.1115498542785645, "learning_rate": 2.2416988416988417e-05, "loss": 0.5549, "step": 10720 }, { "epoch": 2.761549350147986, "grad_norm": 2.30755352973938, "learning_rate": 2.2391248391248392e-05, "loss": 0.5863, "step": 10730 }, { "epoch": 2.7641230214901555, "grad_norm": 2.3531439304351807, "learning_rate": 2.2365508365508367e-05, "loss": 0.5236, "step": 10740 }, { "epoch": 2.7666966928323253, "grad_norm": 1.0621081590652466, "learning_rate": 2.2339768339768342e-05, "loss": 0.5689, "step": 10750 }, { "epoch": 2.7692703641744947, "grad_norm": 3.393990993499756, "learning_rate": 2.2314028314028314e-05, "loss": 0.5736, "step": 10760 }, { "epoch": 2.7718440355166645, "grad_norm": 0.9856056571006775, "learning_rate": 2.228828828828829e-05, "loss": 0.5719, "step": 10770 }, { "epoch": 2.7744177068588343, "grad_norm": 1.3142099380493164, "learning_rate": 2.2262548262548264e-05, "loss": 0.5546, "step": 10780 }, { "epoch": 2.7769913782010036, "grad_norm": 0.969963550567627, "learning_rate": 2.223680823680824e-05, "loss": 0.5425, "step": 10790 }, { "epoch": 2.7795650495431734, "grad_norm": 2.802344560623169, "learning_rate": 2.221106821106821e-05, "loss": 0.5235, "step": 10800 }, { "epoch": 2.782138720885343, "grad_norm": 1.5331348180770874, "learning_rate": 2.2185328185328186e-05, "loss": 0.5436, "step": 10810 }, { "epoch": 2.7847123922275125, "grad_norm": 0.9570822715759277, "learning_rate": 2.215958815958816e-05, "loss": 0.5615, "step": 10820 }, { "epoch": 2.787286063569682, "grad_norm": 3.3715949058532715, "learning_rate": 2.2133848133848136e-05, "loss": 0.5473, "step": 10830 }, { "epoch": 2.7898597349118517, "grad_norm": 1.2080599069595337, "learning_rate": 2.2108108108108107e-05, "loss": 0.6758, "step": 10840 }, { "epoch": 2.7924334062540215, "grad_norm": 0.9372673034667969, "learning_rate": 2.2082368082368082e-05, "loss": 0.5913, "step": 10850 }, { "epoch": 2.795007077596191, "grad_norm": 1.019864559173584, "learning_rate": 2.2056628056628057e-05, "loss": 0.5194, "step": 10860 }, { "epoch": 2.7975807489383606, "grad_norm": 1.1062774658203125, "learning_rate": 2.2030888030888032e-05, "loss": 0.5451, "step": 10870 }, { "epoch": 2.8001544202805304, "grad_norm": 0.8986897468566895, "learning_rate": 2.2005148005148007e-05, "loss": 0.5787, "step": 10880 }, { "epoch": 2.8027280916226998, "grad_norm": 0.9067063331604004, "learning_rate": 2.1979407979407982e-05, "loss": 0.5901, "step": 10890 }, { "epoch": 2.8053017629648695, "grad_norm": 1.141993522644043, "learning_rate": 2.1953667953667957e-05, "loss": 0.5681, "step": 10900 }, { "epoch": 2.807875434307039, "grad_norm": 1.1285045146942139, "learning_rate": 2.192792792792793e-05, "loss": 0.5557, "step": 10910 }, { "epoch": 2.8104491056492087, "grad_norm": 0.9529311656951904, "learning_rate": 2.1902187902187904e-05, "loss": 0.6106, "step": 10920 }, { "epoch": 2.813022776991378, "grad_norm": 0.9633564949035645, "learning_rate": 2.187644787644788e-05, "loss": 0.521, "step": 10930 }, { "epoch": 2.815596448333548, "grad_norm": 1.3736627101898193, "learning_rate": 2.1850707850707854e-05, "loss": 0.4961, "step": 10940 }, { "epoch": 2.8181701196757176, "grad_norm": 1.1988080739974976, "learning_rate": 2.1824967824967825e-05, "loss": 0.5339, "step": 10950 }, { "epoch": 2.820743791017887, "grad_norm": 3.6586461067199707, "learning_rate": 2.17992277992278e-05, "loss": 0.5535, "step": 10960 }, { "epoch": 2.8233174623600568, "grad_norm": 1.6406865119934082, "learning_rate": 2.1773487773487775e-05, "loss": 0.5244, "step": 10970 }, { "epoch": 2.825891133702226, "grad_norm": 1.2448114156723022, "learning_rate": 2.174774774774775e-05, "loss": 0.6124, "step": 10980 }, { "epoch": 2.828464805044396, "grad_norm": 0.9966627955436707, "learning_rate": 2.1722007722007722e-05, "loss": 0.5213, "step": 10990 }, { "epoch": 2.8310384763865653, "grad_norm": 1.5324152708053589, "learning_rate": 2.1696267696267697e-05, "loss": 0.5412, "step": 11000 }, { "epoch": 2.8310384763865653, "eval_loss": 0.7792600989341736, "eval_runtime": 395.6052, "eval_samples_per_second": 49.105, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00037447141262839833, "step": 11000 }, { "epoch": 2.833612147728735, "grad_norm": 3.205904483795166, "learning_rate": 2.1670527670527672e-05, "loss": 0.4943, "step": 11010 }, { "epoch": 2.836185819070905, "grad_norm": 1.0843708515167236, "learning_rate": 2.1644787644787644e-05, "loss": 0.5571, "step": 11020 }, { "epoch": 2.838759490413074, "grad_norm": 0.9089128971099854, "learning_rate": 2.161904761904762e-05, "loss": 0.5604, "step": 11030 }, { "epoch": 2.841333161755244, "grad_norm": 1.2132848501205444, "learning_rate": 2.1593307593307594e-05, "loss": 0.5426, "step": 11040 }, { "epoch": 2.8439068330974133, "grad_norm": 1.2147984504699707, "learning_rate": 2.156756756756757e-05, "loss": 0.5833, "step": 11050 }, { "epoch": 2.846480504439583, "grad_norm": 0.9623033404350281, "learning_rate": 2.154182754182754e-05, "loss": 0.5253, "step": 11060 }, { "epoch": 2.8490541757817525, "grad_norm": 1.1936800479888916, "learning_rate": 2.1516087516087515e-05, "loss": 0.548, "step": 11070 }, { "epoch": 2.8516278471239223, "grad_norm": 3.678395986557007, "learning_rate": 2.1490347490347494e-05, "loss": 0.5867, "step": 11080 }, { "epoch": 2.854201518466092, "grad_norm": 3.4173741340637207, "learning_rate": 2.1464607464607465e-05, "loss": 0.4922, "step": 11090 }, { "epoch": 2.8567751898082614, "grad_norm": 0.8987242579460144, "learning_rate": 2.143886743886744e-05, "loss": 0.5296, "step": 11100 }, { "epoch": 2.859348861150431, "grad_norm": 1.1060086488723755, "learning_rate": 2.1413127413127415e-05, "loss": 0.52, "step": 11110 }, { "epoch": 2.8619225324926005, "grad_norm": 0.9053888320922852, "learning_rate": 2.138738738738739e-05, "loss": 0.5979, "step": 11120 }, { "epoch": 2.8644962038347703, "grad_norm": 0.9858607053756714, "learning_rate": 2.1361647361647362e-05, "loss": 0.5683, "step": 11130 }, { "epoch": 2.8670698751769397, "grad_norm": 2.134103298187256, "learning_rate": 2.1335907335907337e-05, "loss": 0.5358, "step": 11140 }, { "epoch": 2.8696435465191095, "grad_norm": 4.8553009033203125, "learning_rate": 2.1310167310167312e-05, "loss": 0.5975, "step": 11150 }, { "epoch": 2.8722172178612793, "grad_norm": 1.1935124397277832, "learning_rate": 2.1284427284427287e-05, "loss": 0.6338, "step": 11160 }, { "epoch": 2.8747908892034486, "grad_norm": 1.042869210243225, "learning_rate": 2.125868725868726e-05, "loss": 0.5529, "step": 11170 }, { "epoch": 2.8773645605456184, "grad_norm": 0.7498083114624023, "learning_rate": 2.1232947232947233e-05, "loss": 0.6005, "step": 11180 }, { "epoch": 2.879938231887788, "grad_norm": 1.3521268367767334, "learning_rate": 2.120720720720721e-05, "loss": 0.5645, "step": 11190 }, { "epoch": 2.8825119032299575, "grad_norm": 1.2290796041488647, "learning_rate": 2.1181467181467183e-05, "loss": 0.561, "step": 11200 }, { "epoch": 2.885085574572127, "grad_norm": 1.114356517791748, "learning_rate": 2.1155727155727155e-05, "loss": 0.5426, "step": 11210 }, { "epoch": 2.8876592459142967, "grad_norm": 0.7884505391120911, "learning_rate": 2.112998712998713e-05, "loss": 0.5519, "step": 11220 }, { "epoch": 2.8902329172564665, "grad_norm": 1.0232847929000854, "learning_rate": 2.1104247104247105e-05, "loss": 0.5406, "step": 11230 }, { "epoch": 2.892806588598636, "grad_norm": 1.2134202718734741, "learning_rate": 2.107850707850708e-05, "loss": 0.537, "step": 11240 }, { "epoch": 2.8953802599408056, "grad_norm": 1.1819850206375122, "learning_rate": 2.105276705276705e-05, "loss": 0.4655, "step": 11250 }, { "epoch": 2.8979539312829754, "grad_norm": 0.9227721095085144, "learning_rate": 2.1027027027027027e-05, "loss": 0.5465, "step": 11260 }, { "epoch": 2.9005276026251448, "grad_norm": 4.2423858642578125, "learning_rate": 2.1001287001287e-05, "loss": 0.5245, "step": 11270 }, { "epoch": 2.903101273967314, "grad_norm": 1.0948383808135986, "learning_rate": 2.0975546975546977e-05, "loss": 0.5712, "step": 11280 }, { "epoch": 2.905674945309484, "grad_norm": 0.783808171749115, "learning_rate": 2.094980694980695e-05, "loss": 0.5705, "step": 11290 }, { "epoch": 2.9082486166516537, "grad_norm": 0.858773410320282, "learning_rate": 2.0924066924066927e-05, "loss": 0.5632, "step": 11300 }, { "epoch": 2.910822287993823, "grad_norm": 4.24513578414917, "learning_rate": 2.08983268983269e-05, "loss": 0.5613, "step": 11310 }, { "epoch": 2.913395959335993, "grad_norm": 1.336777925491333, "learning_rate": 2.0872586872586873e-05, "loss": 0.6024, "step": 11320 }, { "epoch": 2.9159696306781626, "grad_norm": 0.7360126972198486, "learning_rate": 2.0846846846846848e-05, "loss": 0.5051, "step": 11330 }, { "epoch": 2.918543302020332, "grad_norm": 0.9983276128768921, "learning_rate": 2.0821106821106823e-05, "loss": 0.5366, "step": 11340 }, { "epoch": 2.9211169733625018, "grad_norm": 1.2790330648422241, "learning_rate": 2.0795366795366798e-05, "loss": 0.4972, "step": 11350 }, { "epoch": 2.923690644704671, "grad_norm": 1.0925854444503784, "learning_rate": 2.076962676962677e-05, "loss": 0.4877, "step": 11360 }, { "epoch": 2.926264316046841, "grad_norm": 1.262011170387268, "learning_rate": 2.0743886743886745e-05, "loss": 0.5177, "step": 11370 }, { "epoch": 2.9288379873890102, "grad_norm": 0.7657809853553772, "learning_rate": 2.071814671814672e-05, "loss": 0.5845, "step": 11380 }, { "epoch": 2.93141165873118, "grad_norm": 1.21810781955719, "learning_rate": 2.0692406692406695e-05, "loss": 0.4873, "step": 11390 }, { "epoch": 2.93398533007335, "grad_norm": 1.1129614114761353, "learning_rate": 2.0666666666666666e-05, "loss": 0.5668, "step": 11400 }, { "epoch": 2.936559001415519, "grad_norm": 1.1121175289154053, "learning_rate": 2.064092664092664e-05, "loss": 0.4534, "step": 11410 }, { "epoch": 2.939132672757689, "grad_norm": 1.254805564880371, "learning_rate": 2.0615186615186616e-05, "loss": 0.5647, "step": 11420 }, { "epoch": 2.9417063440998583, "grad_norm": 1.1753339767456055, "learning_rate": 2.058944658944659e-05, "loss": 0.5619, "step": 11430 }, { "epoch": 2.944280015442028, "grad_norm": 1.9952584505081177, "learning_rate": 2.0563706563706563e-05, "loss": 0.5437, "step": 11440 }, { "epoch": 2.9468536867841975, "grad_norm": 2.9970035552978516, "learning_rate": 2.0537966537966538e-05, "loss": 0.6106, "step": 11450 }, { "epoch": 2.9494273581263672, "grad_norm": 1.2098712921142578, "learning_rate": 2.0512226512226513e-05, "loss": 0.5032, "step": 11460 }, { "epoch": 2.952001029468537, "grad_norm": 2.842196464538574, "learning_rate": 2.0486486486486488e-05, "loss": 0.4807, "step": 11470 }, { "epoch": 2.9545747008107064, "grad_norm": 0.9120272397994995, "learning_rate": 2.046074646074646e-05, "loss": 0.5058, "step": 11480 }, { "epoch": 2.957148372152876, "grad_norm": 2.2939398288726807, "learning_rate": 2.0435006435006435e-05, "loss": 0.5666, "step": 11490 }, { "epoch": 2.9597220434950455, "grad_norm": 1.4755898714065552, "learning_rate": 2.0409266409266413e-05, "loss": 0.4845, "step": 11500 }, { "epoch": 2.9622957148372153, "grad_norm": 1.120060920715332, "learning_rate": 2.0383526383526385e-05, "loss": 0.4979, "step": 11510 }, { "epoch": 2.9648693861793847, "grad_norm": 4.3705058097839355, "learning_rate": 2.035778635778636e-05, "loss": 0.6247, "step": 11520 }, { "epoch": 2.9674430575215545, "grad_norm": 1.2147754430770874, "learning_rate": 2.0332046332046335e-05, "loss": 0.512, "step": 11530 }, { "epoch": 2.9700167288637243, "grad_norm": 4.265937328338623, "learning_rate": 2.030630630630631e-05, "loss": 0.533, "step": 11540 }, { "epoch": 2.9725904002058936, "grad_norm": 1.1540546417236328, "learning_rate": 2.028056628056628e-05, "loss": 0.5189, "step": 11550 }, { "epoch": 2.9751640715480634, "grad_norm": 1.3933578729629517, "learning_rate": 2.0254826254826256e-05, "loss": 0.6381, "step": 11560 }, { "epoch": 2.977737742890233, "grad_norm": 1.1027209758758545, "learning_rate": 2.022908622908623e-05, "loss": 0.5235, "step": 11570 }, { "epoch": 2.9803114142324025, "grad_norm": 0.9615407586097717, "learning_rate": 2.0203346203346203e-05, "loss": 0.6184, "step": 11580 }, { "epoch": 2.982885085574572, "grad_norm": 1.2538810968399048, "learning_rate": 2.0177606177606178e-05, "loss": 0.576, "step": 11590 }, { "epoch": 2.9854587569167417, "grad_norm": 1.3477590084075928, "learning_rate": 2.0151866151866153e-05, "loss": 0.5825, "step": 11600 }, { "epoch": 2.9880324282589115, "grad_norm": 4.23989725112915, "learning_rate": 2.0126126126126128e-05, "loss": 0.5581, "step": 11610 }, { "epoch": 2.990606099601081, "grad_norm": 1.2521101236343384, "learning_rate": 2.01003861003861e-05, "loss": 0.6155, "step": 11620 }, { "epoch": 2.9931797709432506, "grad_norm": 1.0629501342773438, "learning_rate": 2.0074646074646074e-05, "loss": 0.5349, "step": 11630 }, { "epoch": 2.9957534422854204, "grad_norm": 1.3439209461212158, "learning_rate": 2.004890604890605e-05, "loss": 0.5551, "step": 11640 }, { "epoch": 2.9983271136275897, "grad_norm": 0.9323076605796814, "learning_rate": 2.0023166023166024e-05, "loss": 0.5558, "step": 11650 }, { "epoch": 3.0009007849697595, "grad_norm": 1.5501611232757568, "learning_rate": 1.9997425997425996e-05, "loss": 0.5076, "step": 11660 }, { "epoch": 3.003474456311929, "grad_norm": 1.0962281227111816, "learning_rate": 1.997168597168597e-05, "loss": 0.4875, "step": 11670 }, { "epoch": 3.0060481276540987, "grad_norm": 1.1923311948776245, "learning_rate": 1.9945945945945946e-05, "loss": 0.4643, "step": 11680 }, { "epoch": 3.008621798996268, "grad_norm": 1.4216543436050415, "learning_rate": 1.992020592020592e-05, "loss": 0.4181, "step": 11690 }, { "epoch": 3.011195470338438, "grad_norm": 1.0901702642440796, "learning_rate": 1.9894465894465893e-05, "loss": 0.4309, "step": 11700 }, { "epoch": 3.013769141680607, "grad_norm": 0.9830016493797302, "learning_rate": 1.986872586872587e-05, "loss": 0.4027, "step": 11710 }, { "epoch": 3.016342813022777, "grad_norm": 0.9537155032157898, "learning_rate": 1.9842985842985846e-05, "loss": 0.4108, "step": 11720 }, { "epoch": 3.0189164843649468, "grad_norm": 1.4666250944137573, "learning_rate": 1.9817245817245818e-05, "loss": 0.4573, "step": 11730 }, { "epoch": 3.021490155707116, "grad_norm": 1.1737228631973267, "learning_rate": 1.9791505791505793e-05, "loss": 0.4616, "step": 11740 }, { "epoch": 3.024063827049286, "grad_norm": 0.9376081824302673, "learning_rate": 1.9765765765765768e-05, "loss": 0.4141, "step": 11750 }, { "epoch": 3.0266374983914552, "grad_norm": 3.225501775741577, "learning_rate": 1.9740025740025743e-05, "loss": 0.4542, "step": 11760 }, { "epoch": 3.029211169733625, "grad_norm": 0.9168078303337097, "learning_rate": 1.9714285714285714e-05, "loss": 0.4518, "step": 11770 }, { "epoch": 3.031784841075795, "grad_norm": 1.606330394744873, "learning_rate": 1.968854568854569e-05, "loss": 0.3725, "step": 11780 }, { "epoch": 3.034358512417964, "grad_norm": 3.959381341934204, "learning_rate": 1.9662805662805664e-05, "loss": 0.4397, "step": 11790 }, { "epoch": 3.036932183760134, "grad_norm": 0.9928016066551208, "learning_rate": 1.963706563706564e-05, "loss": 0.4031, "step": 11800 }, { "epoch": 3.0395058551023033, "grad_norm": 1.3497694730758667, "learning_rate": 1.961132561132561e-05, "loss": 0.4435, "step": 11810 }, { "epoch": 3.042079526444473, "grad_norm": 1.0364716053009033, "learning_rate": 1.9585585585585586e-05, "loss": 0.4138, "step": 11820 }, { "epoch": 3.0446531977866425, "grad_norm": 1.0680062770843506, "learning_rate": 1.955984555984556e-05, "loss": 0.448, "step": 11830 }, { "epoch": 3.0472268691288122, "grad_norm": 0.7505292892456055, "learning_rate": 1.9534105534105536e-05, "loss": 0.4604, "step": 11840 }, { "epoch": 3.049800540470982, "grad_norm": 1.549533724784851, "learning_rate": 1.9508365508365508e-05, "loss": 0.4082, "step": 11850 }, { "epoch": 3.0523742118131514, "grad_norm": 1.0561308860778809, "learning_rate": 1.9482625482625483e-05, "loss": 0.4062, "step": 11860 }, { "epoch": 3.054947883155321, "grad_norm": 1.0458426475524902, "learning_rate": 1.9456885456885458e-05, "loss": 0.4374, "step": 11870 }, { "epoch": 3.0575215544974905, "grad_norm": 0.9172849059104919, "learning_rate": 1.9431145431145433e-05, "loss": 0.4426, "step": 11880 }, { "epoch": 3.0600952258396603, "grad_norm": 1.1287096738815308, "learning_rate": 1.9405405405405404e-05, "loss": 0.4038, "step": 11890 }, { "epoch": 3.0626688971818297, "grad_norm": 1.0460259914398193, "learning_rate": 1.937966537966538e-05, "loss": 0.3982, "step": 11900 }, { "epoch": 3.0652425685239995, "grad_norm": 1.1148087978363037, "learning_rate": 1.9353925353925358e-05, "loss": 0.4112, "step": 11910 }, { "epoch": 3.0678162398661692, "grad_norm": 1.347355604171753, "learning_rate": 1.932818532818533e-05, "loss": 0.4286, "step": 11920 }, { "epoch": 3.0703899112083386, "grad_norm": 1.4332010746002197, "learning_rate": 1.9302445302445304e-05, "loss": 0.3971, "step": 11930 }, { "epoch": 3.0729635825505084, "grad_norm": 2.455965042114258, "learning_rate": 1.927670527670528e-05, "loss": 0.4328, "step": 11940 }, { "epoch": 3.0755372538926777, "grad_norm": 1.319761037826538, "learning_rate": 1.9250965250965254e-05, "loss": 0.4166, "step": 11950 }, { "epoch": 3.0781109252348475, "grad_norm": 1.1820671558380127, "learning_rate": 1.9225225225225226e-05, "loss": 0.4594, "step": 11960 }, { "epoch": 3.0806845965770173, "grad_norm": 0.905479907989502, "learning_rate": 1.91994851994852e-05, "loss": 0.4466, "step": 11970 }, { "epoch": 3.0832582679191867, "grad_norm": 1.3796758651733398, "learning_rate": 1.9173745173745176e-05, "loss": 0.398, "step": 11980 }, { "epoch": 3.0858319392613565, "grad_norm": 1.7083042860031128, "learning_rate": 1.914800514800515e-05, "loss": 0.4554, "step": 11990 }, { "epoch": 3.088405610603526, "grad_norm": 1.0998810529708862, "learning_rate": 1.9122265122265122e-05, "loss": 0.4031, "step": 12000 }, { "epoch": 3.088405610603526, "eval_loss": 0.8223614692687988, "eval_runtime": 395.6102, "eval_samples_per_second": 49.104, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.00035795898845269097, "step": 12000 }, { "epoch": 3.0909792819456956, "grad_norm": 0.8669257760047913, "learning_rate": 1.9096525096525097e-05, "loss": 0.4385, "step": 12010 }, { "epoch": 3.093552953287865, "grad_norm": 1.6666041612625122, "learning_rate": 1.9070785070785072e-05, "loss": 0.4202, "step": 12020 }, { "epoch": 3.0961266246300347, "grad_norm": 1.2697246074676514, "learning_rate": 1.9045045045045047e-05, "loss": 0.4472, "step": 12030 }, { "epoch": 3.0987002959722045, "grad_norm": 1.0847759246826172, "learning_rate": 1.901930501930502e-05, "loss": 0.4642, "step": 12040 }, { "epoch": 3.101273967314374, "grad_norm": 2.803271532058716, "learning_rate": 1.8993564993564994e-05, "loss": 0.4555, "step": 12050 }, { "epoch": 3.1038476386565437, "grad_norm": 1.1774652004241943, "learning_rate": 1.896782496782497e-05, "loss": 0.472, "step": 12060 }, { "epoch": 3.106421309998713, "grad_norm": 3.2696588039398193, "learning_rate": 1.894208494208494e-05, "loss": 0.4239, "step": 12070 }, { "epoch": 3.108994981340883, "grad_norm": 1.0834957361221313, "learning_rate": 1.8916344916344916e-05, "loss": 0.3855, "step": 12080 }, { "epoch": 3.111568652683052, "grad_norm": 1.2990940809249878, "learning_rate": 1.889060489060489e-05, "loss": 0.4416, "step": 12090 }, { "epoch": 3.114142324025222, "grad_norm": 5.679697036743164, "learning_rate": 1.8864864864864866e-05, "loss": 0.4118, "step": 12100 }, { "epoch": 3.1167159953673917, "grad_norm": 1.1892821788787842, "learning_rate": 1.8839124839124837e-05, "loss": 0.3947, "step": 12110 }, { "epoch": 3.119289666709561, "grad_norm": 3.293264389038086, "learning_rate": 1.8813384813384816e-05, "loss": 0.4016, "step": 12120 }, { "epoch": 3.121863338051731, "grad_norm": 1.1677567958831787, "learning_rate": 1.878764478764479e-05, "loss": 0.4077, "step": 12130 }, { "epoch": 3.1244370093939002, "grad_norm": 1.0109992027282715, "learning_rate": 1.8761904761904766e-05, "loss": 0.3985, "step": 12140 }, { "epoch": 3.12701068073607, "grad_norm": 1.0192488431930542, "learning_rate": 1.8736164736164737e-05, "loss": 0.4088, "step": 12150 }, { "epoch": 3.12958435207824, "grad_norm": 2.916017532348633, "learning_rate": 1.8710424710424712e-05, "loss": 0.4372, "step": 12160 }, { "epoch": 3.132158023420409, "grad_norm": 1.1380913257598877, "learning_rate": 1.8684684684684687e-05, "loss": 0.4914, "step": 12170 }, { "epoch": 3.134731694762579, "grad_norm": 1.2204816341400146, "learning_rate": 1.865894465894466e-05, "loss": 0.3692, "step": 12180 }, { "epoch": 3.1373053661047483, "grad_norm": 5.1090803146362305, "learning_rate": 1.8633204633204634e-05, "loss": 0.4168, "step": 12190 }, { "epoch": 3.139879037446918, "grad_norm": 5.300877094268799, "learning_rate": 1.860746460746461e-05, "loss": 0.4533, "step": 12200 }, { "epoch": 3.1424527087890874, "grad_norm": 1.053403377532959, "learning_rate": 1.8581724581724584e-05, "loss": 0.4207, "step": 12210 }, { "epoch": 3.1450263801312572, "grad_norm": 0.8829357028007507, "learning_rate": 1.8555984555984555e-05, "loss": 0.4247, "step": 12220 }, { "epoch": 3.147600051473427, "grad_norm": 1.015411138534546, "learning_rate": 1.853024453024453e-05, "loss": 0.4154, "step": 12230 }, { "epoch": 3.1501737228155964, "grad_norm": 4.939092636108398, "learning_rate": 1.8504504504504505e-05, "loss": 0.3905, "step": 12240 }, { "epoch": 3.152747394157766, "grad_norm": 1.0195497274398804, "learning_rate": 1.847876447876448e-05, "loss": 0.4135, "step": 12250 }, { "epoch": 3.1553210654999355, "grad_norm": 1.3973890542984009, "learning_rate": 1.8453024453024452e-05, "loss": 0.4462, "step": 12260 }, { "epoch": 3.1578947368421053, "grad_norm": 1.9131966829299927, "learning_rate": 1.8427284427284427e-05, "loss": 0.4195, "step": 12270 }, { "epoch": 3.1604684081842747, "grad_norm": 1.4273961782455444, "learning_rate": 1.8401544401544402e-05, "loss": 0.4078, "step": 12280 }, { "epoch": 3.1630420795264445, "grad_norm": 0.850229024887085, "learning_rate": 1.8375804375804377e-05, "loss": 0.4249, "step": 12290 }, { "epoch": 3.1656157508686142, "grad_norm": 1.3435311317443848, "learning_rate": 1.835006435006435e-05, "loss": 0.4236, "step": 12300 }, { "epoch": 3.1681894222107836, "grad_norm": 1.1101402044296265, "learning_rate": 1.8324324324324324e-05, "loss": 0.4806, "step": 12310 }, { "epoch": 3.1707630935529534, "grad_norm": 1.9284107685089111, "learning_rate": 1.82985842985843e-05, "loss": 0.4059, "step": 12320 }, { "epoch": 3.1733367648951227, "grad_norm": 0.9383375644683838, "learning_rate": 1.8272844272844274e-05, "loss": 0.4593, "step": 12330 }, { "epoch": 3.1759104362372925, "grad_norm": 1.1090718507766724, "learning_rate": 1.824710424710425e-05, "loss": 0.451, "step": 12340 }, { "epoch": 3.178484107579462, "grad_norm": 3.4713733196258545, "learning_rate": 1.8221364221364224e-05, "loss": 0.4403, "step": 12350 }, { "epoch": 3.1810577789216317, "grad_norm": 1.001657485961914, "learning_rate": 1.81956241956242e-05, "loss": 0.4593, "step": 12360 }, { "epoch": 3.1836314502638015, "grad_norm": 0.9134547710418701, "learning_rate": 1.816988416988417e-05, "loss": 0.4513, "step": 12370 }, { "epoch": 3.186205121605971, "grad_norm": 1.105063557624817, "learning_rate": 1.8144144144144145e-05, "loss": 0.4476, "step": 12380 }, { "epoch": 3.1887787929481406, "grad_norm": 1.3801063299179077, "learning_rate": 1.811840411840412e-05, "loss": 0.4312, "step": 12390 }, { "epoch": 3.19135246429031, "grad_norm": 5.133249282836914, "learning_rate": 1.8092664092664095e-05, "loss": 0.4124, "step": 12400 }, { "epoch": 3.1939261356324797, "grad_norm": 1.1436964273452759, "learning_rate": 1.8066924066924067e-05, "loss": 0.4242, "step": 12410 }, { "epoch": 3.1964998069746495, "grad_norm": 0.8527979850769043, "learning_rate": 1.8041184041184042e-05, "loss": 0.4159, "step": 12420 }, { "epoch": 3.199073478316819, "grad_norm": 1.0988205671310425, "learning_rate": 1.8015444015444017e-05, "loss": 0.3943, "step": 12430 }, { "epoch": 3.2016471496589887, "grad_norm": 2.3774611949920654, "learning_rate": 1.7989703989703992e-05, "loss": 0.4405, "step": 12440 }, { "epoch": 3.204220821001158, "grad_norm": 1.3425320386886597, "learning_rate": 1.7963963963963963e-05, "loss": 0.4397, "step": 12450 }, { "epoch": 3.206794492343328, "grad_norm": 1.1551008224487305, "learning_rate": 1.793822393822394e-05, "loss": 0.4177, "step": 12460 }, { "epoch": 3.209368163685497, "grad_norm": 1.0739468336105347, "learning_rate": 1.7912483912483913e-05, "loss": 0.4226, "step": 12470 }, { "epoch": 3.211941835027667, "grad_norm": 1.012554407119751, "learning_rate": 1.788674388674389e-05, "loss": 0.4394, "step": 12480 }, { "epoch": 3.2145155063698367, "grad_norm": 1.0329439640045166, "learning_rate": 1.786100386100386e-05, "loss": 0.4219, "step": 12490 }, { "epoch": 3.217089177712006, "grad_norm": 1.340021014213562, "learning_rate": 1.7835263835263835e-05, "loss": 0.4431, "step": 12500 }, { "epoch": 3.219662849054176, "grad_norm": 0.9239644408226013, "learning_rate": 1.780952380952381e-05, "loss": 0.4128, "step": 12510 }, { "epoch": 3.2222365203963452, "grad_norm": 1.1054936647415161, "learning_rate": 1.7783783783783785e-05, "loss": 0.4215, "step": 12520 }, { "epoch": 3.224810191738515, "grad_norm": 4.010881423950195, "learning_rate": 1.7758043758043757e-05, "loss": 0.3813, "step": 12530 }, { "epoch": 3.227383863080685, "grad_norm": 1.6302077770233154, "learning_rate": 1.7732303732303735e-05, "loss": 0.427, "step": 12540 }, { "epoch": 3.229957534422854, "grad_norm": 4.041463851928711, "learning_rate": 1.770656370656371e-05, "loss": 0.4766, "step": 12550 }, { "epoch": 3.232531205765024, "grad_norm": 1.1331210136413574, "learning_rate": 1.768082368082368e-05, "loss": 0.4566, "step": 12560 }, { "epoch": 3.2351048771071933, "grad_norm": 1.1952770948410034, "learning_rate": 1.7655083655083657e-05, "loss": 0.4393, "step": 12570 }, { "epoch": 3.237678548449363, "grad_norm": 1.2376477718353271, "learning_rate": 1.762934362934363e-05, "loss": 0.4346, "step": 12580 }, { "epoch": 3.2402522197915324, "grad_norm": 0.931395947933197, "learning_rate": 1.7603603603603607e-05, "loss": 0.4355, "step": 12590 }, { "epoch": 3.2428258911337022, "grad_norm": 0.7813390493392944, "learning_rate": 1.7577863577863578e-05, "loss": 0.459, "step": 12600 }, { "epoch": 3.245399562475872, "grad_norm": 0.9730493426322937, "learning_rate": 1.7552123552123553e-05, "loss": 0.4541, "step": 12610 }, { "epoch": 3.2479732338180414, "grad_norm": 1.0856324434280396, "learning_rate": 1.7526383526383528e-05, "loss": 0.4046, "step": 12620 }, { "epoch": 3.250546905160211, "grad_norm": 1.2812646627426147, "learning_rate": 1.7500643500643503e-05, "loss": 0.4322, "step": 12630 }, { "epoch": 3.2531205765023805, "grad_norm": 1.2020498514175415, "learning_rate": 1.7474903474903475e-05, "loss": 0.4366, "step": 12640 }, { "epoch": 3.2556942478445503, "grad_norm": 1.462583065032959, "learning_rate": 1.744916344916345e-05, "loss": 0.3911, "step": 12650 }, { "epoch": 3.2582679191867197, "grad_norm": 0.9804076552391052, "learning_rate": 1.7423423423423425e-05, "loss": 0.501, "step": 12660 }, { "epoch": 3.2608415905288894, "grad_norm": 1.153205156326294, "learning_rate": 1.7397683397683396e-05, "loss": 0.4456, "step": 12670 }, { "epoch": 3.2634152618710592, "grad_norm": 0.8798128366470337, "learning_rate": 1.737194337194337e-05, "loss": 0.4622, "step": 12680 }, { "epoch": 3.2659889332132286, "grad_norm": 1.3517348766326904, "learning_rate": 1.7346203346203346e-05, "loss": 0.4705, "step": 12690 }, { "epoch": 3.2685626045553984, "grad_norm": 1.0517302751541138, "learning_rate": 1.732046332046332e-05, "loss": 0.502, "step": 12700 }, { "epoch": 3.2711362758975677, "grad_norm": 0.8883270621299744, "learning_rate": 1.7294723294723293e-05, "loss": 0.4055, "step": 12710 }, { "epoch": 3.2737099472397375, "grad_norm": 0.9611861109733582, "learning_rate": 1.7268983268983268e-05, "loss": 0.4462, "step": 12720 }, { "epoch": 3.276283618581907, "grad_norm": 0.9720549583435059, "learning_rate": 1.7243243243243243e-05, "loss": 0.4339, "step": 12730 }, { "epoch": 3.2788572899240767, "grad_norm": 1.8205825090408325, "learning_rate": 1.7217503217503218e-05, "loss": 0.4815, "step": 12740 }, { "epoch": 3.2814309612662464, "grad_norm": 1.316831111907959, "learning_rate": 1.7191763191763193e-05, "loss": 0.4655, "step": 12750 }, { "epoch": 3.284004632608416, "grad_norm": 1.3125663995742798, "learning_rate": 1.7166023166023168e-05, "loss": 0.4719, "step": 12760 }, { "epoch": 3.2865783039505856, "grad_norm": 1.2112400531768799, "learning_rate": 1.7140283140283143e-05, "loss": 0.4137, "step": 12770 }, { "epoch": 3.289151975292755, "grad_norm": 1.2232820987701416, "learning_rate": 1.7114543114543115e-05, "loss": 0.4392, "step": 12780 }, { "epoch": 3.2917256466349247, "grad_norm": 1.1535100936889648, "learning_rate": 1.708880308880309e-05, "loss": 0.4628, "step": 12790 }, { "epoch": 3.294299317977094, "grad_norm": 1.443505048751831, "learning_rate": 1.7063063063063065e-05, "loss": 0.4385, "step": 12800 }, { "epoch": 3.296872989319264, "grad_norm": 1.0268503427505493, "learning_rate": 1.703732303732304e-05, "loss": 0.4135, "step": 12810 }, { "epoch": 3.2994466606614337, "grad_norm": 0.8816016316413879, "learning_rate": 1.701158301158301e-05, "loss": 0.4897, "step": 12820 }, { "epoch": 3.302020332003603, "grad_norm": 1.4715735912322998, "learning_rate": 1.6985842985842986e-05, "loss": 0.4285, "step": 12830 }, { "epoch": 3.304594003345773, "grad_norm": 1.63077974319458, "learning_rate": 1.696010296010296e-05, "loss": 0.4841, "step": 12840 }, { "epoch": 3.3071676746879426, "grad_norm": 1.157626748085022, "learning_rate": 1.6934362934362936e-05, "loss": 0.4523, "step": 12850 }, { "epoch": 3.309741346030112, "grad_norm": 2.7212443351745605, "learning_rate": 1.6908622908622908e-05, "loss": 0.4134, "step": 12860 }, { "epoch": 3.3123150173722817, "grad_norm": 0.8014114499092102, "learning_rate": 1.6882882882882883e-05, "loss": 0.3669, "step": 12870 }, { "epoch": 3.314888688714451, "grad_norm": 1.2325034141540527, "learning_rate": 1.6857142857142858e-05, "loss": 0.4272, "step": 12880 }, { "epoch": 3.317462360056621, "grad_norm": 0.9247467517852783, "learning_rate": 1.6831402831402833e-05, "loss": 0.4294, "step": 12890 }, { "epoch": 3.32003603139879, "grad_norm": 0.9387899041175842, "learning_rate": 1.6805662805662805e-05, "loss": 0.4372, "step": 12900 }, { "epoch": 3.32260970274096, "grad_norm": 3.1787610054016113, "learning_rate": 1.677992277992278e-05, "loss": 0.4523, "step": 12910 }, { "epoch": 3.32518337408313, "grad_norm": 1.0653231143951416, "learning_rate": 1.6754182754182755e-05, "loss": 0.4646, "step": 12920 }, { "epoch": 3.327757045425299, "grad_norm": 1.1255438327789307, "learning_rate": 1.672844272844273e-05, "loss": 0.4399, "step": 12930 }, { "epoch": 3.330330716767469, "grad_norm": 1.021522879600525, "learning_rate": 1.67027027027027e-05, "loss": 0.4362, "step": 12940 }, { "epoch": 3.3329043881096383, "grad_norm": 1.3337011337280273, "learning_rate": 1.667696267696268e-05, "loss": 0.4598, "step": 12950 }, { "epoch": 3.335478059451808, "grad_norm": 1.0208321809768677, "learning_rate": 1.6651222651222654e-05, "loss": 0.4132, "step": 12960 }, { "epoch": 3.3380517307939774, "grad_norm": 1.224876046180725, "learning_rate": 1.6625482625482626e-05, "loss": 0.412, "step": 12970 }, { "epoch": 3.3406254021361472, "grad_norm": 1.0516496896743774, "learning_rate": 1.65997425997426e-05, "loss": 0.4059, "step": 12980 }, { "epoch": 3.343199073478317, "grad_norm": 1.2751177549362183, "learning_rate": 1.6574002574002576e-05, "loss": 0.4684, "step": 12990 }, { "epoch": 3.3457727448204864, "grad_norm": 1.3460466861724854, "learning_rate": 1.654826254826255e-05, "loss": 0.4497, "step": 13000 }, { "epoch": 3.3457727448204864, "eval_loss": 0.8037166595458984, "eval_runtime": 395.705, "eval_samples_per_second": 49.092, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.00035492222078819304, "step": 13000 }, { "epoch": 3.348346416162656, "grad_norm": 1.3930424451828003, "learning_rate": 1.6522522522522523e-05, "loss": 0.3934, "step": 13010 }, { "epoch": 3.3509200875048255, "grad_norm": 1.0718811750411987, "learning_rate": 1.6496782496782498e-05, "loss": 0.439, "step": 13020 }, { "epoch": 3.3534937588469953, "grad_norm": 0.9547449946403503, "learning_rate": 1.6471042471042473e-05, "loss": 0.4055, "step": 13030 }, { "epoch": 3.3560674301891646, "grad_norm": 1.0712281465530396, "learning_rate": 1.6445302445302448e-05, "loss": 0.4388, "step": 13040 }, { "epoch": 3.3586411015313344, "grad_norm": 1.256049394607544, "learning_rate": 1.641956241956242e-05, "loss": 0.5076, "step": 13050 }, { "epoch": 3.3612147728735042, "grad_norm": 0.8553959131240845, "learning_rate": 1.6393822393822394e-05, "loss": 0.4392, "step": 13060 }, { "epoch": 3.3637884442156736, "grad_norm": 0.9777474403381348, "learning_rate": 1.636808236808237e-05, "loss": 0.4033, "step": 13070 }, { "epoch": 3.3663621155578434, "grad_norm": 1.901464819908142, "learning_rate": 1.6342342342342344e-05, "loss": 0.446, "step": 13080 }, { "epoch": 3.3689357869000127, "grad_norm": 1.1458752155303955, "learning_rate": 1.6316602316602316e-05, "loss": 0.4282, "step": 13090 }, { "epoch": 3.3715094582421825, "grad_norm": 1.075851321220398, "learning_rate": 1.629086229086229e-05, "loss": 0.4591, "step": 13100 }, { "epoch": 3.374083129584352, "grad_norm": 1.1775199174880981, "learning_rate": 1.6265122265122266e-05, "loss": 0.4184, "step": 13110 }, { "epoch": 3.3766568009265217, "grad_norm": 1.3904964923858643, "learning_rate": 1.623938223938224e-05, "loss": 0.4494, "step": 13120 }, { "epoch": 3.3792304722686914, "grad_norm": 1.2566157579421997, "learning_rate": 1.6213642213642213e-05, "loss": 0.483, "step": 13130 }, { "epoch": 3.381804143610861, "grad_norm": 0.7696701288223267, "learning_rate": 1.6187902187902188e-05, "loss": 0.4492, "step": 13140 }, { "epoch": 3.3843778149530306, "grad_norm": 1.138867974281311, "learning_rate": 1.6162162162162163e-05, "loss": 0.448, "step": 13150 }, { "epoch": 3.3869514862952, "grad_norm": 1.0716248750686646, "learning_rate": 1.6136422136422138e-05, "loss": 0.403, "step": 13160 }, { "epoch": 3.3895251576373697, "grad_norm": 1.0408307313919067, "learning_rate": 1.6110682110682113e-05, "loss": 0.4505, "step": 13170 }, { "epoch": 3.392098828979539, "grad_norm": 1.176862120628357, "learning_rate": 1.6084942084942088e-05, "loss": 0.4501, "step": 13180 }, { "epoch": 3.394672500321709, "grad_norm": 1.18982994556427, "learning_rate": 1.6059202059202063e-05, "loss": 0.4152, "step": 13190 }, { "epoch": 3.3972461716638787, "grad_norm": 3.0527262687683105, "learning_rate": 1.6033462033462034e-05, "loss": 0.4094, "step": 13200 }, { "epoch": 3.399819843006048, "grad_norm": 1.14426589012146, "learning_rate": 1.600772200772201e-05, "loss": 0.4099, "step": 13210 }, { "epoch": 3.402393514348218, "grad_norm": 1.081701397895813, "learning_rate": 1.5981981981981984e-05, "loss": 0.428, "step": 13220 }, { "epoch": 3.404967185690387, "grad_norm": 1.6688580513000488, "learning_rate": 1.5956241956241956e-05, "loss": 0.5, "step": 13230 }, { "epoch": 3.407540857032557, "grad_norm": 1.1758854389190674, "learning_rate": 1.593050193050193e-05, "loss": 0.4582, "step": 13240 }, { "epoch": 3.4101145283747267, "grad_norm": 1.9258641004562378, "learning_rate": 1.5904761904761906e-05, "loss": 0.4227, "step": 13250 }, { "epoch": 3.412688199716896, "grad_norm": 0.9658583402633667, "learning_rate": 1.587902187902188e-05, "loss": 0.4391, "step": 13260 }, { "epoch": 3.415261871059066, "grad_norm": 1.136678695678711, "learning_rate": 1.5853281853281852e-05, "loss": 0.4561, "step": 13270 }, { "epoch": 3.417835542401235, "grad_norm": 1.336816668510437, "learning_rate": 1.5827541827541827e-05, "loss": 0.4605, "step": 13280 }, { "epoch": 3.420409213743405, "grad_norm": 1.2746632099151611, "learning_rate": 1.5801801801801802e-05, "loss": 0.4618, "step": 13290 }, { "epoch": 3.422982885085575, "grad_norm": 1.2792760133743286, "learning_rate": 1.5776061776061777e-05, "loss": 0.4101, "step": 13300 }, { "epoch": 3.425556556427744, "grad_norm": 1.0193665027618408, "learning_rate": 1.575032175032175e-05, "loss": 0.4173, "step": 13310 }, { "epoch": 3.428130227769914, "grad_norm": 1.210837721824646, "learning_rate": 1.5724581724581724e-05, "loss": 0.4466, "step": 13320 }, { "epoch": 3.4307038991120833, "grad_norm": 0.9081295132637024, "learning_rate": 1.56988416988417e-05, "loss": 0.4679, "step": 13330 }, { "epoch": 3.433277570454253, "grad_norm": 1.0284429788589478, "learning_rate": 1.5673101673101674e-05, "loss": 0.4082, "step": 13340 }, { "epoch": 3.4358512417964224, "grad_norm": 1.0147459506988525, "learning_rate": 1.5647361647361646e-05, "loss": 0.4096, "step": 13350 }, { "epoch": 3.438424913138592, "grad_norm": 3.9549002647399902, "learning_rate": 1.562162162162162e-05, "loss": 0.4907, "step": 13360 }, { "epoch": 3.440998584480762, "grad_norm": 1.1533993482589722, "learning_rate": 1.55958815958816e-05, "loss": 0.4478, "step": 13370 }, { "epoch": 3.4435722558229314, "grad_norm": 1.3518677949905396, "learning_rate": 1.557014157014157e-05, "loss": 0.4654, "step": 13380 }, { "epoch": 3.446145927165101, "grad_norm": 1.235627293586731, "learning_rate": 1.5544401544401546e-05, "loss": 0.4478, "step": 13390 }, { "epoch": 3.4487195985072705, "grad_norm": 1.4760185480117798, "learning_rate": 1.551866151866152e-05, "loss": 0.4625, "step": 13400 }, { "epoch": 3.4512932698494403, "grad_norm": 1.0546268224716187, "learning_rate": 1.5492921492921496e-05, "loss": 0.3887, "step": 13410 }, { "epoch": 3.4538669411916096, "grad_norm": 0.971731960773468, "learning_rate": 1.5467181467181467e-05, "loss": 0.4381, "step": 13420 }, { "epoch": 3.4564406125337794, "grad_norm": 1.2670224905014038, "learning_rate": 1.5441441441441442e-05, "loss": 0.5249, "step": 13430 }, { "epoch": 3.4590142838759492, "grad_norm": 1.4726381301879883, "learning_rate": 1.5415701415701417e-05, "loss": 0.4644, "step": 13440 }, { "epoch": 3.4615879552181186, "grad_norm": 0.9591072797775269, "learning_rate": 1.5389961389961392e-05, "loss": 0.4573, "step": 13450 }, { "epoch": 3.4641616265602884, "grad_norm": 1.0489442348480225, "learning_rate": 1.5364221364221364e-05, "loss": 0.4467, "step": 13460 }, { "epoch": 3.4667352979024577, "grad_norm": 1.2839806079864502, "learning_rate": 1.533848133848134e-05, "loss": 0.4844, "step": 13470 }, { "epoch": 3.4693089692446275, "grad_norm": 1.2316621541976929, "learning_rate": 1.5312741312741314e-05, "loss": 0.4506, "step": 13480 }, { "epoch": 3.471882640586797, "grad_norm": 1.1504662036895752, "learning_rate": 1.528700128700129e-05, "loss": 0.3969, "step": 13490 }, { "epoch": 3.4744563119289666, "grad_norm": 1.4414169788360596, "learning_rate": 1.526126126126126e-05, "loss": 0.433, "step": 13500 }, { "epoch": 3.4770299832711364, "grad_norm": 1.1712054014205933, "learning_rate": 1.5235521235521235e-05, "loss": 0.4622, "step": 13510 }, { "epoch": 3.479603654613306, "grad_norm": 0.9692360758781433, "learning_rate": 1.520978120978121e-05, "loss": 0.4323, "step": 13520 }, { "epoch": 3.4821773259554756, "grad_norm": 3.664949655532837, "learning_rate": 1.5184041184041184e-05, "loss": 0.4761, "step": 13530 }, { "epoch": 3.484750997297645, "grad_norm": 1.2114914655685425, "learning_rate": 1.5158301158301159e-05, "loss": 0.5089, "step": 13540 }, { "epoch": 3.4873246686398147, "grad_norm": 1.1861014366149902, "learning_rate": 1.5132561132561132e-05, "loss": 0.4537, "step": 13550 }, { "epoch": 3.489898339981984, "grad_norm": 1.2321648597717285, "learning_rate": 1.5106821106821107e-05, "loss": 0.5039, "step": 13560 }, { "epoch": 3.492472011324154, "grad_norm": 1.0023382902145386, "learning_rate": 1.508108108108108e-05, "loss": 0.4347, "step": 13570 }, { "epoch": 3.4950456826663236, "grad_norm": 1.5064911842346191, "learning_rate": 1.5055341055341057e-05, "loss": 0.4311, "step": 13580 }, { "epoch": 3.497619354008493, "grad_norm": 2.946744203567505, "learning_rate": 1.5029601029601032e-05, "loss": 0.4479, "step": 13590 }, { "epoch": 3.500193025350663, "grad_norm": 3.439570188522339, "learning_rate": 1.5003861003861005e-05, "loss": 0.4942, "step": 13600 }, { "epoch": 3.5027666966928326, "grad_norm": 1.228228211402893, "learning_rate": 1.497812097812098e-05, "loss": 0.4554, "step": 13610 }, { "epoch": 3.505340368035002, "grad_norm": 1.0095043182373047, "learning_rate": 1.4952380952380954e-05, "loss": 0.4988, "step": 13620 }, { "epoch": 3.5079140393771713, "grad_norm": 1.0728389024734497, "learning_rate": 1.4926640926640929e-05, "loss": 0.5008, "step": 13630 }, { "epoch": 3.510487710719341, "grad_norm": 1.4836064577102661, "learning_rate": 1.4900900900900902e-05, "loss": 0.4352, "step": 13640 }, { "epoch": 3.513061382061511, "grad_norm": 1.051338791847229, "learning_rate": 1.4875160875160877e-05, "loss": 0.4109, "step": 13650 }, { "epoch": 3.51563505340368, "grad_norm": 1.345146656036377, "learning_rate": 1.484942084942085e-05, "loss": 0.4174, "step": 13660 }, { "epoch": 3.51820872474585, "grad_norm": 1.2961459159851074, "learning_rate": 1.4823680823680825e-05, "loss": 0.474, "step": 13670 }, { "epoch": 3.52078239608802, "grad_norm": 1.370326042175293, "learning_rate": 1.4797940797940799e-05, "loss": 0.4244, "step": 13680 }, { "epoch": 3.523356067430189, "grad_norm": 3.830646514892578, "learning_rate": 1.4772200772200774e-05, "loss": 0.4581, "step": 13690 }, { "epoch": 3.525929738772359, "grad_norm": 1.1295220851898193, "learning_rate": 1.4746460746460747e-05, "loss": 0.435, "step": 13700 }, { "epoch": 3.5285034101145283, "grad_norm": 1.1341050863265991, "learning_rate": 1.472072072072072e-05, "loss": 0.4484, "step": 13710 }, { "epoch": 3.531077081456698, "grad_norm": 1.2032519578933716, "learning_rate": 1.4694980694980695e-05, "loss": 0.4368, "step": 13720 }, { "epoch": 3.5336507527988674, "grad_norm": 2.008488655090332, "learning_rate": 1.4669240669240668e-05, "loss": 0.3941, "step": 13730 }, { "epoch": 3.536224424141037, "grad_norm": 1.0747300386428833, "learning_rate": 1.4643500643500643e-05, "loss": 0.4884, "step": 13740 }, { "epoch": 3.538798095483207, "grad_norm": 2.756009578704834, "learning_rate": 1.4617760617760617e-05, "loss": 0.4298, "step": 13750 }, { "epoch": 3.5413717668253764, "grad_norm": 0.8556985259056091, "learning_rate": 1.4592020592020592e-05, "loss": 0.4827, "step": 13760 }, { "epoch": 3.543945438167546, "grad_norm": 1.360617995262146, "learning_rate": 1.4566280566280565e-05, "loss": 0.4264, "step": 13770 }, { "epoch": 3.5465191095097155, "grad_norm": 1.2899786233901978, "learning_rate": 1.4540540540540543e-05, "loss": 0.4293, "step": 13780 }, { "epoch": 3.5490927808518853, "grad_norm": 1.2246886491775513, "learning_rate": 1.4514800514800517e-05, "loss": 0.4131, "step": 13790 }, { "epoch": 3.5516664521940546, "grad_norm": 1.1932190656661987, "learning_rate": 1.448906048906049e-05, "loss": 0.476, "step": 13800 }, { "epoch": 3.5542401235362244, "grad_norm": 4.828482151031494, "learning_rate": 1.4463320463320465e-05, "loss": 0.465, "step": 13810 }, { "epoch": 3.556813794878394, "grad_norm": 1.4605960845947266, "learning_rate": 1.4437580437580438e-05, "loss": 0.4007, "step": 13820 }, { "epoch": 3.5593874662205636, "grad_norm": 0.9224157929420471, "learning_rate": 1.4411840411840413e-05, "loss": 0.4044, "step": 13830 }, { "epoch": 3.5619611375627334, "grad_norm": 1.017256498336792, "learning_rate": 1.4386100386100387e-05, "loss": 0.4502, "step": 13840 }, { "epoch": 3.5645348089049027, "grad_norm": 1.9395209550857544, "learning_rate": 1.4360360360360362e-05, "loss": 0.3843, "step": 13850 }, { "epoch": 3.5671084802470725, "grad_norm": 1.1960591077804565, "learning_rate": 1.4334620334620335e-05, "loss": 0.4528, "step": 13860 }, { "epoch": 3.569682151589242, "grad_norm": 0.9504403471946716, "learning_rate": 1.430888030888031e-05, "loss": 0.4412, "step": 13870 }, { "epoch": 3.5722558229314116, "grad_norm": 1.2557878494262695, "learning_rate": 1.4283140283140283e-05, "loss": 0.3845, "step": 13880 }, { "epoch": 3.5748294942735814, "grad_norm": 4.165874004364014, "learning_rate": 1.4257400257400258e-05, "loss": 0.3932, "step": 13890 }, { "epoch": 3.577403165615751, "grad_norm": 3.318126678466797, "learning_rate": 1.4231660231660232e-05, "loss": 0.3809, "step": 13900 }, { "epoch": 3.5799768369579206, "grad_norm": 1.0943117141723633, "learning_rate": 1.4205920205920207e-05, "loss": 0.48, "step": 13910 }, { "epoch": 3.58255050830009, "grad_norm": 1.2082642316818237, "learning_rate": 1.418018018018018e-05, "loss": 0.4894, "step": 13920 }, { "epoch": 3.5851241796422597, "grad_norm": 0.6787092089653015, "learning_rate": 1.4154440154440155e-05, "loss": 0.4201, "step": 13930 }, { "epoch": 3.587697850984429, "grad_norm": 1.3796510696411133, "learning_rate": 1.4128700128700128e-05, "loss": 0.5031, "step": 13940 }, { "epoch": 3.590271522326599, "grad_norm": 0.9589558839797974, "learning_rate": 1.4102960102960103e-05, "loss": 0.4299, "step": 13950 }, { "epoch": 3.5928451936687686, "grad_norm": 1.0614103078842163, "learning_rate": 1.4077220077220076e-05, "loss": 0.4265, "step": 13960 }, { "epoch": 3.595418865010938, "grad_norm": 1.4390029907226562, "learning_rate": 1.4051480051480051e-05, "loss": 0.4648, "step": 13970 }, { "epoch": 3.597992536353108, "grad_norm": 4.360158920288086, "learning_rate": 1.4025740025740025e-05, "loss": 0.4575, "step": 13980 }, { "epoch": 3.6005662076952776, "grad_norm": 2.3131937980651855, "learning_rate": 1.4000000000000001e-05, "loss": 0.4087, "step": 13990 }, { "epoch": 3.603139879037447, "grad_norm": 1.0350521802902222, "learning_rate": 1.3974259974259976e-05, "loss": 0.4096, "step": 14000 }, { "epoch": 3.603139879037447, "eval_loss": 0.8108994960784912, "eval_runtime": 395.6175, "eval_samples_per_second": 49.103, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.0003543528268510997, "step": 14000 }, { "epoch": 3.6057135503796163, "grad_norm": 1.2455912828445435, "learning_rate": 1.394851994851995e-05, "loss": 0.4473, "step": 14010 }, { "epoch": 3.608287221721786, "grad_norm": 0.8037226796150208, "learning_rate": 1.3922779922779925e-05, "loss": 0.4257, "step": 14020 }, { "epoch": 3.610860893063956, "grad_norm": 2.2132887840270996, "learning_rate": 1.3897039897039898e-05, "loss": 0.4057, "step": 14030 }, { "epoch": 3.613434564406125, "grad_norm": 1.2261261940002441, "learning_rate": 1.3871299871299873e-05, "loss": 0.4251, "step": 14040 }, { "epoch": 3.616008235748295, "grad_norm": 0.9967421293258667, "learning_rate": 1.3845559845559846e-05, "loss": 0.4117, "step": 14050 }, { "epoch": 3.618581907090465, "grad_norm": 1.0189718008041382, "learning_rate": 1.3819819819819821e-05, "loss": 0.4292, "step": 14060 }, { "epoch": 3.621155578432634, "grad_norm": 1.333105206489563, "learning_rate": 1.3794079794079795e-05, "loss": 0.4989, "step": 14070 }, { "epoch": 3.6237292497748035, "grad_norm": 3.6797611713409424, "learning_rate": 1.376833976833977e-05, "loss": 0.4377, "step": 14080 }, { "epoch": 3.6263029211169733, "grad_norm": 2.8054308891296387, "learning_rate": 1.3742599742599743e-05, "loss": 0.4932, "step": 14090 }, { "epoch": 3.628876592459143, "grad_norm": 2.897303819656372, "learning_rate": 1.3716859716859718e-05, "loss": 0.4816, "step": 14100 }, { "epoch": 3.6314502638013124, "grad_norm": 1.0359746217727661, "learning_rate": 1.3691119691119691e-05, "loss": 0.4421, "step": 14110 }, { "epoch": 3.634023935143482, "grad_norm": 1.1515438556671143, "learning_rate": 1.3665379665379666e-05, "loss": 0.4766, "step": 14120 }, { "epoch": 3.636597606485652, "grad_norm": 1.0731172561645508, "learning_rate": 1.363963963963964e-05, "loss": 0.4578, "step": 14130 }, { "epoch": 3.6391712778278213, "grad_norm": 3.8715970516204834, "learning_rate": 1.3613899613899615e-05, "loss": 0.4549, "step": 14140 }, { "epoch": 3.641744949169991, "grad_norm": 1.386406660079956, "learning_rate": 1.3588159588159588e-05, "loss": 0.5023, "step": 14150 }, { "epoch": 3.6443186205121605, "grad_norm": 1.2222354412078857, "learning_rate": 1.3562419562419563e-05, "loss": 0.4375, "step": 14160 }, { "epoch": 3.6468922918543303, "grad_norm": 1.2102971076965332, "learning_rate": 1.3536679536679536e-05, "loss": 0.4404, "step": 14170 }, { "epoch": 3.6494659631964996, "grad_norm": 0.73658686876297, "learning_rate": 1.3510939510939511e-05, "loss": 0.4004, "step": 14180 }, { "epoch": 3.6520396345386694, "grad_norm": 1.3987421989440918, "learning_rate": 1.3485199485199485e-05, "loss": 0.432, "step": 14190 }, { "epoch": 3.654613305880839, "grad_norm": 1.2989593744277954, "learning_rate": 1.3459459459459461e-05, "loss": 0.4684, "step": 14200 }, { "epoch": 3.6571869772230086, "grad_norm": 1.5131486654281616, "learning_rate": 1.3433719433719436e-05, "loss": 0.5014, "step": 14210 }, { "epoch": 3.6597606485651784, "grad_norm": 0.9460519552230835, "learning_rate": 1.340797940797941e-05, "loss": 0.4686, "step": 14220 }, { "epoch": 3.6623343199073477, "grad_norm": 1.3320332765579224, "learning_rate": 1.3382239382239385e-05, "loss": 0.4612, "step": 14230 }, { "epoch": 3.6649079912495175, "grad_norm": 1.110107421875, "learning_rate": 1.3356499356499358e-05, "loss": 0.4824, "step": 14240 }, { "epoch": 3.667481662591687, "grad_norm": 1.6514250040054321, "learning_rate": 1.3330759330759333e-05, "loss": 0.4678, "step": 14250 }, { "epoch": 3.6700553339338566, "grad_norm": 2.858354330062866, "learning_rate": 1.3305019305019306e-05, "loss": 0.4576, "step": 14260 }, { "epoch": 3.6726290052760264, "grad_norm": 1.5715115070343018, "learning_rate": 1.327927927927928e-05, "loss": 0.4547, "step": 14270 }, { "epoch": 3.6752026766181958, "grad_norm": 0.938371479511261, "learning_rate": 1.3253539253539254e-05, "loss": 0.3987, "step": 14280 }, { "epoch": 3.6777763479603656, "grad_norm": 0.9989467859268188, "learning_rate": 1.3227799227799228e-05, "loss": 0.4403, "step": 14290 }, { "epoch": 3.680350019302535, "grad_norm": 1.117611289024353, "learning_rate": 1.3202059202059203e-05, "loss": 0.4009, "step": 14300 }, { "epoch": 3.6829236906447047, "grad_norm": 1.238559603691101, "learning_rate": 1.3176319176319176e-05, "loss": 0.4198, "step": 14310 }, { "epoch": 3.685497361986874, "grad_norm": 1.2747526168823242, "learning_rate": 1.3150579150579151e-05, "loss": 0.439, "step": 14320 }, { "epoch": 3.688071033329044, "grad_norm": 3.8777308464050293, "learning_rate": 1.3124839124839124e-05, "loss": 0.4573, "step": 14330 }, { "epoch": 3.6906447046712136, "grad_norm": 1.1357238292694092, "learning_rate": 1.30990990990991e-05, "loss": 0.4578, "step": 14340 }, { "epoch": 3.693218376013383, "grad_norm": 1.4202942848205566, "learning_rate": 1.3073359073359073e-05, "loss": 0.4409, "step": 14350 }, { "epoch": 3.695792047355553, "grad_norm": 1.4478346109390259, "learning_rate": 1.3047619047619048e-05, "loss": 0.4707, "step": 14360 }, { "epoch": 3.6983657186977226, "grad_norm": 1.4722789525985718, "learning_rate": 1.3021879021879021e-05, "loss": 0.4445, "step": 14370 }, { "epoch": 3.700939390039892, "grad_norm": 1.0790144205093384, "learning_rate": 1.2996138996138996e-05, "loss": 0.454, "step": 14380 }, { "epoch": 3.7035130613820613, "grad_norm": 2.759157419204712, "learning_rate": 1.297039897039897e-05, "loss": 0.3881, "step": 14390 }, { "epoch": 3.706086732724231, "grad_norm": 1.4952332973480225, "learning_rate": 1.2944658944658944e-05, "loss": 0.4109, "step": 14400 }, { "epoch": 3.708660404066401, "grad_norm": 1.024506688117981, "learning_rate": 1.2918918918918921e-05, "loss": 0.4616, "step": 14410 }, { "epoch": 3.71123407540857, "grad_norm": 5.463751316070557, "learning_rate": 1.2893178893178894e-05, "loss": 0.4293, "step": 14420 }, { "epoch": 3.71380774675074, "grad_norm": 1.2399414777755737, "learning_rate": 1.286743886743887e-05, "loss": 0.4671, "step": 14430 }, { "epoch": 3.71638141809291, "grad_norm": 1.1366684436798096, "learning_rate": 1.2841698841698843e-05, "loss": 0.3873, "step": 14440 }, { "epoch": 3.718955089435079, "grad_norm": 1.025657296180725, "learning_rate": 1.2815958815958818e-05, "loss": 0.4927, "step": 14450 }, { "epoch": 3.7215287607772485, "grad_norm": 1.4381457567214966, "learning_rate": 1.2790218790218791e-05, "loss": 0.399, "step": 14460 }, { "epoch": 3.7241024321194183, "grad_norm": 4.138847351074219, "learning_rate": 1.2764478764478766e-05, "loss": 0.493, "step": 14470 }, { "epoch": 3.726676103461588, "grad_norm": 3.5278961658477783, "learning_rate": 1.273873873873874e-05, "loss": 0.4522, "step": 14480 }, { "epoch": 3.7292497748037574, "grad_norm": 1.029931902885437, "learning_rate": 1.2712998712998714e-05, "loss": 0.3838, "step": 14490 }, { "epoch": 3.731823446145927, "grad_norm": 0.8327284455299377, "learning_rate": 1.268983268983269e-05, "loss": 0.4587, "step": 14500 }, { "epoch": 3.734397117488097, "grad_norm": 2.9695467948913574, "learning_rate": 1.2664092664092665e-05, "loss": 0.4544, "step": 14510 }, { "epoch": 3.7369707888302663, "grad_norm": 1.0291829109191895, "learning_rate": 1.2638352638352638e-05, "loss": 0.4691, "step": 14520 }, { "epoch": 3.739544460172436, "grad_norm": 1.1941965818405151, "learning_rate": 1.2612612612612611e-05, "loss": 0.4585, "step": 14530 }, { "epoch": 3.7421181315146055, "grad_norm": 1.7099372148513794, "learning_rate": 1.2586872586872586e-05, "loss": 0.4164, "step": 14540 }, { "epoch": 3.7446918028567753, "grad_norm": 2.46553373336792, "learning_rate": 1.2561132561132563e-05, "loss": 0.4529, "step": 14550 }, { "epoch": 3.7472654741989446, "grad_norm": 1.0007997751235962, "learning_rate": 1.2535392535392538e-05, "loss": 0.403, "step": 14560 }, { "epoch": 3.7498391455411144, "grad_norm": 1.1691094636917114, "learning_rate": 1.2509652509652511e-05, "loss": 0.4435, "step": 14570 }, { "epoch": 3.752412816883284, "grad_norm": 1.2735754251480103, "learning_rate": 1.2483912483912485e-05, "loss": 0.5126, "step": 14580 }, { "epoch": 3.7549864882254536, "grad_norm": 3.02443265914917, "learning_rate": 1.2458172458172458e-05, "loss": 0.4898, "step": 14590 }, { "epoch": 3.7575601595676233, "grad_norm": 1.1990752220153809, "learning_rate": 1.2432432432432435e-05, "loss": 0.4059, "step": 14600 }, { "epoch": 3.7601338309097927, "grad_norm": 1.180837631225586, "learning_rate": 1.2406692406692408e-05, "loss": 0.4384, "step": 14610 }, { "epoch": 3.7627075022519625, "grad_norm": 0.8962213397026062, "learning_rate": 1.2380952380952381e-05, "loss": 0.4297, "step": 14620 }, { "epoch": 3.765281173594132, "grad_norm": 0.9834825396537781, "learning_rate": 1.2355212355212356e-05, "loss": 0.4057, "step": 14630 }, { "epoch": 3.7678548449363016, "grad_norm": 1.174039363861084, "learning_rate": 1.232947232947233e-05, "loss": 0.4586, "step": 14640 }, { "epoch": 3.7704285162784714, "grad_norm": 0.9779637455940247, "learning_rate": 1.2303732303732304e-05, "loss": 0.4011, "step": 14650 }, { "epoch": 3.7730021876206408, "grad_norm": 1.003485918045044, "learning_rate": 1.2277992277992278e-05, "loss": 0.4899, "step": 14660 }, { "epoch": 3.7755758589628106, "grad_norm": 1.3274792432785034, "learning_rate": 1.2252252252252253e-05, "loss": 0.4584, "step": 14670 }, { "epoch": 3.77814953030498, "grad_norm": 1.5608062744140625, "learning_rate": 1.2226512226512226e-05, "loss": 0.5178, "step": 14680 }, { "epoch": 3.7807232016471497, "grad_norm": 3.1873466968536377, "learning_rate": 1.2200772200772201e-05, "loss": 0.448, "step": 14690 }, { "epoch": 3.783296872989319, "grad_norm": 0.8975459337234497, "learning_rate": 1.2175032175032174e-05, "loss": 0.3988, "step": 14700 }, { "epoch": 3.785870544331489, "grad_norm": 1.3727668523788452, "learning_rate": 1.2149292149292151e-05, "loss": 0.4206, "step": 14710 }, { "epoch": 3.7884442156736586, "grad_norm": 1.0184861421585083, "learning_rate": 1.2123552123552124e-05, "loss": 0.407, "step": 14720 }, { "epoch": 3.791017887015828, "grad_norm": 0.9267153739929199, "learning_rate": 1.20978120978121e-05, "loss": 0.4058, "step": 14730 }, { "epoch": 3.7935915583579978, "grad_norm": 1.0847712755203247, "learning_rate": 1.2072072072072073e-05, "loss": 0.428, "step": 14740 }, { "epoch": 3.7961652297001676, "grad_norm": 1.368102788925171, "learning_rate": 1.2046332046332048e-05, "loss": 0.4002, "step": 14750 }, { "epoch": 3.798738901042337, "grad_norm": 1.2423478364944458, "learning_rate": 1.2020592020592021e-05, "loss": 0.4696, "step": 14760 }, { "epoch": 3.8013125723845063, "grad_norm": 1.3219197988510132, "learning_rate": 1.1994851994851996e-05, "loss": 0.4566, "step": 14770 }, { "epoch": 3.803886243726676, "grad_norm": 0.9948756098747253, "learning_rate": 1.196911196911197e-05, "loss": 0.4422, "step": 14780 }, { "epoch": 3.806459915068846, "grad_norm": 1.5322728157043457, "learning_rate": 1.1943371943371944e-05, "loss": 0.4079, "step": 14790 }, { "epoch": 3.809033586411015, "grad_norm": 1.384191870689392, "learning_rate": 1.1917631917631918e-05, "loss": 0.4617, "step": 14800 }, { "epoch": 3.811607257753185, "grad_norm": 1.3496747016906738, "learning_rate": 1.1891891891891893e-05, "loss": 0.4324, "step": 14810 }, { "epoch": 3.8141809290953548, "grad_norm": 1.0405890941619873, "learning_rate": 1.1866151866151868e-05, "loss": 0.3989, "step": 14820 }, { "epoch": 3.816754600437524, "grad_norm": 0.8725833296775818, "learning_rate": 1.1840411840411841e-05, "loss": 0.4727, "step": 14830 }, { "epoch": 3.8193282717796935, "grad_norm": 1.375089406967163, "learning_rate": 1.1814671814671816e-05, "loss": 0.5229, "step": 14840 }, { "epoch": 3.8219019431218633, "grad_norm": 1.0093263387680054, "learning_rate": 1.178893178893179e-05, "loss": 0.4561, "step": 14850 }, { "epoch": 3.824475614464033, "grad_norm": 1.45624577999115, "learning_rate": 1.1763191763191764e-05, "loss": 0.4518, "step": 14860 }, { "epoch": 3.8270492858062024, "grad_norm": 0.9073015451431274, "learning_rate": 1.1737451737451738e-05, "loss": 0.4404, "step": 14870 }, { "epoch": 3.829622957148372, "grad_norm": 1.0322130918502808, "learning_rate": 1.1711711711711713e-05, "loss": 0.4328, "step": 14880 }, { "epoch": 3.832196628490542, "grad_norm": 1.549871563911438, "learning_rate": 1.1685971685971686e-05, "loss": 0.4882, "step": 14890 }, { "epoch": 3.8347702998327113, "grad_norm": 0.8814623355865479, "learning_rate": 1.166023166023166e-05, "loss": 0.4859, "step": 14900 }, { "epoch": 3.837343971174881, "grad_norm": 1.2014186382293701, "learning_rate": 1.1634491634491634e-05, "loss": 0.4584, "step": 14910 }, { "epoch": 3.8399176425170505, "grad_norm": 1.3947405815124512, "learning_rate": 1.1608751608751609e-05, "loss": 0.4389, "step": 14920 }, { "epoch": 3.8424913138592203, "grad_norm": 1.3492847681045532, "learning_rate": 1.1583011583011584e-05, "loss": 0.4413, "step": 14930 }, { "epoch": 3.8450649852013896, "grad_norm": 0.9952030777931213, "learning_rate": 1.1557271557271557e-05, "loss": 0.4203, "step": 14940 }, { "epoch": 3.8476386565435594, "grad_norm": 1.1566764116287231, "learning_rate": 1.1531531531531532e-05, "loss": 0.4489, "step": 14950 }, { "epoch": 3.850212327885729, "grad_norm": 1.0022921562194824, "learning_rate": 1.1505791505791506e-05, "loss": 0.4422, "step": 14960 }, { "epoch": 3.8527859992278986, "grad_norm": 1.1566778421401978, "learning_rate": 1.148005148005148e-05, "loss": 0.4089, "step": 14970 }, { "epoch": 3.8553596705700683, "grad_norm": 1.3538066148757935, "learning_rate": 1.1454311454311454e-05, "loss": 0.4617, "step": 14980 }, { "epoch": 3.8579333419122377, "grad_norm": 1.1044282913208008, "learning_rate": 1.1428571428571429e-05, "loss": 0.4734, "step": 14990 }, { "epoch": 3.8605070132544075, "grad_norm": 1.195326805114746, "learning_rate": 1.1402831402831402e-05, "loss": 0.4171, "step": 15000 }, { "epoch": 3.8605070132544075, "eval_loss": 0.8098340630531311, "eval_runtime": 395.6098, "eval_samples_per_second": 49.104, "eval_steps_per_second": 2.457, "eval_token_accuracy": 0.0003602365642010644, "step": 15000 }, { "epoch": 3.863080684596577, "grad_norm": 3.7595245838165283, "learning_rate": 1.1377091377091377e-05, "loss": 0.4476, "step": 15010 }, { "epoch": 3.8656543559387466, "grad_norm": 1.0867187976837158, "learning_rate": 1.1351351351351352e-05, "loss": 0.4016, "step": 15020 }, { "epoch": 3.8682280272809164, "grad_norm": 1.2275842428207397, "learning_rate": 1.1325611325611327e-05, "loss": 0.4635, "step": 15030 }, { "epoch": 3.8708016986230858, "grad_norm": 2.140843629837036, "learning_rate": 1.12998712998713e-05, "loss": 0.4431, "step": 15040 }, { "epoch": 3.8733753699652556, "grad_norm": 5.140436172485352, "learning_rate": 1.1274131274131276e-05, "loss": 0.5156, "step": 15050 }, { "epoch": 3.875949041307425, "grad_norm": 1.347815990447998, "learning_rate": 1.1248391248391249e-05, "loss": 0.4836, "step": 15060 }, { "epoch": 3.8785227126495947, "grad_norm": 0.9270965456962585, "learning_rate": 1.1222651222651224e-05, "loss": 0.4724, "step": 15070 }, { "epoch": 3.881096383991764, "grad_norm": 3.1542866230010986, "learning_rate": 1.1196911196911197e-05, "loss": 0.4842, "step": 15080 }, { "epoch": 3.883670055333934, "grad_norm": 3.4730048179626465, "learning_rate": 1.1171171171171172e-05, "loss": 0.4559, "step": 15090 }, { "epoch": 3.8862437266761036, "grad_norm": 1.7973135709762573, "learning_rate": 1.1145431145431146e-05, "loss": 0.4053, "step": 15100 }, { "epoch": 3.888817398018273, "grad_norm": 1.1773796081542969, "learning_rate": 1.1119691119691119e-05, "loss": 0.3808, "step": 15110 }, { "epoch": 3.8913910693604428, "grad_norm": 2.933912754058838, "learning_rate": 1.1093951093951096e-05, "loss": 0.4514, "step": 15120 }, { "epoch": 3.893964740702612, "grad_norm": 1.0887664556503296, "learning_rate": 1.1068211068211069e-05, "loss": 0.4728, "step": 15130 }, { "epoch": 3.896538412044782, "grad_norm": 1.090585708618164, "learning_rate": 1.1042471042471044e-05, "loss": 0.45, "step": 15140 }, { "epoch": 3.8991120833869513, "grad_norm": 2.140028238296509, "learning_rate": 1.1016731016731017e-05, "loss": 0.4672, "step": 15150 }, { "epoch": 3.901685754729121, "grad_norm": 0.912236213684082, "learning_rate": 1.0990990990990992e-05, "loss": 0.4462, "step": 15160 }, { "epoch": 3.904259426071291, "grad_norm": 2.3811190128326416, "learning_rate": 1.0965250965250965e-05, "loss": 0.446, "step": 15170 }, { "epoch": 3.90683309741346, "grad_norm": 1.1356325149536133, "learning_rate": 1.093951093951094e-05, "loss": 0.4317, "step": 15180 }, { "epoch": 3.90940676875563, "grad_norm": 1.289565920829773, "learning_rate": 1.0913770913770914e-05, "loss": 0.4078, "step": 15190 }, { "epoch": 3.9119804400977998, "grad_norm": 0.9899163842201233, "learning_rate": 1.0888030888030889e-05, "loss": 0.4366, "step": 15200 }, { "epoch": 3.914554111439969, "grad_norm": 1.4149761199951172, "learning_rate": 1.0862290862290862e-05, "loss": 0.5163, "step": 15210 }, { "epoch": 3.9171277827821385, "grad_norm": 1.1276021003723145, "learning_rate": 1.0836550836550837e-05, "loss": 0.4242, "step": 15220 }, { "epoch": 3.9197014541243083, "grad_norm": 1.067687749862671, "learning_rate": 1.0810810810810812e-05, "loss": 0.4305, "step": 15230 }, { "epoch": 3.922275125466478, "grad_norm": 2.6914680004119873, "learning_rate": 1.0785070785070785e-05, "loss": 0.3787, "step": 15240 }, { "epoch": 3.9248487968086474, "grad_norm": 1.443280816078186, "learning_rate": 1.075933075933076e-05, "loss": 0.4043, "step": 15250 }, { "epoch": 3.927422468150817, "grad_norm": 3.6616697311401367, "learning_rate": 1.0733590733590734e-05, "loss": 0.4174, "step": 15260 }, { "epoch": 3.929996139492987, "grad_norm": 1.0795118808746338, "learning_rate": 1.0707850707850709e-05, "loss": 0.443, "step": 15270 }, { "epoch": 3.9325698108351563, "grad_norm": 1.2405767440795898, "learning_rate": 1.0682110682110682e-05, "loss": 0.401, "step": 15280 }, { "epoch": 3.9351434821773257, "grad_norm": 1.446061611175537, "learning_rate": 1.0656370656370657e-05, "loss": 0.4593, "step": 15290 }, { "epoch": 3.9377171535194955, "grad_norm": 4.733116626739502, "learning_rate": 1.063063063063063e-05, "loss": 0.4237, "step": 15300 }, { "epoch": 3.9402908248616653, "grad_norm": 1.0354984998703003, "learning_rate": 1.0604890604890605e-05, "loss": 0.4924, "step": 15310 }, { "epoch": 3.9428644962038346, "grad_norm": 4.033681392669678, "learning_rate": 1.0579150579150579e-05, "loss": 0.4451, "step": 15320 }, { "epoch": 3.9454381675460044, "grad_norm": 1.3079330921173096, "learning_rate": 1.0553410553410555e-05, "loss": 0.4257, "step": 15330 }, { "epoch": 3.948011838888174, "grad_norm": 5.547299385070801, "learning_rate": 1.0527670527670529e-05, "loss": 0.4284, "step": 15340 }, { "epoch": 3.9505855102303435, "grad_norm": 1.283423900604248, "learning_rate": 1.0501930501930504e-05, "loss": 0.4863, "step": 15350 }, { "epoch": 3.9531591815725133, "grad_norm": 1.2620773315429688, "learning_rate": 1.0476190476190477e-05, "loss": 0.4162, "step": 15360 }, { "epoch": 3.9557328529146827, "grad_norm": 1.2353228330612183, "learning_rate": 1.0450450450450452e-05, "loss": 0.4916, "step": 15370 }, { "epoch": 3.9583065242568525, "grad_norm": 1.3218683004379272, "learning_rate": 1.0424710424710425e-05, "loss": 0.3846, "step": 15380 }, { "epoch": 3.960880195599022, "grad_norm": 0.9560413956642151, "learning_rate": 1.0398970398970398e-05, "loss": 0.4208, "step": 15390 }, { "epoch": 3.9634538669411916, "grad_norm": 1.0743191242218018, "learning_rate": 1.0373230373230373e-05, "loss": 0.4151, "step": 15400 }, { "epoch": 3.9660275382833614, "grad_norm": 2.8277199268341064, "learning_rate": 1.0347490347490347e-05, "loss": 0.4162, "step": 15410 }, { "epoch": 3.9686012096255308, "grad_norm": 1.490350604057312, "learning_rate": 1.0321750321750322e-05, "loss": 0.486, "step": 15420 }, { "epoch": 3.9711748809677005, "grad_norm": 1.0201480388641357, "learning_rate": 1.0296010296010295e-05, "loss": 0.4567, "step": 15430 }, { "epoch": 3.97374855230987, "grad_norm": 0.8804197311401367, "learning_rate": 1.0270270270270272e-05, "loss": 0.4716, "step": 15440 }, { "epoch": 3.9763222236520397, "grad_norm": 3.7693941593170166, "learning_rate": 1.0244530244530245e-05, "loss": 0.478, "step": 15450 }, { "epoch": 3.978895894994209, "grad_norm": 5.499643325805664, "learning_rate": 1.021879021879022e-05, "loss": 0.4223, "step": 15460 }, { "epoch": 3.981469566336379, "grad_norm": 0.8914560675621033, "learning_rate": 1.0193050193050193e-05, "loss": 0.4079, "step": 15470 }, { "epoch": 3.9840432376785486, "grad_norm": 1.4804257154464722, "learning_rate": 1.0167310167310168e-05, "loss": 0.4416, "step": 15480 }, { "epoch": 3.986616909020718, "grad_norm": 1.3806978464126587, "learning_rate": 1.0141570141570142e-05, "loss": 0.4259, "step": 15490 }, { "epoch": 3.9891905803628878, "grad_norm": 5.6986212730407715, "learning_rate": 1.0115830115830117e-05, "loss": 0.458, "step": 15500 }, { "epoch": 3.991764251705057, "grad_norm": 1.4306472539901733, "learning_rate": 1.009009009009009e-05, "loss": 0.4718, "step": 15510 }, { "epoch": 3.994337923047227, "grad_norm": 0.9893137216567993, "learning_rate": 1.0064350064350065e-05, "loss": 0.3794, "step": 15520 }, { "epoch": 3.9969115943893962, "grad_norm": 1.7962301969528198, "learning_rate": 1.0038610038610038e-05, "loss": 0.4687, "step": 15530 }, { "epoch": 3.999485265731566, "grad_norm": 0.967959463596344, "learning_rate": 1.0012870012870013e-05, "loss": 0.4552, "step": 15540 }, { "epoch": 4.002058937073736, "grad_norm": 0.9137493371963501, "learning_rate": 9.987129987129988e-06, "loss": 0.3108, "step": 15550 }, { "epoch": 4.004632608415905, "grad_norm": 1.2810945510864258, "learning_rate": 9.961389961389962e-06, "loss": 0.3835, "step": 15560 }, { "epoch": 4.0072062797580745, "grad_norm": 0.8937966227531433, "learning_rate": 9.935649935649937e-06, "loss": 0.3403, "step": 15570 }, { "epoch": 4.009779951100245, "grad_norm": 1.049453616142273, "learning_rate": 9.90990990990991e-06, "loss": 0.3809, "step": 15580 }, { "epoch": 4.012353622442414, "grad_norm": 1.1060616970062256, "learning_rate": 9.884169884169885e-06, "loss": 0.3363, "step": 15590 }, { "epoch": 4.0149272937845835, "grad_norm": 0.9460583329200745, "learning_rate": 9.858429858429858e-06, "loss": 0.3511, "step": 15600 }, { "epoch": 4.017500965126754, "grad_norm": 0.9251907467842102, "learning_rate": 9.832689832689833e-06, "loss": 0.3859, "step": 15610 }, { "epoch": 4.020074636468923, "grad_norm": 1.1467247009277344, "learning_rate": 9.806949806949807e-06, "loss": 0.3202, "step": 15620 }, { "epoch": 4.022648307811092, "grad_norm": 1.3818519115447998, "learning_rate": 9.781209781209782e-06, "loss": 0.3372, "step": 15630 }, { "epoch": 4.025221979153262, "grad_norm": 0.9987813234329224, "learning_rate": 9.755469755469757e-06, "loss": 0.3872, "step": 15640 }, { "epoch": 4.027795650495432, "grad_norm": 0.863815188407898, "learning_rate": 9.729729729729732e-06, "loss": 0.3444, "step": 15650 }, { "epoch": 4.030369321837601, "grad_norm": 1.5916346311569214, "learning_rate": 9.703989703989705e-06, "loss": 0.3448, "step": 15660 }, { "epoch": 4.032942993179771, "grad_norm": 0.9141498804092407, "learning_rate": 9.678249678249678e-06, "loss": 0.3513, "step": 15670 }, { "epoch": 4.035516664521941, "grad_norm": 0.990973174571991, "learning_rate": 9.652509652509653e-06, "loss": 0.3584, "step": 15680 }, { "epoch": 4.03809033586411, "grad_norm": 1.2190366983413696, "learning_rate": 9.626769626769626e-06, "loss": 0.3294, "step": 15690 }, { "epoch": 4.04066400720628, "grad_norm": 1.7479588985443115, "learning_rate": 9.601029601029601e-06, "loss": 0.3963, "step": 15700 }, { "epoch": 4.043237678548449, "grad_norm": 1.09687340259552, "learning_rate": 9.575289575289575e-06, "loss": 0.3238, "step": 15710 }, { "epoch": 4.045811349890619, "grad_norm": 0.9358935356140137, "learning_rate": 9.54954954954955e-06, "loss": 0.345, "step": 15720 }, { "epoch": 4.0483850212327885, "grad_norm": 3.78627872467041, "learning_rate": 9.523809523809523e-06, "loss": 0.3594, "step": 15730 }, { "epoch": 4.050958692574958, "grad_norm": 3.2878758907318115, "learning_rate": 9.498069498069498e-06, "loss": 0.3389, "step": 15740 }, { "epoch": 4.053532363917128, "grad_norm": 1.448227882385254, "learning_rate": 9.472329472329473e-06, "loss": 0.3344, "step": 15750 }, { "epoch": 4.0561060352592975, "grad_norm": 1.4450225830078125, "learning_rate": 9.446589446589448e-06, "loss": 0.3224, "step": 15760 }, { "epoch": 4.058679706601467, "grad_norm": 1.051365852355957, "learning_rate": 9.420849420849421e-06, "loss": 0.2675, "step": 15770 }, { "epoch": 4.061253377943636, "grad_norm": 3.0001602172851562, "learning_rate": 9.395109395109396e-06, "loss": 0.353, "step": 15780 }, { "epoch": 4.063827049285806, "grad_norm": 1.0692496299743652, "learning_rate": 9.36936936936937e-06, "loss": 0.3422, "step": 15790 }, { "epoch": 4.066400720627976, "grad_norm": 1.7115063667297363, "learning_rate": 9.343629343629345e-06, "loss": 0.4045, "step": 15800 }, { "epoch": 4.068974391970145, "grad_norm": 4.681840896606445, "learning_rate": 9.317889317889318e-06, "loss": 0.3603, "step": 15810 }, { "epoch": 4.071548063312315, "grad_norm": 1.7194322347640991, "learning_rate": 9.292149292149293e-06, "loss": 0.3498, "step": 15820 }, { "epoch": 4.074121734654485, "grad_norm": 0.9989597797393799, "learning_rate": 9.266409266409266e-06, "loss": 0.3276, "step": 15830 }, { "epoch": 4.076695405996654, "grad_norm": 0.9685651659965515, "learning_rate": 9.240669240669241e-06, "loss": 0.316, "step": 15840 }, { "epoch": 4.079269077338824, "grad_norm": 1.5660096406936646, "learning_rate": 9.214929214929216e-06, "loss": 0.3929, "step": 15850 }, { "epoch": 4.081842748680994, "grad_norm": 1.4593199491500854, "learning_rate": 9.18918918918919e-06, "loss": 0.3564, "step": 15860 }, { "epoch": 4.084416420023163, "grad_norm": 1.3041324615478516, "learning_rate": 9.163449163449165e-06, "loss": 0.2997, "step": 15870 }, { "epoch": 4.086990091365332, "grad_norm": 1.8447177410125732, "learning_rate": 9.137709137709138e-06, "loss": 0.3274, "step": 15880 }, { "epoch": 4.0895637627075025, "grad_norm": 3.268892765045166, "learning_rate": 9.111969111969113e-06, "loss": 0.3505, "step": 15890 }, { "epoch": 4.092137434049672, "grad_norm": 1.6105279922485352, "learning_rate": 9.086229086229086e-06, "loss": 0.3177, "step": 15900 }, { "epoch": 4.094711105391841, "grad_norm": 1.8889609575271606, "learning_rate": 9.060489060489061e-06, "loss": 0.3904, "step": 15910 }, { "epoch": 4.0972847767340115, "grad_norm": 1.7368106842041016, "learning_rate": 9.034749034749034e-06, "loss": 0.339, "step": 15920 }, { "epoch": 4.099858448076181, "grad_norm": 1.3504408597946167, "learning_rate": 9.00900900900901e-06, "loss": 0.3543, "step": 15930 }, { "epoch": 4.10243211941835, "grad_norm": 1.0322836637496948, "learning_rate": 8.983268983268983e-06, "loss": 0.3498, "step": 15940 }, { "epoch": 4.1050057907605195, "grad_norm": 1.1150479316711426, "learning_rate": 8.957528957528958e-06, "loss": 0.3504, "step": 15950 }, { "epoch": 4.10757946210269, "grad_norm": 0.9721202254295349, "learning_rate": 8.931788931788933e-06, "loss": 0.3605, "step": 15960 }, { "epoch": 4.110153133444859, "grad_norm": 1.3478281497955322, "learning_rate": 8.906048906048906e-06, "loss": 0.322, "step": 15970 }, { "epoch": 4.1127268047870285, "grad_norm": 3.2121946811676025, "learning_rate": 8.880308880308881e-06, "loss": 0.3216, "step": 15980 }, { "epoch": 4.115300476129199, "grad_norm": 1.280942440032959, "learning_rate": 8.854568854568854e-06, "loss": 0.3064, "step": 15990 }, { "epoch": 4.117874147471368, "grad_norm": 1.1329562664031982, "learning_rate": 8.82882882882883e-06, "loss": 0.3252, "step": 16000 }, { "epoch": 4.117874147471368, "eval_loss": 0.8655849695205688, "eval_runtime": 395.6994, "eval_samples_per_second": 49.093, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.0003534038369559441, "step": 16000 }, { "epoch": 4.120447818813537, "grad_norm": 1.2837986946105957, "learning_rate": 8.803088803088803e-06, "loss": 0.2988, "step": 16010 }, { "epoch": 4.123021490155707, "grad_norm": 1.140343189239502, "learning_rate": 8.777348777348778e-06, "loss": 0.3398, "step": 16020 }, { "epoch": 4.125595161497877, "grad_norm": 1.1737910509109497, "learning_rate": 8.751608751608751e-06, "loss": 0.3676, "step": 16030 }, { "epoch": 4.128168832840046, "grad_norm": 1.1342344284057617, "learning_rate": 8.725868725868726e-06, "loss": 0.3281, "step": 16040 }, { "epoch": 4.130742504182216, "grad_norm": 1.363948941230774, "learning_rate": 8.7001287001287e-06, "loss": 0.3137, "step": 16050 }, { "epoch": 4.133316175524386, "grad_norm": 0.9117937684059143, "learning_rate": 8.674388674388676e-06, "loss": 0.3463, "step": 16060 }, { "epoch": 4.135889846866555, "grad_norm": 4.627786636352539, "learning_rate": 8.64864864864865e-06, "loss": 0.3507, "step": 16070 }, { "epoch": 4.138463518208725, "grad_norm": 1.0469121932983398, "learning_rate": 8.622908622908624e-06, "loss": 0.3416, "step": 16080 }, { "epoch": 4.141037189550894, "grad_norm": 4.487536907196045, "learning_rate": 8.597168597168598e-06, "loss": 0.3084, "step": 16090 }, { "epoch": 4.143610860893064, "grad_norm": 3.589993953704834, "learning_rate": 8.571428571428573e-06, "loss": 0.3498, "step": 16100 }, { "epoch": 4.1461845322352335, "grad_norm": 0.8773334622383118, "learning_rate": 8.545688545688546e-06, "loss": 0.3859, "step": 16110 }, { "epoch": 4.148758203577403, "grad_norm": 1.286348819732666, "learning_rate": 8.519948519948521e-06, "loss": 0.2771, "step": 16120 }, { "epoch": 4.151331874919573, "grad_norm": 1.3166707754135132, "learning_rate": 8.494208494208494e-06, "loss": 0.4015, "step": 16130 }, { "epoch": 4.1539055462617425, "grad_norm": 4.864274978637695, "learning_rate": 8.46846846846847e-06, "loss": 0.3171, "step": 16140 }, { "epoch": 4.156479217603912, "grad_norm": 1.0414588451385498, "learning_rate": 8.442728442728443e-06, "loss": 0.3396, "step": 16150 }, { "epoch": 4.159052888946081, "grad_norm": 5.264183521270752, "learning_rate": 8.416988416988418e-06, "loss": 0.3569, "step": 16160 }, { "epoch": 4.161626560288251, "grad_norm": 1.0647825002670288, "learning_rate": 8.391248391248393e-06, "loss": 0.3317, "step": 16170 }, { "epoch": 4.164200231630421, "grad_norm": 0.8864617347717285, "learning_rate": 8.365508365508366e-06, "loss": 0.3493, "step": 16180 }, { "epoch": 4.16677390297259, "grad_norm": 1.1304059028625488, "learning_rate": 8.33976833976834e-06, "loss": 0.3038, "step": 16190 }, { "epoch": 4.16934757431476, "grad_norm": 1.0398166179656982, "learning_rate": 8.314028314028314e-06, "loss": 0.3168, "step": 16200 }, { "epoch": 4.17192124565693, "grad_norm": 0.8544594645500183, "learning_rate": 8.288288288288289e-06, "loss": 0.3365, "step": 16210 }, { "epoch": 4.174494916999099, "grad_norm": 0.9230920672416687, "learning_rate": 8.262548262548262e-06, "loss": 0.3184, "step": 16220 }, { "epoch": 4.177068588341269, "grad_norm": 6.984750270843506, "learning_rate": 8.236808236808237e-06, "loss": 0.3553, "step": 16230 }, { "epoch": 4.179642259683439, "grad_norm": 4.589465141296387, "learning_rate": 8.21106821106821e-06, "loss": 0.3263, "step": 16240 }, { "epoch": 4.182215931025608, "grad_norm": 1.004537582397461, "learning_rate": 8.185328185328186e-06, "loss": 0.341, "step": 16250 }, { "epoch": 4.184789602367777, "grad_norm": 2.243464231491089, "learning_rate": 8.159588159588159e-06, "loss": 0.3709, "step": 16260 }, { "epoch": 4.1873632737099475, "grad_norm": 1.2939341068267822, "learning_rate": 8.133848133848134e-06, "loss": 0.2929, "step": 16270 }, { "epoch": 4.189936945052117, "grad_norm": 1.1193703413009644, "learning_rate": 8.108108108108109e-06, "loss": 0.3015, "step": 16280 }, { "epoch": 4.192510616394286, "grad_norm": 0.9649505019187927, "learning_rate": 8.082368082368082e-06, "loss": 0.3044, "step": 16290 }, { "epoch": 4.1950842877364565, "grad_norm": 1.2722036838531494, "learning_rate": 8.056628056628057e-06, "loss": 0.3654, "step": 16300 }, { "epoch": 4.197657959078626, "grad_norm": 1.9029432535171509, "learning_rate": 8.03088803088803e-06, "loss": 0.33, "step": 16310 }, { "epoch": 4.200231630420795, "grad_norm": 0.8544790744781494, "learning_rate": 8.005148005148006e-06, "loss": 0.3368, "step": 16320 }, { "epoch": 4.2028053017629645, "grad_norm": 0.991193950176239, "learning_rate": 7.979407979407979e-06, "loss": 0.3242, "step": 16330 }, { "epoch": 4.205378973105135, "grad_norm": 1.923233151435852, "learning_rate": 7.953667953667954e-06, "loss": 0.2868, "step": 16340 }, { "epoch": 4.207952644447304, "grad_norm": 1.3587863445281982, "learning_rate": 7.927927927927927e-06, "loss": 0.3379, "step": 16350 }, { "epoch": 4.2105263157894735, "grad_norm": 0.9470755457878113, "learning_rate": 7.902187902187902e-06, "loss": 0.311, "step": 16360 }, { "epoch": 4.213099987131644, "grad_norm": 1.3828065395355225, "learning_rate": 7.876447876447877e-06, "loss": 0.3017, "step": 16370 }, { "epoch": 4.215673658473813, "grad_norm": 1.4327831268310547, "learning_rate": 7.850707850707852e-06, "loss": 0.3226, "step": 16380 }, { "epoch": 4.218247329815982, "grad_norm": 1.1601290702819824, "learning_rate": 7.824967824967826e-06, "loss": 0.3352, "step": 16390 }, { "epoch": 4.220821001158152, "grad_norm": 1.0657103061676025, "learning_rate": 7.7992277992278e-06, "loss": 0.332, "step": 16400 }, { "epoch": 4.223394672500322, "grad_norm": 3.6614043712615967, "learning_rate": 7.773487773487774e-06, "loss": 0.342, "step": 16410 }, { "epoch": 4.225968343842491, "grad_norm": 1.095182180404663, "learning_rate": 7.747747747747749e-06, "loss": 0.3488, "step": 16420 }, { "epoch": 4.228542015184661, "grad_norm": 1.0206010341644287, "learning_rate": 7.722007722007722e-06, "loss": 0.3571, "step": 16430 }, { "epoch": 4.231115686526831, "grad_norm": 1.0806519985198975, "learning_rate": 7.696267696267697e-06, "loss": 0.3845, "step": 16440 }, { "epoch": 4.233689357869, "grad_norm": 1.2475837469100952, "learning_rate": 7.67052767052767e-06, "loss": 0.3936, "step": 16450 }, { "epoch": 4.23626302921117, "grad_norm": 2.754807233810425, "learning_rate": 7.644787644787644e-06, "loss": 0.3074, "step": 16460 }, { "epoch": 4.238836700553339, "grad_norm": 2.6300132274627686, "learning_rate": 7.6190476190476205e-06, "loss": 0.3422, "step": 16470 }, { "epoch": 4.241410371895509, "grad_norm": 0.9944770932197571, "learning_rate": 7.593307593307595e-06, "loss": 0.3585, "step": 16480 }, { "epoch": 4.2439840432376785, "grad_norm": 0.8388077616691589, "learning_rate": 7.567567567567568e-06, "loss": 0.3416, "step": 16490 }, { "epoch": 4.246557714579848, "grad_norm": 1.02458918094635, "learning_rate": 7.544401544401544e-06, "loss": 0.3858, "step": 16500 }, { "epoch": 4.249131385922018, "grad_norm": 1.3384264707565308, "learning_rate": 7.518661518661518e-06, "loss": 0.3755, "step": 16510 }, { "epoch": 4.2517050572641875, "grad_norm": 0.9491118788719177, "learning_rate": 7.492921492921494e-06, "loss": 0.3789, "step": 16520 }, { "epoch": 4.254278728606357, "grad_norm": 1.3997193574905396, "learning_rate": 7.467181467181468e-06, "loss": 0.364, "step": 16530 }, { "epoch": 4.256852399948526, "grad_norm": 1.488587498664856, "learning_rate": 7.441441441441442e-06, "loss": 0.3153, "step": 16540 }, { "epoch": 4.259426071290696, "grad_norm": 1.079736590385437, "learning_rate": 7.415701415701416e-06, "loss": 0.3447, "step": 16550 }, { "epoch": 4.261999742632866, "grad_norm": 1.142016887664795, "learning_rate": 7.38996138996139e-06, "loss": 0.325, "step": 16560 }, { "epoch": 4.264573413975035, "grad_norm": 0.9195720553398132, "learning_rate": 7.364221364221364e-06, "loss": 0.3222, "step": 16570 }, { "epoch": 4.267147085317205, "grad_norm": 3.8581366539001465, "learning_rate": 7.338481338481338e-06, "loss": 0.3699, "step": 16580 }, { "epoch": 4.269720756659375, "grad_norm": 0.8989285230636597, "learning_rate": 7.3127413127413125e-06, "loss": 0.3438, "step": 16590 }, { "epoch": 4.272294428001544, "grad_norm": 1.3981261253356934, "learning_rate": 7.287001287001287e-06, "loss": 0.3827, "step": 16600 }, { "epoch": 4.274868099343713, "grad_norm": 1.5550109148025513, "learning_rate": 7.261261261261261e-06, "loss": 0.3755, "step": 16610 }, { "epoch": 4.277441770685884, "grad_norm": 1.4705978631973267, "learning_rate": 7.235521235521237e-06, "loss": 0.3258, "step": 16620 }, { "epoch": 4.280015442028053, "grad_norm": 1.0708073377609253, "learning_rate": 7.209781209781211e-06, "loss": 0.3469, "step": 16630 }, { "epoch": 4.282589113370222, "grad_norm": 4.414210796356201, "learning_rate": 7.184041184041185e-06, "loss": 0.3964, "step": 16640 }, { "epoch": 4.2851627847123925, "grad_norm": 1.2815184593200684, "learning_rate": 7.158301158301159e-06, "loss": 0.3284, "step": 16650 }, { "epoch": 4.287736456054562, "grad_norm": 1.2706879377365112, "learning_rate": 7.132561132561133e-06, "loss": 0.3881, "step": 16660 }, { "epoch": 4.290310127396731, "grad_norm": 0.791149914264679, "learning_rate": 7.106821106821107e-06, "loss": 0.33, "step": 16670 }, { "epoch": 4.292883798738901, "grad_norm": 1.350524663925171, "learning_rate": 7.0810810810810815e-06, "loss": 0.4255, "step": 16680 }, { "epoch": 4.295457470081071, "grad_norm": 1.2001677751541138, "learning_rate": 7.055341055341056e-06, "loss": 0.3677, "step": 16690 }, { "epoch": 4.29803114142324, "grad_norm": 3.5927603244781494, "learning_rate": 7.02960102960103e-06, "loss": 0.377, "step": 16700 }, { "epoch": 4.3006048127654095, "grad_norm": 1.3985093832015991, "learning_rate": 7.003861003861004e-06, "loss": 0.3129, "step": 16710 }, { "epoch": 4.30317848410758, "grad_norm": 2.9167826175689697, "learning_rate": 6.978120978120979e-06, "loss": 0.3275, "step": 16720 }, { "epoch": 4.305752155449749, "grad_norm": 5.278055667877197, "learning_rate": 6.952380952380953e-06, "loss": 0.3816, "step": 16730 }, { "epoch": 4.3083258267919184, "grad_norm": 1.1263563632965088, "learning_rate": 6.926640926640927e-06, "loss": 0.3447, "step": 16740 }, { "epoch": 4.310899498134089, "grad_norm": 1.0541788339614868, "learning_rate": 6.9009009009009014e-06, "loss": 0.3533, "step": 16750 }, { "epoch": 4.313473169476258, "grad_norm": 1.7180668115615845, "learning_rate": 6.875160875160876e-06, "loss": 0.3639, "step": 16760 }, { "epoch": 4.316046840818427, "grad_norm": 1.4886486530303955, "learning_rate": 6.84942084942085e-06, "loss": 0.3239, "step": 16770 }, { "epoch": 4.318620512160597, "grad_norm": 1.4927681684494019, "learning_rate": 6.823680823680824e-06, "loss": 0.327, "step": 16780 }, { "epoch": 4.321194183502767, "grad_norm": 1.2550019025802612, "learning_rate": 6.797940797940798e-06, "loss": 0.3321, "step": 16790 }, { "epoch": 4.323767854844936, "grad_norm": 1.218307614326477, "learning_rate": 6.772200772200772e-06, "loss": 0.3482, "step": 16800 }, { "epoch": 4.326341526187106, "grad_norm": 1.1296604871749878, "learning_rate": 6.746460746460746e-06, "loss": 0.3373, "step": 16810 }, { "epoch": 4.328915197529276, "grad_norm": 3.720290184020996, "learning_rate": 6.7207207207207205e-06, "loss": 0.3701, "step": 16820 }, { "epoch": 4.331488868871445, "grad_norm": 1.146220326423645, "learning_rate": 6.6949806949806955e-06, "loss": 0.3682, "step": 16830 }, { "epoch": 4.334062540213615, "grad_norm": 1.438725233078003, "learning_rate": 6.66924066924067e-06, "loss": 0.337, "step": 16840 }, { "epoch": 4.336636211555784, "grad_norm": 1.264733910560608, "learning_rate": 6.643500643500644e-06, "loss": 0.3546, "step": 16850 }, { "epoch": 4.339209882897954, "grad_norm": 1.1695109605789185, "learning_rate": 6.617760617760618e-06, "loss": 0.37, "step": 16860 }, { "epoch": 4.3417835542401235, "grad_norm": 1.0620770454406738, "learning_rate": 6.592020592020592e-06, "loss": 0.3563, "step": 16870 }, { "epoch": 4.344357225582293, "grad_norm": 4.714122772216797, "learning_rate": 6.566280566280566e-06, "loss": 0.3119, "step": 16880 }, { "epoch": 4.346930896924463, "grad_norm": 1.527566909790039, "learning_rate": 6.54054054054054e-06, "loss": 0.3352, "step": 16890 }, { "epoch": 4.3495045682666325, "grad_norm": 1.4223473072052002, "learning_rate": 6.5148005148005146e-06, "loss": 0.3913, "step": 16900 }, { "epoch": 4.352078239608802, "grad_norm": 1.3415018320083618, "learning_rate": 6.489060489060489e-06, "loss": 0.3759, "step": 16910 }, { "epoch": 4.354651910950971, "grad_norm": 1.1462664604187012, "learning_rate": 6.463320463320463e-06, "loss": 0.3822, "step": 16920 }, { "epoch": 4.357225582293141, "grad_norm": 1.2110233306884766, "learning_rate": 6.437580437580439e-06, "loss": 0.3454, "step": 16930 }, { "epoch": 4.359799253635311, "grad_norm": 4.063595294952393, "learning_rate": 6.411840411840413e-06, "loss": 0.3543, "step": 16940 }, { "epoch": 4.36237292497748, "grad_norm": 0.9260187149047852, "learning_rate": 6.386100386100387e-06, "loss": 0.3595, "step": 16950 }, { "epoch": 4.36494659631965, "grad_norm": 1.3427493572235107, "learning_rate": 6.360360360360361e-06, "loss": 0.366, "step": 16960 }, { "epoch": 4.36752026766182, "grad_norm": 1.0595113039016724, "learning_rate": 6.334620334620335e-06, "loss": 0.2813, "step": 16970 }, { "epoch": 4.370093939003989, "grad_norm": 1.5860631465911865, "learning_rate": 6.3088803088803095e-06, "loss": 0.3394, "step": 16980 }, { "epoch": 4.372667610346159, "grad_norm": 1.2586485147476196, "learning_rate": 6.283140283140284e-06, "loss": 0.3595, "step": 16990 }, { "epoch": 4.375241281688329, "grad_norm": 1.0204342603683472, "learning_rate": 6.257400257400258e-06, "loss": 0.3536, "step": 17000 }, { "epoch": 4.375241281688329, "eval_loss": 0.8610925674438477, "eval_runtime": 395.7656, "eval_samples_per_second": 49.085, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.00034752009960597937, "step": 17000 }, { "epoch": 4.377814953030498, "grad_norm": 0.714625895023346, "learning_rate": 6.231660231660232e-06, "loss": 0.3166, "step": 17010 }, { "epoch": 4.380388624372667, "grad_norm": 1.5543569326400757, "learning_rate": 6.205920205920206e-06, "loss": 0.3606, "step": 17020 }, { "epoch": 4.3829622957148375, "grad_norm": 1.7975174188613892, "learning_rate": 6.18018018018018e-06, "loss": 0.3124, "step": 17030 }, { "epoch": 4.385535967057007, "grad_norm": 4.076536178588867, "learning_rate": 6.154440154440154e-06, "loss": 0.3513, "step": 17040 }, { "epoch": 4.388109638399176, "grad_norm": 1.210050344467163, "learning_rate": 6.1287001287001285e-06, "loss": 0.3522, "step": 17050 }, { "epoch": 4.3906833097413465, "grad_norm": 1.1933529376983643, "learning_rate": 6.1029601029601035e-06, "loss": 0.3474, "step": 17060 }, { "epoch": 4.393256981083516, "grad_norm": 1.305173635482788, "learning_rate": 6.077220077220078e-06, "loss": 0.3271, "step": 17070 }, { "epoch": 4.395830652425685, "grad_norm": 1.0537171363830566, "learning_rate": 6.051480051480052e-06, "loss": 0.3602, "step": 17080 }, { "epoch": 4.3984043237678545, "grad_norm": 1.2690879106521606, "learning_rate": 6.025740025740026e-06, "loss": 0.356, "step": 17090 }, { "epoch": 4.400977995110025, "grad_norm": 1.4055087566375732, "learning_rate": 6e-06, "loss": 0.3896, "step": 17100 }, { "epoch": 4.403551666452194, "grad_norm": 2.5787110328674316, "learning_rate": 5.974259974259975e-06, "loss": 0.3773, "step": 17110 }, { "epoch": 4.406125337794363, "grad_norm": 1.122134804725647, "learning_rate": 5.948519948519949e-06, "loss": 0.3427, "step": 17120 }, { "epoch": 4.408699009136534, "grad_norm": 0.8836435675621033, "learning_rate": 5.9227799227799235e-06, "loss": 0.3142, "step": 17130 }, { "epoch": 4.411272680478703, "grad_norm": 1.6835286617279053, "learning_rate": 5.897039897039898e-06, "loss": 0.3993, "step": 17140 }, { "epoch": 4.413846351820872, "grad_norm": 3.622476577758789, "learning_rate": 5.871299871299872e-06, "loss": 0.331, "step": 17150 }, { "epoch": 4.416420023163042, "grad_norm": 1.1817376613616943, "learning_rate": 5.845559845559846e-06, "loss": 0.356, "step": 17160 }, { "epoch": 4.418993694505212, "grad_norm": 1.0736690759658813, "learning_rate": 5.81981981981982e-06, "loss": 0.4044, "step": 17170 }, { "epoch": 4.421567365847381, "grad_norm": 1.3568885326385498, "learning_rate": 5.794079794079794e-06, "loss": 0.3515, "step": 17180 }, { "epoch": 4.424141037189551, "grad_norm": 1.0532563924789429, "learning_rate": 5.768339768339768e-06, "loss": 0.3542, "step": 17190 }, { "epoch": 4.426714708531721, "grad_norm": 1.5962337255477905, "learning_rate": 5.7425997425997425e-06, "loss": 0.3249, "step": 17200 }, { "epoch": 4.42928837987389, "grad_norm": 1.8830443620681763, "learning_rate": 5.716859716859717e-06, "loss": 0.3256, "step": 17210 }, { "epoch": 4.43186205121606, "grad_norm": 1.0833598375320435, "learning_rate": 5.691119691119692e-06, "loss": 0.3783, "step": 17220 }, { "epoch": 4.434435722558229, "grad_norm": 1.3849422931671143, "learning_rate": 5.665379665379666e-06, "loss": 0.3494, "step": 17230 }, { "epoch": 4.437009393900399, "grad_norm": 2.959022045135498, "learning_rate": 5.63963963963964e-06, "loss": 0.3622, "step": 17240 }, { "epoch": 4.4395830652425685, "grad_norm": 0.9271379113197327, "learning_rate": 5.613899613899614e-06, "loss": 0.3807, "step": 17250 }, { "epoch": 4.442156736584738, "grad_norm": 1.0124372243881226, "learning_rate": 5.588159588159588e-06, "loss": 0.3751, "step": 17260 }, { "epoch": 4.444730407926908, "grad_norm": 3.5358119010925293, "learning_rate": 5.562419562419563e-06, "loss": 0.2874, "step": 17270 }, { "epoch": 4.4473040792690774, "grad_norm": 0.9238685369491577, "learning_rate": 5.5366795366795374e-06, "loss": 0.3796, "step": 17280 }, { "epoch": 4.449877750611247, "grad_norm": 1.2194762229919434, "learning_rate": 5.510939510939512e-06, "loss": 0.3643, "step": 17290 }, { "epoch": 4.452451421953416, "grad_norm": 1.2170203924179077, "learning_rate": 5.485199485199486e-06, "loss": 0.3835, "step": 17300 }, { "epoch": 4.455025093295586, "grad_norm": 1.5656147003173828, "learning_rate": 5.45945945945946e-06, "loss": 0.3385, "step": 17310 }, { "epoch": 4.457598764637756, "grad_norm": 1.351452112197876, "learning_rate": 5.433719433719434e-06, "loss": 0.3646, "step": 17320 }, { "epoch": 4.460172435979925, "grad_norm": 1.1364386081695557, "learning_rate": 5.407979407979408e-06, "loss": 0.3132, "step": 17330 }, { "epoch": 4.462746107322095, "grad_norm": 1.2257845401763916, "learning_rate": 5.382239382239382e-06, "loss": 0.3266, "step": 17340 }, { "epoch": 4.465319778664265, "grad_norm": 1.087573766708374, "learning_rate": 5.3564993564993565e-06, "loss": 0.3644, "step": 17350 }, { "epoch": 4.467893450006434, "grad_norm": 1.2101587057113647, "learning_rate": 5.330759330759331e-06, "loss": 0.3456, "step": 17360 }, { "epoch": 4.470467121348603, "grad_norm": 1.4573382139205933, "learning_rate": 5.305019305019306e-06, "loss": 0.3113, "step": 17370 }, { "epoch": 4.473040792690774, "grad_norm": 0.8906600475311279, "learning_rate": 5.27927927927928e-06, "loss": 0.3345, "step": 17380 }, { "epoch": 4.475614464032943, "grad_norm": 1.079119086265564, "learning_rate": 5.253539253539254e-06, "loss": 0.3668, "step": 17390 }, { "epoch": 4.478188135375112, "grad_norm": 1.0174261331558228, "learning_rate": 5.227799227799228e-06, "loss": 0.2963, "step": 17400 }, { "epoch": 4.4807618067172825, "grad_norm": 5.470022678375244, "learning_rate": 5.202059202059202e-06, "loss": 0.39, "step": 17410 }, { "epoch": 4.483335478059452, "grad_norm": 3.2277028560638428, "learning_rate": 5.176319176319176e-06, "loss": 0.3342, "step": 17420 }, { "epoch": 4.485909149401621, "grad_norm": 1.0484484434127808, "learning_rate": 5.150579150579151e-06, "loss": 0.3536, "step": 17430 }, { "epoch": 4.488482820743791, "grad_norm": 3.3900933265686035, "learning_rate": 5.1248391248391256e-06, "loss": 0.3409, "step": 17440 }, { "epoch": 4.491056492085961, "grad_norm": 1.2388783693313599, "learning_rate": 5.0990990990991e-06, "loss": 0.3389, "step": 17450 }, { "epoch": 4.49363016342813, "grad_norm": 1.4199916124343872, "learning_rate": 5.073359073359073e-06, "loss": 0.3302, "step": 17460 }, { "epoch": 4.4962038347702995, "grad_norm": 1.3507987260818481, "learning_rate": 5.047619047619047e-06, "loss": 0.367, "step": 17470 }, { "epoch": 4.49877750611247, "grad_norm": 1.1383529901504517, "learning_rate": 5.021879021879022e-06, "loss": 0.3361, "step": 17480 }, { "epoch": 4.501351177454639, "grad_norm": 1.3563300371170044, "learning_rate": 4.996138996138996e-06, "loss": 0.3596, "step": 17490 }, { "epoch": 4.503924848796808, "grad_norm": 1.2805308103561401, "learning_rate": 4.9703989703989705e-06, "loss": 0.338, "step": 17500 }, { "epoch": 4.506498520138978, "grad_norm": 1.186610460281372, "learning_rate": 4.944658944658945e-06, "loss": 0.4018, "step": 17510 }, { "epoch": 4.509072191481148, "grad_norm": 1.3856215476989746, "learning_rate": 4.918918918918919e-06, "loss": 0.3387, "step": 17520 }, { "epoch": 4.511645862823317, "grad_norm": 1.4942715167999268, "learning_rate": 4.893178893178894e-06, "loss": 0.3774, "step": 17530 }, { "epoch": 4.514219534165487, "grad_norm": 1.3784438371658325, "learning_rate": 4.867438867438868e-06, "loss": 0.3103, "step": 17540 }, { "epoch": 4.516793205507657, "grad_norm": 1.6347748041152954, "learning_rate": 4.841698841698842e-06, "loss": 0.3187, "step": 17550 }, { "epoch": 4.519366876849826, "grad_norm": 3.5454697608947754, "learning_rate": 4.815958815958816e-06, "loss": 0.3355, "step": 17560 }, { "epoch": 4.521940548191996, "grad_norm": 1.6713000535964966, "learning_rate": 4.79021879021879e-06, "loss": 0.3521, "step": 17570 }, { "epoch": 4.524514219534165, "grad_norm": 1.2210255861282349, "learning_rate": 4.764478764478765e-06, "loss": 0.3532, "step": 17580 }, { "epoch": 4.527087890876335, "grad_norm": 1.9054685831069946, "learning_rate": 4.7387387387387395e-06, "loss": 0.3597, "step": 17590 }, { "epoch": 4.529661562218505, "grad_norm": 1.692408561706543, "learning_rate": 4.712998712998713e-06, "loss": 0.329, "step": 17600 }, { "epoch": 4.532235233560674, "grad_norm": 1.5581741333007812, "learning_rate": 4.687258687258687e-06, "loss": 0.3693, "step": 17610 }, { "epoch": 4.534808904902844, "grad_norm": 1.223633050918579, "learning_rate": 4.661518661518661e-06, "loss": 0.3536, "step": 17620 }, { "epoch": 4.5373825762450135, "grad_norm": 1.719165325164795, "learning_rate": 4.635778635778636e-06, "loss": 0.3352, "step": 17630 }, { "epoch": 4.539956247587183, "grad_norm": 1.2446945905685425, "learning_rate": 4.61003861003861e-06, "loss": 0.305, "step": 17640 }, { "epoch": 4.542529918929353, "grad_norm": 1.1125750541687012, "learning_rate": 4.5842985842985844e-06, "loss": 0.3379, "step": 17650 }, { "epoch": 4.545103590271522, "grad_norm": 1.183079719543457, "learning_rate": 4.558558558558559e-06, "loss": 0.4016, "step": 17660 }, { "epoch": 4.547677261613692, "grad_norm": 1.196256399154663, "learning_rate": 4.532818532818533e-06, "loss": 0.3503, "step": 17670 }, { "epoch": 4.550250932955862, "grad_norm": 2.58664608001709, "learning_rate": 4.507078507078507e-06, "loss": 0.3455, "step": 17680 }, { "epoch": 4.552824604298031, "grad_norm": 1.1959803104400635, "learning_rate": 4.481338481338482e-06, "loss": 0.3916, "step": 17690 }, { "epoch": 4.555398275640201, "grad_norm": 0.9203136563301086, "learning_rate": 4.455598455598456e-06, "loss": 0.3098, "step": 17700 }, { "epoch": 4.55797194698237, "grad_norm": 0.8788485527038574, "learning_rate": 4.42985842985843e-06, "loss": 0.3033, "step": 17710 }, { "epoch": 4.56054561832454, "grad_norm": 0.9711852669715881, "learning_rate": 4.404118404118404e-06, "loss": 0.3704, "step": 17720 }, { "epoch": 4.56311928966671, "grad_norm": 0.947658360004425, "learning_rate": 4.3783783783783785e-06, "loss": 0.2983, "step": 17730 }, { "epoch": 4.565692961008879, "grad_norm": 0.8453003168106079, "learning_rate": 4.3526383526383535e-06, "loss": 0.2935, "step": 17740 }, { "epoch": 4.568266632351049, "grad_norm": 1.273728370666504, "learning_rate": 4.326898326898327e-06, "loss": 0.3509, "step": 17750 }, { "epoch": 4.570840303693219, "grad_norm": 1.0643731355667114, "learning_rate": 4.301158301158301e-06, "loss": 0.3753, "step": 17760 }, { "epoch": 4.573413975035388, "grad_norm": 1.152840256690979, "learning_rate": 4.275418275418275e-06, "loss": 0.3816, "step": 17770 }, { "epoch": 4.575987646377557, "grad_norm": 1.0938299894332886, "learning_rate": 4.249678249678249e-06, "loss": 0.3666, "step": 17780 }, { "epoch": 4.5785613177197275, "grad_norm": 1.659525752067566, "learning_rate": 4.223938223938224e-06, "loss": 0.3586, "step": 17790 }, { "epoch": 4.581134989061897, "grad_norm": 1.3359277248382568, "learning_rate": 4.1981981981981984e-06, "loss": 0.3487, "step": 17800 }, { "epoch": 4.583708660404066, "grad_norm": 1.0927700996398926, "learning_rate": 4.1724581724581726e-06, "loss": 0.3639, "step": 17810 }, { "epoch": 4.5862823317462365, "grad_norm": 0.907636284828186, "learning_rate": 4.146718146718147e-06, "loss": 0.3088, "step": 17820 }, { "epoch": 4.588856003088406, "grad_norm": 1.650233268737793, "learning_rate": 4.120978120978121e-06, "loss": 0.3611, "step": 17830 }, { "epoch": 4.591429674430575, "grad_norm": 0.9985641837120056, "learning_rate": 4.095238095238096e-06, "loss": 0.3119, "step": 17840 }, { "epoch": 4.5940033457727445, "grad_norm": 1.0793542861938477, "learning_rate": 4.06949806949807e-06, "loss": 0.3594, "step": 17850 }, { "epoch": 4.596577017114915, "grad_norm": 1.0795053243637085, "learning_rate": 4.043758043758044e-06, "loss": 0.3464, "step": 17860 }, { "epoch": 4.599150688457084, "grad_norm": 0.9072697758674622, "learning_rate": 4.018018018018018e-06, "loss": 0.3176, "step": 17870 }, { "epoch": 4.601724359799253, "grad_norm": 2.6680212020874023, "learning_rate": 3.9922779922779925e-06, "loss": 0.337, "step": 17880 }, { "epoch": 4.604298031141424, "grad_norm": 1.0229748487472534, "learning_rate": 3.966537966537967e-06, "loss": 0.3516, "step": 17890 }, { "epoch": 4.606871702483593, "grad_norm": 3.6243858337402344, "learning_rate": 3.940797940797941e-06, "loss": 0.3331, "step": 17900 }, { "epoch": 4.609445373825762, "grad_norm": 1.2276082038879395, "learning_rate": 3.915057915057915e-06, "loss": 0.3215, "step": 17910 }, { "epoch": 4.612019045167932, "grad_norm": 1.2664270401000977, "learning_rate": 3.889317889317889e-06, "loss": 0.3782, "step": 17920 }, { "epoch": 4.614592716510102, "grad_norm": 1.5127272605895996, "learning_rate": 3.863577863577863e-06, "loss": 0.3471, "step": 17930 }, { "epoch": 4.617166387852271, "grad_norm": 1.5340853929519653, "learning_rate": 3.837837837837837e-06, "loss": 0.3593, "step": 17940 }, { "epoch": 4.619740059194441, "grad_norm": 0.8411980867385864, "learning_rate": 3.8120978120978124e-06, "loss": 0.3846, "step": 17950 }, { "epoch": 4.622313730536611, "grad_norm": 3.0642356872558594, "learning_rate": 3.7863577863577866e-06, "loss": 0.372, "step": 17960 }, { "epoch": 4.62488740187878, "grad_norm": 2.399491786956787, "learning_rate": 3.7606177606177607e-06, "loss": 0.3127, "step": 17970 }, { "epoch": 4.62746107322095, "grad_norm": 1.3899736404418945, "learning_rate": 3.734877734877735e-06, "loss": 0.3518, "step": 17980 }, { "epoch": 4.630034744563119, "grad_norm": 2.5960450172424316, "learning_rate": 3.709137709137709e-06, "loss": 0.3246, "step": 17990 }, { "epoch": 4.632608415905289, "grad_norm": 1.2150206565856934, "learning_rate": 3.683397683397684e-06, "loss": 0.3655, "step": 18000 }, { "epoch": 4.632608415905289, "eval_loss": 0.8564029932022095, "eval_runtime": 395.6976, "eval_samples_per_second": 49.093, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.00035587121068334865, "step": 18000 }, { "epoch": 4.6351820872474585, "grad_norm": 1.110227108001709, "learning_rate": 3.6576576576576577e-06, "loss": 0.3679, "step": 18010 }, { "epoch": 4.637755758589628, "grad_norm": 1.2694969177246094, "learning_rate": 3.6344916344916345e-06, "loss": 0.4168, "step": 18020 }, { "epoch": 4.640329429931798, "grad_norm": 1.5729743242263794, "learning_rate": 3.6087516087516086e-06, "loss": 0.3952, "step": 18030 }, { "epoch": 4.642903101273967, "grad_norm": 1.1754688024520874, "learning_rate": 3.5830115830115836e-06, "loss": 0.3475, "step": 18040 }, { "epoch": 4.645476772616137, "grad_norm": 1.3123219013214111, "learning_rate": 3.5572715572715578e-06, "loss": 0.3589, "step": 18050 }, { "epoch": 4.648050443958306, "grad_norm": 3.629625082015991, "learning_rate": 3.5315315315315315e-06, "loss": 0.3261, "step": 18060 }, { "epoch": 4.650624115300476, "grad_norm": 1.0868263244628906, "learning_rate": 3.5057915057915057e-06, "loss": 0.3427, "step": 18070 }, { "epoch": 4.653197786642646, "grad_norm": 1.1146104335784912, "learning_rate": 3.48005148005148e-06, "loss": 0.3577, "step": 18080 }, { "epoch": 4.655771457984815, "grad_norm": 1.4053925275802612, "learning_rate": 3.454311454311455e-06, "loss": 0.3473, "step": 18090 }, { "epoch": 4.658345129326985, "grad_norm": 1.5852116346359253, "learning_rate": 3.428571428571429e-06, "loss": 0.3616, "step": 18100 }, { "epoch": 4.660918800669155, "grad_norm": 1.1257933378219604, "learning_rate": 3.402831402831403e-06, "loss": 0.3035, "step": 18110 }, { "epoch": 4.663492472011324, "grad_norm": 1.2144465446472168, "learning_rate": 3.3770913770913773e-06, "loss": 0.3089, "step": 18120 }, { "epoch": 4.666066143353493, "grad_norm": 1.458746314048767, "learning_rate": 3.3513513513513514e-06, "loss": 0.3289, "step": 18130 }, { "epoch": 4.668639814695664, "grad_norm": 1.8117784261703491, "learning_rate": 3.325611325611326e-06, "loss": 0.3345, "step": 18140 }, { "epoch": 4.671213486037833, "grad_norm": 1.3290542364120483, "learning_rate": 3.2998712998713e-06, "loss": 0.2998, "step": 18150 }, { "epoch": 4.673787157380002, "grad_norm": 1.1426622867584229, "learning_rate": 3.2741312741312743e-06, "loss": 0.364, "step": 18160 }, { "epoch": 4.6763608287221725, "grad_norm": 1.2317893505096436, "learning_rate": 3.2483912483912484e-06, "loss": 0.2657, "step": 18170 }, { "epoch": 4.678934500064342, "grad_norm": 1.211256980895996, "learning_rate": 3.2226512226512226e-06, "loss": 0.4204, "step": 18180 }, { "epoch": 4.681508171406511, "grad_norm": 1.5022300481796265, "learning_rate": 3.1969111969111968e-06, "loss": 0.3907, "step": 18190 }, { "epoch": 4.684081842748681, "grad_norm": 1.2132724523544312, "learning_rate": 3.1711711711711718e-06, "loss": 0.3386, "step": 18200 }, { "epoch": 4.686655514090851, "grad_norm": 1.2516412734985352, "learning_rate": 3.1454311454311455e-06, "loss": 0.3616, "step": 18210 }, { "epoch": 4.68922918543302, "grad_norm": 1.1552441120147705, "learning_rate": 3.1196911196911196e-06, "loss": 0.3508, "step": 18220 }, { "epoch": 4.6918028567751895, "grad_norm": 3.639354705810547, "learning_rate": 3.093951093951094e-06, "loss": 0.3425, "step": 18230 }, { "epoch": 4.69437652811736, "grad_norm": 3.8008406162261963, "learning_rate": 3.0682110682110684e-06, "loss": 0.3468, "step": 18240 }, { "epoch": 4.696950199459529, "grad_norm": 1.2858436107635498, "learning_rate": 3.0424710424710425e-06, "loss": 0.3806, "step": 18250 }, { "epoch": 4.699523870801698, "grad_norm": 2.4241316318511963, "learning_rate": 3.016731016731017e-06, "loss": 0.3471, "step": 18260 }, { "epoch": 4.702097542143868, "grad_norm": 1.0600461959838867, "learning_rate": 2.9909909909909912e-06, "loss": 0.301, "step": 18270 }, { "epoch": 4.704671213486038, "grad_norm": 1.1587661504745483, "learning_rate": 2.9652509652509654e-06, "loss": 0.3476, "step": 18280 }, { "epoch": 4.707244884828207, "grad_norm": 2.059361696243286, "learning_rate": 2.9395109395109395e-06, "loss": 0.3625, "step": 18290 }, { "epoch": 4.709818556170377, "grad_norm": 1.0103809833526611, "learning_rate": 2.9137709137709137e-06, "loss": 0.3285, "step": 18300 }, { "epoch": 4.712392227512547, "grad_norm": 3.1036524772644043, "learning_rate": 2.8880308880308883e-06, "loss": 0.4137, "step": 18310 }, { "epoch": 4.714965898854716, "grad_norm": 1.7626713514328003, "learning_rate": 2.8622908622908624e-06, "loss": 0.3824, "step": 18320 }, { "epoch": 4.717539570196886, "grad_norm": 1.179608941078186, "learning_rate": 2.8365508365508366e-06, "loss": 0.3257, "step": 18330 }, { "epoch": 4.720113241539055, "grad_norm": 1.0469659566879272, "learning_rate": 2.810810810810811e-06, "loss": 0.356, "step": 18340 }, { "epoch": 4.722686912881225, "grad_norm": 1.0116140842437744, "learning_rate": 2.7850707850707853e-06, "loss": 0.3352, "step": 18350 }, { "epoch": 4.725260584223395, "grad_norm": 1.530369520187378, "learning_rate": 2.7593307593307595e-06, "loss": 0.4061, "step": 18360 }, { "epoch": 4.727834255565564, "grad_norm": 1.096043586730957, "learning_rate": 2.7335907335907336e-06, "loss": 0.3633, "step": 18370 }, { "epoch": 4.730407926907734, "grad_norm": 2.9259002208709717, "learning_rate": 2.7078507078507078e-06, "loss": 0.3762, "step": 18380 }, { "epoch": 4.7329815982499035, "grad_norm": 2.655984878540039, "learning_rate": 2.6821106821106823e-06, "loss": 0.331, "step": 18390 }, { "epoch": 4.735555269592073, "grad_norm": 1.2596089839935303, "learning_rate": 2.6563706563706565e-06, "loss": 0.314, "step": 18400 }, { "epoch": 4.738128940934243, "grad_norm": 1.149182915687561, "learning_rate": 2.6306306306306306e-06, "loss": 0.3377, "step": 18410 }, { "epoch": 4.740702612276412, "grad_norm": 0.9628931879997253, "learning_rate": 2.6048906048906052e-06, "loss": 0.404, "step": 18420 }, { "epoch": 4.743276283618582, "grad_norm": 1.304672360420227, "learning_rate": 2.5791505791505794e-06, "loss": 0.3934, "step": 18430 }, { "epoch": 4.745849954960751, "grad_norm": 1.1351431608200073, "learning_rate": 2.5534105534105535e-06, "loss": 0.3549, "step": 18440 }, { "epoch": 4.748423626302921, "grad_norm": 3.7342593669891357, "learning_rate": 2.5276705276705277e-06, "loss": 0.3391, "step": 18450 }, { "epoch": 4.750997297645091, "grad_norm": 1.1187039613723755, "learning_rate": 2.501930501930502e-06, "loss": 0.341, "step": 18460 }, { "epoch": 4.75357096898726, "grad_norm": 1.6478817462921143, "learning_rate": 2.4761904761904764e-06, "loss": 0.3171, "step": 18470 }, { "epoch": 4.75614464032943, "grad_norm": 1.0323374271392822, "learning_rate": 2.4504504504504506e-06, "loss": 0.2878, "step": 18480 }, { "epoch": 4.7587183116716, "grad_norm": 1.34542977809906, "learning_rate": 2.424710424710425e-06, "loss": 0.33, "step": 18490 }, { "epoch": 4.761291983013769, "grad_norm": 1.1959524154663086, "learning_rate": 2.3989703989703993e-06, "loss": 0.3031, "step": 18500 }, { "epoch": 4.763865654355939, "grad_norm": 1.4889193773269653, "learning_rate": 2.3732303732303734e-06, "loss": 0.386, "step": 18510 }, { "epoch": 4.766439325698109, "grad_norm": 1.1365255117416382, "learning_rate": 2.3474903474903476e-06, "loss": 0.3173, "step": 18520 }, { "epoch": 4.769012997040278, "grad_norm": 1.1125775575637817, "learning_rate": 2.3217503217503217e-06, "loss": 0.37, "step": 18530 }, { "epoch": 4.771586668382447, "grad_norm": 1.230258584022522, "learning_rate": 2.296010296010296e-06, "loss": 0.3115, "step": 18540 }, { "epoch": 4.7741603397246175, "grad_norm": 1.1539350748062134, "learning_rate": 2.2702702702702705e-06, "loss": 0.3639, "step": 18550 }, { "epoch": 4.776734011066787, "grad_norm": 1.2653287649154663, "learning_rate": 2.2445302445302446e-06, "loss": 0.3541, "step": 18560 }, { "epoch": 4.779307682408956, "grad_norm": 0.9209233522415161, "learning_rate": 2.218790218790219e-06, "loss": 0.3348, "step": 18570 }, { "epoch": 4.781881353751126, "grad_norm": 0.895807683467865, "learning_rate": 2.1930501930501933e-06, "loss": 0.3351, "step": 18580 }, { "epoch": 4.784455025093296, "grad_norm": 0.8402043581008911, "learning_rate": 2.167310167310167e-06, "loss": 0.3198, "step": 18590 }, { "epoch": 4.787028696435465, "grad_norm": 2.4166226387023926, "learning_rate": 2.1415701415701416e-06, "loss": 0.3464, "step": 18600 }, { "epoch": 4.7896023677776345, "grad_norm": 2.75046706199646, "learning_rate": 2.115830115830116e-06, "loss": 0.3046, "step": 18610 }, { "epoch": 4.792176039119805, "grad_norm": 1.4120715856552124, "learning_rate": 2.0900900900900904e-06, "loss": 0.387, "step": 18620 }, { "epoch": 4.794749710461974, "grad_norm": 1.039160132408142, "learning_rate": 2.0643500643500645e-06, "loss": 0.3805, "step": 18630 }, { "epoch": 4.797323381804143, "grad_norm": 3.3216381072998047, "learning_rate": 2.0386100386100387e-06, "loss": 0.3687, "step": 18640 }, { "epoch": 4.799897053146314, "grad_norm": 1.0428948402404785, "learning_rate": 2.0128700128700133e-06, "loss": 0.3146, "step": 18650 }, { "epoch": 4.802470724488483, "grad_norm": 1.2070870399475098, "learning_rate": 1.9871299871299874e-06, "loss": 0.3448, "step": 18660 }, { "epoch": 4.805044395830652, "grad_norm": 0.9999447464942932, "learning_rate": 1.961389961389961e-06, "loss": 0.3139, "step": 18670 }, { "epoch": 4.807618067172822, "grad_norm": 0.7918708920478821, "learning_rate": 1.9356499356499357e-06, "loss": 0.355, "step": 18680 }, { "epoch": 4.810191738514992, "grad_norm": 1.158919334411621, "learning_rate": 1.90990990990991e-06, "loss": 0.3108, "step": 18690 }, { "epoch": 4.812765409857161, "grad_norm": 1.4351342916488647, "learning_rate": 1.8841698841698844e-06, "loss": 0.3367, "step": 18700 }, { "epoch": 4.815339081199331, "grad_norm": 1.1386419534683228, "learning_rate": 1.8584298584298586e-06, "loss": 0.3189, "step": 18710 }, { "epoch": 4.817912752541501, "grad_norm": 1.5254247188568115, "learning_rate": 1.8326898326898327e-06, "loss": 0.3336, "step": 18720 }, { "epoch": 4.82048642388367, "grad_norm": 1.0063539743423462, "learning_rate": 1.8069498069498071e-06, "loss": 0.3185, "step": 18730 }, { "epoch": 4.82306009522584, "grad_norm": 1.1583902835845947, "learning_rate": 1.7812097812097813e-06, "loss": 0.3918, "step": 18740 }, { "epoch": 4.825633766568009, "grad_norm": 1.1044045686721802, "learning_rate": 1.7554697554697556e-06, "loss": 0.2989, "step": 18750 }, { "epoch": 4.828207437910179, "grad_norm": 1.114070177078247, "learning_rate": 1.7297297297297298e-06, "loss": 0.3586, "step": 18760 }, { "epoch": 4.8307811092523485, "grad_norm": 1.4722037315368652, "learning_rate": 1.703989703989704e-06, "loss": 0.364, "step": 18770 }, { "epoch": 4.833354780594518, "grad_norm": 1.1275004148483276, "learning_rate": 1.6782496782496785e-06, "loss": 0.3595, "step": 18780 }, { "epoch": 4.835928451936688, "grad_norm": 1.0008904933929443, "learning_rate": 1.6525096525096527e-06, "loss": 0.3347, "step": 18790 }, { "epoch": 4.838502123278857, "grad_norm": 0.9728559851646423, "learning_rate": 1.6267696267696266e-06, "loss": 0.3359, "step": 18800 }, { "epoch": 4.841075794621027, "grad_norm": 1.343985915184021, "learning_rate": 1.6010296010296012e-06, "loss": 0.3496, "step": 18810 }, { "epoch": 4.843649465963196, "grad_norm": 0.8457748889923096, "learning_rate": 1.5752895752895753e-06, "loss": 0.3495, "step": 18820 }, { "epoch": 4.846223137305366, "grad_norm": 4.120772838592529, "learning_rate": 1.5495495495495495e-06, "loss": 0.3289, "step": 18830 }, { "epoch": 4.848796808647536, "grad_norm": 1.3416566848754883, "learning_rate": 1.5238095238095238e-06, "loss": 0.341, "step": 18840 }, { "epoch": 4.851370479989705, "grad_norm": 2.817333936691284, "learning_rate": 1.4980694980694982e-06, "loss": 0.3213, "step": 18850 }, { "epoch": 4.853944151331875, "grad_norm": 1.17255699634552, "learning_rate": 1.4723294723294726e-06, "loss": 0.3291, "step": 18860 }, { "epoch": 4.856517822674045, "grad_norm": 2.559112787246704, "learning_rate": 1.4465894465894465e-06, "loss": 0.2838, "step": 18870 }, { "epoch": 4.859091494016214, "grad_norm": 1.171962022781372, "learning_rate": 1.4208494208494209e-06, "loss": 0.3366, "step": 18880 }, { "epoch": 4.861665165358383, "grad_norm": 1.125060796737671, "learning_rate": 1.3951093951093952e-06, "loss": 0.3411, "step": 18890 }, { "epoch": 4.864238836700554, "grad_norm": 1.2087688446044922, "learning_rate": 1.3693693693693696e-06, "loss": 0.2909, "step": 18900 }, { "epoch": 4.866812508042723, "grad_norm": 1.3647404909133911, "learning_rate": 1.3436293436293435e-06, "loss": 0.3347, "step": 18910 }, { "epoch": 4.869386179384892, "grad_norm": 1.1150912046432495, "learning_rate": 1.317889317889318e-06, "loss": 0.3467, "step": 18920 }, { "epoch": 4.8719598507270625, "grad_norm": 1.2216849327087402, "learning_rate": 1.2921492921492923e-06, "loss": 0.3459, "step": 18930 }, { "epoch": 4.874533522069232, "grad_norm": 1.0959711074829102, "learning_rate": 1.2664092664092664e-06, "loss": 0.3167, "step": 18940 }, { "epoch": 4.877107193411401, "grad_norm": 1.3392401933670044, "learning_rate": 1.2406692406692406e-06, "loss": 0.2993, "step": 18950 }, { "epoch": 4.8796808647535705, "grad_norm": 1.1086993217468262, "learning_rate": 1.214929214929215e-06, "loss": 0.3165, "step": 18960 }, { "epoch": 4.882254536095741, "grad_norm": 1.0389692783355713, "learning_rate": 1.1891891891891893e-06, "loss": 0.349, "step": 18970 }, { "epoch": 4.88482820743791, "grad_norm": 1.053175449371338, "learning_rate": 1.1634491634491635e-06, "loss": 0.3713, "step": 18980 }, { "epoch": 4.8874018787800795, "grad_norm": 1.4085311889648438, "learning_rate": 1.1377091377091378e-06, "loss": 0.3424, "step": 18990 }, { "epoch": 4.88997555012225, "grad_norm": 1.0388787984848022, "learning_rate": 1.111969111969112e-06, "loss": 0.3223, "step": 19000 }, { "epoch": 4.88997555012225, "eval_loss": 0.8570228219032288, "eval_runtime": 395.7978, "eval_samples_per_second": 49.081, "eval_steps_per_second": 2.456, "eval_token_accuracy": 0.00035359363493497523, "step": 19000 }, { "epoch": 4.892549221464419, "grad_norm": 1.3370487689971924, "learning_rate": 1.0862290862290863e-06, "loss": 0.3524, "step": 19010 }, { "epoch": 4.895122892806588, "grad_norm": 1.1266289949417114, "learning_rate": 1.0604890604890605e-06, "loss": 0.3193, "step": 19020 }, { "epoch": 4.897696564148758, "grad_norm": 1.6796464920043945, "learning_rate": 1.0347490347490348e-06, "loss": 0.3117, "step": 19030 }, { "epoch": 4.900270235490928, "grad_norm": 1.1467808485031128, "learning_rate": 1.009009009009009e-06, "loss": 0.366, "step": 19040 }, { "epoch": 4.902843906833097, "grad_norm": 0.8519312739372253, "learning_rate": 9.832689832689834e-07, "loss": 0.2902, "step": 19050 }, { "epoch": 4.905417578175267, "grad_norm": 1.30669105052948, "learning_rate": 9.575289575289575e-07, "loss": 0.3016, "step": 19060 }, { "epoch": 4.907991249517437, "grad_norm": 1.3413194417953491, "learning_rate": 9.317889317889319e-07, "loss": 0.3301, "step": 19070 }, { "epoch": 4.910564920859606, "grad_norm": 1.2429314851760864, "learning_rate": 9.06048906048906e-07, "loss": 0.328, "step": 19080 }, { "epoch": 4.913138592201776, "grad_norm": 1.04901123046875, "learning_rate": 8.803088803088803e-07, "loss": 0.3483, "step": 19090 }, { "epoch": 4.915712263543945, "grad_norm": 0.9599467515945435, "learning_rate": 8.545688545688547e-07, "loss": 0.3241, "step": 19100 }, { "epoch": 4.918285934886115, "grad_norm": 3.971909284591675, "learning_rate": 8.288288288288289e-07, "loss": 0.3647, "step": 19110 }, { "epoch": 4.9208596062282846, "grad_norm": 1.1139452457427979, "learning_rate": 8.030888030888032e-07, "loss": 0.3326, "step": 19120 }, { "epoch": 4.923433277570454, "grad_norm": 1.015626311302185, "learning_rate": 7.773487773487774e-07, "loss": 0.3422, "step": 19130 }, { "epoch": 4.926006948912624, "grad_norm": 1.0744688510894775, "learning_rate": 7.516087516087516e-07, "loss": 0.3471, "step": 19140 }, { "epoch": 4.9285806202547935, "grad_norm": 1.1969647407531738, "learning_rate": 7.258687258687259e-07, "loss": 0.3139, "step": 19150 }, { "epoch": 4.931154291596963, "grad_norm": 0.8164909482002258, "learning_rate": 7.001287001287001e-07, "loss": 0.3355, "step": 19160 }, { "epoch": 4.933727962939132, "grad_norm": 1.189275860786438, "learning_rate": 6.743886743886745e-07, "loss": 0.3782, "step": 19170 }, { "epoch": 4.936301634281302, "grad_norm": 1.551167368888855, "learning_rate": 6.486486486486486e-07, "loss": 0.3773, "step": 19180 }, { "epoch": 4.938875305623472, "grad_norm": 0.9430525898933411, "learning_rate": 6.22908622908623e-07, "loss": 0.3332, "step": 19190 }, { "epoch": 4.941448976965641, "grad_norm": 1.2511941194534302, "learning_rate": 5.971685971685971e-07, "loss": 0.3804, "step": 19200 }, { "epoch": 4.944022648307811, "grad_norm": 1.0374319553375244, "learning_rate": 5.714285714285715e-07, "loss": 0.3166, "step": 19210 }, { "epoch": 4.946596319649981, "grad_norm": 0.9965717792510986, "learning_rate": 5.456885456885458e-07, "loss": 0.3132, "step": 19220 }, { "epoch": 4.94916999099215, "grad_norm": 1.1020451784133911, "learning_rate": 5.1994851994852e-07, "loss": 0.3717, "step": 19230 }, { "epoch": 4.95174366233432, "grad_norm": 0.7918252944946289, "learning_rate": 4.942084942084943e-07, "loss": 0.2936, "step": 19240 }, { "epoch": 4.95431733367649, "grad_norm": 1.4422928094863892, "learning_rate": 4.6846846846846847e-07, "loss": 0.3374, "step": 19250 }, { "epoch": 4.956891005018659, "grad_norm": 0.949720025062561, "learning_rate": 4.427284427284428e-07, "loss": 0.3771, "step": 19260 }, { "epoch": 4.959464676360829, "grad_norm": 1.0243984460830688, "learning_rate": 4.16988416988417e-07, "loss": 0.3283, "step": 19270 }, { "epoch": 4.962038347702999, "grad_norm": 1.0813188552856445, "learning_rate": 3.912483912483913e-07, "loss": 0.3885, "step": 19280 }, { "epoch": 4.964612019045168, "grad_norm": 4.7011799812316895, "learning_rate": 3.6550836550836556e-07, "loss": 0.3065, "step": 19290 }, { "epoch": 4.967185690387337, "grad_norm": 5.996508598327637, "learning_rate": 3.397683397683398e-07, "loss": 0.3949, "step": 19300 }, { "epoch": 4.9697593617295075, "grad_norm": 1.202574610710144, "learning_rate": 3.1402831402831407e-07, "loss": 0.3265, "step": 19310 }, { "epoch": 4.972333033071677, "grad_norm": 0.9816371202468872, "learning_rate": 2.8828828828828833e-07, "loss": 0.3646, "step": 19320 }, { "epoch": 4.974906704413846, "grad_norm": 1.1760458946228027, "learning_rate": 2.6254826254826254e-07, "loss": 0.3221, "step": 19330 }, { "epoch": 4.977480375756016, "grad_norm": 1.4523731470108032, "learning_rate": 2.3680823680823682e-07, "loss": 0.3247, "step": 19340 }, { "epoch": 4.980054047098186, "grad_norm": 1.1711961030960083, "learning_rate": 2.1106821106821108e-07, "loss": 0.3514, "step": 19350 }, { "epoch": 4.982627718440355, "grad_norm": 0.8875769972801208, "learning_rate": 1.8532818532818534e-07, "loss": 0.3135, "step": 19360 }, { "epoch": 4.9852013897825245, "grad_norm": 1.7097541093826294, "learning_rate": 1.595881595881596e-07, "loss": 0.3252, "step": 19370 }, { "epoch": 4.987775061124695, "grad_norm": 3.204772472381592, "learning_rate": 1.3384813384813385e-07, "loss": 0.2987, "step": 19380 }, { "epoch": 4.990348732466864, "grad_norm": 2.616095542907715, "learning_rate": 1.0810810810810811e-07, "loss": 0.3303, "step": 19390 }, { "epoch": 4.992922403809033, "grad_norm": 1.0616631507873535, "learning_rate": 8.236808236808237e-08, "loss": 0.3565, "step": 19400 }, { "epoch": 4.995496075151204, "grad_norm": 0.8763009905815125, "learning_rate": 5.662805662805663e-08, "loss": 0.3151, "step": 19410 }, { "epoch": 4.998069746493373, "grad_norm": 1.2131026983261108, "learning_rate": 3.088803088803089e-08, "loss": 0.3425, "step": 19420 } ], "logging_steps": 10, "max_steps": 19425, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.4362014238809784e+18, "train_batch_size": 10, "trial_name": null, "trial_params": null }