{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 4466, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005597850425436633, "grad_norm": 9.592120170593262, "learning_rate": 9.999287451410797e-06, "loss": 1.4229, "step": 25 }, { "epoch": 0.011195700850873265, "grad_norm": 10.640639305114746, "learning_rate": 9.997030034001815e-06, "loss": 1.3665, "step": 50 }, { "epoch": 0.016793551276309895, "grad_norm": 11.189779281616211, "learning_rate": 9.993227208820974e-06, "loss": 1.0753, "step": 75 }, { "epoch": 0.02239140170174653, "grad_norm": 15.177507400512695, "learning_rate": 9.987880151949976e-06, "loss": 1.1291, "step": 100 }, { "epoch": 0.027989252127183162, "grad_norm": 6.662464141845703, "learning_rate": 9.980990517047731e-06, "loss": 1.2145, "step": 125 }, { "epoch": 0.03358710255261979, "grad_norm": 9.72287368774414, "learning_rate": 9.972560434838943e-06, "loss": 1.0737, "step": 150 }, { "epoch": 0.03918495297805643, "grad_norm": 8.399821281433105, "learning_rate": 9.96259251245514e-06, "loss": 0.9843, "step": 175 }, { "epoch": 0.04478280340349306, "grad_norm": 10.552536010742188, "learning_rate": 9.951089832628388e-06, "loss": 1.1599, "step": 200 }, { "epoch": 0.05038065382892969, "grad_norm": 8.101223945617676, "learning_rate": 9.938055952737908e-06, "loss": 1.1275, "step": 225 }, { "epoch": 0.055978504254366325, "grad_norm": 6.296974182128906, "learning_rate": 9.923494903709896e-06, "loss": 1.0943, "step": 250 }, { "epoch": 0.06157635467980296, "grad_norm": 8.705899238586426, "learning_rate": 9.90741118877091e-06, "loss": 1.0905, "step": 275 }, { "epoch": 0.06717420510523958, "grad_norm": 13.383705139160156, "learning_rate": 9.88980978205516e-06, "loss": 1.0648, "step": 300 }, { "epoch": 0.07277205553067621, "grad_norm": 6.9907636642456055, "learning_rate": 9.8706961270662e-06, "loss": 1.0228, "step": 325 }, { "epoch": 0.07836990595611286, "grad_norm": 10.480993270874023, "learning_rate": 9.85007613499343e-06, "loss": 1.0825, "step": 350 }, { "epoch": 0.08396775638154949, "grad_norm": 11.11700439453125, "learning_rate": 9.82795618288397e-06, "loss": 1.083, "step": 375 }, { "epoch": 0.08956560680698612, "grad_norm": 10.408477783203125, "learning_rate": 9.804343111670472e-06, "loss": 1.1088, "step": 400 }, { "epoch": 0.09516345723242275, "grad_norm": 8.23421859741211, "learning_rate": 9.779244224055446e-06, "loss": 1.1182, "step": 425 }, { "epoch": 0.10076130765785939, "grad_norm": 11.640494346618652, "learning_rate": 9.752667282252788e-06, "loss": 1.152, "step": 450 }, { "epoch": 0.10635915808329602, "grad_norm": 9.683300971984863, "learning_rate": 9.724620505587205e-06, "loss": 1.1508, "step": 475 }, { "epoch": 0.11195700850873265, "grad_norm": 12.511320114135742, "learning_rate": 9.69511256795226e-06, "loss": 0.9544, "step": 500 }, { "epoch": 0.11755485893416928, "grad_norm": 6.891467094421387, "learning_rate": 9.664152595127834e-06, "loss": 1.0014, "step": 525 }, { "epoch": 0.12315270935960591, "grad_norm": 8.182930946350098, "learning_rate": 9.63175016195784e-06, "loss": 1.0197, "step": 550 }, { "epoch": 0.12875055978504255, "grad_norm": 12.893110275268555, "learning_rate": 9.597915289389067e-06, "loss": 0.9908, "step": 575 }, { "epoch": 0.13434841021047916, "grad_norm": 12.896308898925781, "learning_rate": 9.56265844137203e-06, "loss": 1.0792, "step": 600 }, { "epoch": 0.1399462606359158, "grad_norm": 6.506565093994141, "learning_rate": 9.525990521624855e-06, "loss": 0.996, "step": 625 }, { "epoch": 0.14554411106135243, "grad_norm": 11.659127235412598, "learning_rate": 9.487922870261123e-06, "loss": 1.1131, "step": 650 }, { "epoch": 0.15114196148678907, "grad_norm": 11.150578498840332, "learning_rate": 9.44846726028277e-06, "loss": 1.0725, "step": 675 }, { "epoch": 0.15673981191222572, "grad_norm": 10.530040740966797, "learning_rate": 9.407635893939111e-06, "loss": 1.2278, "step": 700 }, { "epoch": 0.16233766233766234, "grad_norm": 10.855611801147461, "learning_rate": 9.365441398953103e-06, "loss": 0.9453, "step": 725 }, { "epoch": 0.16793551276309898, "grad_norm": 9.590290069580078, "learning_rate": 9.321896824616036e-06, "loss": 1.0567, "step": 750 }, { "epoch": 0.1735333631885356, "grad_norm": 8.801950454711914, "learning_rate": 9.27701563775183e-06, "loss": 1.1515, "step": 775 }, { "epoch": 0.17913121361397225, "grad_norm": 14.007678985595703, "learning_rate": 9.23081171855222e-06, "loss": 0.904, "step": 800 }, { "epoch": 0.18472906403940886, "grad_norm": 11.931122779846191, "learning_rate": 9.183299356284102e-06, "loss": 1.0948, "step": 825 }, { "epoch": 0.1903269144648455, "grad_norm": 11.707269668579102, "learning_rate": 9.134493244870347e-06, "loss": 1.0469, "step": 850 }, { "epoch": 0.19592476489028213, "grad_norm": 7.64641809463501, "learning_rate": 9.084408478345486e-06, "loss": 1.0608, "step": 875 }, { "epoch": 0.20152261531571877, "grad_norm": 11.14941692352295, "learning_rate": 9.033060546187651e-06, "loss": 1.0769, "step": 900 }, { "epoch": 0.2071204657411554, "grad_norm": 7.081947326660156, "learning_rate": 8.98046532852822e-06, "loss": 0.9925, "step": 925 }, { "epoch": 0.21271831616659204, "grad_norm": 8.95167064666748, "learning_rate": 8.926639091240636e-06, "loss": 0.8297, "step": 950 }, { "epoch": 0.21831616659202865, "grad_norm": 8.65690803527832, "learning_rate": 8.871598480909945e-06, "loss": 0.8238, "step": 975 }, { "epoch": 0.2239140170174653, "grad_norm": 8.113872528076172, "learning_rate": 8.815360519684579e-06, "loss": 1.0677, "step": 1000 }, { "epoch": 0.22951186744290192, "grad_norm": 7.616383075714111, "learning_rate": 8.757942600012008e-06, "loss": 1.0761, "step": 1025 }, { "epoch": 0.23510971786833856, "grad_norm": 10.469049453735352, "learning_rate": 8.699362479259847e-06, "loss": 1.0677, "step": 1050 }, { "epoch": 0.24070756829377518, "grad_norm": 10.512489318847656, "learning_rate": 8.639638274224127e-06, "loss": 0.9852, "step": 1075 }, { "epoch": 0.24630541871921183, "grad_norm": 14.909335136413574, "learning_rate": 8.578788455526398e-06, "loss": 0.9895, "step": 1100 }, { "epoch": 0.25190326914464845, "grad_norm": 9.663930892944336, "learning_rate": 8.516831841901406e-06, "loss": 0.9607, "step": 1125 }, { "epoch": 0.2575011195700851, "grad_norm": 10.520059585571289, "learning_rate": 8.453787594377117e-06, "loss": 1.0332, "step": 1150 }, { "epoch": 0.26309896999552174, "grad_norm": 10.005462646484375, "learning_rate": 8.389675210348864e-06, "loss": 1.0453, "step": 1175 }, { "epoch": 0.2686968204209583, "grad_norm": 10.835626602172852, "learning_rate": 8.3245145175495e-06, "loss": 1.0625, "step": 1200 }, { "epoch": 0.274294670846395, "grad_norm": 5.667449951171875, "learning_rate": 8.258325667917355e-06, "loss": 0.8611, "step": 1225 }, { "epoch": 0.2798925212718316, "grad_norm": 13.251060485839844, "learning_rate": 8.191129131363942e-06, "loss": 1.0864, "step": 1250 }, { "epoch": 0.28549037169726826, "grad_norm": 7.886396884918213, "learning_rate": 8.122945689443328e-06, "loss": 0.8904, "step": 1275 }, { "epoch": 0.29108822212270485, "grad_norm": 9.646678924560547, "learning_rate": 8.053796428925123e-06, "loss": 0.9213, "step": 1300 }, { "epoch": 0.2966860725481415, "grad_norm": 10.207015037536621, "learning_rate": 7.98370273527307e-06, "loss": 1.0658, "step": 1325 }, { "epoch": 0.30228392297357815, "grad_norm": 6.621966361999512, "learning_rate": 7.91268628603127e-06, "loss": 1.1713, "step": 1350 }, { "epoch": 0.3078817733990148, "grad_norm": 7.847959518432617, "learning_rate": 7.840769044120067e-06, "loss": 0.9781, "step": 1375 }, { "epoch": 0.31347962382445144, "grad_norm": 9.040787696838379, "learning_rate": 7.76797325104368e-06, "loss": 0.9718, "step": 1400 }, { "epoch": 0.319077474249888, "grad_norm": 6.201238632202148, "learning_rate": 7.694321420011672e-06, "loss": 0.9015, "step": 1425 }, { "epoch": 0.3246753246753247, "grad_norm": 14.064117431640625, "learning_rate": 7.619836328976416e-06, "loss": 1.0073, "step": 1450 }, { "epoch": 0.3302731751007613, "grad_norm": 6.1384663581848145, "learning_rate": 7.5445410135886455e-06, "loss": 0.8406, "step": 1475 }, { "epoch": 0.33587102552619796, "grad_norm": 7.594089508056641, "learning_rate": 7.468458760073334e-06, "loss": 0.8915, "step": 1500 }, { "epoch": 0.34146887595163455, "grad_norm": 9.95975399017334, "learning_rate": 7.391613098028081e-06, "loss": 1.1634, "step": 1525 }, { "epoch": 0.3470667263770712, "grad_norm": 10.708215713500977, "learning_rate": 7.314027793146219e-06, "loss": 0.9969, "step": 1550 }, { "epoch": 0.35266457680250785, "grad_norm": 9.816734313964844, "learning_rate": 7.23572683986691e-06, "loss": 1.1063, "step": 1575 }, { "epoch": 0.3582624272279445, "grad_norm": 10.419496536254883, "learning_rate": 7.156734453954503e-06, "loss": 1.0399, "step": 1600 }, { "epoch": 0.3638602776533811, "grad_norm": 7.747687339782715, "learning_rate": 7.0770750650094335e-06, "loss": 0.8368, "step": 1625 }, { "epoch": 0.3694581280788177, "grad_norm": 9.370880126953125, "learning_rate": 6.996773308913003e-06, "loss": 1.0139, "step": 1650 }, { "epoch": 0.3750559785042544, "grad_norm": 10.095951080322266, "learning_rate": 6.915854020208358e-06, "loss": 1.0617, "step": 1675 }, { "epoch": 0.380653828929691, "grad_norm": 11.371135711669922, "learning_rate": 6.834342224420016e-06, "loss": 1.0399, "step": 1700 }, { "epoch": 0.3862516793551276, "grad_norm": 9.077759742736816, "learning_rate": 6.752263130314361e-06, "loss": 1.0934, "step": 1725 }, { "epoch": 0.39184952978056425, "grad_norm": 8.414709091186523, "learning_rate": 6.669642122103423e-06, "loss": 0.9078, "step": 1750 }, { "epoch": 0.3974473802060009, "grad_norm": 7.037852764129639, "learning_rate": 6.586504751594451e-06, "loss": 1.0392, "step": 1775 }, { "epoch": 0.40304523063143755, "grad_norm": 12.552664756774902, "learning_rate": 6.5028767302875974e-06, "loss": 0.9487, "step": 1800 }, { "epoch": 0.40864308105687414, "grad_norm": 6.407071113586426, "learning_rate": 6.418783921424269e-06, "loss": 0.9555, "step": 1825 }, { "epoch": 0.4142409314823108, "grad_norm": 5.643894672393799, "learning_rate": 6.334252331988515e-06, "loss": 1.074, "step": 1850 }, { "epoch": 0.4198387819077474, "grad_norm": 7.254835605621338, "learning_rate": 6.249308104663966e-06, "loss": 0.9496, "step": 1875 }, { "epoch": 0.4254366323331841, "grad_norm": 8.206978797912598, "learning_rate": 6.1639775097488155e-06, "loss": 1.0772, "step": 1900 }, { "epoch": 0.43103448275862066, "grad_norm": 4.874852657318115, "learning_rate": 6.07828693703133e-06, "loss": 0.8274, "step": 1925 }, { "epoch": 0.4366323331840573, "grad_norm": 14.283883094787598, "learning_rate": 5.992262887628405e-06, "loss": 1.1204, "step": 1950 }, { "epoch": 0.44223018360949395, "grad_norm": 12.45128059387207, "learning_rate": 5.905931965789688e-06, "loss": 1.0209, "step": 1975 }, { "epoch": 0.4478280340349306, "grad_norm": 9.155681610107422, "learning_rate": 5.819320870669806e-06, "loss": 1.0208, "step": 2000 }, { "epoch": 0.45342588446036725, "grad_norm": 7.600436210632324, "learning_rate": 5.732456388071247e-06, "loss": 0.9395, "step": 2025 }, { "epoch": 0.45902373488580384, "grad_norm": 10.021013259887695, "learning_rate": 5.645365382160446e-06, "loss": 0.9514, "step": 2050 }, { "epoch": 0.4646215853112405, "grad_norm": 10.850826263427734, "learning_rate": 5.558074787159629e-06, "loss": 1.0362, "step": 2075 }, { "epoch": 0.4702194357366771, "grad_norm": 8.926505088806152, "learning_rate": 5.470611599016986e-06, "loss": 0.9716, "step": 2100 }, { "epoch": 0.4758172861621138, "grad_norm": 6.5484185218811035, "learning_rate": 5.383002867057778e-06, "loss": 1.0704, "step": 2125 }, { "epoch": 0.48141513658755036, "grad_norm": 7.495320796966553, "learning_rate": 5.295275685618905e-06, "loss": 0.9566, "step": 2150 }, { "epoch": 0.487012987012987, "grad_norm": 8.97859001159668, "learning_rate": 5.2074571856695776e-06, "loss": 0.9744, "step": 2175 }, { "epoch": 0.49261083743842365, "grad_norm": 8.450396537780762, "learning_rate": 5.119574526420652e-06, "loss": 1.0064, "step": 2200 }, { "epoch": 0.4982086878638603, "grad_norm": 15.110372543334961, "learning_rate": 5.0316548869252205e-06, "loss": 0.9724, "step": 2225 }, { "epoch": 0.5038065382892969, "grad_norm": 11.507658004760742, "learning_rate": 4.943725457673086e-06, "loss": 1.0204, "step": 2250 }, { "epoch": 0.5094043887147336, "grad_norm": 5.711903095245361, "learning_rate": 4.8558134321816816e-06, "loss": 1.0455, "step": 2275 }, { "epoch": 0.5150022391401702, "grad_norm": 7.272406101226807, "learning_rate": 4.76794599858606e-06, "loss": 1.0133, "step": 2300 }, { "epoch": 0.5206000895656068, "grad_norm": 6.9339680671691895, "learning_rate": 4.680150331230552e-06, "loss": 1.1114, "step": 2325 }, { "epoch": 0.5261979399910435, "grad_norm": 6.307779788970947, "learning_rate": 4.592453582264684e-06, "loss": 0.8829, "step": 2350 }, { "epoch": 0.5317957904164801, "grad_norm": 4.144500255584717, "learning_rate": 4.504882873245961e-06, "loss": 1.0952, "step": 2375 }, { "epoch": 0.5373936408419167, "grad_norm": 8.365829467773438, "learning_rate": 4.417465286752109e-06, "loss": 1.0237, "step": 2400 }, { "epoch": 0.5429914912673534, "grad_norm": 9.530734062194824, "learning_rate": 4.330227858005379e-06, "loss": 1.0088, "step": 2425 }, { "epoch": 0.54858934169279, "grad_norm": 9.499034881591797, "learning_rate": 4.243197566511493e-06, "loss": 1.0404, "step": 2450 }, { "epoch": 0.5541871921182266, "grad_norm": 6.624443054199219, "learning_rate": 4.15640132771581e-06, "loss": 0.802, "step": 2475 }, { "epoch": 0.5597850425436632, "grad_norm": 4.885298252105713, "learning_rate": 4.069865984679332e-06, "loss": 0.9824, "step": 2500 }, { "epoch": 0.5653828929690998, "grad_norm": 10.335043907165527, "learning_rate": 3.9836182997770586e-06, "loss": 1.0334, "step": 2525 }, { "epoch": 0.5709807433945365, "grad_norm": 6.98168420791626, "learning_rate": 3.897684946421326e-06, "loss": 1.0047, "step": 2550 }, { "epoch": 0.5765785938199731, "grad_norm": 7.940051555633545, "learning_rate": 3.8120925008126457e-06, "loss": 0.8648, "step": 2575 }, { "epoch": 0.5821764442454097, "grad_norm": 7.1400885581970215, "learning_rate": 3.7268674337206025e-06, "loss": 0.8521, "step": 2600 }, { "epoch": 0.5877742946708464, "grad_norm": 9.360673904418945, "learning_rate": 3.6420361022973784e-06, "loss": 0.8878, "step": 2625 }, { "epoch": 0.593372145096283, "grad_norm": 6.841546058654785, "learning_rate": 3.5576247419263854e-06, "loss": 0.9828, "step": 2650 }, { "epoch": 0.5989699955217197, "grad_norm": 9.706862449645996, "learning_rate": 3.4736594581085837e-06, "loss": 1.1297, "step": 2675 }, { "epoch": 0.6045678459471563, "grad_norm": 9.606383323669434, "learning_rate": 3.390166218388956e-06, "loss": 1.0531, "step": 2700 }, { "epoch": 0.6101656963725929, "grad_norm": 5.520534515380859, "learning_rate": 3.3071708443256414e-06, "loss": 0.9844, "step": 2725 }, { "epoch": 0.6157635467980296, "grad_norm": 9.2130765914917, "learning_rate": 3.224699003504236e-06, "loss": 0.9696, "step": 2750 }, { "epoch": 0.6213613972234662, "grad_norm": 8.891196250915527, "learning_rate": 3.142776201599689e-06, "loss": 0.8803, "step": 2775 }, { "epoch": 0.6269592476489029, "grad_norm": 9.511999130249023, "learning_rate": 3.061427774488287e-06, "loss": 1.0108, "step": 2800 }, { "epoch": 0.6325570980743395, "grad_norm": 5.509415626525879, "learning_rate": 2.9806788804121562e-06, "loss": 0.8862, "step": 2825 }, { "epoch": 0.638154948499776, "grad_norm": 5.889389991760254, "learning_rate": 2.9005544921986774e-06, "loss": 0.9829, "step": 2850 }, { "epoch": 0.6437527989252128, "grad_norm": 8.13467788696289, "learning_rate": 2.8210793895372722e-06, "loss": 0.8431, "step": 2875 }, { "epoch": 0.6493506493506493, "grad_norm": 10.089922904968262, "learning_rate": 2.742278151315898e-06, "loss": 1.1452, "step": 2900 }, { "epoch": 0.6549484997760859, "grad_norm": 9.242680549621582, "learning_rate": 2.6641751480196485e-06, "loss": 0.906, "step": 2925 }, { "epoch": 0.6605463502015226, "grad_norm": 11.566295623779297, "learning_rate": 2.5867945341938255e-06, "loss": 0.941, "step": 2950 }, { "epoch": 0.6661442006269592, "grad_norm": 8.217988014221191, "learning_rate": 2.510160240973757e-06, "loss": 0.9058, "step": 2975 }, { "epoch": 0.6717420510523959, "grad_norm": 13.650288581848145, "learning_rate": 2.434295968683741e-06, "loss": 1.1236, "step": 3000 }, { "epoch": 0.6773399014778325, "grad_norm": 9.068547248840332, "learning_rate": 2.3592251795073564e-06, "loss": 0.9533, "step": 3025 }, { "epoch": 0.6829377519032691, "grad_norm": 10.866704940795898, "learning_rate": 2.2849710902314205e-06, "loss": 1.0299, "step": 3050 }, { "epoch": 0.6885356023287058, "grad_norm": 9.991189002990723, "learning_rate": 2.211556665065854e-06, "loss": 0.9484, "step": 3075 }, { "epoch": 0.6941334527541424, "grad_norm": 7.939027786254883, "learning_rate": 2.1390046085416356e-06, "loss": 0.8995, "step": 3100 }, { "epoch": 0.699731303179579, "grad_norm": 10.984658241271973, "learning_rate": 2.0673373584890847e-06, "loss": 1.0449, "step": 3125 }, { "epoch": 0.7053291536050157, "grad_norm": 6.965851306915283, "learning_rate": 1.996577079098628e-06, "loss": 0.976, "step": 3150 }, { "epoch": 0.7109270040304523, "grad_norm": 10.873515129089355, "learning_rate": 1.9267456540661723e-06, "loss": 1.0484, "step": 3175 }, { "epoch": 0.716524854455889, "grad_norm": 8.086957931518555, "learning_rate": 1.8578646798252432e-06, "loss": 0.8454, "step": 3200 }, { "epoch": 0.7221227048813256, "grad_norm": 9.924606323242188, "learning_rate": 1.7899554588679636e-06, "loss": 1.0399, "step": 3225 }, { "epoch": 0.7277205553067622, "grad_norm": 7.758438587188721, "learning_rate": 1.7230389931569242e-06, "loss": 0.9454, "step": 3250 }, { "epoch": 0.7333184057321989, "grad_norm": 6.910051345825195, "learning_rate": 1.657135977630015e-06, "loss": 0.9082, "step": 3275 }, { "epoch": 0.7389162561576355, "grad_norm": 8.083282470703125, "learning_rate": 1.5922667938001913e-06, "loss": 0.958, "step": 3300 }, { "epoch": 0.7445141065830722, "grad_norm": 10.176769256591797, "learning_rate": 1.5284515034521858e-06, "loss": 0.9261, "step": 3325 }, { "epoch": 0.7501119570085087, "grad_norm": 8.147079467773438, "learning_rate": 1.4657098424380945e-06, "loss": 0.9271, "step": 3350 }, { "epoch": 0.7557098074339453, "grad_norm": 9.081329345703125, "learning_rate": 1.4040612145737608e-06, "loss": 1.031, "step": 3375 }, { "epoch": 0.761307657859382, "grad_norm": 8.017196655273438, "learning_rate": 1.3435246856378524e-06, "loss": 1.0238, "step": 3400 }, { "epoch": 0.7669055082848186, "grad_norm": 9.929486274719238, "learning_rate": 1.284118977475481e-06, "loss": 0.954, "step": 3425 }, { "epoch": 0.7725033587102552, "grad_norm": 12.716034889221191, "learning_rate": 1.2258624622081755e-06, "loss": 0.9619, "step": 3450 }, { "epoch": 0.7781012091356919, "grad_norm": 8.163155555725098, "learning_rate": 1.1687731565520372e-06, "loss": 1.1041, "step": 3475 }, { "epoch": 0.7836990595611285, "grad_norm": 9.499692916870117, "learning_rate": 1.112868716245783e-06, "loss": 0.8927, "step": 3500 }, { "epoch": 0.7892969099865652, "grad_norm": 6.03558874130249, "learning_rate": 1.058166430590446e-06, "loss": 0.9804, "step": 3525 }, { "epoch": 0.7948947604120018, "grad_norm": 8.300068855285645, "learning_rate": 1.0046832171023952e-06, "loss": 1.0007, "step": 3550 }, { "epoch": 0.8004926108374384, "grad_norm": 7.682162761688232, "learning_rate": 9.524356162813326e-07, "loss": 0.9482, "step": 3575 }, { "epoch": 0.8060904612628751, "grad_norm": 8.57835578918457, "learning_rate": 9.014397864948921e-07, "loss": 0.9541, "step": 3600 }, { "epoch": 0.8116883116883117, "grad_norm": 11.44987678527832, "learning_rate": 8.51711498981419e-07, "loss": 0.9518, "step": 3625 }, { "epoch": 0.8172861621137483, "grad_norm": 7.918177604675293, "learning_rate": 8.032661329724717e-07, "loss": 1.0321, "step": 3650 }, { "epoch": 0.822884012539185, "grad_norm": 10.909501075744629, "learning_rate": 7.561186709365653e-07, "loss": 0.9651, "step": 3675 }, { "epoch": 0.8284818629646216, "grad_norm": 14.997469902038574, "learning_rate": 7.102836939456071e-07, "loss": 1.0462, "step": 3700 }, { "epoch": 0.8340797133900583, "grad_norm": 5.220563888549805, "learning_rate": 6.657753771654812e-07, "loss": 0.8815, "step": 3725 }, { "epoch": 0.8396775638154949, "grad_norm": 8.45197868347168, "learning_rate": 6.226074854721653e-07, "loss": 1.0041, "step": 3750 }, { "epoch": 0.8452754142409314, "grad_norm": 11.861766815185547, "learning_rate": 5.807933691947248e-07, "loss": 0.8721, "step": 3775 }, { "epoch": 0.8508732646663681, "grad_norm": 6.297611236572266, "learning_rate": 5.403459599865307e-07, "loss": 1.0609, "step": 3800 }, { "epoch": 0.8564711150918047, "grad_norm": 9.576262474060059, "learning_rate": 5.012777668259378e-07, "loss": 0.9142, "step": 3825 }, { "epoch": 0.8620689655172413, "grad_norm": 8.582077980041504, "learning_rate": 4.6360087214769923e-07, "loss": 0.9383, "step": 3850 }, { "epoch": 0.867666815942678, "grad_norm": 9.107322692871094, "learning_rate": 4.2732692810628583e-07, "loss": 1.1366, "step": 3875 }, { "epoch": 0.8732646663681146, "grad_norm": 5.725574493408203, "learning_rate": 3.9246715297228176e-07, "loss": 1.0077, "step": 3900 }, { "epoch": 0.8788625167935513, "grad_norm": 6.691014766693115, "learning_rate": 3.59032327662962e-07, "loss": 1.0128, "step": 3925 }, { "epoch": 0.8844603672189879, "grad_norm": 7.009856700897217, "learning_rate": 3.270327924081301e-07, "loss": 1.0404, "step": 3950 }, { "epoch": 0.8900582176444245, "grad_norm": 9.100337982177734, "learning_rate": 2.964784435522422e-07, "loss": 0.9983, "step": 3975 }, { "epoch": 0.8956560680698612, "grad_norm": 11.176060676574707, "learning_rate": 2.6737873049381523e-07, "loss": 0.9483, "step": 4000 }, { "epoch": 0.9012539184952978, "grad_norm": 6.4081268310546875, "learning_rate": 2.3974265276305253e-07, "loss": 1.0171, "step": 4025 }, { "epoch": 0.9068517689207345, "grad_norm": 5.774712562561035, "learning_rate": 2.1357875723860222e-07, "loss": 0.8993, "step": 4050 }, { "epoch": 0.9124496193461711, "grad_norm": 9.53131103515625, "learning_rate": 1.8889513550430892e-07, "loss": 0.9881, "step": 4075 }, { "epoch": 0.9180474697716077, "grad_norm": 7.965620040893555, "learning_rate": 1.656994213467622e-07, "loss": 0.9479, "step": 4100 }, { "epoch": 0.9236453201970444, "grad_norm": 9.653414726257324, "learning_rate": 1.439987883944355e-07, "loss": 0.9613, "step": 4125 }, { "epoch": 0.929243170622481, "grad_norm": 9.446654319763184, "learning_rate": 1.237999478991303e-07, "loss": 1.1298, "step": 4150 }, { "epoch": 0.9348410210479176, "grad_norm": 9.220303535461426, "learning_rate": 1.0510914666041927e-07, "loss": 0.9897, "step": 4175 }, { "epoch": 0.9404388714733543, "grad_norm": 9.508859634399414, "learning_rate": 8.793216509373038e-08, "loss": 1.0607, "step": 4200 }, { "epoch": 0.9460367218987908, "grad_norm": 8.090096473693848, "learning_rate": 7.227431544266194e-08, "loss": 1.0712, "step": 4225 }, { "epoch": 0.9516345723242275, "grad_norm": 5.4540510177612305, "learning_rate": 5.8140440136091326e-08, "loss": 1.0226, "step": 4250 }, { "epoch": 0.9572324227496641, "grad_norm": 5.08390474319458, "learning_rate": 4.553491029058221e-08, "loss": 0.8554, "step": 4275 }, { "epoch": 0.9628302731751007, "grad_norm": 11.732171058654785, "learning_rate": 3.4461624358546056e-08, "loss": 1.0033, "step": 4300 }, { "epoch": 0.9684281236005374, "grad_norm": 8.629095077514648, "learning_rate": 2.4924006922590338e-08, "loss": 0.973, "step": 4325 }, { "epoch": 0.974025974025974, "grad_norm": 6.797715663909912, "learning_rate": 1.6925007636411362e-08, "loss": 0.8825, "step": 4350 }, { "epoch": 0.9796238244514106, "grad_norm": 10.467803001403809, "learning_rate": 1.0467100312568923e-08, "loss": 1.0412, "step": 4375 }, { "epoch": 0.9852216748768473, "grad_norm": 10.413917541503906, "learning_rate": 5.552282157424427e-09, "loss": 0.9493, "step": 4400 }, { "epoch": 0.9908195253022839, "grad_norm": 6.7092719078063965, "learning_rate": 2.182073153471631e-09, "loss": 0.986, "step": 4425 }, { "epoch": 0.9964173757277206, "grad_norm": 6.736852645874023, "learning_rate": 3.575155892604487e-10, "loss": 0.9939, "step": 4450 } ], "logging_steps": 25, "max_steps": 4466, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.4169358435549184e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }