{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2111, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.014214641080312722, "grad_norm": 2.834075681079323, "learning_rate": 4.2452830188679244e-07, "loss": 0.8888, "step": 10 }, { "epoch": 0.028429282160625444, "grad_norm": 2.382206917176712, "learning_rate": 8.962264150943397e-07, "loss": 0.8863, "step": 20 }, { "epoch": 0.042643923240938165, "grad_norm": 1.5553139981482849, "learning_rate": 1.3679245283018869e-06, "loss": 0.8307, "step": 30 }, { "epoch": 0.05685856432125089, "grad_norm": 1.2537358335844357, "learning_rate": 1.839622641509434e-06, "loss": 0.7826, "step": 40 }, { "epoch": 0.07107320540156362, "grad_norm": 0.766804883609826, "learning_rate": 2.3113207547169815e-06, "loss": 0.7347, "step": 50 }, { "epoch": 0.08528784648187633, "grad_norm": 0.7667121953750998, "learning_rate": 2.7830188679245286e-06, "loss": 0.6931, "step": 60 }, { "epoch": 0.09950248756218906, "grad_norm": 0.6131128302034111, "learning_rate": 3.2547169811320758e-06, "loss": 0.6639, "step": 70 }, { "epoch": 0.11371712864250177, "grad_norm": 0.6988128095327644, "learning_rate": 3.726415094339623e-06, "loss": 0.6375, "step": 80 }, { "epoch": 0.1279317697228145, "grad_norm": 0.6317415794697651, "learning_rate": 4.19811320754717e-06, "loss": 0.6236, "step": 90 }, { "epoch": 0.14214641080312723, "grad_norm": 0.6086478743343868, "learning_rate": 4.6698113207547175e-06, "loss": 0.6138, "step": 100 }, { "epoch": 0.15636105188343993, "grad_norm": 0.5886423623895793, "learning_rate": 5.1415094339622655e-06, "loss": 0.6086, "step": 110 }, { "epoch": 0.17057569296375266, "grad_norm": 0.6688530721380745, "learning_rate": 5.613207547169813e-06, "loss": 0.603, "step": 120 }, { "epoch": 0.1847903340440654, "grad_norm": 0.6884356457949358, "learning_rate": 6.08490566037736e-06, "loss": 0.5903, "step": 130 }, { "epoch": 0.19900497512437812, "grad_norm": 0.609548931554606, "learning_rate": 6.556603773584907e-06, "loss": 0.5966, "step": 140 }, { "epoch": 0.21321961620469082, "grad_norm": 0.5949327566115654, "learning_rate": 7.028301886792454e-06, "loss": 0.5904, "step": 150 }, { "epoch": 0.22743425728500355, "grad_norm": 0.6242334967031955, "learning_rate": 7.500000000000001e-06, "loss": 0.579, "step": 160 }, { "epoch": 0.24164889836531628, "grad_norm": 0.6241786288297929, "learning_rate": 7.971698113207547e-06, "loss": 0.5791, "step": 170 }, { "epoch": 0.255863539445629, "grad_norm": 0.6521923298448227, "learning_rate": 8.443396226415095e-06, "loss": 0.5772, "step": 180 }, { "epoch": 0.27007818052594174, "grad_norm": 0.72406379743765, "learning_rate": 8.915094339622642e-06, "loss": 0.5783, "step": 190 }, { "epoch": 0.28429282160625446, "grad_norm": 0.6865581822637807, "learning_rate": 9.38679245283019e-06, "loss": 0.5667, "step": 200 }, { "epoch": 0.29850746268656714, "grad_norm": 0.7370553596384785, "learning_rate": 9.858490566037736e-06, "loss": 0.5787, "step": 210 }, { "epoch": 0.31272210376687987, "grad_norm": 0.680822415396935, "learning_rate": 9.999665093340166e-06, "loss": 0.5693, "step": 220 }, { "epoch": 0.3269367448471926, "grad_norm": 0.7111908345205171, "learning_rate": 9.998024842193876e-06, "loss": 0.5513, "step": 230 }, { "epoch": 0.3411513859275053, "grad_norm": 0.7236426819082362, "learning_rate": 9.995018180960701e-06, "loss": 0.5713, "step": 240 }, { "epoch": 0.35536602700781805, "grad_norm": 0.5852089481414202, "learning_rate": 9.990645931631796e-06, "loss": 0.5538, "step": 250 }, { "epoch": 0.3695806680881308, "grad_norm": 0.6824815552672151, "learning_rate": 9.984909289536473e-06, "loss": 0.5532, "step": 260 }, { "epoch": 0.3837953091684435, "grad_norm": 0.6415502778484192, "learning_rate": 9.9778098230154e-06, "loss": 0.5651, "step": 270 }, { "epoch": 0.39800995024875624, "grad_norm": 0.6951097230627817, "learning_rate": 9.969349472991838e-06, "loss": 0.5545, "step": 280 }, { "epoch": 0.41222459132906897, "grad_norm": 0.6940350949391038, "learning_rate": 9.959530552441006e-06, "loss": 0.5618, "step": 290 }, { "epoch": 0.42643923240938164, "grad_norm": 0.6795280607455072, "learning_rate": 9.94835574575774e-06, "loss": 0.5678, "step": 300 }, { "epoch": 0.44065387348969437, "grad_norm": 0.740615761131801, "learning_rate": 9.93582810802261e-06, "loss": 0.5589, "step": 310 }, { "epoch": 0.4548685145700071, "grad_norm": 0.6190142422707086, "learning_rate": 9.921951064166685e-06, "loss": 0.5545, "step": 320 }, { "epoch": 0.4690831556503198, "grad_norm": 0.6133098155280035, "learning_rate": 9.90672840803519e-06, "loss": 0.5424, "step": 330 }, { "epoch": 0.48329779673063256, "grad_norm": 0.6671428298287642, "learning_rate": 9.890164301350318e-06, "loss": 0.5543, "step": 340 }, { "epoch": 0.4975124378109453, "grad_norm": 0.6195386955465261, "learning_rate": 9.872263272573443e-06, "loss": 0.5436, "step": 350 }, { "epoch": 0.511727078891258, "grad_norm": 0.5930299278047146, "learning_rate": 9.853030215667095e-06, "loss": 0.5419, "step": 360 }, { "epoch": 0.5259417199715707, "grad_norm": 0.5808813232166387, "learning_rate": 9.832470388756987e-06, "loss": 0.5355, "step": 370 }, { "epoch": 0.5401563610518835, "grad_norm": 0.6451836622188146, "learning_rate": 9.81058941269451e-06, "loss": 0.5411, "step": 380 }, { "epoch": 0.5543710021321961, "grad_norm": 0.6430253547664657, "learning_rate": 9.787393269520039e-06, "loss": 0.5424, "step": 390 }, { "epoch": 0.5685856432125089, "grad_norm": 0.6434468036531783, "learning_rate": 9.762888300827507e-06, "loss": 0.5375, "step": 400 }, { "epoch": 0.5828002842928216, "grad_norm": 0.6545071562387185, "learning_rate": 9.737081206030671e-06, "loss": 0.5392, "step": 410 }, { "epoch": 0.5970149253731343, "grad_norm": 0.6520857107905386, "learning_rate": 9.709979040531568e-06, "loss": 0.5394, "step": 420 }, { "epoch": 0.6112295664534471, "grad_norm": 0.6850212115229948, "learning_rate": 9.681589213791633e-06, "loss": 0.5466, "step": 430 }, { "epoch": 0.6254442075337597, "grad_norm": 0.6269054102432758, "learning_rate": 9.651919487306025e-06, "loss": 0.5322, "step": 440 }, { "epoch": 0.6396588486140725, "grad_norm": 0.736787330144809, "learning_rate": 9.620977972481715e-06, "loss": 0.5429, "step": 450 }, { "epoch": 0.6538734896943852, "grad_norm": 0.662925500196175, "learning_rate": 9.588773128419907e-06, "loss": 0.5312, "step": 460 }, { "epoch": 0.668088130774698, "grad_norm": 0.6148143038929397, "learning_rate": 9.555313759603403e-06, "loss": 0.5316, "step": 470 }, { "epoch": 0.6823027718550106, "grad_norm": 0.6077969318281708, "learning_rate": 9.520609013489548e-06, "loss": 0.5223, "step": 480 }, { "epoch": 0.6965174129353234, "grad_norm": 0.6168040167150479, "learning_rate": 9.484668378009407e-06, "loss": 0.5405, "step": 490 }, { "epoch": 0.7107320540156361, "grad_norm": 0.6128096552430267, "learning_rate": 9.447501678973853e-06, "loss": 0.5338, "step": 500 }, { "epoch": 0.7107320540156361, "eval_loss": 0.5341117978096008, "eval_runtime": 234.2375, "eval_samples_per_second": 21.346, "eval_steps_per_second": 2.668, "step": 500 }, { "epoch": 0.7249466950959488, "grad_norm": 0.6234640184402841, "learning_rate": 9.409119077387295e-06, "loss": 0.5377, "step": 510 }, { "epoch": 0.7391613361762616, "grad_norm": 0.6131203844453222, "learning_rate": 9.369531066669759e-06, "loss": 0.5259, "step": 520 }, { "epoch": 0.7533759772565742, "grad_norm": 0.6262740746023064, "learning_rate": 9.328748469788094e-06, "loss": 0.5328, "step": 530 }, { "epoch": 0.767590618336887, "grad_norm": 0.6262806652099724, "learning_rate": 9.286782436297072e-06, "loss": 0.5344, "step": 540 }, { "epoch": 0.7818052594171997, "grad_norm": 0.6615647825526851, "learning_rate": 9.243644439291223e-06, "loss": 0.5257, "step": 550 }, { "epoch": 0.7960199004975125, "grad_norm": 0.6072216321225935, "learning_rate": 9.1993462722682e-06, "loss": 0.5339, "step": 560 }, { "epoch": 0.8102345415778252, "grad_norm": 0.5963884112129639, "learning_rate": 9.15390004590455e-06, "loss": 0.5235, "step": 570 }, { "epoch": 0.8244491826581379, "grad_norm": 0.6479032652506073, "learning_rate": 9.107318184744782e-06, "loss": 0.5244, "step": 580 }, { "epoch": 0.8386638237384506, "grad_norm": 0.6241228532450154, "learning_rate": 9.059613423804623e-06, "loss": 0.5279, "step": 590 }, { "epoch": 0.8528784648187633, "grad_norm": 0.6346658666864172, "learning_rate": 9.010798805089385e-06, "loss": 0.5315, "step": 600 }, { "epoch": 0.8670931058990761, "grad_norm": 0.6189100879810271, "learning_rate": 8.960887674028411e-06, "loss": 0.516, "step": 610 }, { "epoch": 0.8813077469793887, "grad_norm": 0.6310100666801625, "learning_rate": 8.909893675826575e-06, "loss": 0.5343, "step": 620 }, { "epoch": 0.8955223880597015, "grad_norm": 0.5705654281594257, "learning_rate": 8.857830751733815e-06, "loss": 0.5196, "step": 630 }, { "epoch": 0.9097370291400142, "grad_norm": 0.581798609611747, "learning_rate": 8.80471313523373e-06, "loss": 0.5252, "step": 640 }, { "epoch": 0.923951670220327, "grad_norm": 0.6021206642948125, "learning_rate": 8.750555348152299e-06, "loss": 0.5301, "step": 650 }, { "epoch": 0.9381663113006397, "grad_norm": 0.6258455776720466, "learning_rate": 8.695372196687743e-06, "loss": 0.5247, "step": 660 }, { "epoch": 0.9523809523809523, "grad_norm": 0.5984715139941876, "learning_rate": 8.639178767362677e-06, "loss": 0.5166, "step": 670 }, { "epoch": 0.9665955934612651, "grad_norm": 0.6174130149894701, "learning_rate": 8.581990422899586e-06, "loss": 0.5186, "step": 680 }, { "epoch": 0.9808102345415778, "grad_norm": 0.669919825928536, "learning_rate": 8.523822798020827e-06, "loss": 0.5232, "step": 690 }, { "epoch": 0.9950248756218906, "grad_norm": 0.5789470998307686, "learning_rate": 8.46469179517424e-06, "loss": 0.5128, "step": 700 }, { "epoch": 1.0085287846481876, "grad_norm": 0.6391934862223132, "learning_rate": 8.404613580185586e-06, "loss": 0.4571, "step": 710 }, { "epoch": 1.0227434257285004, "grad_norm": 0.646088830119043, "learning_rate": 8.343604577838965e-06, "loss": 0.4844, "step": 720 }, { "epoch": 1.036958066808813, "grad_norm": 0.5785525652959986, "learning_rate": 8.281681467386447e-06, "loss": 0.4731, "step": 730 }, { "epoch": 1.0511727078891258, "grad_norm": 0.6609376724143528, "learning_rate": 8.21886117798813e-06, "loss": 0.483, "step": 740 }, { "epoch": 1.0653873489694385, "grad_norm": 0.6092591984384597, "learning_rate": 8.155160884083881e-06, "loss": 0.4868, "step": 750 }, { "epoch": 1.0796019900497513, "grad_norm": 0.6374135491483588, "learning_rate": 8.090598000698009e-06, "loss": 0.492, "step": 760 }, { "epoch": 1.0938166311300639, "grad_norm": 0.6008076415379893, "learning_rate": 8.025190178678175e-06, "loss": 0.4814, "step": 770 }, { "epoch": 1.1080312722103767, "grad_norm": 0.6018194277584734, "learning_rate": 7.958955299869826e-06, "loss": 0.4768, "step": 780 }, { "epoch": 1.1222459132906895, "grad_norm": 0.5917444038650317, "learning_rate": 7.891911472227478e-06, "loss": 0.4841, "step": 790 }, { "epoch": 1.136460554371002, "grad_norm": 0.6728799540384178, "learning_rate": 7.82407702486418e-06, "loss": 0.4797, "step": 800 }, { "epoch": 1.1506751954513148, "grad_norm": 0.5791378894590185, "learning_rate": 7.755470503040516e-06, "loss": 0.4842, "step": 810 }, { "epoch": 1.1648898365316276, "grad_norm": 0.6318987068220235, "learning_rate": 7.686110663094527e-06, "loss": 0.4752, "step": 820 }, { "epoch": 1.1791044776119404, "grad_norm": 0.5836896353182887, "learning_rate": 7.616016467313891e-06, "loss": 0.4783, "step": 830 }, { "epoch": 1.193319118692253, "grad_norm": 0.6133166239448045, "learning_rate": 7.545207078751858e-06, "loss": 0.4804, "step": 840 }, { "epoch": 1.2075337597725657, "grad_norm": 0.6647464026501058, "learning_rate": 7.473701855988227e-06, "loss": 0.4837, "step": 850 }, { "epoch": 1.2217484008528785, "grad_norm": 0.6037242762628189, "learning_rate": 7.4015203478369266e-06, "loss": 0.48, "step": 860 }, { "epoch": 1.235963041933191, "grad_norm": 0.6060908786113385, "learning_rate": 7.328682288001561e-06, "loss": 0.4828, "step": 870 }, { "epoch": 1.2501776830135038, "grad_norm": 0.5748244127007777, "learning_rate": 7.255207589680403e-06, "loss": 0.4737, "step": 880 }, { "epoch": 1.2643923240938166, "grad_norm": 0.5739667849410425, "learning_rate": 7.181116340122336e-06, "loss": 0.4692, "step": 890 }, { "epoch": 1.2786069651741294, "grad_norm": 0.5356752158374642, "learning_rate": 7.10642879513519e-06, "loss": 0.467, "step": 900 }, { "epoch": 1.2928216062544422, "grad_norm": 0.6231787580072905, "learning_rate": 7.0311653735480136e-06, "loss": 0.4787, "step": 910 }, { "epoch": 1.3070362473347548, "grad_norm": 0.5539031871696426, "learning_rate": 6.95534665162877e-06, "loss": 0.4803, "step": 920 }, { "epoch": 1.3212508884150675, "grad_norm": 0.5545024040236839, "learning_rate": 6.878993357458986e-06, "loss": 0.4727, "step": 930 }, { "epoch": 1.33546552949538, "grad_norm": 0.5526343563080722, "learning_rate": 6.8021263652669055e-06, "loss": 0.4749, "step": 940 }, { "epoch": 1.349680170575693, "grad_norm": 0.5949065425774154, "learning_rate": 6.7247666897206795e-06, "loss": 0.466, "step": 950 }, { "epoch": 1.3638948116560057, "grad_norm": 0.6109499791395212, "learning_rate": 6.646935480183173e-06, "loss": 0.4721, "step": 960 }, { "epoch": 1.3781094527363185, "grad_norm": 0.5761776478178481, "learning_rate": 6.568654014929933e-06, "loss": 0.4773, "step": 970 }, { "epoch": 1.3923240938166312, "grad_norm": 0.6309785557491375, "learning_rate": 6.4899436953319235e-06, "loss": 0.4783, "step": 980 }, { "epoch": 1.4065387348969438, "grad_norm": 0.6079839523261087, "learning_rate": 6.410826040004607e-06, "loss": 0.4874, "step": 990 }, { "epoch": 1.4207533759772566, "grad_norm": 0.5694417979836612, "learning_rate": 6.331322678924963e-06, "loss": 0.4883, "step": 1000 }, { "epoch": 1.4207533759772566, "eval_loss": 0.5142297148704529, "eval_runtime": 233.9812, "eval_samples_per_second": 21.369, "eval_steps_per_second": 2.671, "step": 1000 }, { "epoch": 1.4349680170575694, "grad_norm": 0.590623858748834, "learning_rate": 6.251455347518074e-06, "loss": 0.4674, "step": 1010 }, { "epoch": 1.449182658137882, "grad_norm": 0.6492265919143085, "learning_rate": 6.1712458807148804e-06, "loss": 0.4566, "step": 1020 }, { "epoch": 1.4633972992181947, "grad_norm": 0.6690825078821129, "learning_rate": 6.090716206982714e-06, "loss": 0.4415, "step": 1030 }, { "epoch": 1.4776119402985075, "grad_norm": 0.6207866027112856, "learning_rate": 6.009888342330292e-06, "loss": 0.4419, "step": 1040 }, { "epoch": 1.4918265813788203, "grad_norm": 0.5983146438640481, "learning_rate": 5.92878438428875e-06, "loss": 0.4529, "step": 1050 }, { "epoch": 1.5060412224591329, "grad_norm": 0.6251595437559933, "learning_rate": 5.847426505870399e-06, "loss": 0.4433, "step": 1060 }, { "epoch": 1.5202558635394456, "grad_norm": 0.5973997328290039, "learning_rate": 5.765836949506843e-06, "loss": 0.4465, "step": 1070 }, { "epoch": 1.5344705046197582, "grad_norm": 0.61605264507367, "learning_rate": 5.684038020968126e-06, "loss": 0.4354, "step": 1080 }, { "epoch": 1.548685145700071, "grad_norm": 0.6356804350566382, "learning_rate": 5.6020520832645555e-06, "loss": 0.4474, "step": 1090 }, { "epoch": 1.5628997867803838, "grad_norm": 0.5409307423191971, "learning_rate": 5.519901550532871e-06, "loss": 0.4442, "step": 1100 }, { "epoch": 1.5771144278606966, "grad_norm": 0.5628516264470729, "learning_rate": 5.437608881908456e-06, "loss": 0.4469, "step": 1110 }, { "epoch": 1.5913290689410093, "grad_norm": 0.5646776831646609, "learning_rate": 5.3551965753852255e-06, "loss": 0.4444, "step": 1120 }, { "epoch": 1.6055437100213221, "grad_norm": 0.6070095692660494, "learning_rate": 5.2726871616649e-06, "loss": 0.4546, "step": 1130 }, { "epoch": 1.6197583511016347, "grad_norm": 0.5955403854226203, "learning_rate": 5.190103197997339e-06, "loss": 0.4453, "step": 1140 }, { "epoch": 1.6339729921819472, "grad_norm": 0.5847049727426836, "learning_rate": 5.107467262013614e-06, "loss": 0.4456, "step": 1150 }, { "epoch": 1.64818763326226, "grad_norm": 0.6464335190947482, "learning_rate": 5.02480194555351e-06, "loss": 0.4431, "step": 1160 }, { "epoch": 1.6624022743425728, "grad_norm": 0.5976768465351757, "learning_rate": 4.942129848489137e-06, "loss": 0.4459, "step": 1170 }, { "epoch": 1.6766169154228856, "grad_norm": 0.5759461226301025, "learning_rate": 4.8594735725463575e-06, "loss": 0.4331, "step": 1180 }, { "epoch": 1.6908315565031984, "grad_norm": 0.5618161675831925, "learning_rate": 4.776855715125694e-06, "loss": 0.4459, "step": 1190 }, { "epoch": 1.7050461975835112, "grad_norm": 0.5665638890009612, "learning_rate": 4.694298863124435e-06, "loss": 0.4434, "step": 1200 }, { "epoch": 1.7192608386638237, "grad_norm": 0.5897127590620418, "learning_rate": 4.611825586761591e-06, "loss": 0.4526, "step": 1210 }, { "epoch": 1.7334754797441365, "grad_norm": 0.5964155647113037, "learning_rate": 4.529458433407429e-06, "loss": 0.4404, "step": 1220 }, { "epoch": 1.747690120824449, "grad_norm": 0.557832612527273, "learning_rate": 4.447219921419244e-06, "loss": 0.4466, "step": 1230 }, { "epoch": 1.7619047619047619, "grad_norm": 0.6542060047375909, "learning_rate": 4.365132533985071e-06, "loss": 0.4456, "step": 1240 }, { "epoch": 1.7761194029850746, "grad_norm": 0.5917352620945386, "learning_rate": 4.283218712976992e-06, "loss": 0.4452, "step": 1250 }, { "epoch": 1.7903340440653874, "grad_norm": 0.561379370088737, "learning_rate": 4.201500852815769e-06, "loss": 0.442, "step": 1260 }, { "epoch": 1.8045486851457002, "grad_norm": 0.5491467540898044, "learning_rate": 4.12000129434842e-06, "loss": 0.4444, "step": 1270 }, { "epoch": 1.8187633262260128, "grad_norm": 0.5872206166430523, "learning_rate": 4.0387423187404656e-06, "loss": 0.4375, "step": 1280 }, { "epoch": 1.8329779673063256, "grad_norm": 0.5927050301786448, "learning_rate": 3.957746141384469e-06, "loss": 0.4471, "step": 1290 }, { "epoch": 1.8471926083866381, "grad_norm": 0.5414882779351676, "learning_rate": 3.877034905826577e-06, "loss": 0.4466, "step": 1300 }, { "epoch": 1.861407249466951, "grad_norm": 0.5612180051513147, "learning_rate": 3.796630677712697e-06, "loss": 0.4388, "step": 1310 }, { "epoch": 1.8756218905472637, "grad_norm": 0.5529411457142649, "learning_rate": 3.716555438755961e-06, "loss": 0.4472, "step": 1320 }, { "epoch": 1.8898365316275765, "grad_norm": 0.5616746991079874, "learning_rate": 3.6368310807271546e-06, "loss": 0.4446, "step": 1330 }, { "epoch": 1.9040511727078893, "grad_norm": 0.5799752713225496, "learning_rate": 3.557479399469721e-06, "loss": 0.4493, "step": 1340 }, { "epoch": 1.9182658137882018, "grad_norm": 0.6619814912951459, "learning_rate": 3.4785220889409934e-06, "loss": 0.4405, "step": 1350 }, { "epoch": 1.9324804548685146, "grad_norm": 0.5613562889297112, "learning_rate": 3.3999807352812862e-06, "loss": 0.4541, "step": 1360 }, { "epoch": 1.9466950959488272, "grad_norm": 0.5713920903616314, "learning_rate": 3.321876810912461e-06, "loss": 0.4428, "step": 1370 }, { "epoch": 1.96090973702914, "grad_norm": 0.5665905645368096, "learning_rate": 3.2442316686675783e-06, "loss": 0.4437, "step": 1380 }, { "epoch": 1.9751243781094527, "grad_norm": 0.595196616315948, "learning_rate": 3.1670665359532415e-06, "loss": 0.4438, "step": 1390 }, { "epoch": 1.9893390191897655, "grad_norm": 0.5920175988784884, "learning_rate": 3.090402508946249e-06, "loss": 0.4373, "step": 1400 }, { "epoch": 2.0042643923240937, "grad_norm": 0.5358117816177491, "learning_rate": 3.0142605468260976e-06, "loss": 0.4572, "step": 1410 }, { "epoch": 2.0184790334044065, "grad_norm": 0.5576529014623839, "learning_rate": 2.9386614660449598e-06, "loss": 0.4263, "step": 1420 }, { "epoch": 2.0326936744847193, "grad_norm": 0.5822207756082265, "learning_rate": 2.8636259346366666e-06, "loss": 0.4385, "step": 1430 }, { "epoch": 2.046908315565032, "grad_norm": 0.5829696226017161, "learning_rate": 2.7891744665662824e-06, "loss": 0.4218, "step": 1440 }, { "epoch": 2.061122956645345, "grad_norm": 0.581038802974167, "learning_rate": 2.7153274161217847e-06, "loss": 0.4404, "step": 1450 }, { "epoch": 2.0753375977256576, "grad_norm": 0.5347733423294494, "learning_rate": 2.642104972349403e-06, "loss": 0.4198, "step": 1460 }, { "epoch": 2.08955223880597, "grad_norm": 0.543588717662816, "learning_rate": 2.5695271535341443e-06, "loss": 0.4411, "step": 1470 }, { "epoch": 2.1037668798862827, "grad_norm": 0.5720716950974386, "learning_rate": 2.4976138017269906e-06, "loss": 0.4412, "step": 1480 }, { "epoch": 2.1179815209665955, "grad_norm": 0.5603303814513405, "learning_rate": 2.4263845773202738e-06, "loss": 0.4298, "step": 1490 }, { "epoch": 2.1321961620469083, "grad_norm": 0.5598020164471934, "learning_rate": 2.355858953672728e-06, "loss": 0.4397, "step": 1500 }, { "epoch": 2.1321961620469083, "eval_loss": 0.5144844651222229, "eval_runtime": 231.3372, "eval_samples_per_second": 21.613, "eval_steps_per_second": 2.702, "step": 1500 }, { "epoch": 2.146410803127221, "grad_norm": 0.552503902368828, "learning_rate": 2.286056211785665e-06, "loss": 0.4339, "step": 1510 }, { "epoch": 2.160625444207534, "grad_norm": 0.5606006250953287, "learning_rate": 2.2169954350317372e-06, "loss": 0.4408, "step": 1520 }, { "epoch": 2.1748400852878467, "grad_norm": 0.5478812953721691, "learning_rate": 2.148695503937745e-06, "loss": 0.4387, "step": 1530 }, { "epoch": 2.189054726368159, "grad_norm": 0.5327946624830445, "learning_rate": 2.081175091022877e-06, "loss": 0.4395, "step": 1540 }, { "epoch": 2.203269367448472, "grad_norm": 0.5793441763139854, "learning_rate": 2.014452655693839e-06, "loss": 0.4394, "step": 1550 }, { "epoch": 2.2174840085287846, "grad_norm": 0.5656062787295659, "learning_rate": 1.9485464391982282e-06, "loss": 0.4357, "step": 1560 }, { "epoch": 2.2316986496090974, "grad_norm": 0.575605417624959, "learning_rate": 1.8834744596375664e-06, "loss": 0.4364, "step": 1570 }, { "epoch": 2.24591329068941, "grad_norm": 0.5511255544375402, "learning_rate": 1.8192545070413281e-06, "loss": 0.4278, "step": 1580 }, { "epoch": 2.260127931769723, "grad_norm": 0.5527437049458982, "learning_rate": 1.755904138503316e-06, "loss": 0.4281, "step": 1590 }, { "epoch": 2.2743425728500357, "grad_norm": 0.5567466633261947, "learning_rate": 1.6934406733817417e-06, "loss": 0.4326, "step": 1600 }, { "epoch": 2.288557213930348, "grad_norm": 0.5542280645431442, "learning_rate": 1.6318811885642749e-06, "loss": 0.4426, "step": 1610 }, { "epoch": 2.302771855010661, "grad_norm": 0.5385541713578466, "learning_rate": 1.5712425137993976e-06, "loss": 0.4363, "step": 1620 }, { "epoch": 2.3169864960909736, "grad_norm": 0.5568902806764913, "learning_rate": 1.5115412270953166e-06, "loss": 0.4346, "step": 1630 }, { "epoch": 2.3312011371712864, "grad_norm": 0.6058817336949707, "learning_rate": 1.4527936501877033e-06, "loss": 0.4321, "step": 1640 }, { "epoch": 2.345415778251599, "grad_norm": 0.5318061336854996, "learning_rate": 1.3950158440774958e-06, "loss": 0.418, "step": 1650 }, { "epoch": 2.359630419331912, "grad_norm": 0.5553934307354601, "learning_rate": 1.3382236046399722e-06, "loss": 0.4143, "step": 1660 }, { "epoch": 2.3738450604122248, "grad_norm": 0.5429114693602697, "learning_rate": 1.2824324583063303e-06, "loss": 0.4413, "step": 1670 }, { "epoch": 2.388059701492537, "grad_norm": 0.5286255307861315, "learning_rate": 1.2276576578189065e-06, "loss": 0.4218, "step": 1680 }, { "epoch": 2.40227434257285, "grad_norm": 0.5513151927730859, "learning_rate": 1.1739141780612306e-06, "loss": 0.4278, "step": 1690 }, { "epoch": 2.4164889836531627, "grad_norm": 0.5523983447772243, "learning_rate": 1.1212167119640439e-06, "loss": 0.4364, "step": 1700 }, { "epoch": 2.4307036247334755, "grad_norm": 0.5707718364766191, "learning_rate": 1.069579666488395e-06, "loss": 0.429, "step": 1710 }, { "epoch": 2.4449182658137882, "grad_norm": 0.5620343463380818, "learning_rate": 1.0190171586869258e-06, "loss": 0.4239, "step": 1720 }, { "epoch": 2.459132906894101, "grad_norm": 0.5335024558606225, "learning_rate": 9.695430118444049e-07, "loss": 0.4199, "step": 1730 }, { "epoch": 2.473347547974414, "grad_norm": 0.5088660962728315, "learning_rate": 9.21170751698583e-07, "loss": 0.4294, "step": 1740 }, { "epoch": 2.487562189054726, "grad_norm": 0.5854213827338935, "learning_rate": 8.739136027423894e-07, "loss": 0.4407, "step": 1750 }, { "epoch": 2.501776830135039, "grad_norm": 0.570296065993969, "learning_rate": 8.277844846084898e-07, "loss": 0.4273, "step": 1760 }, { "epoch": 2.5159914712153517, "grad_norm": 0.5432303322927625, "learning_rate": 7.827960085371855e-07, "loss": 0.4372, "step": 1770 }, { "epoch": 2.5302061122956645, "grad_norm": 0.5462735332010124, "learning_rate": 7.389604739286271e-07, "loss": 0.4384, "step": 1780 }, { "epoch": 2.5444207533759773, "grad_norm": 0.5514867453544212, "learning_rate": 6.962898649802824e-07, "loss": 0.4279, "step": 1790 }, { "epoch": 2.55863539445629, "grad_norm": 0.5374953523760396, "learning_rate": 6.547958474105726e-07, "loss": 0.4207, "step": 1800 }, { "epoch": 2.572850035536603, "grad_norm": 0.5368534822557964, "learning_rate": 6.144897652695864e-07, "loss": 0.4425, "step": 1810 }, { "epoch": 2.587064676616915, "grad_norm": 0.5331851936220864, "learning_rate": 5.753826378377287e-07, "loss": 0.4227, "step": 1820 }, { "epoch": 2.6012793176972284, "grad_norm": 0.532292119043563, "learning_rate": 5.374851566131561e-07, "loss": 0.4332, "step": 1830 }, { "epoch": 2.6154939587775408, "grad_norm": 0.522004918322321, "learning_rate": 5.008076823888319e-07, "loss": 0.4363, "step": 1840 }, { "epoch": 2.6297085998578535, "grad_norm": 0.5747875880967467, "learning_rate": 4.653602424199877e-07, "loss": 0.441, "step": 1850 }, { "epoch": 2.6439232409381663, "grad_norm": 0.5701129333105889, "learning_rate": 4.3115252768276827e-07, "loss": 0.4281, "step": 1860 }, { "epoch": 2.658137882018479, "grad_norm": 0.5171567622519524, "learning_rate": 3.9819389022482226e-07, "loss": 0.439, "step": 1870 }, { "epoch": 2.672352523098792, "grad_norm": 0.5506453302407677, "learning_rate": 3.6649334060854027e-07, "loss": 0.4261, "step": 1880 }, { "epoch": 2.6865671641791042, "grad_norm": 0.5611217928624682, "learning_rate": 3.360595454476595e-07, "loss": 0.4296, "step": 1890 }, { "epoch": 2.7007818052594175, "grad_norm": 0.5693324256729151, "learning_rate": 3.069008250378974e-07, "loss": 0.4349, "step": 1900 }, { "epoch": 2.71499644633973, "grad_norm": 0.5448625830798527, "learning_rate": 2.7902515108226613e-07, "loss": 0.4167, "step": 1910 }, { "epoch": 2.7292110874200426, "grad_norm": 0.5178772865930678, "learning_rate": 2.5244014451168863e-07, "loss": 0.4335, "step": 1920 }, { "epoch": 2.7434257285003554, "grad_norm": 0.555692357887426, "learning_rate": 2.271530734015104e-07, "loss": 0.435, "step": 1930 }, { "epoch": 2.757640369580668, "grad_norm": 0.5575021836492726, "learning_rate": 2.0317085098448373e-07, "loss": 0.4222, "step": 1940 }, { "epoch": 2.771855010660981, "grad_norm": 0.5354458107562179, "learning_rate": 1.8050003376075708e-07, "loss": 0.4426, "step": 1950 }, { "epoch": 2.7860696517412933, "grad_norm": 0.5182277398666688, "learning_rate": 1.591468197053919e-07, "loss": 0.4359, "step": 1960 }, { "epoch": 2.8002842928216065, "grad_norm": 0.5422707453993397, "learning_rate": 1.3911704657390113e-07, "loss": 0.4243, "step": 1970 }, { "epoch": 2.814498933901919, "grad_norm": 0.5242061992724598, "learning_rate": 1.2041619030626283e-07, "loss": 0.4323, "step": 1980 }, { "epoch": 2.8287135749822316, "grad_norm": 0.5113276517453622, "learning_rate": 1.0304936352985351e-07, "loss": 0.4273, "step": 1990 }, { "epoch": 2.8429282160625444, "grad_norm": 0.5355864223633939, "learning_rate": 8.702131416170657e-08, "loss": 0.436, "step": 2000 }, { "epoch": 2.8429282160625444, "eval_loss": 0.5102300047874451, "eval_runtime": 231.9212, "eval_samples_per_second": 21.559, "eval_steps_per_second": 2.695, "step": 2000 }, { "epoch": 2.857142857142857, "grad_norm": 0.5388940284634288, "learning_rate": 7.233642411048014e-08, "loss": 0.437, "step": 2010 }, { "epoch": 2.87135749822317, "grad_norm": 0.5557321412758267, "learning_rate": 5.899870807848762e-08, "loss": 0.4304, "step": 2020 }, { "epoch": 2.8855721393034823, "grad_norm": 0.5448610951969772, "learning_rate": 4.701181246411501e-08, "loss": 0.4275, "step": 2030 }, { "epoch": 2.8997867803837956, "grad_norm": 0.5737836534540497, "learning_rate": 3.6379014364935075e-08, "loss": 0.4349, "step": 2040 }, { "epoch": 2.914001421464108, "grad_norm": 0.55622828041129, "learning_rate": 2.7103220681780616e-08, "loss": 0.4291, "step": 2050 }, { "epoch": 2.9282160625444207, "grad_norm": 0.5234759202957545, "learning_rate": 1.9186967324026364e-08, "loss": 0.4283, "step": 2060 }, { "epoch": 2.9424307036247335, "grad_norm": 0.533033568944861, "learning_rate": 1.2632418516296263e-08, "loss": 0.4269, "step": 2070 }, { "epoch": 2.9566453447050463, "grad_norm": 0.5413416159064122, "learning_rate": 7.4413662067884806e-09, "loss": 0.4291, "step": 2080 }, { "epoch": 2.970859985785359, "grad_norm": 0.5523224114708861, "learning_rate": 3.615229577371149e-09, "loss": 0.432, "step": 2090 }, { "epoch": 2.9850746268656714, "grad_norm": 0.5328203745561761, "learning_rate": 1.1550546555960662e-09, "loss": 0.4398, "step": 2100 }, { "epoch": 2.9992892679459846, "grad_norm": 0.5266787263911724, "learning_rate": 6.151402872134337e-11, "loss": 0.422, "step": 2110 }, { "epoch": 3.0, "step": 2111, "total_flos": 1256820975730688.0, "train_loss": 0.22985298537010831, "train_runtime": 4731.3341, "train_samples_per_second": 28.533, "train_steps_per_second": 0.446 } ], "logging_steps": 10, "max_steps": 2112, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1256820975730688.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }