{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 9213, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0162813415825464, "grad_norm": 1.3684147596359253, "learning_rate": 4.973407142081841e-05, "loss": 4.405337219238281, "step": 50 }, { "epoch": 0.0325626831650928, "grad_norm": 2.8155977725982666, "learning_rate": 4.946271572777597e-05, "loss": 3.2570425415039064, "step": 100 }, { "epoch": 0.04884402474763921, "grad_norm": 1.134498119354248, "learning_rate": 4.919136003473353e-05, "loss": 2.13068603515625, "step": 150 }, { "epoch": 0.0651253663301856, "grad_norm": 1.3554809093475342, "learning_rate": 4.892000434169109e-05, "loss": 1.8829190063476562, "step": 200 }, { "epoch": 0.08140670791273201, "grad_norm": 1.3557366132736206, "learning_rate": 4.8648648648648654e-05, "loss": 1.819012908935547, "step": 250 }, { "epoch": 0.09768804949527841, "grad_norm": 1.4770574569702148, "learning_rate": 4.837729295560621e-05, "loss": 1.7141534423828124, "step": 300 }, { "epoch": 0.11396939107782482, "grad_norm": 1.4904377460479736, "learning_rate": 4.810593726256377e-05, "loss": 1.6724388122558593, "step": 350 }, { "epoch": 0.1302507326603712, "grad_norm": 1.4447667598724365, "learning_rate": 4.783458156952133e-05, "loss": 1.5954844665527343, "step": 400 }, { "epoch": 0.14653207424291761, "grad_norm": 1.3709800243377686, "learning_rate": 4.756322587647889e-05, "loss": 1.585284423828125, "step": 450 }, { "epoch": 0.16281341582546402, "grad_norm": 1.9350112676620483, "learning_rate": 4.729187018343645e-05, "loss": 1.4481756591796875, "step": 500 }, { "epoch": 0.17909475740801042, "grad_norm": 1.8580759763717651, "learning_rate": 4.7020514490394014e-05, "loss": 1.5084706115722657, "step": 550 }, { "epoch": 0.19537609899055683, "grad_norm": 2.5280864238739014, "learning_rate": 4.674915879735157e-05, "loss": 1.3583644104003907, "step": 600 }, { "epoch": 0.21165744057310323, "grad_norm": 1.7610465288162231, "learning_rate": 4.647780310430913e-05, "loss": 1.3752433776855468, "step": 650 }, { "epoch": 0.22793878215564964, "grad_norm": 1.9367973804473877, "learning_rate": 4.620644741126669e-05, "loss": 1.3308279418945312, "step": 700 }, { "epoch": 0.24422012373819602, "grad_norm": 1.8170645236968994, "learning_rate": 4.593509171822425e-05, "loss": 1.2124520111083985, "step": 750 }, { "epoch": 0.2605014653207424, "grad_norm": 2.4546971321105957, "learning_rate": 4.566373602518181e-05, "loss": 1.2327165222167968, "step": 800 }, { "epoch": 0.2767828069032888, "grad_norm": 2.476984977722168, "learning_rate": 4.5392380332139374e-05, "loss": 1.2717761993408203, "step": 850 }, { "epoch": 0.29306414848583523, "grad_norm": 2.4676401615142822, "learning_rate": 4.512102463909693e-05, "loss": 1.1377059173583985, "step": 900 }, { "epoch": 0.30934549006838163, "grad_norm": 3.8174469470977783, "learning_rate": 4.484966894605449e-05, "loss": 1.195997314453125, "step": 950 }, { "epoch": 0.32562683165092804, "grad_norm": 2.879594564437866, "learning_rate": 4.457831325301205e-05, "loss": 1.160465316772461, "step": 1000 }, { "epoch": 0.34190817323347444, "grad_norm": 2.7231075763702393, "learning_rate": 4.430695755996961e-05, "loss": 1.1307568359375, "step": 1050 }, { "epoch": 0.35818951481602085, "grad_norm": 2.86122989654541, "learning_rate": 4.403560186692717e-05, "loss": 1.1008319854736328, "step": 1100 }, { "epoch": 0.37447085639856725, "grad_norm": 5.22833251953125, "learning_rate": 4.3764246173884733e-05, "loss": 1.075364456176758, "step": 1150 }, { "epoch": 0.39075219798111366, "grad_norm": 2.75107479095459, "learning_rate": 4.349289048084229e-05, "loss": 1.02243408203125, "step": 1200 }, { "epoch": 0.40703353956366006, "grad_norm": 5.390212535858154, "learning_rate": 4.322153478779985e-05, "loss": 0.9791134643554688, "step": 1250 }, { "epoch": 0.42331488114620647, "grad_norm": 2.980733871459961, "learning_rate": 4.295017909475741e-05, "loss": 0.9889099884033203, "step": 1300 }, { "epoch": 0.4395962227287529, "grad_norm": 2.0358238220214844, "learning_rate": 4.267882340171497e-05, "loss": 1.0286465454101563, "step": 1350 }, { "epoch": 0.4558775643112993, "grad_norm": 3.465447187423706, "learning_rate": 4.2407467708672534e-05, "loss": 0.921280517578125, "step": 1400 }, { "epoch": 0.4721589058938456, "grad_norm": 3.2969017028808594, "learning_rate": 4.213611201563009e-05, "loss": 0.9863865661621094, "step": 1450 }, { "epoch": 0.48844024747639203, "grad_norm": 2.22483229637146, "learning_rate": 4.186475632258765e-05, "loss": 0.9354537963867188, "step": 1500 }, { "epoch": 0.5047215890589385, "grad_norm": 2.548597812652588, "learning_rate": 4.159340062954521e-05, "loss": 0.9002230834960937, "step": 1550 }, { "epoch": 0.5210029306414848, "grad_norm": 2.0713655948638916, "learning_rate": 4.132204493650277e-05, "loss": 0.9603328704833984, "step": 1600 }, { "epoch": 0.5372842722240313, "grad_norm": 2.6993308067321777, "learning_rate": 4.105068924346033e-05, "loss": 0.91706787109375, "step": 1650 }, { "epoch": 0.5535656138065776, "grad_norm": 2.801042079925537, "learning_rate": 4.0779333550417894e-05, "loss": 0.9315050506591797, "step": 1700 }, { "epoch": 0.5698469553891241, "grad_norm": 2.2465991973876953, "learning_rate": 4.050797785737545e-05, "loss": 0.9338973236083984, "step": 1750 }, { "epoch": 0.5861282969716705, "grad_norm": 2.300447702407837, "learning_rate": 4.023662216433301e-05, "loss": 0.8962403106689453, "step": 1800 }, { "epoch": 0.6024096385542169, "grad_norm": 3.7357118129730225, "learning_rate": 3.996526647129057e-05, "loss": 0.8952316284179688, "step": 1850 }, { "epoch": 0.6186909801367633, "grad_norm": 2.987694263458252, "learning_rate": 3.969391077824813e-05, "loss": 0.8913738250732421, "step": 1900 }, { "epoch": 0.6349723217193096, "grad_norm": 2.678509473800659, "learning_rate": 3.942255508520569e-05, "loss": 0.8106181335449218, "step": 1950 }, { "epoch": 0.6512536633018561, "grad_norm": 2.694267511367798, "learning_rate": 3.9151199392163254e-05, "loss": 0.8525662994384766, "step": 2000 }, { "epoch": 0.6675350048844024, "grad_norm": 2.4209225177764893, "learning_rate": 3.887984369912081e-05, "loss": 0.8165419769287109, "step": 2050 }, { "epoch": 0.6838163464669489, "grad_norm": 3.41129994392395, "learning_rate": 3.860848800607837e-05, "loss": 0.7995121765136719, "step": 2100 }, { "epoch": 0.7000976880494952, "grad_norm": 2.8171026706695557, "learning_rate": 3.833713231303593e-05, "loss": 0.8041473388671875, "step": 2150 }, { "epoch": 0.7163790296320417, "grad_norm": 2.9240710735321045, "learning_rate": 3.806577661999349e-05, "loss": 0.9764547729492188, "step": 2200 }, { "epoch": 0.732660371214588, "grad_norm": 2.051103353500366, "learning_rate": 3.7794420926951055e-05, "loss": 0.8336094665527344, "step": 2250 }, { "epoch": 0.7489417127971345, "grad_norm": 3.2116966247558594, "learning_rate": 3.7523065233908613e-05, "loss": 0.7777658081054688, "step": 2300 }, { "epoch": 0.7652230543796809, "grad_norm": 3.8302836418151855, "learning_rate": 3.725170954086617e-05, "loss": 0.848994140625, "step": 2350 }, { "epoch": 0.7815043959622273, "grad_norm": 2.0940141677856445, "learning_rate": 3.698035384782373e-05, "loss": 0.6995362091064453, "step": 2400 }, { "epoch": 0.7977857375447737, "grad_norm": 3.155562400817871, "learning_rate": 3.670899815478129e-05, "loss": 0.7727601623535156, "step": 2450 }, { "epoch": 0.8140670791273201, "grad_norm": 2.886545419692993, "learning_rate": 3.643764246173885e-05, "loss": 0.7873483276367188, "step": 2500 }, { "epoch": 0.8303484207098665, "grad_norm": 2.5638420581817627, "learning_rate": 3.6166286768696414e-05, "loss": 0.7987093353271484, "step": 2550 }, { "epoch": 0.8466297622924129, "grad_norm": 2.3368115425109863, "learning_rate": 3.589493107565397e-05, "loss": 0.7667920684814453, "step": 2600 }, { "epoch": 0.8629111038749593, "grad_norm": 2.953003168106079, "learning_rate": 3.562357538261153e-05, "loss": 0.6812684631347656, "step": 2650 }, { "epoch": 0.8791924454575057, "grad_norm": 3.382375955581665, "learning_rate": 3.535221968956909e-05, "loss": 0.7589307403564454, "step": 2700 }, { "epoch": 0.8954737870400521, "grad_norm": 2.6374568939208984, "learning_rate": 3.508086399652665e-05, "loss": 0.7288359069824218, "step": 2750 }, { "epoch": 0.9117551286225986, "grad_norm": 2.135319709777832, "learning_rate": 3.480950830348421e-05, "loss": 0.7487136077880859, "step": 2800 }, { "epoch": 0.9280364702051449, "grad_norm": 3.235957145690918, "learning_rate": 3.4538152610441774e-05, "loss": 0.7287068176269531, "step": 2850 }, { "epoch": 0.9443178117876913, "grad_norm": 3.254149913787842, "learning_rate": 3.426679691739933e-05, "loss": 0.7420393371582031, "step": 2900 }, { "epoch": 0.9605991533702377, "grad_norm": 2.4111130237579346, "learning_rate": 3.399544122435689e-05, "loss": 0.7257329559326172, "step": 2950 }, { "epoch": 0.9768804949527841, "grad_norm": 3.764932870864868, "learning_rate": 3.372408553131445e-05, "loss": 0.6783822631835937, "step": 3000 }, { "epoch": 0.9931618365353305, "grad_norm": 3.1191906929016113, "learning_rate": 3.345272983827201e-05, "loss": 0.7785000610351562, "step": 3050 }, { "epoch": 1.0, "eval_bertscore_f1": 0.9666035647731817, "eval_bleu": 0.6286042636210436, "eval_loss": 0.4953671097755432, "eval_meteor": 0.727270993632129, "eval_rouge1": 0.8535427677296903, "eval_rouge2": 0.7442740134144676, "eval_runtime": 125.3536, "eval_samples_per_second": 10.307, "eval_steps_per_second": 1.292, "step": 3071 }, { "epoch": 1.009443178117877, "grad_norm": 3.273244857788086, "learning_rate": 3.318137414522957e-05, "loss": 0.6082234954833985, "step": 3100 }, { "epoch": 1.0257245197004232, "grad_norm": 3.4152684211730957, "learning_rate": 3.291001845218713e-05, "loss": 0.5953511047363281, "step": 3150 }, { "epoch": 1.0420058612829697, "grad_norm": 3.5387468338012695, "learning_rate": 3.263866275914469e-05, "loss": 0.6991500091552735, "step": 3200 }, { "epoch": 1.0582872028655161, "grad_norm": 4.7770280838012695, "learning_rate": 3.236730706610225e-05, "loss": 0.6681033325195312, "step": 3250 }, { "epoch": 1.0745685444480626, "grad_norm": 3.492780923843384, "learning_rate": 3.209595137305981e-05, "loss": 0.6578762054443359, "step": 3300 }, { "epoch": 1.0908498860306088, "grad_norm": 3.3742964267730713, "learning_rate": 3.182459568001737e-05, "loss": 0.6371722030639648, "step": 3350 }, { "epoch": 1.1071312276131553, "grad_norm": 2.9794905185699463, "learning_rate": 3.155323998697493e-05, "loss": 0.6163125228881836, "step": 3400 }, { "epoch": 1.1234125691957018, "grad_norm": 4.077469825744629, "learning_rate": 3.128188429393249e-05, "loss": 0.6614529418945313, "step": 3450 }, { "epoch": 1.1396939107782482, "grad_norm": 2.614112615585327, "learning_rate": 3.1010528600890045e-05, "loss": 0.6072891998291016, "step": 3500 }, { "epoch": 1.1559752523607945, "grad_norm": 2.661522150039673, "learning_rate": 3.0739172907847604e-05, "loss": 0.5901923370361328, "step": 3550 }, { "epoch": 1.172256593943341, "grad_norm": 3.5029804706573486, "learning_rate": 3.0467817214805167e-05, "loss": 0.7141423034667969, "step": 3600 }, { "epoch": 1.1885379355258874, "grad_norm": 2.75795578956604, "learning_rate": 3.0196461521762725e-05, "loss": 0.6581143951416015, "step": 3650 }, { "epoch": 1.2048192771084336, "grad_norm": 3.335984230041504, "learning_rate": 2.992510582872029e-05, "loss": 0.6345047378540039, "step": 3700 }, { "epoch": 1.22110061869098, "grad_norm": 2.886397361755371, "learning_rate": 2.965375013567785e-05, "loss": 0.6183316802978516, "step": 3750 }, { "epoch": 1.2373819602735265, "grad_norm": 3.5690371990203857, "learning_rate": 2.938239444263541e-05, "loss": 0.6330132293701172, "step": 3800 }, { "epoch": 1.253663301856073, "grad_norm": 2.857058525085449, "learning_rate": 2.9111038749592967e-05, "loss": 0.5762019729614258, "step": 3850 }, { "epoch": 1.2699446434386195, "grad_norm": 3.4295766353607178, "learning_rate": 2.8839683056550526e-05, "loss": 0.6346803665161133, "step": 3900 }, { "epoch": 1.2862259850211657, "grad_norm": 3.578397274017334, "learning_rate": 2.8568327363508085e-05, "loss": 0.6218410491943359, "step": 3950 }, { "epoch": 1.3025073266037122, "grad_norm": 4.432509422302246, "learning_rate": 2.829697167046565e-05, "loss": 0.6320372009277344, "step": 4000 }, { "epoch": 1.3187886681862586, "grad_norm": 2.806671142578125, "learning_rate": 2.802561597742321e-05, "loss": 0.652824478149414, "step": 4050 }, { "epoch": 1.3350700097688049, "grad_norm": 3.3935513496398926, "learning_rate": 2.775426028438077e-05, "loss": 0.6719865417480468, "step": 4100 }, { "epoch": 1.3513513513513513, "grad_norm": 2.2870659828186035, "learning_rate": 2.7482904591338327e-05, "loss": 0.586540756225586, "step": 4150 }, { "epoch": 1.3676326929338978, "grad_norm": 3.7163562774658203, "learning_rate": 2.7211548898295886e-05, "loss": 0.5858817672729493, "step": 4200 }, { "epoch": 1.3839140345164442, "grad_norm": 2.730625629425049, "learning_rate": 2.694019320525345e-05, "loss": 0.5543305206298829, "step": 4250 }, { "epoch": 1.4001953760989905, "grad_norm": 3.728361129760742, "learning_rate": 2.666883751221101e-05, "loss": 0.6442169952392578, "step": 4300 }, { "epoch": 1.416476717681537, "grad_norm": 2.2651383876800537, "learning_rate": 2.639748181916857e-05, "loss": 0.582794075012207, "step": 4350 }, { "epoch": 1.4327580592640834, "grad_norm": 2.430931806564331, "learning_rate": 2.6126126126126128e-05, "loss": 0.6022589874267578, "step": 4400 }, { "epoch": 1.4490394008466296, "grad_norm": 3.222708225250244, "learning_rate": 2.5854770433083687e-05, "loss": 0.6583724975585937, "step": 4450 }, { "epoch": 1.465320742429176, "grad_norm": 4.678814888000488, "learning_rate": 2.5583414740041246e-05, "loss": 0.5943536376953125, "step": 4500 }, { "epoch": 1.4816020840117226, "grad_norm": 3.0403647422790527, "learning_rate": 2.5312059046998808e-05, "loss": 0.5958936309814453, "step": 4550 }, { "epoch": 1.497883425594269, "grad_norm": 2.538402795791626, "learning_rate": 2.504070335395637e-05, "loss": 0.6167163467407226, "step": 4600 }, { "epoch": 1.5141647671768155, "grad_norm": 5.17458438873291, "learning_rate": 2.476934766091393e-05, "loss": 0.5508831405639648, "step": 4650 }, { "epoch": 1.530446108759362, "grad_norm": 2.4245738983154297, "learning_rate": 2.4497991967871488e-05, "loss": 0.6157744598388671, "step": 4700 }, { "epoch": 1.5467274503419082, "grad_norm": 2.481553792953491, "learning_rate": 2.4226636274829047e-05, "loss": 0.6096940231323242, "step": 4750 }, { "epoch": 1.5630087919244544, "grad_norm": 2.797117233276367, "learning_rate": 2.395528058178661e-05, "loss": 0.6150970458984375, "step": 4800 }, { "epoch": 1.5792901335070009, "grad_norm": 4.05893611907959, "learning_rate": 2.3683924888744168e-05, "loss": 0.6198799133300781, "step": 4850 }, { "epoch": 1.5955714750895473, "grad_norm": 3.5510406494140625, "learning_rate": 2.3412569195701726e-05, "loss": 0.5494784545898438, "step": 4900 }, { "epoch": 1.6118528166720938, "grad_norm": 4.9277424812316895, "learning_rate": 2.3141213502659285e-05, "loss": 0.500348129272461, "step": 4950 }, { "epoch": 1.6281341582546403, "grad_norm": 2.930206775665283, "learning_rate": 2.2869857809616847e-05, "loss": 0.5013648986816406, "step": 5000 }, { "epoch": 1.6444154998371867, "grad_norm": 10.043802261352539, "learning_rate": 2.2598502116574406e-05, "loss": 0.5953195190429688, "step": 5050 }, { "epoch": 1.660696841419733, "grad_norm": 3.334196090698242, "learning_rate": 2.2327146423531965e-05, "loss": 0.5374015426635742, "step": 5100 }, { "epoch": 1.6769781830022794, "grad_norm": 3.292771339416504, "learning_rate": 2.2055790730489524e-05, "loss": 0.49036331176757814, "step": 5150 }, { "epoch": 1.6932595245848256, "grad_norm": 3.440549373626709, "learning_rate": 2.1784435037447086e-05, "loss": 0.5311320114135742, "step": 5200 }, { "epoch": 1.709540866167372, "grad_norm": 3.720184803009033, "learning_rate": 2.1513079344404645e-05, "loss": 0.5855265045166016, "step": 5250 }, { "epoch": 1.7258222077499186, "grad_norm": 3.9988839626312256, "learning_rate": 2.1241723651362204e-05, "loss": 0.5410572052001953, "step": 5300 }, { "epoch": 1.742103549332465, "grad_norm": 3.244662046432495, "learning_rate": 2.0970367958319766e-05, "loss": 0.553615951538086, "step": 5350 }, { "epoch": 1.7583848909150115, "grad_norm": 3.0616097450256348, "learning_rate": 2.0699012265277325e-05, "loss": 0.4832605361938477, "step": 5400 }, { "epoch": 1.774666232497558, "grad_norm": 3.586770534515381, "learning_rate": 2.0427656572234887e-05, "loss": 0.5060377883911132, "step": 5450 }, { "epoch": 1.7909475740801042, "grad_norm": 2.99122953414917, "learning_rate": 2.0156300879192446e-05, "loss": 0.510421142578125, "step": 5500 }, { "epoch": 1.8072289156626506, "grad_norm": 2.647064685821533, "learning_rate": 1.9884945186150005e-05, "loss": 0.5255117034912109, "step": 5550 }, { "epoch": 1.8235102572451969, "grad_norm": 3.2572667598724365, "learning_rate": 1.9613589493107567e-05, "loss": 0.48496082305908206, "step": 5600 }, { "epoch": 1.8397915988277433, "grad_norm": 4.511913299560547, "learning_rate": 1.9342233800065126e-05, "loss": 0.5061034393310547, "step": 5650 }, { "epoch": 1.8560729404102898, "grad_norm": 2.9536867141723633, "learning_rate": 1.9070878107022685e-05, "loss": 0.5593103790283203, "step": 5700 }, { "epoch": 1.8723542819928363, "grad_norm": 2.071770429611206, "learning_rate": 1.8799522413980247e-05, "loss": 0.5091780471801758, "step": 5750 }, { "epoch": 1.8886356235753827, "grad_norm": 4.718142986297607, "learning_rate": 1.8528166720937806e-05, "loss": 0.4754158401489258, "step": 5800 }, { "epoch": 1.904916965157929, "grad_norm": 4.707818508148193, "learning_rate": 1.8256811027895364e-05, "loss": 0.5697180557250977, "step": 5850 }, { "epoch": 1.9211983067404754, "grad_norm": 4.631124973297119, "learning_rate": 1.7985455334852927e-05, "loss": 0.4917443084716797, "step": 5900 }, { "epoch": 1.9374796483230217, "grad_norm": 3.7037010192871094, "learning_rate": 1.7714099641810485e-05, "loss": 0.577647933959961, "step": 5950 }, { "epoch": 1.9537609899055681, "grad_norm": 4.4254302978515625, "learning_rate": 1.7442743948768044e-05, "loss": 0.49387569427490235, "step": 6000 }, { "epoch": 1.9700423314881146, "grad_norm": 4.429046630859375, "learning_rate": 1.7171388255725606e-05, "loss": 0.555143165588379, "step": 6050 }, { "epoch": 1.986323673070661, "grad_norm": 3.211913824081421, "learning_rate": 1.6900032562683165e-05, "loss": 0.5266495132446289, "step": 6100 }, { "epoch": 2.0, "eval_bertscore_f1": 0.9751379909253342, "eval_bleu": 0.7283162211638717, "eval_loss": 0.3667888939380646, "eval_meteor": 0.806130448591728, "eval_rouge1": 0.8830570311221204, "eval_rouge2": 0.803848501332357, "eval_runtime": 126.2611, "eval_samples_per_second": 10.233, "eval_steps_per_second": 1.283, "step": 6142 }, { "epoch": 2.0026050146532075, "grad_norm": 2.130920886993408, "learning_rate": 1.6628676869640724e-05, "loss": 0.49579532623291017, "step": 6150 }, { "epoch": 2.018886356235754, "grad_norm": 3.622119665145874, "learning_rate": 1.6357321176598286e-05, "loss": 0.5926795959472656, "step": 6200 }, { "epoch": 2.0351676978183004, "grad_norm": 4.096598148345947, "learning_rate": 1.6085965483555845e-05, "loss": 0.511412124633789, "step": 6250 }, { "epoch": 2.0514490394008464, "grad_norm": 3.3541550636291504, "learning_rate": 1.5814609790513404e-05, "loss": 0.5026980590820312, "step": 6300 }, { "epoch": 2.067730380983393, "grad_norm": 3.797450304031372, "learning_rate": 1.5543254097470966e-05, "loss": 0.48105335235595703, "step": 6350 }, { "epoch": 2.0840117225659394, "grad_norm": 4.0247626304626465, "learning_rate": 1.5271898404428525e-05, "loss": 0.48235019683837893, "step": 6400 }, { "epoch": 2.100293064148486, "grad_norm": 4.341481685638428, "learning_rate": 1.5000542711386087e-05, "loss": 0.5079761123657227, "step": 6450 }, { "epoch": 2.1165744057310323, "grad_norm": 2.415269374847412, "learning_rate": 1.4729187018343646e-05, "loss": 0.5332447814941407, "step": 6500 }, { "epoch": 2.1328557473135787, "grad_norm": 4.66851282119751, "learning_rate": 1.4457831325301205e-05, "loss": 0.5538092041015625, "step": 6550 }, { "epoch": 2.149137088896125, "grad_norm": 3.448925495147705, "learning_rate": 1.4186475632258767e-05, "loss": 0.5555255889892579, "step": 6600 }, { "epoch": 2.165418430478671, "grad_norm": 4.12534236907959, "learning_rate": 1.3915119939216326e-05, "loss": 0.49463138580322263, "step": 6650 }, { "epoch": 2.1816997720612177, "grad_norm": 2.726137638092041, "learning_rate": 1.3643764246173885e-05, "loss": 0.4911256408691406, "step": 6700 }, { "epoch": 2.197981113643764, "grad_norm": 2.9859631061553955, "learning_rate": 1.3372408553131447e-05, "loss": 0.5317694091796875, "step": 6750 }, { "epoch": 2.2142624552263106, "grad_norm": 3.0363500118255615, "learning_rate": 1.3101052860089006e-05, "loss": 0.48204010009765624, "step": 6800 }, { "epoch": 2.230543796808857, "grad_norm": 2.7553327083587646, "learning_rate": 1.2829697167046565e-05, "loss": 0.5275654602050781, "step": 6850 }, { "epoch": 2.2468251383914035, "grad_norm": 3.37378191947937, "learning_rate": 1.2558341474004127e-05, "loss": 0.49252197265625, "step": 6900 }, { "epoch": 2.26310647997395, "grad_norm": 3.5572731494903564, "learning_rate": 1.2286985780961686e-05, "loss": 0.5376744079589844, "step": 6950 }, { "epoch": 2.2793878215564964, "grad_norm": 4.323084354400635, "learning_rate": 1.2015630087919246e-05, "loss": 0.5603115081787109, "step": 7000 }, { "epoch": 2.295669163139043, "grad_norm": 3.9723100662231445, "learning_rate": 1.1744274394876805e-05, "loss": 0.4530460739135742, "step": 7050 }, { "epoch": 2.311950504721589, "grad_norm": 4.0148491859436035, "learning_rate": 1.1472918701834365e-05, "loss": 0.4808524703979492, "step": 7100 }, { "epoch": 2.3282318463041354, "grad_norm": 3.357252597808838, "learning_rate": 1.1201563008791926e-05, "loss": 0.5342596817016602, "step": 7150 }, { "epoch": 2.344513187886682, "grad_norm": 1.4084240198135376, "learning_rate": 1.0930207315749485e-05, "loss": 0.5188829040527344, "step": 7200 }, { "epoch": 2.3607945294692283, "grad_norm": 4.223776817321777, "learning_rate": 1.0658851622707045e-05, "loss": 0.5551160049438476, "step": 7250 }, { "epoch": 2.3770758710517748, "grad_norm": 3.2388432025909424, "learning_rate": 1.0387495929664604e-05, "loss": 0.4678123092651367, "step": 7300 }, { "epoch": 2.393357212634321, "grad_norm": 2.2621917724609375, "learning_rate": 1.0116140236622165e-05, "loss": 0.49837650299072267, "step": 7350 }, { "epoch": 2.4096385542168672, "grad_norm": 3.9080207347869873, "learning_rate": 9.844784543579723e-06, "loss": 0.4679309844970703, "step": 7400 }, { "epoch": 2.4259198957994137, "grad_norm": 2.2795565128326416, "learning_rate": 9.573428850537284e-06, "loss": 0.4956610870361328, "step": 7450 }, { "epoch": 2.44220123738196, "grad_norm": 2.872938632965088, "learning_rate": 9.302073157494844e-06, "loss": 0.48609119415283203, "step": 7500 }, { "epoch": 2.4584825789645066, "grad_norm": 3.869072675704956, "learning_rate": 9.030717464452403e-06, "loss": 0.5468455886840821, "step": 7550 }, { "epoch": 2.474763920547053, "grad_norm": 3.9535272121429443, "learning_rate": 8.759361771409964e-06, "loss": 0.4581578826904297, "step": 7600 }, { "epoch": 2.4910452621295995, "grad_norm": 2.2123703956604004, "learning_rate": 8.488006078367524e-06, "loss": 0.48407554626464844, "step": 7650 }, { "epoch": 2.507326603712146, "grad_norm": 3.5734164714813232, "learning_rate": 8.216650385325085e-06, "loss": 0.4096232986450195, "step": 7700 }, { "epoch": 2.5236079452946925, "grad_norm": 5.056736946105957, "learning_rate": 7.945294692282644e-06, "loss": 0.48154861450195313, "step": 7750 }, { "epoch": 2.539889286877239, "grad_norm": 5.742955207824707, "learning_rate": 7.673938999240204e-06, "loss": 0.46851539611816406, "step": 7800 }, { "epoch": 2.556170628459785, "grad_norm": 3.6929991245269775, "learning_rate": 7.402583306197765e-06, "loss": 0.4527290725708008, "step": 7850 }, { "epoch": 2.5724519700423314, "grad_norm": 4.624508857727051, "learning_rate": 7.1312276131553235e-06, "loss": 0.487193717956543, "step": 7900 }, { "epoch": 2.588733311624878, "grad_norm": 3.526402711868286, "learning_rate": 6.859871920112884e-06, "loss": 0.4767793273925781, "step": 7950 }, { "epoch": 2.6050146532074243, "grad_norm": 2.557279586791992, "learning_rate": 6.5885162270704446e-06, "loss": 0.47788654327392577, "step": 8000 }, { "epoch": 2.6212959947899708, "grad_norm": 3.1233720779418945, "learning_rate": 6.317160534028005e-06, "loss": 0.43118988037109374, "step": 8050 }, { "epoch": 2.6375773363725172, "grad_norm": 3.1848602294921875, "learning_rate": 6.045804840985564e-06, "loss": 0.4824806594848633, "step": 8100 }, { "epoch": 2.6538586779550632, "grad_norm": 2.1851677894592285, "learning_rate": 5.774449147943124e-06, "loss": 0.5195888900756835, "step": 8150 }, { "epoch": 2.6701400195376097, "grad_norm": 4.459860801696777, "learning_rate": 5.503093454900684e-06, "loss": 0.5190325164794922, "step": 8200 }, { "epoch": 2.686421361120156, "grad_norm": 2.845478057861328, "learning_rate": 5.231737761858245e-06, "loss": 0.5278813171386719, "step": 8250 }, { "epoch": 2.7027027027027026, "grad_norm": 3.6786937713623047, "learning_rate": 4.960382068815804e-06, "loss": 0.4795223236083984, "step": 8300 }, { "epoch": 2.718984044285249, "grad_norm": 3.6952383518218994, "learning_rate": 4.689026375773364e-06, "loss": 0.5154057312011718, "step": 8350 }, { "epoch": 2.7352653858677956, "grad_norm": 2.0930325984954834, "learning_rate": 4.417670682730924e-06, "loss": 0.4847321319580078, "step": 8400 }, { "epoch": 2.751546727450342, "grad_norm": 6.89557409286499, "learning_rate": 4.146314989688484e-06, "loss": 0.39144508361816405, "step": 8450 }, { "epoch": 2.7678280690328885, "grad_norm": 4.23341703414917, "learning_rate": 3.874959296646044e-06, "loss": 0.47240074157714845, "step": 8500 }, { "epoch": 2.784109410615435, "grad_norm": 5.129818439483643, "learning_rate": 3.603603603603604e-06, "loss": 0.4966455841064453, "step": 8550 }, { "epoch": 2.800390752197981, "grad_norm": 2.9496281147003174, "learning_rate": 3.3322479105611635e-06, "loss": 0.48110599517822267, "step": 8600 }, { "epoch": 2.8166720937805274, "grad_norm": 5.308284759521484, "learning_rate": 3.0608922175187236e-06, "loss": 0.4717050552368164, "step": 8650 }, { "epoch": 2.832953435363074, "grad_norm": 3.275146722793579, "learning_rate": 2.7895365244762837e-06, "loss": 0.4382866668701172, "step": 8700 }, { "epoch": 2.8492347769456203, "grad_norm": 2.5721969604492188, "learning_rate": 2.5181808314338434e-06, "loss": 0.46353874206542967, "step": 8750 }, { "epoch": 2.865516118528167, "grad_norm": 3.426837921142578, "learning_rate": 2.2468251383914035e-06, "loss": 0.5103521347045898, "step": 8800 }, { "epoch": 2.8817974601107132, "grad_norm": 2.3799993991851807, "learning_rate": 1.9754694453489636e-06, "loss": 0.47350223541259767, "step": 8850 }, { "epoch": 2.8980788016932593, "grad_norm": 3.316828966140747, "learning_rate": 1.7041137523065235e-06, "loss": 0.4602887725830078, "step": 8900 }, { "epoch": 2.9143601432758057, "grad_norm": 5.089015007019043, "learning_rate": 1.4327580592640834e-06, "loss": 0.4908058547973633, "step": 8950 }, { "epoch": 2.930641484858352, "grad_norm": 5.589182376861572, "learning_rate": 1.1614023662216434e-06, "loss": 0.4539414215087891, "step": 9000 }, { "epoch": 2.9469228264408986, "grad_norm": 2.4032421112060547, "learning_rate": 8.900466731792033e-07, "loss": 0.42325592041015625, "step": 9050 }, { "epoch": 2.963204168023445, "grad_norm": 3.4075100421905518, "learning_rate": 6.186909801367633e-07, "loss": 0.4070619201660156, "step": 9100 }, { "epoch": 2.9794855096059916, "grad_norm": 1.6805068254470825, "learning_rate": 3.4733528709432325e-07, "loss": 0.4350212860107422, "step": 9150 }, { "epoch": 2.995766851188538, "grad_norm": 3.386293888092041, "learning_rate": 7.597959405188321e-08, "loss": 0.4917270660400391, "step": 9200 }, { "epoch": 3.0, "eval_bertscore_f1": 0.9777762908776847, "eval_bleu": 0.7589549870918858, "eval_loss": 0.3381657600402832, "eval_meteor": 0.8307886222747983, "eval_rouge1": 0.893313778543083, "eval_rouge2": 0.821781747009253, "eval_runtime": 130.6802, "eval_samples_per_second": 9.887, "eval_steps_per_second": 1.24, "step": 9213 } ], "logging_steps": 50, "max_steps": 9213, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.263887217557504e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }