{ "best_global_step": 17094, "best_metric": 0.9845099349219312, "best_model_checkpoint": "./xlmr-language-identification/checkpoint-17094", "epoch": 2.0, "eval_steps": 2500, "global_step": 17094, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.011701304695473546, "grad_norm": 0.9905334711074829, "learning_rate": 4.971042471042471e-05, "loss": 16.156219482421875, "step": 100 }, { "epoch": 0.023402609390947092, "grad_norm": 1.069915533065796, "learning_rate": 4.941792441792442e-05, "loss": 5.045557556152343, "step": 200 }, { "epoch": 0.035103914086420634, "grad_norm": 1.8720062971115112, "learning_rate": 4.912542412542413e-05, "loss": 4.602024841308594, "step": 300 }, { "epoch": 0.046805218781894184, "grad_norm": 2.246952772140503, "learning_rate": 4.883292383292383e-05, "loss": 2.9484735107421876, "step": 400 }, { "epoch": 0.058506523477367726, "grad_norm": 1.7273783683776855, "learning_rate": 4.8540423540423544e-05, "loss": 1.8604541015625, "step": 500 }, { "epoch": 0.07020782817284127, "grad_norm": 1.871302843093872, "learning_rate": 4.824792324792325e-05, "loss": 1.2988777160644531, "step": 600 }, { "epoch": 0.08190913286831482, "grad_norm": 1.9515539407730103, "learning_rate": 4.7955422955422954e-05, "loss": 0.9935964965820312, "step": 700 }, { "epoch": 0.09361043756378837, "grad_norm": 1.469983696937561, "learning_rate": 4.7662922662922666e-05, "loss": 0.803764877319336, "step": 800 }, { "epoch": 0.1053117422592619, "grad_norm": 1.6788742542266846, "learning_rate": 4.737042237042237e-05, "loss": 0.6818081665039063, "step": 900 }, { "epoch": 0.11701304695473545, "grad_norm": 1.4318715333938599, "learning_rate": 4.707792207792208e-05, "loss": 0.5820805740356445, "step": 1000 }, { "epoch": 0.128714351650209, "grad_norm": 1.9161852598190308, "learning_rate": 4.678542178542179e-05, "loss": 0.5121846389770508, "step": 1100 }, { "epoch": 0.14041565634568254, "grad_norm": 1.6554639339447021, "learning_rate": 4.6492921492921494e-05, "loss": 0.4561897277832031, "step": 1200 }, { "epoch": 0.1521169610411561, "grad_norm": 1.839687466621399, "learning_rate": 4.62004212004212e-05, "loss": 0.4188547134399414, "step": 1300 }, { "epoch": 0.16381826573662964, "grad_norm": 1.735112190246582, "learning_rate": 4.590792090792091e-05, "loss": 0.3874264144897461, "step": 1400 }, { "epoch": 0.17551957043210317, "grad_norm": 2.1364924907684326, "learning_rate": 4.5615420615420616e-05, "loss": 0.36660758972167967, "step": 1500 }, { "epoch": 0.18722087512757674, "grad_norm": 1.5190038681030273, "learning_rate": 4.532292032292033e-05, "loss": 0.3320538330078125, "step": 1600 }, { "epoch": 0.19892217982305027, "grad_norm": 1.4698094129562378, "learning_rate": 4.503042003042003e-05, "loss": 0.3180769729614258, "step": 1700 }, { "epoch": 0.2106234845185238, "grad_norm": 1.4362446069717407, "learning_rate": 4.4737919737919745e-05, "loss": 0.3039104461669922, "step": 1800 }, { "epoch": 0.22232478921399737, "grad_norm": 1.5884244441986084, "learning_rate": 4.444541944541944e-05, "loss": 0.2853180694580078, "step": 1900 }, { "epoch": 0.2340260939094709, "grad_norm": 1.14271879196167, "learning_rate": 4.4152919152919155e-05, "loss": 0.2749899673461914, "step": 2000 }, { "epoch": 0.24572739860494444, "grad_norm": 1.3998445272445679, "learning_rate": 4.386041886041886e-05, "loss": 0.2584847450256348, "step": 2100 }, { "epoch": 0.257428703300418, "grad_norm": 1.126080870628357, "learning_rate": 4.356791856791857e-05, "loss": 0.2443878746032715, "step": 2200 }, { "epoch": 0.26913000799589154, "grad_norm": 1.6860474348068237, "learning_rate": 4.327541827541828e-05, "loss": 0.23169761657714844, "step": 2300 }, { "epoch": 0.2808313126913651, "grad_norm": 1.8752473592758179, "learning_rate": 4.298291798291799e-05, "loss": 0.22600336074829103, "step": 2400 }, { "epoch": 0.2925326173868386, "grad_norm": 1.6066241264343262, "learning_rate": 4.2690417690417694e-05, "loss": 0.2185952568054199, "step": 2500 }, { "epoch": 0.2925326173868386, "eval_accuracy": 0.8560363636363636, "eval_f1": 0.9651478556858426, "eval_loss": 0.039476945996284485, "eval_precision": 0.9777755239169846, "eval_recall": 0.9528421936966432, "eval_runtime": 165.9297, "eval_samples_per_second": 165.733, "eval_steps_per_second": 5.183, "step": 2500 }, { "epoch": 0.3042339220823122, "grad_norm": 1.2747637033462524, "learning_rate": 4.2397917397917406e-05, "loss": 0.22376758575439454, "step": 2600 }, { "epoch": 0.31593522677778574, "grad_norm": 1.5281429290771484, "learning_rate": 4.2105417105417104e-05, "loss": 0.21567138671875, "step": 2700 }, { "epoch": 0.3276365314732593, "grad_norm": 1.0241914987564087, "learning_rate": 4.1812916812916816e-05, "loss": 0.20220571517944336, "step": 2800 }, { "epoch": 0.3393378361687328, "grad_norm": 1.9281812906265259, "learning_rate": 4.152041652041652e-05, "loss": 0.19900651931762695, "step": 2900 }, { "epoch": 0.35103914086420634, "grad_norm": 1.0900094509124756, "learning_rate": 4.122791622791623e-05, "loss": 0.19108434677124023, "step": 3000 }, { "epoch": 0.3627404455596799, "grad_norm": 1.9392260313034058, "learning_rate": 4.093541593541594e-05, "loss": 0.1862166404724121, "step": 3100 }, { "epoch": 0.37444175025515347, "grad_norm": 1.4907201528549194, "learning_rate": 4.064291564291564e-05, "loss": 0.18224323272705079, "step": 3200 }, { "epoch": 0.386143054950627, "grad_norm": 1.0142017602920532, "learning_rate": 4.0350415350415355e-05, "loss": 0.17916433334350587, "step": 3300 }, { "epoch": 0.39784435964610054, "grad_norm": 1.3136364221572876, "learning_rate": 4.005791505791506e-05, "loss": 0.17482011795043945, "step": 3400 }, { "epoch": 0.4095456643415741, "grad_norm": 1.7887938022613525, "learning_rate": 3.9765414765414765e-05, "loss": 0.16918930053710937, "step": 3500 }, { "epoch": 0.4212469690370476, "grad_norm": 1.368762731552124, "learning_rate": 3.947291447291447e-05, "loss": 0.1643056869506836, "step": 3600 }, { "epoch": 0.43294827373252115, "grad_norm": 1.666864275932312, "learning_rate": 3.918041418041418e-05, "loss": 0.16255182266235352, "step": 3700 }, { "epoch": 0.44464957842799474, "grad_norm": 1.4107720851898193, "learning_rate": 3.888791388791389e-05, "loss": 0.15760359764099122, "step": 3800 }, { "epoch": 0.4563508831234683, "grad_norm": 1.6475497484207153, "learning_rate": 3.85954135954136e-05, "loss": 0.1564300537109375, "step": 3900 }, { "epoch": 0.4680521878189418, "grad_norm": 1.3566536903381348, "learning_rate": 3.8302913302913305e-05, "loss": 0.15231061935424806, "step": 4000 }, { "epoch": 0.47975349251441535, "grad_norm": 0.9237350225448608, "learning_rate": 3.801041301041301e-05, "loss": 0.14665995597839354, "step": 4100 }, { "epoch": 0.4914547972098889, "grad_norm": 2.0160303115844727, "learning_rate": 3.7717912717912715e-05, "loss": 0.1439968776702881, "step": 4200 }, { "epoch": 0.5031561019053624, "grad_norm": 1.495821237564087, "learning_rate": 3.742541242541243e-05, "loss": 0.14349424362182617, "step": 4300 }, { "epoch": 0.514857406600836, "grad_norm": 1.1643633842468262, "learning_rate": 3.713291213291213e-05, "loss": 0.1439281463623047, "step": 4400 }, { "epoch": 0.5265587112963095, "grad_norm": 1.2707788944244385, "learning_rate": 3.6840411840411844e-05, "loss": 0.14324712753295898, "step": 4500 }, { "epoch": 0.5382600159917831, "grad_norm": 0.9467172026634216, "learning_rate": 3.654791154791155e-05, "loss": 0.136934118270874, "step": 4600 }, { "epoch": 0.5499613206872567, "grad_norm": 1.4451957941055298, "learning_rate": 3.625541125541126e-05, "loss": 0.13631460189819336, "step": 4700 }, { "epoch": 0.5616626253827302, "grad_norm": 1.1038591861724854, "learning_rate": 3.5962910962910966e-05, "loss": 0.13634085655212402, "step": 4800 }, { "epoch": 0.5733639300782037, "grad_norm": 1.2149685621261597, "learning_rate": 3.567041067041067e-05, "loss": 0.13158455848693848, "step": 4900 }, { "epoch": 0.5850652347736772, "grad_norm": 1.421316146850586, "learning_rate": 3.5377910377910376e-05, "loss": 0.13308174133300782, "step": 5000 }, { "epoch": 0.5850652347736772, "eval_accuracy": 0.9069818181818182, "eval_f1": 0.9759635668726578, "eval_loss": 0.02319982275366783, "eval_precision": 0.9802819359100597, "eval_recall": 0.971683077797408, "eval_runtime": 164.1216, "eval_samples_per_second": 167.559, "eval_steps_per_second": 5.24, "step": 5000 }, { "epoch": 0.5967665394691508, "grad_norm": 1.9884291887283325, "learning_rate": 3.508541008541009e-05, "loss": 0.12942585945129395, "step": 5100 }, { "epoch": 0.6084678441646244, "grad_norm": 1.1202914714813232, "learning_rate": 3.479290979290979e-05, "loss": 0.12913961410522462, "step": 5200 }, { "epoch": 0.6201691488600979, "grad_norm": 1.470561146736145, "learning_rate": 3.4500409500409505e-05, "loss": 0.12901023864746095, "step": 5300 }, { "epoch": 0.6318704535555715, "grad_norm": 0.930363655090332, "learning_rate": 3.420790920790921e-05, "loss": 0.12278788566589355, "step": 5400 }, { "epoch": 0.643571758251045, "grad_norm": 1.7242929935455322, "learning_rate": 3.391540891540892e-05, "loss": 0.12592774391174316, "step": 5500 }, { "epoch": 0.6552730629465185, "grad_norm": 1.8954977989196777, "learning_rate": 3.362290862290862e-05, "loss": 0.12447975158691406, "step": 5600 }, { "epoch": 0.666974367641992, "grad_norm": 1.352124571800232, "learning_rate": 3.333040833040833e-05, "loss": 0.1232685947418213, "step": 5700 }, { "epoch": 0.6786756723374656, "grad_norm": 1.2466479539871216, "learning_rate": 3.303790803790804e-05, "loss": 0.11894026756286621, "step": 5800 }, { "epoch": 0.6903769770329392, "grad_norm": 1.429012656211853, "learning_rate": 3.274540774540775e-05, "loss": 0.11655033111572266, "step": 5900 }, { "epoch": 0.7020782817284127, "grad_norm": 1.4804608821868896, "learning_rate": 3.2452907452907454e-05, "loss": 0.11837745666503906, "step": 6000 }, { "epoch": 0.7137795864238863, "grad_norm": 1.3687437772750854, "learning_rate": 3.2160407160407166e-05, "loss": 0.11587305068969726, "step": 6100 }, { "epoch": 0.7254808911193598, "grad_norm": 1.4826337099075317, "learning_rate": 3.186790686790687e-05, "loss": 0.11750144958496093, "step": 6200 }, { "epoch": 0.7371821958148334, "grad_norm": 0.959021270275116, "learning_rate": 3.1575406575406577e-05, "loss": 0.11341402053833008, "step": 6300 }, { "epoch": 0.7488835005103069, "grad_norm": 0.8908078670501709, "learning_rate": 3.128290628290628e-05, "loss": 0.11524638175964355, "step": 6400 }, { "epoch": 0.7605848052057804, "grad_norm": 0.9967349171638489, "learning_rate": 3.0990405990405994e-05, "loss": 0.1124759292602539, "step": 6500 }, { "epoch": 0.772286109901254, "grad_norm": 1.4200743436813354, "learning_rate": 3.06979056979057e-05, "loss": 0.10863906860351563, "step": 6600 }, { "epoch": 0.7839874145967275, "grad_norm": 1.6694557666778564, "learning_rate": 3.0405405405405407e-05, "loss": 0.11165953636169433, "step": 6700 }, { "epoch": 0.7956887192922011, "grad_norm": 1.0781575441360474, "learning_rate": 3.0112905112905116e-05, "loss": 0.1116695213317871, "step": 6800 }, { "epoch": 0.8073900239876747, "grad_norm": 1.3060591220855713, "learning_rate": 2.9820404820404824e-05, "loss": 0.1075587272644043, "step": 6900 }, { "epoch": 0.8190913286831482, "grad_norm": 1.448541283607483, "learning_rate": 2.9527904527904533e-05, "loss": 0.11105222702026367, "step": 7000 }, { "epoch": 0.8307926333786217, "grad_norm": 1.8385331630706787, "learning_rate": 2.9235404235404234e-05, "loss": 0.10690691947937012, "step": 7100 }, { "epoch": 0.8424939380740952, "grad_norm": 1.4097715616226196, "learning_rate": 2.8942903942903943e-05, "loss": 0.10494094848632812, "step": 7200 }, { "epoch": 0.8541952427695688, "grad_norm": 1.7280794382095337, "learning_rate": 2.865040365040365e-05, "loss": 0.10908853530883789, "step": 7300 }, { "epoch": 0.8658965474650423, "grad_norm": 1.0693798065185547, "learning_rate": 2.835790335790336e-05, "loss": 0.10775123596191406, "step": 7400 }, { "epoch": 0.8775978521605159, "grad_norm": 1.5537015199661255, "learning_rate": 2.806540306540307e-05, "loss": 0.10442270278930664, "step": 7500 }, { "epoch": 0.8775978521605159, "eval_accuracy": 0.9217818181818181, "eval_f1": 0.9801159592655914, "eval_loss": 0.017201833426952362, "eval_precision": 0.982831231753525, "eval_recall": 0.9774156484352801, "eval_runtime": 185.5814, "eval_samples_per_second": 148.183, "eval_steps_per_second": 4.634, "step": 7500 }, { "epoch": 0.8892991568559895, "grad_norm": 1.0237928628921509, "learning_rate": 2.7772902772902777e-05, "loss": 0.10509194374084473, "step": 7600 }, { "epoch": 0.901000461551463, "grad_norm": 1.5233043432235718, "learning_rate": 2.7480402480402486e-05, "loss": 0.10594425201416016, "step": 7700 }, { "epoch": 0.9127017662469366, "grad_norm": 0.8517800569534302, "learning_rate": 2.7187902187902187e-05, "loss": 0.10539012908935547, "step": 7800 }, { "epoch": 0.92440307094241, "grad_norm": 1.3082163333892822, "learning_rate": 2.6895401895401896e-05, "loss": 0.09960749626159668, "step": 7900 }, { "epoch": 0.9361043756378836, "grad_norm": 0.8803576827049255, "learning_rate": 2.6602901602901604e-05, "loss": 0.10325809478759766, "step": 8000 }, { "epoch": 0.9478056803333572, "grad_norm": 1.5242176055908203, "learning_rate": 2.6310401310401313e-05, "loss": 0.10169939041137695, "step": 8100 }, { "epoch": 0.9595069850288307, "grad_norm": 1.2510749101638794, "learning_rate": 2.601790101790102e-05, "loss": 0.100986909866333, "step": 8200 }, { "epoch": 0.9712082897243043, "grad_norm": 1.4342858791351318, "learning_rate": 2.5725400725400726e-05, "loss": 0.09906567573547363, "step": 8300 }, { "epoch": 0.9829095944197778, "grad_norm": 1.1943491697311401, "learning_rate": 2.5432900432900435e-05, "loss": 0.09981002807617187, "step": 8400 }, { "epoch": 0.9946108991152514, "grad_norm": 1.0533969402313232, "learning_rate": 2.5140400140400143e-05, "loss": 0.10099024772644043, "step": 8500 }, { "epoch": 1.006201691488601, "grad_norm": 0.952810525894165, "learning_rate": 2.484789984789985e-05, "loss": 0.09148699760437012, "step": 8600 }, { "epoch": 1.0179029961840744, "grad_norm": 0.9051135182380676, "learning_rate": 2.4555399555399554e-05, "loss": 0.08493613243103028, "step": 8700 }, { "epoch": 1.0296043008795481, "grad_norm": 1.4660611152648926, "learning_rate": 2.4262899262899262e-05, "loss": 0.08283645629882813, "step": 8800 }, { "epoch": 1.0413056055750216, "grad_norm": 1.2308636903762817, "learning_rate": 2.397039897039897e-05, "loss": 0.08224470138549805, "step": 8900 }, { "epoch": 1.053006910270495, "grad_norm": 1.180936574935913, "learning_rate": 2.367789867789868e-05, "loss": 0.08456890106201172, "step": 9000 }, { "epoch": 1.0647082149659688, "grad_norm": 1.1681982278823853, "learning_rate": 2.3385398385398384e-05, "loss": 0.08683476448059083, "step": 9100 }, { "epoch": 1.0764095196614423, "grad_norm": 1.1853655576705933, "learning_rate": 2.3092898092898093e-05, "loss": 0.08339482307434082, "step": 9200 }, { "epoch": 1.0881108243569158, "grad_norm": 0.7993655204772949, "learning_rate": 2.28003978003978e-05, "loss": 0.08154037475585937, "step": 9300 }, { "epoch": 1.0998121290523892, "grad_norm": 1.3015902042388916, "learning_rate": 2.250789750789751e-05, "loss": 0.0843494701385498, "step": 9400 }, { "epoch": 1.111513433747863, "grad_norm": 0.927474319934845, "learning_rate": 2.2215397215397215e-05, "loss": 0.08107766151428222, "step": 9500 }, { "epoch": 1.1232147384433364, "grad_norm": 1.2510508298873901, "learning_rate": 2.1922896922896923e-05, "loss": 0.08390594482421875, "step": 9600 }, { "epoch": 1.13491604313881, "grad_norm": 1.012722373008728, "learning_rate": 2.1630396630396632e-05, "loss": 0.08134162902832032, "step": 9700 }, { "epoch": 1.1466173478342836, "grad_norm": 1.8160477876663208, "learning_rate": 2.1337896337896337e-05, "loss": 0.08689288139343261, "step": 9800 }, { "epoch": 1.158318652529757, "grad_norm": 1.6144487857818604, "learning_rate": 2.1045396045396046e-05, "loss": 0.08774213790893555, "step": 9900 }, { "epoch": 1.1700199572252306, "grad_norm": 1.088478922843933, "learning_rate": 2.0752895752895754e-05, "loss": 0.08510435104370118, "step": 10000 }, { "epoch": 1.1700199572252306, "eval_accuracy": 0.9310545454545455, "eval_f1": 0.9822452239157081, "eval_loss": 0.0149807995185256, "eval_precision": 0.9843658878562654, "eval_recall": 0.9801336776170297, "eval_runtime": 187.3637, "eval_samples_per_second": 146.773, "eval_steps_per_second": 4.59, "step": 10000 }, { "epoch": 1.1817212619207043, "grad_norm": 1.2410812377929688, "learning_rate": 2.0460395460395463e-05, "loss": 0.08356905937194824, "step": 10100 }, { "epoch": 1.1934225666161777, "grad_norm": 0.9291555881500244, "learning_rate": 2.0167895167895168e-05, "loss": 0.08259629249572754, "step": 10200 }, { "epoch": 1.2051238713116512, "grad_norm": 1.146966576576233, "learning_rate": 1.9875394875394876e-05, "loss": 0.08215347290039063, "step": 10300 }, { "epoch": 1.2168251760071247, "grad_norm": 0.9125510454177856, "learning_rate": 1.9582894582894585e-05, "loss": 0.08209040641784668, "step": 10400 }, { "epoch": 1.2285264807025984, "grad_norm": 0.906428337097168, "learning_rate": 1.9290394290394293e-05, "loss": 0.08041071891784668, "step": 10500 }, { "epoch": 1.2402277853980719, "grad_norm": 1.0091209411621094, "learning_rate": 1.8997893997894e-05, "loss": 0.07910086631774903, "step": 10600 }, { "epoch": 1.2519290900935454, "grad_norm": 0.9152646660804749, "learning_rate": 1.8705393705393707e-05, "loss": 0.08119074821472168, "step": 10700 }, { "epoch": 1.2636303947890188, "grad_norm": 1.0613765716552734, "learning_rate": 1.8412893412893415e-05, "loss": 0.07813576221466065, "step": 10800 }, { "epoch": 1.2753316994844925, "grad_norm": 0.7284146547317505, "learning_rate": 1.8120393120393124e-05, "loss": 0.08263915061950683, "step": 10900 }, { "epoch": 1.287033004179966, "grad_norm": 0.8245161771774292, "learning_rate": 1.782789282789283e-05, "loss": 0.07991621494293213, "step": 11000 }, { "epoch": 1.2987343088754395, "grad_norm": 0.9131597876548767, "learning_rate": 1.7535392535392538e-05, "loss": 0.08203693389892579, "step": 11100 }, { "epoch": 1.3104356135709132, "grad_norm": 1.4919288158416748, "learning_rate": 1.7242892242892246e-05, "loss": 0.07703531742095947, "step": 11200 }, { "epoch": 1.3221369182663867, "grad_norm": 0.9880580306053162, "learning_rate": 1.695039195039195e-05, "loss": 0.07978516101837158, "step": 11300 }, { "epoch": 1.3338382229618602, "grad_norm": 1.483494758605957, "learning_rate": 1.665789165789166e-05, "loss": 0.0792800235748291, "step": 11400 }, { "epoch": 1.3455395276573339, "grad_norm": 1.2706217765808105, "learning_rate": 1.6365391365391368e-05, "loss": 0.08085798263549805, "step": 11500 }, { "epoch": 1.3572408323528073, "grad_norm": 0.9506115913391113, "learning_rate": 1.6072891072891073e-05, "loss": 0.0732752513885498, "step": 11600 }, { "epoch": 1.3689421370482808, "grad_norm": 0.7671661972999573, "learning_rate": 1.5780390780390782e-05, "loss": 0.07732769966125488, "step": 11700 }, { "epoch": 1.3806434417437545, "grad_norm": 1.3610262870788574, "learning_rate": 1.5487890487890487e-05, "loss": 0.07633553028106689, "step": 11800 }, { "epoch": 1.392344746439228, "grad_norm": 1.1439802646636963, "learning_rate": 1.5195390195390197e-05, "loss": 0.07713084220886231, "step": 11900 }, { "epoch": 1.4040460511347015, "grad_norm": 1.389791488647461, "learning_rate": 1.4902889902889906e-05, "loss": 0.0749136209487915, "step": 12000 }, { "epoch": 1.4157473558301752, "grad_norm": 0.6515536308288574, "learning_rate": 1.461038961038961e-05, "loss": 0.07383899688720703, "step": 12100 }, { "epoch": 1.4274486605256487, "grad_norm": 0.7534876465797424, "learning_rate": 1.431788931788932e-05, "loss": 0.07853510379791259, "step": 12200 }, { "epoch": 1.4391499652211222, "grad_norm": 0.6730746626853943, "learning_rate": 1.4025389025389026e-05, "loss": 0.0697617483139038, "step": 12300 }, { "epoch": 1.4508512699165956, "grad_norm": 1.8400371074676514, "learning_rate": 1.3732888732888733e-05, "loss": 0.07848044872283935, "step": 12400 }, { "epoch": 1.4625525746120691, "grad_norm": 1.2114777565002441, "learning_rate": 1.3440388440388441e-05, "loss": 0.0783261775970459, "step": 12500 }, { "epoch": 1.4625525746120691, "eval_accuracy": 0.9353818181818182, "eval_f1": 0.9834085799751036, "eval_loss": 0.013636507093906403, "eval_precision": 0.9859303596264654, "eval_recall": 0.9808996676591591, "eval_runtime": 167.3633, "eval_samples_per_second": 164.313, "eval_steps_per_second": 5.139, "step": 12500 }, { "epoch": 1.4742538793075428, "grad_norm": 1.7295293807983398, "learning_rate": 1.3147888147888148e-05, "loss": 0.074767746925354, "step": 12600 }, { "epoch": 1.4859551840030163, "grad_norm": 1.08072030544281, "learning_rate": 1.2855387855387857e-05, "loss": 0.07684030532836914, "step": 12700 }, { "epoch": 1.4976564886984898, "grad_norm": 1.334511637687683, "learning_rate": 1.2562887562887562e-05, "loss": 0.07815152645111084, "step": 12800 }, { "epoch": 1.5093577933939635, "grad_norm": 0.8732834458351135, "learning_rate": 1.227038727038727e-05, "loss": 0.07617097854614258, "step": 12900 }, { "epoch": 1.521059098089437, "grad_norm": 1.038492202758789, "learning_rate": 1.1977886977886979e-05, "loss": 0.07482788562774659, "step": 13000 }, { "epoch": 1.5327604027849104, "grad_norm": 1.1699777841567993, "learning_rate": 1.1685386685386686e-05, "loss": 0.07500550746917725, "step": 13100 }, { "epoch": 1.5444617074803841, "grad_norm": 1.1473757028579712, "learning_rate": 1.1392886392886394e-05, "loss": 0.07690254688262939, "step": 13200 }, { "epoch": 1.5561630121758576, "grad_norm": 1.6335910558700562, "learning_rate": 1.1100386100386101e-05, "loss": 0.07375136375427246, "step": 13300 }, { "epoch": 1.567864316871331, "grad_norm": 0.9858669638633728, "learning_rate": 1.0807885807885808e-05, "loss": 0.07300055027008057, "step": 13400 }, { "epoch": 1.5795656215668048, "grad_norm": 1.1223580837249756, "learning_rate": 1.0515385515385516e-05, "loss": 0.07348923683166504, "step": 13500 }, { "epoch": 1.5912669262622783, "grad_norm": 0.8121886253356934, "learning_rate": 1.0222885222885223e-05, "loss": 0.0721654510498047, "step": 13600 }, { "epoch": 1.6029682309577518, "grad_norm": 1.125582218170166, "learning_rate": 9.930384930384932e-06, "loss": 0.07356025695800782, "step": 13700 }, { "epoch": 1.6146695356532255, "grad_norm": 0.9660710692405701, "learning_rate": 9.637884637884638e-06, "loss": 0.07352369308471679, "step": 13800 }, { "epoch": 1.6263708403486987, "grad_norm": 0.9546118378639221, "learning_rate": 9.345384345384347e-06, "loss": 0.07006223201751709, "step": 13900 }, { "epoch": 1.6380721450441724, "grad_norm": 1.2664750814437866, "learning_rate": 9.052884052884054e-06, "loss": 0.07232100963592529, "step": 14000 }, { "epoch": 1.6497734497396461, "grad_norm": 1.087778925895691, "learning_rate": 8.76038376038376e-06, "loss": 0.07518599510192871, "step": 14100 }, { "epoch": 1.6614747544351194, "grad_norm": 1.0009450912475586, "learning_rate": 8.467883467883467e-06, "loss": 0.07395988464355469, "step": 14200 }, { "epoch": 1.673176059130593, "grad_norm": 1.052869200706482, "learning_rate": 8.175383175383176e-06, "loss": 0.07117973327636719, "step": 14300 }, { "epoch": 1.6848773638260666, "grad_norm": 0.9935372471809387, "learning_rate": 7.882882882882883e-06, "loss": 0.07157835006713867, "step": 14400 }, { "epoch": 1.69657866852154, "grad_norm": 0.8762065768241882, "learning_rate": 7.59038259038259e-06, "loss": 0.07200119018554688, "step": 14500 }, { "epoch": 1.7082799732170137, "grad_norm": 1.1860989332199097, "learning_rate": 7.297882297882298e-06, "loss": 0.06994849681854248, "step": 14600 }, { "epoch": 1.7199812779124872, "grad_norm": 0.7438328862190247, "learning_rate": 7.005382005382005e-06, "loss": 0.06839815139770508, "step": 14700 }, { "epoch": 1.7316825826079607, "grad_norm": 1.6703062057495117, "learning_rate": 6.712881712881713e-06, "loss": 0.07001969337463379, "step": 14800 }, { "epoch": 1.7433838873034344, "grad_norm": 1.143515706062317, "learning_rate": 6.42038142038142e-06, "loss": 0.07166263103485107, "step": 14900 }, { "epoch": 1.755085191998908, "grad_norm": 1.0320720672607422, "learning_rate": 6.127881127881129e-06, "loss": 0.0704725456237793, "step": 15000 }, { "epoch": 1.755085191998908, "eval_accuracy": 0.9399272727272727, "eval_f1": 0.9843185306203277, "eval_loss": 0.012572239153087139, "eval_precision": 0.9860882072137978, "eval_recall": 0.9825551945244066, "eval_runtime": 207.4736, "eval_samples_per_second": 132.547, "eval_steps_per_second": 4.145, "step": 15000 }, { "epoch": 1.7667864966943814, "grad_norm": 0.8305580019950867, "learning_rate": 5.8353808353808354e-06, "loss": 0.07156490802764892, "step": 15100 }, { "epoch": 1.778487801389855, "grad_norm": 1.298708438873291, "learning_rate": 5.542880542880543e-06, "loss": 0.06973484516143799, "step": 15200 }, { "epoch": 1.7901891060853286, "grad_norm": 0.8593277931213379, "learning_rate": 5.250380250380251e-06, "loss": 0.06879038333892823, "step": 15300 }, { "epoch": 1.801890410780802, "grad_norm": 0.732769787311554, "learning_rate": 4.9578799578799576e-06, "loss": 0.06908240318298339, "step": 15400 }, { "epoch": 1.8135917154762757, "grad_norm": 1.3583112955093384, "learning_rate": 4.665379665379665e-06, "loss": 0.07114370822906495, "step": 15500 }, { "epoch": 1.825293020171749, "grad_norm": 0.9253562092781067, "learning_rate": 4.372879372879373e-06, "loss": 0.07082613945007324, "step": 15600 }, { "epoch": 1.8369943248672227, "grad_norm": 1.340505599975586, "learning_rate": 4.0803790803790806e-06, "loss": 0.06995931625366211, "step": 15700 }, { "epoch": 1.8486956295626964, "grad_norm": 0.9078701734542847, "learning_rate": 3.7878787878787882e-06, "loss": 0.06670016288757324, "step": 15800 }, { "epoch": 1.8603969342581697, "grad_norm": 1.2157623767852783, "learning_rate": 3.495378495378496e-06, "loss": 0.0690723991394043, "step": 15900 }, { "epoch": 1.8720982389536434, "grad_norm": 1.7708772420883179, "learning_rate": 3.202878202878203e-06, "loss": 0.06770940780639649, "step": 16000 }, { "epoch": 1.8837995436491168, "grad_norm": 0.976672887802124, "learning_rate": 2.9103779103779103e-06, "loss": 0.0646241569519043, "step": 16100 }, { "epoch": 1.8955008483445903, "grad_norm": 1.7383304834365845, "learning_rate": 2.617877617877618e-06, "loss": 0.06627600193023682, "step": 16200 }, { "epoch": 1.907202153040064, "grad_norm": 0.8907257318496704, "learning_rate": 2.3253773253773257e-06, "loss": 0.06847190380096435, "step": 16300 }, { "epoch": 1.9189034577355375, "grad_norm": 0.7403037548065186, "learning_rate": 2.032877032877033e-06, "loss": 0.06623115539550781, "step": 16400 }, { "epoch": 1.930604762431011, "grad_norm": 1.0598572492599487, "learning_rate": 1.7403767403767406e-06, "loss": 0.06684101104736329, "step": 16500 }, { "epoch": 1.9423060671264847, "grad_norm": 1.2067943811416626, "learning_rate": 1.447876447876448e-06, "loss": 0.06935319423675537, "step": 16600 }, { "epoch": 1.9540073718219582, "grad_norm": 0.9936091899871826, "learning_rate": 1.1553761553761555e-06, "loss": 0.06889093399047852, "step": 16700 }, { "epoch": 1.9657086765174316, "grad_norm": 1.5107425451278687, "learning_rate": 8.628758628758629e-07, "loss": 0.07209495544433593, "step": 16800 }, { "epoch": 1.9774099812129053, "grad_norm": 1.0668072700500488, "learning_rate": 5.703755703755704e-07, "loss": 0.06988236904144288, "step": 16900 }, { "epoch": 1.9891112859083788, "grad_norm": 1.113166093826294, "learning_rate": 2.7787527787527786e-07, "loss": 0.06924228668212891, "step": 17000 }, { "epoch": 2.0, "eval_accuracy": 0.9411636363636363, "eval_f1": 0.9845099349219312, "eval_loss": 0.012292231433093548, "eval_precision": 0.9859127008710089, "eval_recall": 0.9831111550388555, "eval_runtime": 183.6292, "eval_samples_per_second": 149.758, "eval_steps_per_second": 4.683, "step": 17094 }, { "epoch": 2.0, "step": 17094, "total_flos": 1.9402797892099978e+18, "train_loss": 0.08505657710317173, "train_runtime": 156175.5955, "train_samples_per_second": 63.038, "train_steps_per_second": 0.109 } ], "logging_steps": 100, "max_steps": 17094, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 2500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.9402797892099978e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }