{ "best_metric": 0.8973993679198728, "best_model_checkpoint": "./nlu_finetuned_models/qqp/roberta-base_lr1e-05/checkpoint-204670", "epoch": 10.0, "eval_steps": 500, "global_step": 204670, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.024429569550984513, "grad_norm": 10.54019546508789, "learning_rate": 4.0713296962788053e-07, "loss": 0.6758, "step": 500 }, { "epoch": 0.048859139101969026, "grad_norm": 7.6315178871154785, "learning_rate": 8.142659392557611e-07, "loss": 0.6406, "step": 1000 }, { "epoch": 0.07328870865295353, "grad_norm": 25.296419143676758, "learning_rate": 1.2213989088836414e-06, "loss": 0.528, "step": 1500 }, { "epoch": 0.09771827820393805, "grad_norm": 134.22874450683594, "learning_rate": 1.6285318785115221e-06, "loss": 0.4561, "step": 2000 }, { "epoch": 0.12214784775492256, "grad_norm": 34.93831253051758, "learning_rate": 2.0356648481394024e-06, "loss": 0.4151, "step": 2500 }, { "epoch": 0.14657741730590707, "grad_norm": 15.966097831726074, "learning_rate": 2.442797817767283e-06, "loss": 0.3976, "step": 3000 }, { "epoch": 0.17100698685689159, "grad_norm": 19.514278411865234, "learning_rate": 2.8499307873951637e-06, "loss": 0.377, "step": 3500 }, { "epoch": 0.1954365564078761, "grad_norm": 15.483463287353516, "learning_rate": 3.2570637570230442e-06, "loss": 0.3688, "step": 4000 }, { "epoch": 0.2198661259588606, "grad_norm": 17.194320678710938, "learning_rate": 3.6641967266509243e-06, "loss": 0.373, "step": 4500 }, { "epoch": 0.24429569550984512, "grad_norm": 7.317613124847412, "learning_rate": 4.071329696278805e-06, "loss": 0.3518, "step": 5000 }, { "epoch": 0.2687252650608296, "grad_norm": 15.082125663757324, "learning_rate": 4.478462665906685e-06, "loss": 0.3587, "step": 5500 }, { "epoch": 0.29315483461181413, "grad_norm": 49.410526275634766, "learning_rate": 4.885595635534566e-06, "loss": 0.3513, "step": 6000 }, { "epoch": 0.31758440416279865, "grad_norm": 52.64762878417969, "learning_rate": 5.292728605162446e-06, "loss": 0.3426, "step": 6500 }, { "epoch": 0.34201397371378317, "grad_norm": 21.950075149536133, "learning_rate": 5.6998615747903275e-06, "loss": 0.3401, "step": 7000 }, { "epoch": 0.3664435432647677, "grad_norm": 14.155437469482422, "learning_rate": 6.106994544418208e-06, "loss": 0.3343, "step": 7500 }, { "epoch": 0.3908731128157522, "grad_norm": 12.090170860290527, "learning_rate": 6.5141275140460884e-06, "loss": 0.3163, "step": 8000 }, { "epoch": 0.4153026823667367, "grad_norm": 17.292564392089844, "learning_rate": 6.921260483673968e-06, "loss": 0.33, "step": 8500 }, { "epoch": 0.4397322519177212, "grad_norm": 29.74077033996582, "learning_rate": 7.3283934533018485e-06, "loss": 0.3326, "step": 9000 }, { "epoch": 0.4641618214687057, "grad_norm": 16.858999252319336, "learning_rate": 7.73552642292973e-06, "loss": 0.3229, "step": 9500 }, { "epoch": 0.48859139101969024, "grad_norm": 16.417327880859375, "learning_rate": 8.14265939255761e-06, "loss": 0.3207, "step": 10000 }, { "epoch": 0.5130209605706747, "grad_norm": 18.313976287841797, "learning_rate": 8.54979236218549e-06, "loss": 0.3154, "step": 10500 }, { "epoch": 0.5374505301216592, "grad_norm": 26.479129791259766, "learning_rate": 8.95692533181337e-06, "loss": 0.3199, "step": 11000 }, { "epoch": 0.5618800996726437, "grad_norm": 11.478958129882812, "learning_rate": 9.364058301441251e-06, "loss": 0.3041, "step": 11500 }, { "epoch": 0.5863096692236283, "grad_norm": 18.56060028076172, "learning_rate": 9.771191271069131e-06, "loss": 0.3027, "step": 12000 }, { "epoch": 0.6107392387746128, "grad_norm": 7.155969142913818, "learning_rate": 9.988616812811544e-06, "loss": 0.3122, "step": 12500 }, { "epoch": 0.6351688083255973, "grad_norm": 8.636488914489746, "learning_rate": 9.962627800965753e-06, "loss": 0.3218, "step": 13000 }, { "epoch": 0.6595983778765818, "grad_norm": 23.951885223388672, "learning_rate": 9.936638789119961e-06, "loss": 0.3109, "step": 13500 }, { "epoch": 0.6840279474275663, "grad_norm": 12.705269813537598, "learning_rate": 9.91064977727417e-06, "loss": 0.3054, "step": 14000 }, { "epoch": 0.7084575169785509, "grad_norm": 19.542213439941406, "learning_rate": 9.884660765428378e-06, "loss": 0.2983, "step": 14500 }, { "epoch": 0.7328870865295354, "grad_norm": 18.7448673248291, "learning_rate": 9.858671753582586e-06, "loss": 0.2882, "step": 15000 }, { "epoch": 0.7573166560805199, "grad_norm": 12.782718658447266, "learning_rate": 9.832682741736795e-06, "loss": 0.2956, "step": 15500 }, { "epoch": 0.7817462256315044, "grad_norm": 24.18721580505371, "learning_rate": 9.806693729891003e-06, "loss": 0.2958, "step": 16000 }, { "epoch": 0.8061757951824888, "grad_norm": 34.30314636230469, "learning_rate": 9.78070471804521e-06, "loss": 0.2972, "step": 16500 }, { "epoch": 0.8306053647334734, "grad_norm": 52.42458724975586, "learning_rate": 9.75471570619942e-06, "loss": 0.2831, "step": 17000 }, { "epoch": 0.8550349342844579, "grad_norm": 15.497232437133789, "learning_rate": 9.728726694353629e-06, "loss": 0.2976, "step": 17500 }, { "epoch": 0.8794645038354424, "grad_norm": 30.495269775390625, "learning_rate": 9.702737682507836e-06, "loss": 0.2977, "step": 18000 }, { "epoch": 0.9038940733864269, "grad_norm": 2.9801318645477295, "learning_rate": 9.676748670662046e-06, "loss": 0.2834, "step": 18500 }, { "epoch": 0.9283236429374114, "grad_norm": 12.599749565124512, "learning_rate": 9.650759658816253e-06, "loss": 0.2919, "step": 19000 }, { "epoch": 0.952753212488396, "grad_norm": 8.34764575958252, "learning_rate": 9.624770646970461e-06, "loss": 0.28, "step": 19500 }, { "epoch": 0.9771827820393805, "grad_norm": 16.70107650756836, "learning_rate": 9.59878163512467e-06, "loss": 0.2802, "step": 20000 }, { "epoch": 1.0, "eval_accuracy": 0.8791809811735606, "eval_averaged_scores": 0.8627373437336334, "eval_f1": 0.8462937062937063, "eval_loss": 0.2766279876232147, "eval_runtime": 20.1271, "eval_samples_per_second": 1807.757, "eval_steps_per_second": 113.031, "step": 20467 }, { "epoch": 1.001612351590365, "grad_norm": 2.792717218399048, "learning_rate": 9.572792623278878e-06, "loss": 0.2782, "step": 20500 }, { "epoch": 1.0260419211413494, "grad_norm": 8.300065040588379, "learning_rate": 9.546803611433086e-06, "loss": 0.2515, "step": 21000 }, { "epoch": 1.050471490692334, "grad_norm": 29.82993507385254, "learning_rate": 9.520814599587295e-06, "loss": 0.2532, "step": 21500 }, { "epoch": 1.0749010602433184, "grad_norm": 51.71006393432617, "learning_rate": 9.494825587741504e-06, "loss": 0.2411, "step": 22000 }, { "epoch": 1.099330629794303, "grad_norm": 12.74963092803955, "learning_rate": 9.468836575895712e-06, "loss": 0.2545, "step": 22500 }, { "epoch": 1.1237601993452875, "grad_norm": 11.051137924194336, "learning_rate": 9.442847564049921e-06, "loss": 0.2597, "step": 23000 }, { "epoch": 1.1481897688962721, "grad_norm": 16.989593505859375, "learning_rate": 9.416858552204129e-06, "loss": 0.2508, "step": 23500 }, { "epoch": 1.1726193384472565, "grad_norm": 18.592727661132812, "learning_rate": 9.390869540358337e-06, "loss": 0.2518, "step": 24000 }, { "epoch": 1.1970489079982412, "grad_norm": 14.818976402282715, "learning_rate": 9.364880528512546e-06, "loss": 0.2497, "step": 24500 }, { "epoch": 1.2214784775492256, "grad_norm": 11.762436866760254, "learning_rate": 9.338891516666754e-06, "loss": 0.2477, "step": 25000 }, { "epoch": 1.24590804710021, "grad_norm": 3.5736145973205566, "learning_rate": 9.312902504820961e-06, "loss": 0.2557, "step": 25500 }, { "epoch": 1.2703376166511946, "grad_norm": 17.63505744934082, "learning_rate": 9.28691349297517e-06, "loss": 0.251, "step": 26000 }, { "epoch": 1.2947671862021792, "grad_norm": 13.431313514709473, "learning_rate": 9.26092448112938e-06, "loss": 0.2541, "step": 26500 }, { "epoch": 1.3191967557531636, "grad_norm": 29.016206741333008, "learning_rate": 9.234935469283588e-06, "loss": 0.2417, "step": 27000 }, { "epoch": 1.343626325304148, "grad_norm": 21.277286529541016, "learning_rate": 9.208946457437797e-06, "loss": 0.2466, "step": 27500 }, { "epoch": 1.3680558948551327, "grad_norm": 12.927133560180664, "learning_rate": 9.182957445592005e-06, "loss": 0.2484, "step": 28000 }, { "epoch": 1.392485464406117, "grad_norm": 14.714786529541016, "learning_rate": 9.156968433746212e-06, "loss": 0.2543, "step": 28500 }, { "epoch": 1.4169150339571017, "grad_norm": 4.885636806488037, "learning_rate": 9.130979421900422e-06, "loss": 0.2554, "step": 29000 }, { "epoch": 1.4413446035080861, "grad_norm": 13.911579132080078, "learning_rate": 9.10499041005463e-06, "loss": 0.2504, "step": 29500 }, { "epoch": 1.4657741730590708, "grad_norm": 8.80788803100586, "learning_rate": 9.079001398208837e-06, "loss": 0.2379, "step": 30000 }, { "epoch": 1.4902037426100552, "grad_norm": 14.809807777404785, "learning_rate": 9.053012386363046e-06, "loss": 0.2462, "step": 30500 }, { "epoch": 1.5146333121610396, "grad_norm": 2.741872549057007, "learning_rate": 9.027023374517256e-06, "loss": 0.248, "step": 31000 }, { "epoch": 1.5390628817120242, "grad_norm": 15.03188419342041, "learning_rate": 9.001034362671463e-06, "loss": 0.2523, "step": 31500 }, { "epoch": 1.5634924512630088, "grad_norm": 13.394209861755371, "learning_rate": 8.975045350825672e-06, "loss": 0.2445, "step": 32000 }, { "epoch": 1.5879220208139933, "grad_norm": 17.654428482055664, "learning_rate": 8.94905633897988e-06, "loss": 0.2419, "step": 32500 }, { "epoch": 1.6123515903649777, "grad_norm": 2.1160449981689453, "learning_rate": 8.923067327134088e-06, "loss": 0.2423, "step": 33000 }, { "epoch": 1.6367811599159623, "grad_norm": 9.247222900390625, "learning_rate": 8.897078315288297e-06, "loss": 0.2371, "step": 33500 }, { "epoch": 1.661210729466947, "grad_norm": 12.957727432250977, "learning_rate": 8.871089303442505e-06, "loss": 0.245, "step": 34000 }, { "epoch": 1.6856402990179313, "grad_norm": 18.7705020904541, "learning_rate": 8.845100291596712e-06, "loss": 0.2409, "step": 34500 }, { "epoch": 1.7100698685689157, "grad_norm": 17.920528411865234, "learning_rate": 8.819111279750922e-06, "loss": 0.2369, "step": 35000 }, { "epoch": 1.7344994381199004, "grad_norm": 6.508107662200928, "learning_rate": 8.793122267905131e-06, "loss": 0.2445, "step": 35500 }, { "epoch": 1.758929007670885, "grad_norm": 15.203210830688477, "learning_rate": 8.767133256059339e-06, "loss": 0.2441, "step": 36000 }, { "epoch": 1.7833585772218692, "grad_norm": 3.705578327178955, "learning_rate": 8.741144244213548e-06, "loss": 0.2359, "step": 36500 }, { "epoch": 1.8077881467728538, "grad_norm": 26.21458625793457, "learning_rate": 8.715155232367756e-06, "loss": 0.2339, "step": 37000 }, { "epoch": 1.8322177163238385, "grad_norm": 3.0099897384643555, "learning_rate": 8.689166220521963e-06, "loss": 0.2376, "step": 37500 }, { "epoch": 1.8566472858748229, "grad_norm": 17.3660888671875, "learning_rate": 8.663177208676173e-06, "loss": 0.2415, "step": 38000 }, { "epoch": 1.8810768554258073, "grad_norm": 23.040613174438477, "learning_rate": 8.63718819683038e-06, "loss": 0.236, "step": 38500 }, { "epoch": 1.905506424976792, "grad_norm": 7.123594760894775, "learning_rate": 8.61119918498459e-06, "loss": 0.2372, "step": 39000 }, { "epoch": 1.9299359945277765, "grad_norm": 34.456966400146484, "learning_rate": 8.585210173138797e-06, "loss": 0.2391, "step": 39500 }, { "epoch": 1.954365564078761, "grad_norm": 23.899415969848633, "learning_rate": 8.559221161293007e-06, "loss": 0.2353, "step": 40000 }, { "epoch": 1.9787951336297454, "grad_norm": 19.077669143676758, "learning_rate": 8.533232149447214e-06, "loss": 0.2369, "step": 40500 }, { "epoch": 2.0, "eval_accuracy": 0.8980623883468463, "eval_averaged_scores": 0.8814449118745056, "eval_f1": 0.8648274354021648, "eval_loss": 0.28552600741386414, "eval_runtime": 20.2684, "eval_samples_per_second": 1795.163, "eval_steps_per_second": 112.244, "step": 40934 }, { "epoch": 2.00322470318073, "grad_norm": 20.173015594482422, "learning_rate": 8.507243137601424e-06, "loss": 0.2315, "step": 41000 }, { "epoch": 2.0276542727317146, "grad_norm": 39.991127014160156, "learning_rate": 8.481254125755631e-06, "loss": 0.1966, "step": 41500 }, { "epoch": 2.052083842282699, "grad_norm": 30.70036506652832, "learning_rate": 8.455265113909839e-06, "loss": 0.2143, "step": 42000 }, { "epoch": 2.0765134118336834, "grad_norm": 17.682048797607422, "learning_rate": 8.429276102064048e-06, "loss": 0.2001, "step": 42500 }, { "epoch": 2.100942981384668, "grad_norm": 20.595016479492188, "learning_rate": 8.403287090218256e-06, "loss": 0.2088, "step": 43000 }, { "epoch": 2.1253725509356527, "grad_norm": 4.142117977142334, "learning_rate": 8.377298078372465e-06, "loss": 0.2159, "step": 43500 }, { "epoch": 2.149802120486637, "grad_norm": 1.1202857494354248, "learning_rate": 8.351309066526673e-06, "loss": 0.2031, "step": 44000 }, { "epoch": 2.1742316900376215, "grad_norm": 26.033584594726562, "learning_rate": 8.325320054680882e-06, "loss": 0.216, "step": 44500 }, { "epoch": 2.198661259588606, "grad_norm": 5.992620468139648, "learning_rate": 8.29933104283509e-06, "loss": 0.2099, "step": 45000 }, { "epoch": 2.2230908291395908, "grad_norm": 1.6579467058181763, "learning_rate": 8.2733420309893e-06, "loss": 0.2048, "step": 45500 }, { "epoch": 2.247520398690575, "grad_norm": 30.13793182373047, "learning_rate": 8.247353019143507e-06, "loss": 0.2148, "step": 46000 }, { "epoch": 2.2719499682415596, "grad_norm": 15.029723167419434, "learning_rate": 8.221364007297714e-06, "loss": 0.199, "step": 46500 }, { "epoch": 2.2963795377925442, "grad_norm": 7.096563816070557, "learning_rate": 8.195374995451924e-06, "loss": 0.2133, "step": 47000 }, { "epoch": 2.3208091073435284, "grad_norm": 21.75437355041504, "learning_rate": 8.169385983606131e-06, "loss": 0.2149, "step": 47500 }, { "epoch": 2.345238676894513, "grad_norm": 17.925037384033203, "learning_rate": 8.14339697176034e-06, "loss": 0.2006, "step": 48000 }, { "epoch": 2.3696682464454977, "grad_norm": 3.024029493331909, "learning_rate": 8.117407959914548e-06, "loss": 0.2001, "step": 48500 }, { "epoch": 2.3940978159964823, "grad_norm": 17.959518432617188, "learning_rate": 8.091418948068758e-06, "loss": 0.2068, "step": 49000 }, { "epoch": 2.4185273855474665, "grad_norm": 12.003142356872559, "learning_rate": 8.065429936222965e-06, "loss": 0.2113, "step": 49500 }, { "epoch": 2.442956955098451, "grad_norm": 18.196792602539062, "learning_rate": 8.039440924377175e-06, "loss": 0.2031, "step": 50000 }, { "epoch": 2.4673865246494358, "grad_norm": 19.81887435913086, "learning_rate": 8.013451912531382e-06, "loss": 0.2156, "step": 50500 }, { "epoch": 2.49181609420042, "grad_norm": 20.83788299560547, "learning_rate": 7.98746290068559e-06, "loss": 0.2036, "step": 51000 }, { "epoch": 2.5162456637514046, "grad_norm": 28.70236587524414, "learning_rate": 7.9614738888398e-06, "loss": 0.2109, "step": 51500 }, { "epoch": 2.540675233302389, "grad_norm": 28.118547439575195, "learning_rate": 7.935484876994007e-06, "loss": 0.2022, "step": 52000 }, { "epoch": 2.565104802853374, "grad_norm": 5.818374156951904, "learning_rate": 7.909495865148216e-06, "loss": 0.2061, "step": 52500 }, { "epoch": 2.5895343724043585, "grad_norm": 11.427398681640625, "learning_rate": 7.883506853302424e-06, "loss": 0.209, "step": 53000 }, { "epoch": 2.6139639419553427, "grad_norm": 16.268341064453125, "learning_rate": 7.857517841456633e-06, "loss": 0.2123, "step": 53500 }, { "epoch": 2.6383935115063273, "grad_norm": 23.71199607849121, "learning_rate": 7.831528829610841e-06, "loss": 0.2125, "step": 54000 }, { "epoch": 2.662823081057312, "grad_norm": 27.364391326904297, "learning_rate": 7.80553981776505e-06, "loss": 0.2014, "step": 54500 }, { "epoch": 2.687252650608296, "grad_norm": 27.82223129272461, "learning_rate": 7.779550805919258e-06, "loss": 0.1981, "step": 55000 }, { "epoch": 2.7116822201592807, "grad_norm": 19.91096305847168, "learning_rate": 7.753561794073467e-06, "loss": 0.2048, "step": 55500 }, { "epoch": 2.7361117897102654, "grad_norm": 22.202478408813477, "learning_rate": 7.727572782227675e-06, "loss": 0.2077, "step": 56000 }, { "epoch": 2.76054135926125, "grad_norm": 15.403483390808105, "learning_rate": 7.701583770381883e-06, "loss": 0.2057, "step": 56500 }, { "epoch": 2.784970928812234, "grad_norm": 2.367913007736206, "learning_rate": 7.675594758536092e-06, "loss": 0.2097, "step": 57000 }, { "epoch": 2.809400498363219, "grad_norm": 30.33095932006836, "learning_rate": 7.6496057466903e-06, "loss": 0.2073, "step": 57500 }, { "epoch": 2.8338300679142034, "grad_norm": 16.65831184387207, "learning_rate": 7.623616734844508e-06, "loss": 0.2109, "step": 58000 }, { "epoch": 2.8582596374651876, "grad_norm": 26.70164680480957, "learning_rate": 7.597627722998717e-06, "loss": 0.1967, "step": 58500 }, { "epoch": 2.8826892070161723, "grad_norm": 15.616362571716309, "learning_rate": 7.571638711152926e-06, "loss": 0.21, "step": 59000 }, { "epoch": 2.907118776567157, "grad_norm": 1.022119164466858, "learning_rate": 7.5456496993071335e-06, "loss": 0.2154, "step": 59500 }, { "epoch": 2.9315483461181415, "grad_norm": 23.004011154174805, "learning_rate": 7.519660687461342e-06, "loss": 0.2081, "step": 60000 }, { "epoch": 2.955977915669126, "grad_norm": 49.54454040527344, "learning_rate": 7.4936716756155505e-06, "loss": 0.2066, "step": 60500 }, { "epoch": 2.9804074852201103, "grad_norm": 22.06740379333496, "learning_rate": 7.467682663769759e-06, "loss": 0.2111, "step": 61000 }, { "epoch": 3.0, "eval_accuracy": 0.9057578672529889, "eval_averaged_scores": 0.8894141918163889, "eval_f1": 0.873070516379789, "eval_loss": 0.2923884093761444, "eval_runtime": 20.2639, "eval_samples_per_second": 1795.555, "eval_steps_per_second": 112.268, "step": 61401 }, { "epoch": 3.004837054771095, "grad_norm": 11.454232215881348, "learning_rate": 7.441693651923967e-06, "loss": 0.2018, "step": 61500 }, { "epoch": 3.0292666243220796, "grad_norm": 21.45491600036621, "learning_rate": 7.415704640078175e-06, "loss": 0.1833, "step": 62000 }, { "epoch": 3.053696193873064, "grad_norm": 0.8129162192344666, "learning_rate": 7.389715628232384e-06, "loss": 0.1801, "step": 62500 }, { "epoch": 3.0781257634240484, "grad_norm": 0.3495664894580841, "learning_rate": 7.363726616386593e-06, "loss": 0.1776, "step": 63000 }, { "epoch": 3.102555332975033, "grad_norm": 0.13657505810260773, "learning_rate": 7.337737604540801e-06, "loss": 0.182, "step": 63500 }, { "epoch": 3.1269849025260177, "grad_norm": 86.07280731201172, "learning_rate": 7.311748592695009e-06, "loss": 0.1861, "step": 64000 }, { "epoch": 3.151414472077002, "grad_norm": 24.988689422607422, "learning_rate": 7.2857595808492175e-06, "loss": 0.1776, "step": 64500 }, { "epoch": 3.1758440416279865, "grad_norm": 37.55842208862305, "learning_rate": 7.259770569003426e-06, "loss": 0.1854, "step": 65000 }, { "epoch": 3.200273611178971, "grad_norm": 69.79829406738281, "learning_rate": 7.2337815571576345e-06, "loss": 0.1747, "step": 65500 }, { "epoch": 3.2247031807299553, "grad_norm": 28.17424201965332, "learning_rate": 7.207792545311842e-06, "loss": 0.1938, "step": 66000 }, { "epoch": 3.24913275028094, "grad_norm": 56.655364990234375, "learning_rate": 7.181803533466051e-06, "loss": 0.1912, "step": 66500 }, { "epoch": 3.2735623198319246, "grad_norm": 0.3666965067386627, "learning_rate": 7.155814521620259e-06, "loss": 0.1915, "step": 67000 }, { "epoch": 3.297991889382909, "grad_norm": 9.555418014526367, "learning_rate": 7.1298255097744685e-06, "loss": 0.188, "step": 67500 }, { "epoch": 3.3224214589338934, "grad_norm": 21.880718231201172, "learning_rate": 7.103836497928677e-06, "loss": 0.1824, "step": 68000 }, { "epoch": 3.346851028484878, "grad_norm": 34.551021575927734, "learning_rate": 7.077847486082885e-06, "loss": 0.1886, "step": 68500 }, { "epoch": 3.3712805980358627, "grad_norm": 18.153520584106445, "learning_rate": 7.051858474237093e-06, "loss": 0.1908, "step": 69000 }, { "epoch": 3.3957101675868473, "grad_norm": 0.3916667103767395, "learning_rate": 7.025869462391302e-06, "loss": 0.182, "step": 69500 }, { "epoch": 3.4201397371378315, "grad_norm": 28.306758880615234, "learning_rate": 6.99988045054551e-06, "loss": 0.1852, "step": 70000 }, { "epoch": 3.444569306688816, "grad_norm": 29.383926391601562, "learning_rate": 6.973891438699718e-06, "loss": 0.1763, "step": 70500 }, { "epoch": 3.4689988762398007, "grad_norm": 16.787750244140625, "learning_rate": 6.947902426853926e-06, "loss": 0.188, "step": 71000 }, { "epoch": 3.4934284457907854, "grad_norm": 22.620866775512695, "learning_rate": 6.921913415008135e-06, "loss": 0.193, "step": 71500 }, { "epoch": 3.5178580153417696, "grad_norm": 0.9201492667198181, "learning_rate": 6.895924403162344e-06, "loss": 0.186, "step": 72000 }, { "epoch": 3.542287584892754, "grad_norm": 46.33378601074219, "learning_rate": 6.8699353913165525e-06, "loss": 0.1973, "step": 72500 }, { "epoch": 3.566717154443739, "grad_norm": 3.13004207611084, "learning_rate": 6.84394637947076e-06, "loss": 0.2003, "step": 73000 }, { "epoch": 3.591146723994723, "grad_norm": 29.041542053222656, "learning_rate": 6.817957367624969e-06, "loss": 0.1874, "step": 73500 }, { "epoch": 3.6155762935457076, "grad_norm": 9.451910018920898, "learning_rate": 6.791968355779177e-06, "loss": 0.1757, "step": 74000 }, { "epoch": 3.6400058630966923, "grad_norm": 0.4655057191848755, "learning_rate": 6.765979343933386e-06, "loss": 0.1895, "step": 74500 }, { "epoch": 3.664435432647677, "grad_norm": 31.763437271118164, "learning_rate": 6.739990332087593e-06, "loss": 0.1822, "step": 75000 }, { "epoch": 3.6888650021986615, "grad_norm": 38.27184295654297, "learning_rate": 6.714001320241802e-06, "loss": 0.1849, "step": 75500 }, { "epoch": 3.7132945717496457, "grad_norm": 6.707937717437744, "learning_rate": 6.68801230839601e-06, "loss": 0.2007, "step": 76000 }, { "epoch": 3.7377241413006304, "grad_norm": 30.213747024536133, "learning_rate": 6.66202329655022e-06, "loss": 0.1903, "step": 76500 }, { "epoch": 3.7621537108516145, "grad_norm": 51.97517776489258, "learning_rate": 6.636034284704428e-06, "loss": 0.1875, "step": 77000 }, { "epoch": 3.786583280402599, "grad_norm": 28.538976669311523, "learning_rate": 6.610045272858637e-06, "loss": 0.1857, "step": 77500 }, { "epoch": 3.811012849953584, "grad_norm": 0.3994407057762146, "learning_rate": 6.584056261012844e-06, "loss": 0.1819, "step": 78000 }, { "epoch": 3.8354424195045684, "grad_norm": 28.58909797668457, "learning_rate": 6.558067249167053e-06, "loss": 0.1787, "step": 78500 }, { "epoch": 3.859871989055553, "grad_norm": 0.4431411325931549, "learning_rate": 6.532078237321261e-06, "loss": 0.1864, "step": 79000 }, { "epoch": 3.8843015586065373, "grad_norm": 65.34574890136719, "learning_rate": 6.50608922547547e-06, "loss": 0.1844, "step": 79500 }, { "epoch": 3.908731128157522, "grad_norm": 12.750852584838867, "learning_rate": 6.480100213629677e-06, "loss": 0.1824, "step": 80000 }, { "epoch": 3.9331606977085065, "grad_norm": 39.058040618896484, "learning_rate": 6.454111201783886e-06, "loss": 0.1849, "step": 80500 }, { "epoch": 3.9575902672594907, "grad_norm": 2.713413715362549, "learning_rate": 6.428122189938095e-06, "loss": 0.1768, "step": 81000 }, { "epoch": 3.9820198368104753, "grad_norm": 1.6995596885681152, "learning_rate": 6.402133178092304e-06, "loss": 0.1835, "step": 81500 }, { "epoch": 4.0, "eval_accuracy": 0.9077641885392332, "eval_averaged_scores": 0.8907519212071636, "eval_f1": 0.873739653875094, "eval_loss": 0.33108219504356384, "eval_runtime": 20.2844, "eval_samples_per_second": 1793.747, "eval_steps_per_second": 112.155, "step": 81868 }, { "epoch": 4.00644940636146, "grad_norm": 0.7421801686286926, "learning_rate": 6.376144166246512e-06, "loss": 0.1796, "step": 82000 }, { "epoch": 4.030878975912445, "grad_norm": 30.367263793945312, "learning_rate": 6.35015515440072e-06, "loss": 0.1649, "step": 82500 }, { "epoch": 4.055308545463429, "grad_norm": 14.93786907196045, "learning_rate": 6.324166142554928e-06, "loss": 0.164, "step": 83000 }, { "epoch": 4.079738115014414, "grad_norm": 0.6399831175804138, "learning_rate": 6.298177130709137e-06, "loss": 0.1605, "step": 83500 }, { "epoch": 4.104167684565398, "grad_norm": 23.02483558654785, "learning_rate": 6.272188118863345e-06, "loss": 0.1659, "step": 84000 }, { "epoch": 4.128597254116382, "grad_norm": 0.35685986280441284, "learning_rate": 6.246199107017553e-06, "loss": 0.1573, "step": 84500 }, { "epoch": 4.153026823667367, "grad_norm": 52.071285247802734, "learning_rate": 6.220210095171761e-06, "loss": 0.1683, "step": 85000 }, { "epoch": 4.1774563932183515, "grad_norm": 4.686119556427002, "learning_rate": 6.194221083325971e-06, "loss": 0.1641, "step": 85500 }, { "epoch": 4.201885962769336, "grad_norm": 0.708002507686615, "learning_rate": 6.168232071480179e-06, "loss": 0.1603, "step": 86000 }, { "epoch": 4.226315532320321, "grad_norm": 1.475074291229248, "learning_rate": 6.142243059634388e-06, "loss": 0.1671, "step": 86500 }, { "epoch": 4.250745101871305, "grad_norm": 1.395318627357483, "learning_rate": 6.116254047788595e-06, "loss": 0.1575, "step": 87000 }, { "epoch": 4.275174671422289, "grad_norm": 0.3956087827682495, "learning_rate": 6.090265035942804e-06, "loss": 0.1591, "step": 87500 }, { "epoch": 4.299604240973274, "grad_norm": 0.6056288480758667, "learning_rate": 6.064276024097012e-06, "loss": 0.1614, "step": 88000 }, { "epoch": 4.324033810524258, "grad_norm": 1.0418305397033691, "learning_rate": 6.038287012251221e-06, "loss": 0.1715, "step": 88500 }, { "epoch": 4.348463380075243, "grad_norm": 1.0520128011703491, "learning_rate": 6.0122980004054285e-06, "loss": 0.1762, "step": 89000 }, { "epoch": 4.372892949626228, "grad_norm": 0.8157211542129517, "learning_rate": 5.986308988559637e-06, "loss": 0.1702, "step": 89500 }, { "epoch": 4.397322519177212, "grad_norm": 21.881351470947266, "learning_rate": 5.9603199767138455e-06, "loss": 0.1649, "step": 90000 }, { "epoch": 4.421752088728197, "grad_norm": 21.02885627746582, "learning_rate": 5.934330964868055e-06, "loss": 0.1681, "step": 90500 }, { "epoch": 4.4461816582791815, "grad_norm": 17.863567352294922, "learning_rate": 5.908341953022263e-06, "loss": 0.1705, "step": 91000 }, { "epoch": 4.470611227830165, "grad_norm": 66.63270568847656, "learning_rate": 5.882352941176471e-06, "loss": 0.1674, "step": 91500 }, { "epoch": 4.49504079738115, "grad_norm": 14.567200660705566, "learning_rate": 5.856363929330679e-06, "loss": 0.1612, "step": 92000 }, { "epoch": 4.5194703669321346, "grad_norm": 2.0841243267059326, "learning_rate": 5.830374917484888e-06, "loss": 0.1588, "step": 92500 }, { "epoch": 4.543899936483119, "grad_norm": 21.195947647094727, "learning_rate": 5.804385905639096e-06, "loss": 0.1686, "step": 93000 }, { "epoch": 4.568329506034104, "grad_norm": 19.75173568725586, "learning_rate": 5.778396893793304e-06, "loss": 0.1718, "step": 93500 }, { "epoch": 4.5927590755850884, "grad_norm": 34.48356628417969, "learning_rate": 5.7524078819475125e-06, "loss": 0.1693, "step": 94000 }, { "epoch": 4.617188645136073, "grad_norm": 66.5281982421875, "learning_rate": 5.726418870101721e-06, "loss": 0.1654, "step": 94500 }, { "epoch": 4.641618214687057, "grad_norm": 21.60127830505371, "learning_rate": 5.70042985825593e-06, "loss": 0.1525, "step": 95000 }, { "epoch": 4.6660477842380415, "grad_norm": 53.85319900512695, "learning_rate": 5.674440846410139e-06, "loss": 0.1683, "step": 95500 }, { "epoch": 4.690477353789026, "grad_norm": 21.81485366821289, "learning_rate": 5.6484518345643465e-06, "loss": 0.1597, "step": 96000 }, { "epoch": 4.714906923340011, "grad_norm": 3.4359285831451416, "learning_rate": 5.622462822718555e-06, "loss": 0.1744, "step": 96500 }, { "epoch": 4.739336492890995, "grad_norm": 37.74496078491211, "learning_rate": 5.5964738108727635e-06, "loss": 0.1549, "step": 97000 }, { "epoch": 4.76376606244198, "grad_norm": 46.64603042602539, "learning_rate": 5.570484799026972e-06, "loss": 0.1628, "step": 97500 }, { "epoch": 4.788195631992965, "grad_norm": 1.0236244201660156, "learning_rate": 5.54449578718118e-06, "loss": 0.1543, "step": 98000 }, { "epoch": 4.812625201543948, "grad_norm": 75.760009765625, "learning_rate": 5.518506775335388e-06, "loss": 0.1788, "step": 98500 }, { "epoch": 4.837054771094933, "grad_norm": 1.3948771953582764, "learning_rate": 5.4925177634895966e-06, "loss": 0.1675, "step": 99000 }, { "epoch": 4.861484340645918, "grad_norm": 25.799287796020508, "learning_rate": 5.466528751643806e-06, "loss": 0.1756, "step": 99500 }, { "epoch": 4.885913910196902, "grad_norm": 10.39145278930664, "learning_rate": 5.440539739798014e-06, "loss": 0.168, "step": 100000 }, { "epoch": 4.910343479747887, "grad_norm": 0.5542752742767334, "learning_rate": 5.414550727952223e-06, "loss": 0.158, "step": 100500 }, { "epoch": 4.9347730492988715, "grad_norm": 2.8913373947143555, "learning_rate": 5.3885617161064305e-06, "loss": 0.1703, "step": 101000 }, { "epoch": 4.959202618849856, "grad_norm": 76.17244720458984, "learning_rate": 5.362572704260639e-06, "loss": 0.1571, "step": 101500 }, { "epoch": 4.98363218840084, "grad_norm": 29.392118453979492, "learning_rate": 5.3365836924148475e-06, "loss": 0.1671, "step": 102000 }, { "epoch": 5.0, "eval_accuracy": 0.9093857358801704, "eval_averaged_scores": 0.893972399414541, "eval_f1": 0.8785590629489115, "eval_loss": 0.37774011492729187, "eval_runtime": 20.1329, "eval_samples_per_second": 1807.24, "eval_steps_per_second": 112.999, "step": 102335 }, { "epoch": 5.0080617579518245, "grad_norm": 0.33907392621040344, "learning_rate": 5.310594680569056e-06, "loss": 0.1594, "step": 102500 }, { "epoch": 5.032491327502809, "grad_norm": 0.20179282128810883, "learning_rate": 5.284605668723264e-06, "loss": 0.1234, "step": 103000 }, { "epoch": 5.056920897053794, "grad_norm": 0.2239135205745697, "learning_rate": 5.258616656877472e-06, "loss": 0.1398, "step": 103500 }, { "epoch": 5.081350466604778, "grad_norm": 0.2501750886440277, "learning_rate": 5.2326276450316815e-06, "loss": 0.141, "step": 104000 }, { "epoch": 5.105780036155763, "grad_norm": 27.644052505493164, "learning_rate": 5.20663863318589e-06, "loss": 0.1447, "step": 104500 }, { "epoch": 5.130209605706748, "grad_norm": 0.7906270623207092, "learning_rate": 5.1806496213400985e-06, "loss": 0.1368, "step": 105000 }, { "epoch": 5.154639175257732, "grad_norm": 0.17316019535064697, "learning_rate": 5.154660609494306e-06, "loss": 0.131, "step": 105500 }, { "epoch": 5.179068744808716, "grad_norm": 15.188929557800293, "learning_rate": 5.128671597648515e-06, "loss": 0.1351, "step": 106000 }, { "epoch": 5.203498314359701, "grad_norm": 0.39822831749916077, "learning_rate": 5.102682585802723e-06, "loss": 0.1424, "step": 106500 }, { "epoch": 5.227927883910685, "grad_norm": 0.22537678480148315, "learning_rate": 5.0766935739569316e-06, "loss": 0.1369, "step": 107000 }, { "epoch": 5.25235745346167, "grad_norm": 1.028478741645813, "learning_rate": 5.050704562111139e-06, "loss": 0.1358, "step": 107500 }, { "epoch": 5.276787023012655, "grad_norm": 23.18048858642578, "learning_rate": 5.024715550265348e-06, "loss": 0.1335, "step": 108000 }, { "epoch": 5.301216592563639, "grad_norm": 22.66480255126953, "learning_rate": 4.998726538419556e-06, "loss": 0.1358, "step": 108500 }, { "epoch": 5.325646162114624, "grad_norm": 1.0610641241073608, "learning_rate": 4.972737526573765e-06, "loss": 0.1385, "step": 109000 }, { "epoch": 5.3500757316656085, "grad_norm": 0.12008700519800186, "learning_rate": 4.946748514727974e-06, "loss": 0.1462, "step": 109500 }, { "epoch": 5.374505301216592, "grad_norm": 0.47721561789512634, "learning_rate": 4.920759502882182e-06, "loss": 0.1407, "step": 110000 }, { "epoch": 5.398934870767577, "grad_norm": 0.9414154291152954, "learning_rate": 4.89477049103639e-06, "loss": 0.1435, "step": 110500 }, { "epoch": 5.4233644403185615, "grad_norm": 260.96466064453125, "learning_rate": 4.868781479190599e-06, "loss": 0.1319, "step": 111000 }, { "epoch": 5.447794009869546, "grad_norm": 0.2496846616268158, "learning_rate": 4.842792467344807e-06, "loss": 0.1544, "step": 111500 }, { "epoch": 5.472223579420531, "grad_norm": 57.217315673828125, "learning_rate": 4.816803455499016e-06, "loss": 0.1328, "step": 112000 }, { "epoch": 5.496653148971515, "grad_norm": 0.7028847932815552, "learning_rate": 4.790814443653224e-06, "loss": 0.1396, "step": 112500 }, { "epoch": 5.5210827185225, "grad_norm": 0.15381655097007751, "learning_rate": 4.764825431807433e-06, "loss": 0.1384, "step": 113000 }, { "epoch": 5.545512288073484, "grad_norm": 1.1577835083007812, "learning_rate": 4.73883641996164e-06, "loss": 0.1447, "step": 113500 }, { "epoch": 5.569941857624468, "grad_norm": 0.1358271986246109, "learning_rate": 4.712847408115849e-06, "loss": 0.1311, "step": 114000 }, { "epoch": 5.594371427175453, "grad_norm": 44.117523193359375, "learning_rate": 4.686858396270057e-06, "loss": 0.1533, "step": 114500 }, { "epoch": 5.618800996726438, "grad_norm": 1.6837427616119385, "learning_rate": 4.660869384424266e-06, "loss": 0.1392, "step": 115000 }, { "epoch": 5.643230566277422, "grad_norm": 0.5804646611213684, "learning_rate": 4.634880372578474e-06, "loss": 0.1416, "step": 115500 }, { "epoch": 5.667660135828407, "grad_norm": 6.186709880828857, "learning_rate": 4.608891360732683e-06, "loss": 0.1377, "step": 116000 }, { "epoch": 5.6920897053793915, "grad_norm": 3.6101438999176025, "learning_rate": 4.582902348886891e-06, "loss": 0.1355, "step": 116500 }, { "epoch": 5.716519274930375, "grad_norm": 59.04292297363281, "learning_rate": 4.5569133370411e-06, "loss": 0.1418, "step": 117000 }, { "epoch": 5.74094884448136, "grad_norm": 103.19622802734375, "learning_rate": 4.530924325195308e-06, "loss": 0.1385, "step": 117500 }, { "epoch": 5.7653784140323445, "grad_norm": 0.046471111476421356, "learning_rate": 4.504935313349516e-06, "loss": 0.1437, "step": 118000 }, { "epoch": 5.789807983583329, "grad_norm": 14.723210334777832, "learning_rate": 4.478946301503724e-06, "loss": 0.1489, "step": 118500 }, { "epoch": 5.814237553134314, "grad_norm": 0.23389513790607452, "learning_rate": 4.452957289657933e-06, "loss": 0.1368, "step": 119000 }, { "epoch": 5.838667122685298, "grad_norm": 52.34353256225586, "learning_rate": 4.426968277812141e-06, "loss": 0.1366, "step": 119500 }, { "epoch": 5.863096692236283, "grad_norm": 0.24756839871406555, "learning_rate": 4.40097926596635e-06, "loss": 0.1466, "step": 120000 }, { "epoch": 5.887526261787268, "grad_norm": 12.284710884094238, "learning_rate": 4.374990254120558e-06, "loss": 0.1477, "step": 120500 }, { "epoch": 5.911955831338252, "grad_norm": 4.85374116897583, "learning_rate": 4.349001242274767e-06, "loss": 0.1406, "step": 121000 }, { "epoch": 5.936385400889236, "grad_norm": 23.76519012451172, "learning_rate": 4.323012230428975e-06, "loss": 0.1434, "step": 121500 }, { "epoch": 5.960814970440221, "grad_norm": 0.7099828720092773, "learning_rate": 4.297023218583184e-06, "loss": 0.1524, "step": 122000 }, { "epoch": 5.985244539991205, "grad_norm": 30.665578842163086, "learning_rate": 4.271034206737391e-06, "loss": 0.1381, "step": 122500 }, { "epoch": 6.0, "eval_accuracy": 0.9105950254225642, "eval_averaged_scores": 0.8952502575704396, "eval_f1": 0.8799054897183151, "eval_loss": 0.44903886318206787, "eval_runtime": 20.1055, "eval_samples_per_second": 1809.706, "eval_steps_per_second": 113.153, "step": 122802 }, { "epoch": 6.00967410954219, "grad_norm": 21.405393600463867, "learning_rate": 4.2450451948916e-06, "loss": 0.1166, "step": 123000 }, { "epoch": 6.034103679093175, "grad_norm": 8.164361953735352, "learning_rate": 4.219056183045809e-06, "loss": 0.1086, "step": 123500 }, { "epoch": 6.058533248644159, "grad_norm": 0.2232547402381897, "learning_rate": 4.193067171200017e-06, "loss": 0.1152, "step": 124000 }, { "epoch": 6.082962818195144, "grad_norm": 72.73697662353516, "learning_rate": 4.167078159354225e-06, "loss": 0.109, "step": 124500 }, { "epoch": 6.107392387746128, "grad_norm": 0.5123595595359802, "learning_rate": 4.141089147508434e-06, "loss": 0.1183, "step": 125000 }, { "epoch": 6.131821957297112, "grad_norm": 0.04295356571674347, "learning_rate": 4.115100135662642e-06, "loss": 0.1102, "step": 125500 }, { "epoch": 6.156251526848097, "grad_norm": 17.82163429260254, "learning_rate": 4.089111123816851e-06, "loss": 0.1059, "step": 126000 }, { "epoch": 6.1806810963990815, "grad_norm": 0.28495362401008606, "learning_rate": 4.063122111971059e-06, "loss": 0.1053, "step": 126500 }, { "epoch": 6.205110665950066, "grad_norm": 9.597000122070312, "learning_rate": 4.037133100125267e-06, "loss": 0.1244, "step": 127000 }, { "epoch": 6.229540235501051, "grad_norm": 0.0872587263584137, "learning_rate": 4.011144088279475e-06, "loss": 0.1191, "step": 127500 }, { "epoch": 6.253969805052035, "grad_norm": 22.000926971435547, "learning_rate": 3.985155076433685e-06, "loss": 0.1243, "step": 128000 }, { "epoch": 6.278399374603019, "grad_norm": 9.766866683959961, "learning_rate": 3.959166064587892e-06, "loss": 0.1202, "step": 128500 }, { "epoch": 6.302828944154004, "grad_norm": 12.51276969909668, "learning_rate": 3.933177052742101e-06, "loss": 0.1167, "step": 129000 }, { "epoch": 6.327258513704988, "grad_norm": 0.99669349193573, "learning_rate": 3.907188040896309e-06, "loss": 0.1157, "step": 129500 }, { "epoch": 6.351688083255973, "grad_norm": 0.4419931471347809, "learning_rate": 3.881199029050518e-06, "loss": 0.113, "step": 130000 }, { "epoch": 6.376117652806958, "grad_norm": 0.0723571702837944, "learning_rate": 3.855210017204726e-06, "loss": 0.1239, "step": 130500 }, { "epoch": 6.400547222357942, "grad_norm": 67.760009765625, "learning_rate": 3.829221005358935e-06, "loss": 0.1206, "step": 131000 }, { "epoch": 6.424976791908927, "grad_norm": 145.9121551513672, "learning_rate": 3.803231993513143e-06, "loss": 0.1153, "step": 131500 }, { "epoch": 6.449406361459911, "grad_norm": 14.276485443115234, "learning_rate": 3.777242981667351e-06, "loss": 0.1255, "step": 132000 }, { "epoch": 6.473835931010895, "grad_norm": 0.9386702179908752, "learning_rate": 3.75125396982156e-06, "loss": 0.1196, "step": 132500 }, { "epoch": 6.49826550056188, "grad_norm": 0.18407511711120605, "learning_rate": 3.7252649579757684e-06, "loss": 0.1138, "step": 133000 }, { "epoch": 6.5226950701128645, "grad_norm": 0.8368340730667114, "learning_rate": 3.6992759461299764e-06, "loss": 0.1273, "step": 133500 }, { "epoch": 6.547124639663849, "grad_norm": 8.838400840759277, "learning_rate": 3.673286934284185e-06, "loss": 0.1103, "step": 134000 }, { "epoch": 6.571554209214834, "grad_norm": 0.08394081890583038, "learning_rate": 3.647297922438393e-06, "loss": 0.1205, "step": 134500 }, { "epoch": 6.595983778765818, "grad_norm": 0.11461492627859116, "learning_rate": 3.621308910592602e-06, "loss": 0.1103, "step": 135000 }, { "epoch": 6.620413348316802, "grad_norm": 0.16645289957523346, "learning_rate": 3.5953198987468104e-06, "loss": 0.1154, "step": 135500 }, { "epoch": 6.644842917867787, "grad_norm": 288.46868896484375, "learning_rate": 3.5693308869010185e-06, "loss": 0.1281, "step": 136000 }, { "epoch": 6.669272487418771, "grad_norm": 0.03742976859211922, "learning_rate": 3.543341875055227e-06, "loss": 0.1225, "step": 136500 }, { "epoch": 6.693702056969756, "grad_norm": 0.650506854057312, "learning_rate": 3.5173528632094355e-06, "loss": 0.1271, "step": 137000 }, { "epoch": 6.718131626520741, "grad_norm": 54.199398040771484, "learning_rate": 3.491363851363644e-06, "loss": 0.1315, "step": 137500 }, { "epoch": 6.742561196071725, "grad_norm": 23.973060607910156, "learning_rate": 3.465374839517852e-06, "loss": 0.1241, "step": 138000 }, { "epoch": 6.76699076562271, "grad_norm": 0.1303454339504242, "learning_rate": 3.4393858276720605e-06, "loss": 0.1214, "step": 138500 }, { "epoch": 6.791420335173695, "grad_norm": 0.046723511070013046, "learning_rate": 3.4133968158262686e-06, "loss": 0.1143, "step": 139000 }, { "epoch": 6.815849904724679, "grad_norm": 6.235154628753662, "learning_rate": 3.3874078039804775e-06, "loss": 0.124, "step": 139500 }, { "epoch": 6.840279474275663, "grad_norm": 0.06593639403581619, "learning_rate": 3.361418792134686e-06, "loss": 0.122, "step": 140000 }, { "epoch": 6.864709043826648, "grad_norm": 0.6418370604515076, "learning_rate": 3.335429780288894e-06, "loss": 0.1232, "step": 140500 }, { "epoch": 6.889138613377632, "grad_norm": 0.2320944368839264, "learning_rate": 3.3094407684431025e-06, "loss": 0.1232, "step": 141000 }, { "epoch": 6.913568182928617, "grad_norm": 30.253524780273438, "learning_rate": 3.283451756597311e-06, "loss": 0.1298, "step": 141500 }, { "epoch": 6.9379977524796015, "grad_norm": 261.870849609375, "learning_rate": 3.2574627447515195e-06, "loss": 0.1192, "step": 142000 }, { "epoch": 6.962427322030586, "grad_norm": 0.349097341299057, "learning_rate": 3.2314737329057276e-06, "loss": 0.1236, "step": 142500 }, { "epoch": 6.986856891581571, "grad_norm": 0.3783528208732605, "learning_rate": 3.205484721059936e-06, "loss": 0.1148, "step": 143000 }, { "epoch": 7.0, "eval_accuracy": 0.9117218634052494, "eval_averaged_scores": 0.8971449328725574, "eval_f1": 0.8825680023398654, "eval_loss": 0.467751681804657, "eval_runtime": 20.1427, "eval_samples_per_second": 1806.36, "eval_steps_per_second": 112.944, "step": 143269 }, { "epoch": 7.0112864611325545, "grad_norm": 1.4202431440353394, "learning_rate": 3.179495709214144e-06, "loss": 0.1088, "step": 143500 }, { "epoch": 7.035716030683539, "grad_norm": 0.1592591255903244, "learning_rate": 3.153506697368353e-06, "loss": 0.0945, "step": 144000 }, { "epoch": 7.060145600234524, "grad_norm": 8.698958396911621, "learning_rate": 3.1275176855225615e-06, "loss": 0.0984, "step": 144500 }, { "epoch": 7.084575169785508, "grad_norm": 0.0554330088198185, "learning_rate": 3.1015286736767696e-06, "loss": 0.0925, "step": 145000 }, { "epoch": 7.109004739336493, "grad_norm": 0.3783748745918274, "learning_rate": 3.075539661830978e-06, "loss": 0.0914, "step": 145500 }, { "epoch": 7.133434308887478, "grad_norm": 0.33608478307724, "learning_rate": 3.049550649985186e-06, "loss": 0.0978, "step": 146000 }, { "epoch": 7.157863878438462, "grad_norm": 0.36914902925491333, "learning_rate": 3.023561638139395e-06, "loss": 0.0923, "step": 146500 }, { "epoch": 7.182293447989446, "grad_norm": 0.05221285670995712, "learning_rate": 2.9975726262936036e-06, "loss": 0.0939, "step": 147000 }, { "epoch": 7.206723017540431, "grad_norm": 0.09707313030958176, "learning_rate": 2.9715836144478116e-06, "loss": 0.0925, "step": 147500 }, { "epoch": 7.231152587091415, "grad_norm": 0.21917091310024261, "learning_rate": 2.94559460260202e-06, "loss": 0.0947, "step": 148000 }, { "epoch": 7.2555821566424, "grad_norm": 0.08681820333003998, "learning_rate": 2.9196055907562286e-06, "loss": 0.0965, "step": 148500 }, { "epoch": 7.2800117261933845, "grad_norm": 0.27925539016723633, "learning_rate": 2.893616578910437e-06, "loss": 0.1044, "step": 149000 }, { "epoch": 7.304441295744369, "grad_norm": 33.41240692138672, "learning_rate": 2.867627567064645e-06, "loss": 0.0968, "step": 149500 }, { "epoch": 7.328870865295354, "grad_norm": 142.32989501953125, "learning_rate": 2.8416385552188536e-06, "loss": 0.0998, "step": 150000 }, { "epoch": 7.3533004348463376, "grad_norm": 0.03225061669945717, "learning_rate": 2.8156495433730617e-06, "loss": 0.1013, "step": 150500 }, { "epoch": 7.377730004397322, "grad_norm": 0.08403612673282623, "learning_rate": 2.7896605315272706e-06, "loss": 0.1009, "step": 151000 }, { "epoch": 7.402159573948307, "grad_norm": 0.3820052742958069, "learning_rate": 2.763671519681479e-06, "loss": 0.0963, "step": 151500 }, { "epoch": 7.4265891434992914, "grad_norm": 16.261816024780273, "learning_rate": 2.737682507835687e-06, "loss": 0.1029, "step": 152000 }, { "epoch": 7.451018713050276, "grad_norm": 0.331718385219574, "learning_rate": 2.7116934959898957e-06, "loss": 0.1041, "step": 152500 }, { "epoch": 7.475448282601261, "grad_norm": 1.3978043794631958, "learning_rate": 2.685704484144104e-06, "loss": 0.0921, "step": 153000 }, { "epoch": 7.499877852152245, "grad_norm": 0.04950063303112984, "learning_rate": 2.6597154722983127e-06, "loss": 0.0959, "step": 153500 }, { "epoch": 7.52430742170323, "grad_norm": 0.14950354397296906, "learning_rate": 2.6337264604525207e-06, "loss": 0.0886, "step": 154000 }, { "epoch": 7.548736991254214, "grad_norm": 37.28718566894531, "learning_rate": 2.607737448606729e-06, "loss": 0.1035, "step": 154500 }, { "epoch": 7.573166560805198, "grad_norm": 0.17789612710475922, "learning_rate": 2.5817484367609373e-06, "loss": 0.0945, "step": 155000 }, { "epoch": 7.597596130356183, "grad_norm": 0.2735101580619812, "learning_rate": 2.555759424915146e-06, "loss": 0.0996, "step": 155500 }, { "epoch": 7.622025699907168, "grad_norm": 162.47628784179688, "learning_rate": 2.5297704130693547e-06, "loss": 0.1001, "step": 156000 }, { "epoch": 7.646455269458152, "grad_norm": 0.21928128600120544, "learning_rate": 2.5037814012235627e-06, "loss": 0.1058, "step": 156500 }, { "epoch": 7.670884839009137, "grad_norm": 8.425532341003418, "learning_rate": 2.4777923893777712e-06, "loss": 0.0884, "step": 157000 }, { "epoch": 7.6953144085601215, "grad_norm": 0.1196184754371643, "learning_rate": 2.4518033775319797e-06, "loss": 0.1073, "step": 157500 }, { "epoch": 7.719743978111106, "grad_norm": 18.718048095703125, "learning_rate": 2.425814365686188e-06, "loss": 0.0884, "step": 158000 }, { "epoch": 7.74417354766209, "grad_norm": 37.17155075073242, "learning_rate": 2.3998253538403967e-06, "loss": 0.0945, "step": 158500 }, { "epoch": 7.7686031172130745, "grad_norm": 0.38690435886383057, "learning_rate": 2.3738363419946048e-06, "loss": 0.0976, "step": 159000 }, { "epoch": 7.793032686764059, "grad_norm": 13.13317584991455, "learning_rate": 2.3478473301488133e-06, "loss": 0.0986, "step": 159500 }, { "epoch": 7.817462256315044, "grad_norm": 0.05163729190826416, "learning_rate": 2.3218583183030217e-06, "loss": 0.1, "step": 160000 }, { "epoch": 7.841891825866028, "grad_norm": 25.491252899169922, "learning_rate": 2.2958693064572302e-06, "loss": 0.0874, "step": 160500 }, { "epoch": 7.866321395417013, "grad_norm": 0.3773481547832489, "learning_rate": 2.2698802946114383e-06, "loss": 0.1027, "step": 161000 }, { "epoch": 7.890750964967998, "grad_norm": 13.568100929260254, "learning_rate": 2.243891282765647e-06, "loss": 0.1062, "step": 161500 }, { "epoch": 7.915180534518981, "grad_norm": 24.115978240966797, "learning_rate": 2.2179022709198553e-06, "loss": 0.0934, "step": 162000 }, { "epoch": 7.939610104069966, "grad_norm": 0.012055924162268639, "learning_rate": 2.1919132590740634e-06, "loss": 0.0854, "step": 162500 }, { "epoch": 7.964039673620951, "grad_norm": 0.3433281481266022, "learning_rate": 2.1659242472282723e-06, "loss": 0.1038, "step": 163000 }, { "epoch": 7.988469243171935, "grad_norm": 0.2032276839017868, "learning_rate": 2.1399352353824803e-06, "loss": 0.0992, "step": 163500 }, { "epoch": 8.0, "eval_accuracy": 0.9103476707434383, "eval_averaged_scores": 0.8959594874193519, "eval_f1": 0.8815713040952657, "eval_loss": 0.5451585650444031, "eval_runtime": 20.179, "eval_samples_per_second": 1803.108, "eval_steps_per_second": 112.741, "step": 163736 }, { "epoch": 8.01289881272292, "grad_norm": 0.012172297574579716, "learning_rate": 2.113946223536689e-06, "loss": 0.0873, "step": 164000 }, { "epoch": 8.037328382273904, "grad_norm": 0.4155900180339813, "learning_rate": 2.0879572116908973e-06, "loss": 0.0777, "step": 164500 }, { "epoch": 8.06175795182489, "grad_norm": 0.19762279093265533, "learning_rate": 2.061968199845106e-06, "loss": 0.0829, "step": 165000 }, { "epoch": 8.086187521375873, "grad_norm": 0.1475783735513687, "learning_rate": 2.035979187999314e-06, "loss": 0.0707, "step": 165500 }, { "epoch": 8.110617090926858, "grad_norm": 145.56173706054688, "learning_rate": 2.0099901761535224e-06, "loss": 0.0751, "step": 166000 }, { "epoch": 8.135046660477842, "grad_norm": 0.041457872837781906, "learning_rate": 1.984001164307731e-06, "loss": 0.0671, "step": 166500 }, { "epoch": 8.159476230028828, "grad_norm": 5.121572971343994, "learning_rate": 1.9580121524619393e-06, "loss": 0.078, "step": 167000 }, { "epoch": 8.183905799579811, "grad_norm": 0.2376221865415573, "learning_rate": 1.932023140616148e-06, "loss": 0.0842, "step": 167500 }, { "epoch": 8.208335369130795, "grad_norm": 27.775150299072266, "learning_rate": 1.9060341287703559e-06, "loss": 0.0767, "step": 168000 }, { "epoch": 8.23276493868178, "grad_norm": 0.18645720183849335, "learning_rate": 1.8800451169245646e-06, "loss": 0.0802, "step": 168500 }, { "epoch": 8.257194508232764, "grad_norm": 7.003711700439453, "learning_rate": 1.8540561050787729e-06, "loss": 0.0803, "step": 169000 }, { "epoch": 8.28162407778375, "grad_norm": 44.9241828918457, "learning_rate": 1.8280670932329814e-06, "loss": 0.08, "step": 169500 }, { "epoch": 8.306053647334734, "grad_norm": 0.3100529909133911, "learning_rate": 1.8020780813871896e-06, "loss": 0.0807, "step": 170000 }, { "epoch": 8.33048321688572, "grad_norm": 1.1184957027435303, "learning_rate": 1.776089069541398e-06, "loss": 0.0744, "step": 170500 }, { "epoch": 8.354912786436703, "grad_norm": 0.22304557263851166, "learning_rate": 1.7501000576956064e-06, "loss": 0.0784, "step": 171000 }, { "epoch": 8.379342355987687, "grad_norm": 0.036572836339473724, "learning_rate": 1.7241110458498147e-06, "loss": 0.077, "step": 171500 }, { "epoch": 8.403771925538672, "grad_norm": 0.1464405059814453, "learning_rate": 1.6981220340040234e-06, "loss": 0.0834, "step": 172000 }, { "epoch": 8.428201495089656, "grad_norm": 0.08662763237953186, "learning_rate": 1.6721330221582317e-06, "loss": 0.0787, "step": 172500 }, { "epoch": 8.452631064640642, "grad_norm": 1.44623601436615, "learning_rate": 1.6461440103124402e-06, "loss": 0.0726, "step": 173000 }, { "epoch": 8.477060634191625, "grad_norm": 14.024663925170898, "learning_rate": 1.6201549984666484e-06, "loss": 0.0907, "step": 173500 }, { "epoch": 8.50149020374261, "grad_norm": 0.02955411933362484, "learning_rate": 1.5941659866208567e-06, "loss": 0.0836, "step": 174000 }, { "epoch": 8.525919773293595, "grad_norm": 0.1095781922340393, "learning_rate": 1.5681769747750652e-06, "loss": 0.0881, "step": 174500 }, { "epoch": 8.550349342844578, "grad_norm": 0.8175545930862427, "learning_rate": 1.5421879629292735e-06, "loss": 0.0858, "step": 175000 }, { "epoch": 8.574778912395564, "grad_norm": 0.0535324402153492, "learning_rate": 1.5161989510834822e-06, "loss": 0.0781, "step": 175500 }, { "epoch": 8.599208481946548, "grad_norm": 0.0196294616907835, "learning_rate": 1.4902099392376905e-06, "loss": 0.0991, "step": 176000 }, { "epoch": 8.623638051497533, "grad_norm": 3.7140159606933594, "learning_rate": 1.464220927391899e-06, "loss": 0.0747, "step": 176500 }, { "epoch": 8.648067621048517, "grad_norm": 101.7074966430664, "learning_rate": 1.4382319155461072e-06, "loss": 0.0914, "step": 177000 }, { "epoch": 8.672497190599502, "grad_norm": 0.013971562497317791, "learning_rate": 1.4122429037003157e-06, "loss": 0.0681, "step": 177500 }, { "epoch": 8.696926760150486, "grad_norm": 164.79086303710938, "learning_rate": 1.386253891854524e-06, "loss": 0.0789, "step": 178000 }, { "epoch": 8.721356329701472, "grad_norm": 0.1486426293849945, "learning_rate": 1.3602648800087323e-06, "loss": 0.0847, "step": 178500 }, { "epoch": 8.745785899252455, "grad_norm": 32.686614990234375, "learning_rate": 1.3342758681629408e-06, "loss": 0.0914, "step": 179000 }, { "epoch": 8.770215468803439, "grad_norm": 0.8406556844711304, "learning_rate": 1.308286856317149e-06, "loss": 0.0742, "step": 179500 }, { "epoch": 8.794645038354425, "grad_norm": 0.07689370214939117, "learning_rate": 1.2822978444713577e-06, "loss": 0.0791, "step": 180000 }, { "epoch": 8.819074607905408, "grad_norm": 17.384950637817383, "learning_rate": 1.256308832625566e-06, "loss": 0.0764, "step": 180500 }, { "epoch": 8.843504177456394, "grad_norm": 29.185670852661133, "learning_rate": 1.2303198207797743e-06, "loss": 0.0862, "step": 181000 }, { "epoch": 8.867933747007378, "grad_norm": 58.11179733276367, "learning_rate": 1.2043308089339828e-06, "loss": 0.0862, "step": 181500 }, { "epoch": 8.892363316558363, "grad_norm": 0.12057217955589294, "learning_rate": 1.1783417970881913e-06, "loss": 0.08, "step": 182000 }, { "epoch": 8.916792886109347, "grad_norm": 67.25086212158203, "learning_rate": 1.1523527852423996e-06, "loss": 0.0881, "step": 182500 }, { "epoch": 8.94122245566033, "grad_norm": 0.04171622171998024, "learning_rate": 1.126363773396608e-06, "loss": 0.0831, "step": 183000 }, { "epoch": 8.965652025211316, "grad_norm": 0.3835027515888214, "learning_rate": 1.1003747615508165e-06, "loss": 0.0768, "step": 183500 }, { "epoch": 8.9900815947623, "grad_norm": 0.16164089739322662, "learning_rate": 1.0743857497050248e-06, "loss": 0.0847, "step": 184000 }, { "epoch": 9.0, "eval_accuracy": 0.9122440566167377, "eval_averaged_scores": 0.8971997179189981, "eval_f1": 0.8821553792212585, "eval_loss": 0.5808023810386658, "eval_runtime": 20.2131, "eval_samples_per_second": 1800.067, "eval_steps_per_second": 112.551, "step": 184203 }, { "epoch": 9.014511164313285, "grad_norm": 0.031089797616004944, "learning_rate": 1.048396737859233e-06, "loss": 0.072, "step": 184500 }, { "epoch": 9.038940733864269, "grad_norm": 0.0398230142891407, "learning_rate": 1.0224077260134416e-06, "loss": 0.0608, "step": 185000 }, { "epoch": 9.063370303415255, "grad_norm": 0.15601158142089844, "learning_rate": 9.9641871416765e-07, "loss": 0.0689, "step": 185500 }, { "epoch": 9.087799872966238, "grad_norm": 1.7415249347686768, "learning_rate": 9.704297023218584e-07, "loss": 0.0676, "step": 186000 }, { "epoch": 9.112229442517222, "grad_norm": 0.036363422870635986, "learning_rate": 9.444406904760668e-07, "loss": 0.0696, "step": 186500 }, { "epoch": 9.136659012068208, "grad_norm": 16.439847946166992, "learning_rate": 9.184516786302752e-07, "loss": 0.0704, "step": 187000 }, { "epoch": 9.161088581619191, "grad_norm": 0.16665898263454437, "learning_rate": 8.924626667844836e-07, "loss": 0.0706, "step": 187500 }, { "epoch": 9.185518151170177, "grad_norm": 0.006615887396037579, "learning_rate": 8.664736549386919e-07, "loss": 0.0718, "step": 188000 }, { "epoch": 9.20994772072116, "grad_norm": 0.015520376153290272, "learning_rate": 8.404846430929004e-07, "loss": 0.0639, "step": 188500 }, { "epoch": 9.234377290272146, "grad_norm": 0.18699295818805695, "learning_rate": 8.144956312471088e-07, "loss": 0.0671, "step": 189000 }, { "epoch": 9.25880685982313, "grad_norm": 208.81222534179688, "learning_rate": 7.885066194013171e-07, "loss": 0.0577, "step": 189500 }, { "epoch": 9.283236429374114, "grad_norm": 35.29654312133789, "learning_rate": 7.625176075555256e-07, "loss": 0.0754, "step": 190000 }, { "epoch": 9.3076659989251, "grad_norm": 0.024813145399093628, "learning_rate": 7.36528595709734e-07, "loss": 0.073, "step": 190500 }, { "epoch": 9.332095568476083, "grad_norm": 0.0019479466136544943, "learning_rate": 7.105395838639424e-07, "loss": 0.0621, "step": 191000 }, { "epoch": 9.356525138027068, "grad_norm": 0.034291163086891174, "learning_rate": 6.845505720181508e-07, "loss": 0.0646, "step": 191500 }, { "epoch": 9.380954707578052, "grad_norm": 21.998003005981445, "learning_rate": 6.585615601723591e-07, "loss": 0.058, "step": 192000 }, { "epoch": 9.405384277129038, "grad_norm": 0.04920468479394913, "learning_rate": 6.325725483265676e-07, "loss": 0.0834, "step": 192500 }, { "epoch": 9.429813846680021, "grad_norm": 0.0892450362443924, "learning_rate": 6.065835364807759e-07, "loss": 0.0666, "step": 193000 }, { "epoch": 9.454243416231005, "grad_norm": 0.2844100892543793, "learning_rate": 5.805945246349843e-07, "loss": 0.0583, "step": 193500 }, { "epoch": 9.47867298578199, "grad_norm": 0.26199427247047424, "learning_rate": 5.546055127891928e-07, "loss": 0.0719, "step": 194000 }, { "epoch": 9.503102555332974, "grad_norm": 0.016245270147919655, "learning_rate": 5.286165009434012e-07, "loss": 0.072, "step": 194500 }, { "epoch": 9.52753212488396, "grad_norm": 0.2348903864622116, "learning_rate": 5.026274890976096e-07, "loss": 0.0594, "step": 195000 }, { "epoch": 9.551961694434944, "grad_norm": 0.00560354720801115, "learning_rate": 4.7663847725181796e-07, "loss": 0.0601, "step": 195500 }, { "epoch": 9.57639126398593, "grad_norm": 0.20233391225337982, "learning_rate": 4.5064946540602635e-07, "loss": 0.0597, "step": 196000 }, { "epoch": 9.600820833536913, "grad_norm": 0.028764836490154266, "learning_rate": 4.246604535602348e-07, "loss": 0.0748, "step": 196500 }, { "epoch": 9.625250403087897, "grad_norm": 0.03489026054739952, "learning_rate": 3.986714417144431e-07, "loss": 0.0593, "step": 197000 }, { "epoch": 9.649679972638882, "grad_norm": 0.17677338421344757, "learning_rate": 3.7268242986865155e-07, "loss": 0.0575, "step": 197500 }, { "epoch": 9.674109542189866, "grad_norm": 0.006993240211158991, "learning_rate": 3.4669341802286e-07, "loss": 0.059, "step": 198000 }, { "epoch": 9.698539111740851, "grad_norm": 0.5008031129837036, "learning_rate": 3.207044061770684e-07, "loss": 0.0728, "step": 198500 }, { "epoch": 9.722968681291835, "grad_norm": 0.20838424563407898, "learning_rate": 2.9471539433127676e-07, "loss": 0.0706, "step": 199000 }, { "epoch": 9.74739825084282, "grad_norm": 0.023256618529558182, "learning_rate": 2.6872638248548514e-07, "loss": 0.0704, "step": 199500 }, { "epoch": 9.771827820393804, "grad_norm": 0.2845751643180847, "learning_rate": 2.427373706396936e-07, "loss": 0.0665, "step": 200000 }, { "epoch": 9.796257389944788, "grad_norm": 0.04077767953276634, "learning_rate": 2.1674835879390194e-07, "loss": 0.0609, "step": 200500 }, { "epoch": 9.820686959495774, "grad_norm": 0.01134885661303997, "learning_rate": 1.9075934694811035e-07, "loss": 0.0593, "step": 201000 }, { "epoch": 9.845116529046757, "grad_norm": 0.018247269093990326, "learning_rate": 1.6477033510231873e-07, "loss": 0.0657, "step": 201500 }, { "epoch": 9.869546098597743, "grad_norm": 0.008613000623881817, "learning_rate": 1.3878132325652717e-07, "loss": 0.0563, "step": 202000 }, { "epoch": 9.893975668148727, "grad_norm": 0.2736217975616455, "learning_rate": 1.1279231141073555e-07, "loss": 0.0709, "step": 202500 }, { "epoch": 9.918405237699712, "grad_norm": 0.02775813639163971, "learning_rate": 8.680329956494395e-08, "loss": 0.0677, "step": 203000 }, { "epoch": 9.942834807250696, "grad_norm": 0.06576963514089584, "learning_rate": 6.081428771915235e-08, "loss": 0.0685, "step": 203500 }, { "epoch": 9.967264376801682, "grad_norm": 0.1034393236041069, "learning_rate": 3.4825275873360744e-08, "loss": 0.0677, "step": 204000 }, { "epoch": 9.991693946352665, "grad_norm": 0.14732515811920166, "learning_rate": 8.836264027569144e-09, "loss": 0.0649, "step": 204500 }, { "epoch": 10.0, "eval_accuracy": 0.9122165727635014, "eval_averaged_scores": 0.8973993679198728, "eval_f1": 0.8825821630762444, "eval_loss": 0.6054433584213257, "eval_runtime": 20.1824, "eval_samples_per_second": 1802.808, "eval_steps_per_second": 112.722, "step": 204670 }, { "epoch": 10.0, "step": 204670, "total_flos": 1.0020319530941755e+17, "train_loss": 0.16519966338430658, "train_runtime": 10940.7303, "train_samples_per_second": 299.305, "train_steps_per_second": 18.707 } ], "logging_steps": 500, "max_steps": 204670, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0020319530941755e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }