{ "best_metric": 0.8667464541264547, "best_model_checkpoint": "./nlu_finetuned_models/mnli/roberta-base_lr1e-05/checkpoint-110450", "epoch": 10.0, "eval_steps": 500, "global_step": 220900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.022634676324128564, "grad_norm": 1.6427655220031738, "learning_rate": 3.772446054021428e-07, "loss": 1.0998, "step": 500 }, { "epoch": 0.04526935264825713, "grad_norm": 3.6388282775878906, "learning_rate": 7.544892108042856e-07, "loss": 1.0989, "step": 1000 }, { "epoch": 0.06790402897238569, "grad_norm": 5.199594974517822, "learning_rate": 1.1317338162064282e-06, "loss": 1.0951, "step": 1500 }, { "epoch": 0.09053870529651425, "grad_norm": 16.297880172729492, "learning_rate": 1.5089784216085712e-06, "loss": 0.9832, "step": 2000 }, { "epoch": 0.11317338162064282, "grad_norm": 30.214738845825195, "learning_rate": 1.886223027010714e-06, "loss": 0.7491, "step": 2500 }, { "epoch": 0.13580805794477138, "grad_norm": 27.916301727294922, "learning_rate": 2.2634676324128565e-06, "loss": 0.6467, "step": 3000 }, { "epoch": 0.15844273426889996, "grad_norm": 42.154232025146484, "learning_rate": 2.6407122378149996e-06, "loss": 0.6092, "step": 3500 }, { "epoch": 0.1810774105930285, "grad_norm": 29.020992279052734, "learning_rate": 3.0179568432171424e-06, "loss": 0.5782, "step": 4000 }, { "epoch": 0.2037120869171571, "grad_norm": 31.89137840270996, "learning_rate": 3.395201448619285e-06, "loss": 0.5567, "step": 4500 }, { "epoch": 0.22634676324128564, "grad_norm": 46.158416748046875, "learning_rate": 3.772446054021428e-06, "loss": 0.5382, "step": 5000 }, { "epoch": 0.24898143956541421, "grad_norm": 16.78737449645996, "learning_rate": 4.149690659423571e-06, "loss": 0.5278, "step": 5500 }, { "epoch": 0.27161611588954276, "grad_norm": 13.616703033447266, "learning_rate": 4.526935264825713e-06, "loss": 0.5233, "step": 6000 }, { "epoch": 0.29425079221367134, "grad_norm": 21.825986862182617, "learning_rate": 4.904179870227856e-06, "loss": 0.5042, "step": 6500 }, { "epoch": 0.3168854685377999, "grad_norm": 24.68229103088379, "learning_rate": 5.281424475629999e-06, "loss": 0.4967, "step": 7000 }, { "epoch": 0.3395201448619285, "grad_norm": 17.55514907836914, "learning_rate": 5.658669081032142e-06, "loss": 0.4979, "step": 7500 }, { "epoch": 0.362154821186057, "grad_norm": 20.74059295654297, "learning_rate": 6.035913686434285e-06, "loss": 0.4983, "step": 8000 }, { "epoch": 0.3847894975101856, "grad_norm": 12.430363655090332, "learning_rate": 6.4131582918364275e-06, "loss": 0.4773, "step": 8500 }, { "epoch": 0.4074241738343142, "grad_norm": 30.890859603881836, "learning_rate": 6.79040289723857e-06, "loss": 0.4802, "step": 9000 }, { "epoch": 0.43005885015844275, "grad_norm": 27.11601448059082, "learning_rate": 7.167647502640713e-06, "loss": 0.4631, "step": 9500 }, { "epoch": 0.4526935264825713, "grad_norm": 22.201452255249023, "learning_rate": 7.544892108042856e-06, "loss": 0.4575, "step": 10000 }, { "epoch": 0.47532820280669985, "grad_norm": 22.29547119140625, "learning_rate": 7.922136713445e-06, "loss": 0.4643, "step": 10500 }, { "epoch": 0.49796287913082843, "grad_norm": 15.072667121887207, "learning_rate": 8.299381318847142e-06, "loss": 0.466, "step": 11000 }, { "epoch": 0.520597555454957, "grad_norm": 28.1677188873291, "learning_rate": 8.676625924249283e-06, "loss": 0.4619, "step": 11500 }, { "epoch": 0.5432322317790855, "grad_norm": 17.43030548095703, "learning_rate": 9.053870529651426e-06, "loss": 0.4542, "step": 12000 }, { "epoch": 0.5658669081032142, "grad_norm": 23.5450382232666, "learning_rate": 9.431115135053569e-06, "loss": 0.449, "step": 12500 }, { "epoch": 0.5885015844273427, "grad_norm": 21.659141540527344, "learning_rate": 9.808359740455711e-06, "loss": 0.4434, "step": 13000 }, { "epoch": 0.6111362607514712, "grad_norm": 19.54996681213379, "learning_rate": 9.98815291409418e-06, "loss": 0.4555, "step": 13500 }, { "epoch": 0.6337709370755998, "grad_norm": 11.85847282409668, "learning_rate": 9.964073471196171e-06, "loss": 0.4511, "step": 14000 }, { "epoch": 0.6564056133997284, "grad_norm": 26.537872314453125, "learning_rate": 9.939994028298163e-06, "loss": 0.4424, "step": 14500 }, { "epoch": 0.679040289723857, "grad_norm": 16.38085174560547, "learning_rate": 9.915914585400153e-06, "loss": 0.4488, "step": 15000 }, { "epoch": 0.7016749660479855, "grad_norm": 20.928190231323242, "learning_rate": 9.891835142502145e-06, "loss": 0.4368, "step": 15500 }, { "epoch": 0.724309642372114, "grad_norm": 16.260562896728516, "learning_rate": 9.867755699604135e-06, "loss": 0.4481, "step": 16000 }, { "epoch": 0.7469443186962427, "grad_norm": 9.784706115722656, "learning_rate": 9.843676256706126e-06, "loss": 0.4344, "step": 16500 }, { "epoch": 0.7695789950203712, "grad_norm": 6.384799480438232, "learning_rate": 9.819596813808116e-06, "loss": 0.4416, "step": 17000 }, { "epoch": 0.7922136713444998, "grad_norm": 14.344380378723145, "learning_rate": 9.795517370910108e-06, "loss": 0.429, "step": 17500 }, { "epoch": 0.8148483476686283, "grad_norm": 33.43516159057617, "learning_rate": 9.771437928012098e-06, "loss": 0.4271, "step": 18000 }, { "epoch": 0.8374830239927569, "grad_norm": 23.591716766357422, "learning_rate": 9.74735848511409e-06, "loss": 0.44, "step": 18500 }, { "epoch": 0.8601177003168855, "grad_norm": 20.75270652770996, "learning_rate": 9.723279042216081e-06, "loss": 0.423, "step": 19000 }, { "epoch": 0.882752376641014, "grad_norm": 24.833736419677734, "learning_rate": 9.699199599318071e-06, "loss": 0.4212, "step": 19500 }, { "epoch": 0.9053870529651425, "grad_norm": 18.03992462158203, "learning_rate": 9.675120156420061e-06, "loss": 0.4164, "step": 20000 }, { "epoch": 0.9280217292892712, "grad_norm": 12.502860069274902, "learning_rate": 9.651040713522053e-06, "loss": 0.4213, "step": 20500 }, { "epoch": 0.9506564056133997, "grad_norm": 13.808119773864746, "learning_rate": 9.626961270624043e-06, "loss": 0.4197, "step": 21000 }, { "epoch": 0.9732910819375283, "grad_norm": 15.191283226013184, "learning_rate": 9.602881827726035e-06, "loss": 0.4219, "step": 21500 }, { "epoch": 0.9959257582616569, "grad_norm": 27.065793991088867, "learning_rate": 9.578802384828026e-06, "loss": 0.3965, "step": 22000 }, { "epoch": 1.0, "eval_accuracy": 0.8536579155101729, "eval_loss": 0.4072297513484955, "eval_runtime": 25.9705, "eval_samples_per_second": 1512.137, "eval_steps_per_second": 94.53, "step": 22090 }, { "epoch": 1.0185604345857855, "grad_norm": 40.00636672973633, "learning_rate": 9.554722941930016e-06, "loss": 0.371, "step": 22500 }, { "epoch": 1.041195110909914, "grad_norm": 14.272022247314453, "learning_rate": 9.530643499032008e-06, "loss": 0.357, "step": 23000 }, { "epoch": 1.0638297872340425, "grad_norm": 14.10964298248291, "learning_rate": 9.506564056133998e-06, "loss": 0.3584, "step": 23500 }, { "epoch": 1.086464463558171, "grad_norm": 49.96983337402344, "learning_rate": 9.48248461323599e-06, "loss": 0.3567, "step": 24000 }, { "epoch": 1.1090991398822996, "grad_norm": 15.82590389251709, "learning_rate": 9.45840517033798e-06, "loss": 0.3682, "step": 24500 }, { "epoch": 1.1317338162064283, "grad_norm": 17.939584732055664, "learning_rate": 9.43432572743997e-06, "loss": 0.3589, "step": 25000 }, { "epoch": 1.1543684925305568, "grad_norm": 9.916324615478516, "learning_rate": 9.410246284541961e-06, "loss": 0.3654, "step": 25500 }, { "epoch": 1.1770031688546854, "grad_norm": 13.99868392944336, "learning_rate": 9.386166841643953e-06, "loss": 0.3609, "step": 26000 }, { "epoch": 1.1996378451788139, "grad_norm": 19.407920837402344, "learning_rate": 9.362087398745945e-06, "loss": 0.3615, "step": 26500 }, { "epoch": 1.2222725215029424, "grad_norm": 22.240596771240234, "learning_rate": 9.338007955847935e-06, "loss": 0.3628, "step": 27000 }, { "epoch": 1.2449071978270712, "grad_norm": 27.17753028869629, "learning_rate": 9.313928512949925e-06, "loss": 0.3629, "step": 27500 }, { "epoch": 1.2675418741511997, "grad_norm": 16.045578002929688, "learning_rate": 9.289849070051916e-06, "loss": 0.3735, "step": 28000 }, { "epoch": 1.2901765504753282, "grad_norm": 16.507614135742188, "learning_rate": 9.265769627153906e-06, "loss": 0.3646, "step": 28500 }, { "epoch": 1.3128112267994567, "grad_norm": 16.56389808654785, "learning_rate": 9.241690184255898e-06, "loss": 0.3459, "step": 29000 }, { "epoch": 1.3354459031235852, "grad_norm": 21.800291061401367, "learning_rate": 9.217610741357888e-06, "loss": 0.3552, "step": 29500 }, { "epoch": 1.358080579447714, "grad_norm": 19.036035537719727, "learning_rate": 9.19353129845988e-06, "loss": 0.3569, "step": 30000 }, { "epoch": 1.3807152557718425, "grad_norm": 20.2823543548584, "learning_rate": 9.169451855561871e-06, "loss": 0.3635, "step": 30500 }, { "epoch": 1.403349932095971, "grad_norm": 9.948273658752441, "learning_rate": 9.145372412663861e-06, "loss": 0.3626, "step": 31000 }, { "epoch": 1.4259846084200996, "grad_norm": 24.87046241760254, "learning_rate": 9.121292969765853e-06, "loss": 0.3625, "step": 31500 }, { "epoch": 1.448619284744228, "grad_norm": 7.488998889923096, "learning_rate": 9.097213526867843e-06, "loss": 0.3598, "step": 32000 }, { "epoch": 1.4712539610683568, "grad_norm": 22.464574813842773, "learning_rate": 9.073134083969835e-06, "loss": 0.354, "step": 32500 }, { "epoch": 1.4938886373924853, "grad_norm": 18.735248565673828, "learning_rate": 9.049054641071825e-06, "loss": 0.3559, "step": 33000 }, { "epoch": 1.5165233137166139, "grad_norm": 17.83134651184082, "learning_rate": 9.024975198173815e-06, "loss": 0.3458, "step": 33500 }, { "epoch": 1.5391579900407424, "grad_norm": 22.070232391357422, "learning_rate": 9.000895755275806e-06, "loss": 0.3665, "step": 34000 }, { "epoch": 1.561792666364871, "grad_norm": 6.331955432891846, "learning_rate": 8.976816312377798e-06, "loss": 0.3607, "step": 34500 }, { "epoch": 1.5844273426889997, "grad_norm": 9.2369966506958, "learning_rate": 8.95273686947979e-06, "loss": 0.3549, "step": 35000 }, { "epoch": 1.607062019013128, "grad_norm": 14.87072467803955, "learning_rate": 8.92865742658178e-06, "loss": 0.3577, "step": 35500 }, { "epoch": 1.6296966953372567, "grad_norm": 21.532787322998047, "learning_rate": 8.90457798368377e-06, "loss": 0.353, "step": 36000 }, { "epoch": 1.6523313716613852, "grad_norm": 21.97151756286621, "learning_rate": 8.880498540785761e-06, "loss": 0.3706, "step": 36500 }, { "epoch": 1.6749660479855137, "grad_norm": 17.71976661682129, "learning_rate": 8.856419097887751e-06, "loss": 0.3648, "step": 37000 }, { "epoch": 1.6976007243096425, "grad_norm": 21.98705291748047, "learning_rate": 8.832339654989743e-06, "loss": 0.3591, "step": 37500 }, { "epoch": 1.7202354006337708, "grad_norm": 26.871360778808594, "learning_rate": 8.808260212091733e-06, "loss": 0.3454, "step": 38000 }, { "epoch": 1.7428700769578995, "grad_norm": 12.331396102905273, "learning_rate": 8.784180769193725e-06, "loss": 0.3418, "step": 38500 }, { "epoch": 1.765504753282028, "grad_norm": 7.1670756340026855, "learning_rate": 8.760101326295716e-06, "loss": 0.3454, "step": 39000 }, { "epoch": 1.7881394296061566, "grad_norm": 18.57856559753418, "learning_rate": 8.736021883397706e-06, "loss": 0.345, "step": 39500 }, { "epoch": 1.8107741059302853, "grad_norm": 36.80974578857422, "learning_rate": 8.711942440499698e-06, "loss": 0.3475, "step": 40000 }, { "epoch": 1.8334087822544136, "grad_norm": 19.742300033569336, "learning_rate": 8.687862997601688e-06, "loss": 0.3499, "step": 40500 }, { "epoch": 1.8560434585785424, "grad_norm": 13.53554630279541, "learning_rate": 8.663783554703678e-06, "loss": 0.3625, "step": 41000 }, { "epoch": 1.8786781349026709, "grad_norm": 12.660998344421387, "learning_rate": 8.63970411180567e-06, "loss": 0.355, "step": 41500 }, { "epoch": 1.9013128112267994, "grad_norm": 10.94740104675293, "learning_rate": 8.615624668907661e-06, "loss": 0.3495, "step": 42000 }, { "epoch": 1.9239474875509281, "grad_norm": 39.87784957885742, "learning_rate": 8.591545226009653e-06, "loss": 0.3484, "step": 42500 }, { "epoch": 1.9465821638750564, "grad_norm": 15.916511535644531, "learning_rate": 8.567465783111643e-06, "loss": 0.3511, "step": 43000 }, { "epoch": 1.9692168401991852, "grad_norm": 19.070566177368164, "learning_rate": 8.543386340213633e-06, "loss": 0.334, "step": 43500 }, { "epoch": 1.9918515165233137, "grad_norm": 21.015113830566406, "learning_rate": 8.519306897315625e-06, "loss": 0.3463, "step": 44000 }, { "epoch": 2.0, "eval_accuracy": 0.8632324106847292, "eval_loss": 0.37872129678726196, "eval_runtime": 26.0123, "eval_samples_per_second": 1509.711, "eval_steps_per_second": 94.379, "step": 44180 }, { "epoch": 2.0144861928474422, "grad_norm": 12.006568908691406, "learning_rate": 8.495227454417615e-06, "loss": 0.2925, "step": 44500 }, { "epoch": 2.037120869171571, "grad_norm": 16.82537078857422, "learning_rate": 8.471148011519606e-06, "loss": 0.2821, "step": 45000 }, { "epoch": 2.0597555454956993, "grad_norm": 14.019499778747559, "learning_rate": 8.447068568621596e-06, "loss": 0.2759, "step": 45500 }, { "epoch": 2.082390221819828, "grad_norm": 11.033167839050293, "learning_rate": 8.422989125723588e-06, "loss": 0.2778, "step": 46000 }, { "epoch": 2.1050248981439568, "grad_norm": 55.102169036865234, "learning_rate": 8.39890968282558e-06, "loss": 0.2759, "step": 46500 }, { "epoch": 2.127659574468085, "grad_norm": 22.346426010131836, "learning_rate": 8.37483023992757e-06, "loss": 0.2793, "step": 47000 }, { "epoch": 2.150294250792214, "grad_norm": 9.00412368774414, "learning_rate": 8.350750797029561e-06, "loss": 0.2696, "step": 47500 }, { "epoch": 2.172928927116342, "grad_norm": 13.098092079162598, "learning_rate": 8.326671354131551e-06, "loss": 0.2882, "step": 48000 }, { "epoch": 2.195563603440471, "grad_norm": 16.290449142456055, "learning_rate": 8.302591911233543e-06, "loss": 0.2807, "step": 48500 }, { "epoch": 2.218198279764599, "grad_norm": 36.5540771484375, "learning_rate": 8.278512468335533e-06, "loss": 0.2761, "step": 49000 }, { "epoch": 2.240832956088728, "grad_norm": 6.7274065017700195, "learning_rate": 8.254433025437523e-06, "loss": 0.2727, "step": 49500 }, { "epoch": 2.2634676324128566, "grad_norm": 10.446264266967773, "learning_rate": 8.230353582539515e-06, "loss": 0.2798, "step": 50000 }, { "epoch": 2.286102308736985, "grad_norm": 14.677602767944336, "learning_rate": 8.206274139641506e-06, "loss": 0.2677, "step": 50500 }, { "epoch": 2.3087369850611137, "grad_norm": 23.758255004882812, "learning_rate": 8.182194696743498e-06, "loss": 0.2917, "step": 51000 }, { "epoch": 2.3313716613852424, "grad_norm": 15.98766803741455, "learning_rate": 8.158115253845488e-06, "loss": 0.2918, "step": 51500 }, { "epoch": 2.3540063377093707, "grad_norm": 2.399088144302368, "learning_rate": 8.134035810947478e-06, "loss": 0.28, "step": 52000 }, { "epoch": 2.3766410140334995, "grad_norm": 15.759695053100586, "learning_rate": 8.10995636804947e-06, "loss": 0.2775, "step": 52500 }, { "epoch": 2.3992756903576278, "grad_norm": 46.171875, "learning_rate": 8.08587692515146e-06, "loss": 0.2799, "step": 53000 }, { "epoch": 2.4219103666817565, "grad_norm": 41.83917236328125, "learning_rate": 8.061797482253451e-06, "loss": 0.2905, "step": 53500 }, { "epoch": 2.444545043005885, "grad_norm": 28.17142677307129, "learning_rate": 8.037718039355441e-06, "loss": 0.2897, "step": 54000 }, { "epoch": 2.4671797193300136, "grad_norm": 30.47774314880371, "learning_rate": 8.013638596457433e-06, "loss": 0.2687, "step": 54500 }, { "epoch": 2.4898143956541423, "grad_norm": 24.639873504638672, "learning_rate": 7.989559153559425e-06, "loss": 0.2965, "step": 55000 }, { "epoch": 2.5124490719782706, "grad_norm": 24.947662353515625, "learning_rate": 7.965479710661415e-06, "loss": 0.2893, "step": 55500 }, { "epoch": 2.5350837483023994, "grad_norm": 28.483293533325195, "learning_rate": 7.941400267763406e-06, "loss": 0.2757, "step": 56000 }, { "epoch": 2.557718424626528, "grad_norm": 45.0990104675293, "learning_rate": 7.917320824865396e-06, "loss": 0.2682, "step": 56500 }, { "epoch": 2.5803531009506564, "grad_norm": 22.60608673095703, "learning_rate": 7.893241381967386e-06, "loss": 0.2779, "step": 57000 }, { "epoch": 2.6029877772747847, "grad_norm": 31.8905029296875, "learning_rate": 7.869161939069378e-06, "loss": 0.2857, "step": 57500 }, { "epoch": 2.6256224535989134, "grad_norm": 10.05256175994873, "learning_rate": 7.84508249617137e-06, "loss": 0.2765, "step": 58000 }, { "epoch": 2.648257129923042, "grad_norm": 18.466182708740234, "learning_rate": 7.82100305327336e-06, "loss": 0.2743, "step": 58500 }, { "epoch": 2.6708918062471705, "grad_norm": 22.704708099365234, "learning_rate": 7.796923610375351e-06, "loss": 0.2718, "step": 59000 }, { "epoch": 2.6935264825712992, "grad_norm": 38.51487731933594, "learning_rate": 7.772844167477341e-06, "loss": 0.2733, "step": 59500 }, { "epoch": 2.716161158895428, "grad_norm": 25.67682647705078, "learning_rate": 7.748764724579333e-06, "loss": 0.2803, "step": 60000 }, { "epoch": 2.7387958352195563, "grad_norm": 39.251068115234375, "learning_rate": 7.724685281681323e-06, "loss": 0.2757, "step": 60500 }, { "epoch": 2.761430511543685, "grad_norm": 20.692581176757812, "learning_rate": 7.700605838783315e-06, "loss": 0.2853, "step": 61000 }, { "epoch": 2.7840651878678138, "grad_norm": 22.915571212768555, "learning_rate": 7.676526395885305e-06, "loss": 0.2899, "step": 61500 }, { "epoch": 2.806699864191942, "grad_norm": 20.167299270629883, "learning_rate": 7.652446952987296e-06, "loss": 0.2802, "step": 62000 }, { "epoch": 2.8293345405160704, "grad_norm": 31.358797073364258, "learning_rate": 7.628367510089287e-06, "loss": 0.2775, "step": 62500 }, { "epoch": 2.851969216840199, "grad_norm": 16.160572052001953, "learning_rate": 7.604288067191278e-06, "loss": 0.2755, "step": 63000 }, { "epoch": 2.874603893164328, "grad_norm": 24.560882568359375, "learning_rate": 7.58020862429327e-06, "loss": 0.3024, "step": 63500 }, { "epoch": 2.897238569488456, "grad_norm": 30.818029403686523, "learning_rate": 7.55612918139526e-06, "loss": 0.2878, "step": 64000 }, { "epoch": 2.919873245812585, "grad_norm": 31.68956756591797, "learning_rate": 7.532049738497251e-06, "loss": 0.2793, "step": 64500 }, { "epoch": 2.9425079221367136, "grad_norm": 17.775924682617188, "learning_rate": 7.507970295599241e-06, "loss": 0.2824, "step": 65000 }, { "epoch": 2.965142598460842, "grad_norm": 26.35023307800293, "learning_rate": 7.483890852701232e-06, "loss": 0.275, "step": 65500 }, { "epoch": 2.9877772747849707, "grad_norm": 13.411957740783691, "learning_rate": 7.459811409803224e-06, "loss": 0.2867, "step": 66000 }, { "epoch": 3.0, "eval_accuracy": 0.8651422169030583, "eval_loss": 0.39841848611831665, "eval_runtime": 26.0093, "eval_samples_per_second": 1509.882, "eval_steps_per_second": 94.389, "step": 66270 }, { "epoch": 3.010411951109099, "grad_norm": 31.959758758544922, "learning_rate": 7.435731966905214e-06, "loss": 0.2628, "step": 66500 }, { "epoch": 3.0330466274332277, "grad_norm": 9.174257278442383, "learning_rate": 7.4116525240072056e-06, "loss": 0.2354, "step": 67000 }, { "epoch": 3.0556813037573565, "grad_norm": 41.27067565917969, "learning_rate": 7.387573081109196e-06, "loss": 0.231, "step": 67500 }, { "epoch": 3.0783159800814848, "grad_norm": 37.20170211791992, "learning_rate": 7.363493638211186e-06, "loss": 0.2168, "step": 68000 }, { "epoch": 3.1009506564056135, "grad_norm": 37.324825286865234, "learning_rate": 7.339414195313178e-06, "loss": 0.2244, "step": 68500 }, { "epoch": 3.123585332729742, "grad_norm": 39.479610443115234, "learning_rate": 7.315334752415169e-06, "loss": 0.2199, "step": 69000 }, { "epoch": 3.1462200090538706, "grad_norm": 38.33029556274414, "learning_rate": 7.29125530951716e-06, "loss": 0.2285, "step": 69500 }, { "epoch": 3.1688546853779993, "grad_norm": 34.06528091430664, "learning_rate": 7.2671758666191506e-06, "loss": 0.224, "step": 70000 }, { "epoch": 3.1914893617021276, "grad_norm": 36.66078186035156, "learning_rate": 7.2430964237211406e-06, "loss": 0.2224, "step": 70500 }, { "epoch": 3.2141240380262563, "grad_norm": 13.858600616455078, "learning_rate": 7.219016980823132e-06, "loss": 0.2163, "step": 71000 }, { "epoch": 3.2367587143503846, "grad_norm": 2.7164244651794434, "learning_rate": 7.194937537925123e-06, "loss": 0.2356, "step": 71500 }, { "epoch": 3.2593933906745134, "grad_norm": 57.87660598754883, "learning_rate": 7.170858095027115e-06, "loss": 0.2271, "step": 72000 }, { "epoch": 3.2820280669986417, "grad_norm": 90.21813201904297, "learning_rate": 7.146778652129105e-06, "loss": 0.2177, "step": 72500 }, { "epoch": 3.3046627433227704, "grad_norm": 14.034249305725098, "learning_rate": 7.1226992092310956e-06, "loss": 0.2225, "step": 73000 }, { "epoch": 3.327297419646899, "grad_norm": 46.05585861206055, "learning_rate": 7.098619766333087e-06, "loss": 0.2242, "step": 73500 }, { "epoch": 3.3499320959710275, "grad_norm": 37.766517639160156, "learning_rate": 7.074540323435077e-06, "loss": 0.2384, "step": 74000 }, { "epoch": 3.3725667722951562, "grad_norm": 9.106913566589355, "learning_rate": 7.050460880537069e-06, "loss": 0.2485, "step": 74500 }, { "epoch": 3.395201448619285, "grad_norm": 15.67898178100586, "learning_rate": 7.026381437639059e-06, "loss": 0.2373, "step": 75000 }, { "epoch": 3.4178361249434133, "grad_norm": 26.127885818481445, "learning_rate": 7.00230199474105e-06, "loss": 0.2333, "step": 75500 }, { "epoch": 3.440470801267542, "grad_norm": 14.250904083251953, "learning_rate": 6.9782225518430414e-06, "loss": 0.2189, "step": 76000 }, { "epoch": 3.4631054775916703, "grad_norm": 67.55126190185547, "learning_rate": 6.954143108945031e-06, "loss": 0.2378, "step": 76500 }, { "epoch": 3.485740153915799, "grad_norm": 12.584871292114258, "learning_rate": 6.930063666047023e-06, "loss": 0.2448, "step": 77000 }, { "epoch": 3.5083748302399274, "grad_norm": 26.13035011291504, "learning_rate": 6.905984223149014e-06, "loss": 0.2302, "step": 77500 }, { "epoch": 3.531009506564056, "grad_norm": 19.708215713500977, "learning_rate": 6.881904780251004e-06, "loss": 0.2239, "step": 78000 }, { "epoch": 3.553644182888185, "grad_norm": 9.146390914916992, "learning_rate": 6.857825337352996e-06, "loss": 0.2303, "step": 78500 }, { "epoch": 3.576278859212313, "grad_norm": 69.38152313232422, "learning_rate": 6.8337458944549864e-06, "loss": 0.2366, "step": 79000 }, { "epoch": 3.598913535536442, "grad_norm": 43.6939811706543, "learning_rate": 6.809666451556978e-06, "loss": 0.2263, "step": 79500 }, { "epoch": 3.6215482118605706, "grad_norm": 12.009560585021973, "learning_rate": 6.785587008658968e-06, "loss": 0.2251, "step": 80000 }, { "epoch": 3.644182888184699, "grad_norm": 21.58733558654785, "learning_rate": 6.761507565760959e-06, "loss": 0.2447, "step": 80500 }, { "epoch": 3.6668175645088277, "grad_norm": 15.762284278869629, "learning_rate": 6.73742812286295e-06, "loss": 0.2349, "step": 81000 }, { "epoch": 3.689452240832956, "grad_norm": 32.47639083862305, "learning_rate": 6.713348679964941e-06, "loss": 0.2413, "step": 81500 }, { "epoch": 3.7120869171570847, "grad_norm": 36.506526947021484, "learning_rate": 6.689269237066932e-06, "loss": 0.2311, "step": 82000 }, { "epoch": 3.734721593481213, "grad_norm": 30.49101448059082, "learning_rate": 6.665189794168922e-06, "loss": 0.2399, "step": 82500 }, { "epoch": 3.7573562698053418, "grad_norm": 10.327536582946777, "learning_rate": 6.641110351270914e-06, "loss": 0.2332, "step": 83000 }, { "epoch": 3.7799909461294705, "grad_norm": 41.280303955078125, "learning_rate": 6.617030908372905e-06, "loss": 0.2274, "step": 83500 }, { "epoch": 3.802625622453599, "grad_norm": 35.73218536376953, "learning_rate": 6.592951465474895e-06, "loss": 0.2363, "step": 84000 }, { "epoch": 3.8252602987777276, "grad_norm": 19.240692138671875, "learning_rate": 6.5688720225768865e-06, "loss": 0.2408, "step": 84500 }, { "epoch": 3.8478949751018563, "grad_norm": 55.575645446777344, "learning_rate": 6.544792579678877e-06, "loss": 0.2281, "step": 85000 }, { "epoch": 3.8705296514259846, "grad_norm": 5.29152250289917, "learning_rate": 6.520713136780868e-06, "loss": 0.2316, "step": 85500 }, { "epoch": 3.893164327750113, "grad_norm": 74.08470153808594, "learning_rate": 6.496633693882859e-06, "loss": 0.2361, "step": 86000 }, { "epoch": 3.9157990040742416, "grad_norm": 23.3903751373291, "learning_rate": 6.472554250984849e-06, "loss": 0.2425, "step": 86500 }, { "epoch": 3.9384336803983704, "grad_norm": 37.02583694458008, "learning_rate": 6.448474808086841e-06, "loss": 0.2377, "step": 87000 }, { "epoch": 3.9610683567224987, "grad_norm": 13.134513854980469, "learning_rate": 6.4243953651888315e-06, "loss": 0.2424, "step": 87500 }, { "epoch": 3.9837030330466274, "grad_norm": 10.808335304260254, "learning_rate": 6.400315922290823e-06, "loss": 0.2339, "step": 88000 }, { "epoch": 4.0, "eval_accuracy": 0.8661353161365893, "eval_loss": 0.4954204857349396, "eval_runtime": 26.0621, "eval_samples_per_second": 1506.824, "eval_steps_per_second": 94.198, "step": 88360 }, { "epoch": 4.006337709370756, "grad_norm": 9.569930076599121, "learning_rate": 6.376236479392813e-06, "loss": 0.2142, "step": 88500 }, { "epoch": 4.0289723856948845, "grad_norm": 25.4268856048584, "learning_rate": 6.352157036494804e-06, "loss": 0.1827, "step": 89000 }, { "epoch": 4.051607062019013, "grad_norm": 60.39373779296875, "learning_rate": 6.328077593596796e-06, "loss": 0.197, "step": 89500 }, { "epoch": 4.074241738343142, "grad_norm": 63.3898811340332, "learning_rate": 6.303998150698786e-06, "loss": 0.196, "step": 90000 }, { "epoch": 4.09687641466727, "grad_norm": 61.60245132446289, "learning_rate": 6.279918707800777e-06, "loss": 0.1904, "step": 90500 }, { "epoch": 4.119511090991399, "grad_norm": 12.662140846252441, "learning_rate": 6.255839264902767e-06, "loss": 0.1933, "step": 91000 }, { "epoch": 4.142145767315528, "grad_norm": 15.43615436553955, "learning_rate": 6.231759822004758e-06, "loss": 0.1922, "step": 91500 }, { "epoch": 4.164780443639656, "grad_norm": 64.14022064208984, "learning_rate": 6.20768037910675e-06, "loss": 0.1969, "step": 92000 }, { "epoch": 4.187415119963784, "grad_norm": 46.96083068847656, "learning_rate": 6.18360093620874e-06, "loss": 0.2086, "step": 92500 }, { "epoch": 4.2100497962879135, "grad_norm": 28.856672286987305, "learning_rate": 6.1595214933107315e-06, "loss": 0.2022, "step": 93000 }, { "epoch": 4.232684472612042, "grad_norm": 0.08567750453948975, "learning_rate": 6.135442050412722e-06, "loss": 0.1966, "step": 93500 }, { "epoch": 4.25531914893617, "grad_norm": 23.097795486450195, "learning_rate": 6.111362607514712e-06, "loss": 0.1994, "step": 94000 }, { "epoch": 4.277953825260298, "grad_norm": 124.21513366699219, "learning_rate": 6.087283164616704e-06, "loss": 0.2014, "step": 94500 }, { "epoch": 4.300588501584428, "grad_norm": 106.59521484375, "learning_rate": 6.063203721718695e-06, "loss": 0.1999, "step": 95000 }, { "epoch": 4.323223177908556, "grad_norm": 9.190028190612793, "learning_rate": 6.039124278820686e-06, "loss": 0.1989, "step": 95500 }, { "epoch": 4.345857854232684, "grad_norm": 0.6034038066864014, "learning_rate": 6.0150448359226765e-06, "loss": 0.2003, "step": 96000 }, { "epoch": 4.368492530556813, "grad_norm": 28.348718643188477, "learning_rate": 5.990965393024667e-06, "loss": 0.1944, "step": 96500 }, { "epoch": 4.391127206880942, "grad_norm": 34.211814880371094, "learning_rate": 5.966885950126658e-06, "loss": 0.214, "step": 97000 }, { "epoch": 4.41376188320507, "grad_norm": 78.27364349365234, "learning_rate": 5.942806507228649e-06, "loss": 0.1948, "step": 97500 }, { "epoch": 4.436396559529198, "grad_norm": 39.985931396484375, "learning_rate": 5.918727064330641e-06, "loss": 0.1979, "step": 98000 }, { "epoch": 4.4590312358533275, "grad_norm": 125.82061004638672, "learning_rate": 5.894647621432631e-06, "loss": 0.2057, "step": 98500 }, { "epoch": 4.481665912177456, "grad_norm": 24.297237396240234, "learning_rate": 5.8705681785346215e-06, "loss": 0.1873, "step": 99000 }, { "epoch": 4.504300588501584, "grad_norm": 14.736886024475098, "learning_rate": 5.846488735636613e-06, "loss": 0.2043, "step": 99500 }, { "epoch": 4.526935264825713, "grad_norm": 47.609375, "learning_rate": 5.822409292738603e-06, "loss": 0.2043, "step": 100000 }, { "epoch": 4.549569941149842, "grad_norm": 32.71791076660156, "learning_rate": 5.798329849840595e-06, "loss": 0.1981, "step": 100500 }, { "epoch": 4.57220461747397, "grad_norm": 32.31149673461914, "learning_rate": 5.774250406942586e-06, "loss": 0.1904, "step": 101000 }, { "epoch": 4.594839293798099, "grad_norm": 15.635351181030273, "learning_rate": 5.7501709640445765e-06, "loss": 0.2027, "step": 101500 }, { "epoch": 4.617473970122227, "grad_norm": 0.519290566444397, "learning_rate": 5.726091521146567e-06, "loss": 0.2102, "step": 102000 }, { "epoch": 4.640108646446356, "grad_norm": 54.78620910644531, "learning_rate": 5.702012078248557e-06, "loss": 0.2028, "step": 102500 }, { "epoch": 4.662743322770485, "grad_norm": 15.834295272827148, "learning_rate": 5.677932635350549e-06, "loss": 0.2068, "step": 103000 }, { "epoch": 4.685377999094613, "grad_norm": 18.294235229492188, "learning_rate": 5.65385319245254e-06, "loss": 0.2194, "step": 103500 }, { "epoch": 4.7080126754187415, "grad_norm": 9.605391502380371, "learning_rate": 5.6297737495545315e-06, "loss": 0.1946, "step": 104000 }, { "epoch": 4.73064735174287, "grad_norm": 50.544219970703125, "learning_rate": 5.6056943066565215e-06, "loss": 0.2007, "step": 104500 }, { "epoch": 4.753282028066999, "grad_norm": 19.007843017578125, "learning_rate": 5.581614863758512e-06, "loss": 0.2192, "step": 105000 }, { "epoch": 4.775916704391127, "grad_norm": 1.8254756927490234, "learning_rate": 5.557535420860504e-06, "loss": 0.1972, "step": 105500 }, { "epoch": 4.7985513807152556, "grad_norm": 32.10722732543945, "learning_rate": 5.533455977962494e-06, "loss": 0.2088, "step": 106000 }, { "epoch": 4.821186057039384, "grad_norm": 54.646392822265625, "learning_rate": 5.509376535064486e-06, "loss": 0.2111, "step": 106500 }, { "epoch": 4.843820733363513, "grad_norm": 0.40658873319625854, "learning_rate": 5.485297092166476e-06, "loss": 0.2114, "step": 107000 }, { "epoch": 4.866455409687641, "grad_norm": 12.083222389221191, "learning_rate": 5.4612176492684665e-06, "loss": 0.1959, "step": 107500 }, { "epoch": 4.88909008601177, "grad_norm": 0.27834174036979675, "learning_rate": 5.437138206370458e-06, "loss": 0.1994, "step": 108000 }, { "epoch": 4.911724762335899, "grad_norm": 7.066097259521484, "learning_rate": 5.413058763472448e-06, "loss": 0.2121, "step": 108500 }, { "epoch": 4.934359438660027, "grad_norm": 39.164085388183594, "learning_rate": 5.38897932057444e-06, "loss": 0.1952, "step": 109000 }, { "epoch": 4.956994114984155, "grad_norm": 27.279882431030273, "learning_rate": 5.364899877676431e-06, "loss": 0.1997, "step": 109500 }, { "epoch": 4.979628791308285, "grad_norm": 54.53019332885742, "learning_rate": 5.340820434778421e-06, "loss": 0.2037, "step": 110000 }, { "epoch": 5.0, "eval_accuracy": 0.8667464541264547, "eval_loss": 0.6144042611122131, "eval_runtime": 26.0348, "eval_samples_per_second": 1508.405, "eval_steps_per_second": 94.297, "step": 110450 }, { "epoch": 5.002263467632413, "grad_norm": 24.079191207885742, "learning_rate": 5.316740991880412e-06, "loss": 0.197, "step": 110500 }, { "epoch": 5.024898143956541, "grad_norm": 0.3425958752632141, "learning_rate": 5.292661548982403e-06, "loss": 0.1472, "step": 111000 }, { "epoch": 5.04753282028067, "grad_norm": 0.11918644607067108, "learning_rate": 5.268582106084394e-06, "loss": 0.1614, "step": 111500 }, { "epoch": 5.070167496604799, "grad_norm": 0.0681818500161171, "learning_rate": 5.244502663186385e-06, "loss": 0.1781, "step": 112000 }, { "epoch": 5.092802172928927, "grad_norm": 62.686737060546875, "learning_rate": 5.220423220288376e-06, "loss": 0.1781, "step": 112500 }, { "epoch": 5.115436849253055, "grad_norm": 73.86345672607422, "learning_rate": 5.1963437773903666e-06, "loss": 0.1679, "step": 113000 }, { "epoch": 5.1380715255771845, "grad_norm": 0.13405387103557587, "learning_rate": 5.172264334492357e-06, "loss": 0.184, "step": 113500 }, { "epoch": 5.160706201901313, "grad_norm": 38.33438491821289, "learning_rate": 5.148184891594349e-06, "loss": 0.1679, "step": 114000 }, { "epoch": 5.183340878225441, "grad_norm": 0.36470118165016174, "learning_rate": 5.124105448696339e-06, "loss": 0.154, "step": 114500 }, { "epoch": 5.20597555454957, "grad_norm": 31.240108489990234, "learning_rate": 5.10002600579833e-06, "loss": 0.1718, "step": 115000 }, { "epoch": 5.228610230873699, "grad_norm": 0.2006056010723114, "learning_rate": 5.075946562900322e-06, "loss": 0.1624, "step": 115500 }, { "epoch": 5.251244907197827, "grad_norm": 80.91893768310547, "learning_rate": 5.0518671200023116e-06, "loss": 0.1751, "step": 116000 }, { "epoch": 5.273879583521955, "grad_norm": 39.18518829345703, "learning_rate": 5.027787677104303e-06, "loss": 0.1768, "step": 116500 }, { "epoch": 5.296514259846084, "grad_norm": 10.39631175994873, "learning_rate": 5.003708234206294e-06, "loss": 0.186, "step": 117000 }, { "epoch": 5.319148936170213, "grad_norm": 2.1696979999542236, "learning_rate": 4.979628791308285e-06, "loss": 0.175, "step": 117500 }, { "epoch": 5.341783612494341, "grad_norm": 21.064584732055664, "learning_rate": 4.955549348410276e-06, "loss": 0.1625, "step": 118000 }, { "epoch": 5.36441828881847, "grad_norm": 0.3759268522262573, "learning_rate": 4.931469905512267e-06, "loss": 0.1834, "step": 118500 }, { "epoch": 5.3870529651425985, "grad_norm": 35.439117431640625, "learning_rate": 4.9073904626142574e-06, "loss": 0.1859, "step": 119000 }, { "epoch": 5.409687641466727, "grad_norm": 63.224666595458984, "learning_rate": 4.883311019716248e-06, "loss": 0.1722, "step": 119500 }, { "epoch": 5.432322317790856, "grad_norm": 2.553009033203125, "learning_rate": 4.859231576818239e-06, "loss": 0.173, "step": 120000 }, { "epoch": 5.454956994114984, "grad_norm": 15.771255493164062, "learning_rate": 4.83515213392023e-06, "loss": 0.1922, "step": 120500 }, { "epoch": 5.4775916704391125, "grad_norm": 108.56519317626953, "learning_rate": 4.811072691022221e-06, "loss": 0.1717, "step": 121000 }, { "epoch": 5.500226346763242, "grad_norm": 78.23528289794922, "learning_rate": 4.7869932481242124e-06, "loss": 0.1822, "step": 121500 }, { "epoch": 5.52286102308737, "grad_norm": 62.75898742675781, "learning_rate": 4.7629138052262024e-06, "loss": 0.1697, "step": 122000 }, { "epoch": 5.545495699411498, "grad_norm": 136.13113403320312, "learning_rate": 4.738834362328193e-06, "loss": 0.1807, "step": 122500 }, { "epoch": 5.568130375735627, "grad_norm": 52.2840461730957, "learning_rate": 4.714754919430184e-06, "loss": 0.1762, "step": 123000 }, { "epoch": 5.590765052059756, "grad_norm": 4.957085609436035, "learning_rate": 4.690675476532176e-06, "loss": 0.1875, "step": 123500 }, { "epoch": 5.613399728383884, "grad_norm": 39.328086853027344, "learning_rate": 4.666596033634167e-06, "loss": 0.1732, "step": 124000 }, { "epoch": 5.636034404708012, "grad_norm": 3.6397218704223633, "learning_rate": 4.642516590736157e-06, "loss": 0.175, "step": 124500 }, { "epoch": 5.658669081032142, "grad_norm": 32.74443435668945, "learning_rate": 4.618437147838148e-06, "loss": 0.1667, "step": 125000 }, { "epoch": 5.68130375735627, "grad_norm": 32.64069366455078, "learning_rate": 4.594357704940139e-06, "loss": 0.1691, "step": 125500 }, { "epoch": 5.703938433680398, "grad_norm": 21.668283462524414, "learning_rate": 4.57027826204213e-06, "loss": 0.1785, "step": 126000 }, { "epoch": 5.7265731100045265, "grad_norm": 8.382264137268066, "learning_rate": 4.546198819144121e-06, "loss": 0.1829, "step": 126500 }, { "epoch": 5.749207786328656, "grad_norm": 24.240978240966797, "learning_rate": 4.522119376246112e-06, "loss": 0.1573, "step": 127000 }, { "epoch": 5.771842462652784, "grad_norm": 98.30403900146484, "learning_rate": 4.4980399333481025e-06, "loss": 0.1962, "step": 127500 }, { "epoch": 5.794477138976912, "grad_norm": 0.6671485304832458, "learning_rate": 4.473960490450093e-06, "loss": 0.1813, "step": 128000 }, { "epoch": 5.8171118153010415, "grad_norm": 71.27288055419922, "learning_rate": 4.449881047552084e-06, "loss": 0.1747, "step": 128500 }, { "epoch": 5.83974649162517, "grad_norm": 148.5382537841797, "learning_rate": 4.425801604654075e-06, "loss": 0.1676, "step": 129000 }, { "epoch": 5.862381167949298, "grad_norm": 0.19661898910999298, "learning_rate": 4.401722161756066e-06, "loss": 0.1771, "step": 129500 }, { "epoch": 5.885015844273427, "grad_norm": 16.245052337646484, "learning_rate": 4.377642718858057e-06, "loss": 0.1864, "step": 130000 }, { "epoch": 5.907650520597556, "grad_norm": 0.5395733118057251, "learning_rate": 4.3535632759600475e-06, "loss": 0.1775, "step": 130500 }, { "epoch": 5.930285196921684, "grad_norm": 13.2942533493042, "learning_rate": 4.329483833062038e-06, "loss": 0.1669, "step": 131000 }, { "epoch": 5.952919873245813, "grad_norm": 12.363393783569336, "learning_rate": 4.30540439016403e-06, "loss": 0.1747, "step": 131500 }, { "epoch": 5.975554549569941, "grad_norm": 0.334881067276001, "learning_rate": 4.281324947266021e-06, "loss": 0.1941, "step": 132000 }, { "epoch": 5.99818922589407, "grad_norm": 9.071168899536133, "learning_rate": 4.257245504368011e-06, "loss": 0.1745, "step": 132500 }, { "epoch": 6.0, "eval_accuracy": 0.8641491176695272, "eval_loss": 0.6998937726020813, "eval_runtime": 26.0201, "eval_samples_per_second": 1509.256, "eval_steps_per_second": 94.35, "step": 132540 }, { "epoch": 6.020823902218198, "grad_norm": 79.37480163574219, "learning_rate": 4.2331660614700025e-06, "loss": 0.1323, "step": 133000 }, { "epoch": 6.043458578542327, "grad_norm": 97.10159301757812, "learning_rate": 4.209086618571993e-06, "loss": 0.1373, "step": 133500 }, { "epoch": 6.0660932548664555, "grad_norm": 9.43271541595459, "learning_rate": 4.185007175673984e-06, "loss": 0.1422, "step": 134000 }, { "epoch": 6.088727931190584, "grad_norm": 0.20963682234287262, "learning_rate": 4.160927732775975e-06, "loss": 0.1457, "step": 134500 }, { "epoch": 6.111362607514713, "grad_norm": 55.66864776611328, "learning_rate": 4.136848289877966e-06, "loss": 0.1508, "step": 135000 }, { "epoch": 6.133997283838841, "grad_norm": 84.8683090209961, "learning_rate": 4.112768846979957e-06, "loss": 0.1347, "step": 135500 }, { "epoch": 6.1566319601629695, "grad_norm": 77.05133819580078, "learning_rate": 4.0886894040819475e-06, "loss": 0.1424, "step": 136000 }, { "epoch": 6.179266636487098, "grad_norm": 0.16760210692882538, "learning_rate": 4.064609961183938e-06, "loss": 0.1401, "step": 136500 }, { "epoch": 6.201901312811227, "grad_norm": 6.680587291717529, "learning_rate": 4.040530518285929e-06, "loss": 0.1579, "step": 137000 }, { "epoch": 6.224535989135355, "grad_norm": 10.419951438903809, "learning_rate": 4.01645107538792e-06, "loss": 0.1441, "step": 137500 }, { "epoch": 6.247170665459484, "grad_norm": 75.0548095703125, "learning_rate": 3.992371632489911e-06, "loss": 0.1489, "step": 138000 }, { "epoch": 6.269805341783613, "grad_norm": 1.2689846754074097, "learning_rate": 3.968292189591902e-06, "loss": 0.1527, "step": 138500 }, { "epoch": 6.292440018107741, "grad_norm": 23.291440963745117, "learning_rate": 3.9442127466938925e-06, "loss": 0.1508, "step": 139000 }, { "epoch": 6.315074694431869, "grad_norm": 99.6236343383789, "learning_rate": 3.920133303795884e-06, "loss": 0.1561, "step": 139500 }, { "epoch": 6.337709370755999, "grad_norm": 62.657745361328125, "learning_rate": 3.896053860897875e-06, "loss": 0.1316, "step": 140000 }, { "epoch": 6.360344047080127, "grad_norm": 1.494821310043335, "learning_rate": 3.871974417999865e-06, "loss": 0.1371, "step": 140500 }, { "epoch": 6.382978723404255, "grad_norm": 54.55570602416992, "learning_rate": 3.847894975101857e-06, "loss": 0.1281, "step": 141000 }, { "epoch": 6.4056133997283835, "grad_norm": 39.3503303527832, "learning_rate": 3.8238155322038475e-06, "loss": 0.153, "step": 141500 }, { "epoch": 6.428248076052513, "grad_norm": 211.0976104736328, "learning_rate": 3.7997360893058384e-06, "loss": 0.1529, "step": 142000 }, { "epoch": 6.450882752376641, "grad_norm": 36.55986404418945, "learning_rate": 3.775656646407829e-06, "loss": 0.1446, "step": 142500 }, { "epoch": 6.473517428700769, "grad_norm": 14.937396049499512, "learning_rate": 3.7515772035098196e-06, "loss": 0.157, "step": 143000 }, { "epoch": 6.4961521050248985, "grad_norm": 0.12284702807664871, "learning_rate": 3.727497760611811e-06, "loss": 0.1576, "step": 143500 }, { "epoch": 6.518786781349027, "grad_norm": 169.0521697998047, "learning_rate": 3.7034183177138017e-06, "loss": 0.1516, "step": 144000 }, { "epoch": 6.541421457673155, "grad_norm": 119.7757339477539, "learning_rate": 3.6793388748157925e-06, "loss": 0.142, "step": 144500 }, { "epoch": 6.564056133997283, "grad_norm": 0.5702412128448486, "learning_rate": 3.6552594319177838e-06, "loss": 0.1519, "step": 145000 }, { "epoch": 6.586690810321413, "grad_norm": 0.6575600504875183, "learning_rate": 3.631179989019774e-06, "loss": 0.1494, "step": 145500 }, { "epoch": 6.609325486645541, "grad_norm": 104.18098449707031, "learning_rate": 3.607100546121765e-06, "loss": 0.1431, "step": 146000 }, { "epoch": 6.631960162969669, "grad_norm": 0.18219584226608276, "learning_rate": 3.583021103223756e-06, "loss": 0.1397, "step": 146500 }, { "epoch": 6.654594839293798, "grad_norm": 39.80546569824219, "learning_rate": 3.558941660325747e-06, "loss": 0.1384, "step": 147000 }, { "epoch": 6.677229515617927, "grad_norm": 70.61176300048828, "learning_rate": 3.534862217427738e-06, "loss": 0.1452, "step": 147500 }, { "epoch": 6.699864191942055, "grad_norm": 0.11137774586677551, "learning_rate": 3.5107827745297292e-06, "loss": 0.1649, "step": 148000 }, { "epoch": 6.722498868266184, "grad_norm": 1.3033461570739746, "learning_rate": 3.4867033316317196e-06, "loss": 0.1468, "step": 148500 }, { "epoch": 6.7451335445903124, "grad_norm": 188.11358642578125, "learning_rate": 3.4626238887337105e-06, "loss": 0.1396, "step": 149000 }, { "epoch": 6.767768220914441, "grad_norm": 186.4955596923828, "learning_rate": 3.4385444458357013e-06, "loss": 0.1415, "step": 149500 }, { "epoch": 6.79040289723857, "grad_norm": 10.157150268554688, "learning_rate": 3.4144650029376926e-06, "loss": 0.1446, "step": 150000 }, { "epoch": 6.813037573562698, "grad_norm": 14.647910118103027, "learning_rate": 3.3903855600396834e-06, "loss": 0.1363, "step": 150500 }, { "epoch": 6.8356722498868265, "grad_norm": 0.07332862168550491, "learning_rate": 3.366306117141674e-06, "loss": 0.1622, "step": 151000 }, { "epoch": 6.858306926210955, "grad_norm": 125.91682434082031, "learning_rate": 3.342226674243665e-06, "loss": 0.1408, "step": 151500 }, { "epoch": 6.880941602535084, "grad_norm": 0.21492162346839905, "learning_rate": 3.318147231345656e-06, "loss": 0.1423, "step": 152000 }, { "epoch": 6.903576278859212, "grad_norm": 27.321796417236328, "learning_rate": 3.2940677884476467e-06, "loss": 0.1562, "step": 152500 }, { "epoch": 6.926210955183341, "grad_norm": 0.10927353799343109, "learning_rate": 3.269988345549638e-06, "loss": 0.1475, "step": 153000 }, { "epoch": 6.94884563150747, "grad_norm": 97.3139877319336, "learning_rate": 3.2459089026516284e-06, "loss": 0.1438, "step": 153500 }, { "epoch": 6.971480307831598, "grad_norm": 159.5480499267578, "learning_rate": 3.2218294597536192e-06, "loss": 0.1481, "step": 154000 }, { "epoch": 6.994114984155726, "grad_norm": 0.06921840459108353, "learning_rate": 3.19775001685561e-06, "loss": 0.1671, "step": 154500 }, { "epoch": 7.0, "eval_accuracy": 0.8639454050062387, "eval_loss": 0.7750576734542847, "eval_runtime": 26.0647, "eval_samples_per_second": 1506.673, "eval_steps_per_second": 94.189, "step": 154630 }, { "epoch": 7.016749660479855, "grad_norm": 1.394852638244629, "learning_rate": 3.1736705739576013e-06, "loss": 0.1232, "step": 155000 }, { "epoch": 7.039384336803984, "grad_norm": 0.33521416783332825, "learning_rate": 3.149591131059592e-06, "loss": 0.108, "step": 155500 }, { "epoch": 7.062019013128112, "grad_norm": 0.06891336292028427, "learning_rate": 3.1255116881615826e-06, "loss": 0.1155, "step": 156000 }, { "epoch": 7.0846536894522405, "grad_norm": 15.24691390991211, "learning_rate": 3.101432245263574e-06, "loss": 0.1266, "step": 156500 }, { "epoch": 7.10728836577637, "grad_norm": 0.06932001560926437, "learning_rate": 3.0773528023655647e-06, "loss": 0.1114, "step": 157000 }, { "epoch": 7.129923042100498, "grad_norm": 0.0540509857237339, "learning_rate": 3.0532733594675555e-06, "loss": 0.1279, "step": 157500 }, { "epoch": 7.152557718424626, "grad_norm": 29.716217041015625, "learning_rate": 3.0291939165695468e-06, "loss": 0.115, "step": 158000 }, { "epoch": 7.1751923947487555, "grad_norm": 0.0442744679749012, "learning_rate": 3.0051144736715376e-06, "loss": 0.1252, "step": 158500 }, { "epoch": 7.197827071072884, "grad_norm": 7.542829513549805, "learning_rate": 2.981035030773528e-06, "loss": 0.1156, "step": 159000 }, { "epoch": 7.220461747397012, "grad_norm": 11.190882682800293, "learning_rate": 2.9569555878755193e-06, "loss": 0.1065, "step": 159500 }, { "epoch": 7.24309642372114, "grad_norm": 0.10643190145492554, "learning_rate": 2.93287614497751e-06, "loss": 0.1316, "step": 160000 }, { "epoch": 7.26573110004527, "grad_norm": 0.16745133697986603, "learning_rate": 2.908796702079501e-06, "loss": 0.1101, "step": 160500 }, { "epoch": 7.288365776369398, "grad_norm": 0.16076330840587616, "learning_rate": 2.884717259181492e-06, "loss": 0.1314, "step": 161000 }, { "epoch": 7.311000452693526, "grad_norm": 0.10047034919261932, "learning_rate": 2.8606378162834826e-06, "loss": 0.1181, "step": 161500 }, { "epoch": 7.333635129017655, "grad_norm": 0.331920862197876, "learning_rate": 2.8365583733854734e-06, "loss": 0.1259, "step": 162000 }, { "epoch": 7.356269805341784, "grad_norm": 0.1453462541103363, "learning_rate": 2.8124789304874643e-06, "loss": 0.1249, "step": 162500 }, { "epoch": 7.378904481665912, "grad_norm": 0.8490937948226929, "learning_rate": 2.7883994875894555e-06, "loss": 0.1209, "step": 163000 }, { "epoch": 7.401539157990041, "grad_norm": 0.22700923681259155, "learning_rate": 2.7643200446914464e-06, "loss": 0.1244, "step": 163500 }, { "epoch": 7.424173834314169, "grad_norm": 0.19261109828948975, "learning_rate": 2.7402406017934368e-06, "loss": 0.135, "step": 164000 }, { "epoch": 7.446808510638298, "grad_norm": 75.31595611572266, "learning_rate": 2.716161158895428e-06, "loss": 0.1262, "step": 164500 }, { "epoch": 7.469443186962426, "grad_norm": 65.7965087890625, "learning_rate": 2.692081715997419e-06, "loss": 0.1139, "step": 165000 }, { "epoch": 7.492077863286555, "grad_norm": 114.45712280273438, "learning_rate": 2.6680022730994097e-06, "loss": 0.1335, "step": 165500 }, { "epoch": 7.5147125396106835, "grad_norm": 0.08683761209249496, "learning_rate": 2.643922830201401e-06, "loss": 0.1257, "step": 166000 }, { "epoch": 7.537347215934812, "grad_norm": 91.00257873535156, "learning_rate": 2.6198433873033918e-06, "loss": 0.1307, "step": 166500 }, { "epoch": 7.559981892258941, "grad_norm": 0.0967201367020607, "learning_rate": 2.595763944405382e-06, "loss": 0.1163, "step": 167000 }, { "epoch": 7.582616568583069, "grad_norm": 179.25857543945312, "learning_rate": 2.5716845015073735e-06, "loss": 0.1207, "step": 167500 }, { "epoch": 7.605251244907198, "grad_norm": 0.09339158982038498, "learning_rate": 2.5476050586093643e-06, "loss": 0.1284, "step": 168000 }, { "epoch": 7.627885921231327, "grad_norm": 0.0970580130815506, "learning_rate": 2.523525615711355e-06, "loss": 0.1222, "step": 168500 }, { "epoch": 7.650520597555455, "grad_norm": 0.26078376173973083, "learning_rate": 2.499446172813346e-06, "loss": 0.1225, "step": 169000 }, { "epoch": 7.673155273879583, "grad_norm": 29.06781005859375, "learning_rate": 2.475366729915337e-06, "loss": 0.1238, "step": 169500 }, { "epoch": 7.695789950203712, "grad_norm": 4.162774085998535, "learning_rate": 2.4512872870173276e-06, "loss": 0.1323, "step": 170000 }, { "epoch": 7.718424626527841, "grad_norm": 0.9394495487213135, "learning_rate": 2.4272078441193185e-06, "loss": 0.1173, "step": 170500 }, { "epoch": 7.741059302851969, "grad_norm": 0.06645090132951736, "learning_rate": 2.4031284012213097e-06, "loss": 0.1275, "step": 171000 }, { "epoch": 7.7636939791760975, "grad_norm": 0.12068886309862137, "learning_rate": 2.3790489583233006e-06, "loss": 0.1103, "step": 171500 }, { "epoch": 7.786328655500227, "grad_norm": 0.18481621146202087, "learning_rate": 2.3549695154252914e-06, "loss": 0.1225, "step": 172000 }, { "epoch": 7.808963331824355, "grad_norm": 0.0315103605389595, "learning_rate": 2.3308900725272822e-06, "loss": 0.1061, "step": 172500 }, { "epoch": 7.831598008148483, "grad_norm": 2.336836814880371, "learning_rate": 2.306810629629273e-06, "loss": 0.1226, "step": 173000 }, { "epoch": 7.854232684472612, "grad_norm": 32.957130432128906, "learning_rate": 2.282731186731264e-06, "loss": 0.1228, "step": 173500 }, { "epoch": 7.876867360796741, "grad_norm": 0.14461065828800201, "learning_rate": 2.2586517438332547e-06, "loss": 0.1172, "step": 174000 }, { "epoch": 7.899502037120869, "grad_norm": 0.13647380471229553, "learning_rate": 2.2345723009352456e-06, "loss": 0.1242, "step": 174500 }, { "epoch": 7.922136713444997, "grad_norm": 0.14755909144878387, "learning_rate": 2.210492858037237e-06, "loss": 0.1167, "step": 175000 }, { "epoch": 7.944771389769127, "grad_norm": 0.16207629442214966, "learning_rate": 2.1864134151392277e-06, "loss": 0.127, "step": 175500 }, { "epoch": 7.967406066093255, "grad_norm": 15.389420509338379, "learning_rate": 2.1623339722412185e-06, "loss": 0.1163, "step": 176000 }, { "epoch": 7.990040742417383, "grad_norm": 202.52561950683594, "learning_rate": 2.1382545293432093e-06, "loss": 0.121, "step": 176500 }, { "epoch": 8.0, "eval_accuracy": 0.8648621119910367, "eval_loss": 0.8655109405517578, "eval_runtime": 26.0286, "eval_samples_per_second": 1508.765, "eval_steps_per_second": 94.319, "step": 176720 }, { "epoch": 8.012675418741512, "grad_norm": 1.5518616437911987, "learning_rate": 2.1141750864452e-06, "loss": 0.1058, "step": 177000 }, { "epoch": 8.03531009506564, "grad_norm": 0.015344664454460144, "learning_rate": 2.090095643547191e-06, "loss": 0.0809, "step": 177500 }, { "epoch": 8.057944771389769, "grad_norm": 100.70498657226562, "learning_rate": 2.066016200649182e-06, "loss": 0.094, "step": 178000 }, { "epoch": 8.080579447713898, "grad_norm": 0.47632962465286255, "learning_rate": 2.0419367577511727e-06, "loss": 0.1115, "step": 178500 }, { "epoch": 8.103214124038026, "grad_norm": 6.641209125518799, "learning_rate": 2.017857314853164e-06, "loss": 0.0856, "step": 179000 }, { "epoch": 8.125848800362155, "grad_norm": 27.82591438293457, "learning_rate": 1.9937778719551548e-06, "loss": 0.0983, "step": 179500 }, { "epoch": 8.148483476686284, "grad_norm": 314.1797790527344, "learning_rate": 1.9696984290571456e-06, "loss": 0.1075, "step": 180000 }, { "epoch": 8.171118153010411, "grad_norm": 254.616455078125, "learning_rate": 1.9456189861591364e-06, "loss": 0.1016, "step": 180500 }, { "epoch": 8.19375282933454, "grad_norm": 0.07589972764253616, "learning_rate": 1.9215395432611273e-06, "loss": 0.0924, "step": 181000 }, { "epoch": 8.21638750565867, "grad_norm": 77.42697143554688, "learning_rate": 1.8974601003631183e-06, "loss": 0.0965, "step": 181500 }, { "epoch": 8.239022181982797, "grad_norm": 0.12723857164382935, "learning_rate": 1.873380657465109e-06, "loss": 0.1024, "step": 182000 }, { "epoch": 8.261656858306926, "grad_norm": 7.569960594177246, "learning_rate": 1.8493012145671e-06, "loss": 0.1201, "step": 182500 }, { "epoch": 8.284291534631055, "grad_norm": 92.7170181274414, "learning_rate": 1.825221771669091e-06, "loss": 0.0976, "step": 183000 }, { "epoch": 8.306926210955183, "grad_norm": 0.009992193430662155, "learning_rate": 1.8011423287710816e-06, "loss": 0.1025, "step": 183500 }, { "epoch": 8.329560887279312, "grad_norm": 0.11167449504137039, "learning_rate": 1.7770628858730727e-06, "loss": 0.09, "step": 184000 }, { "epoch": 8.352195563603441, "grad_norm": 54.87889099121094, "learning_rate": 1.7529834429750633e-06, "loss": 0.0993, "step": 184500 }, { "epoch": 8.374830239927569, "grad_norm": 0.07053136825561523, "learning_rate": 1.7289040000770544e-06, "loss": 0.0987, "step": 185000 }, { "epoch": 8.397464916251698, "grad_norm": 4.904270648956299, "learning_rate": 1.7048245571790454e-06, "loss": 0.105, "step": 185500 }, { "epoch": 8.420099592575827, "grad_norm": 0.10817304253578186, "learning_rate": 1.680745114281036e-06, "loss": 0.0977, "step": 186000 }, { "epoch": 8.442734268899954, "grad_norm": 0.03593330830335617, "learning_rate": 1.656665671383027e-06, "loss": 0.0954, "step": 186500 }, { "epoch": 8.465368945224084, "grad_norm": 105.52520751953125, "learning_rate": 1.6325862284850181e-06, "loss": 0.1065, "step": 187000 }, { "epoch": 8.488003621548211, "grad_norm": 0.19925498962402344, "learning_rate": 1.6085067855870087e-06, "loss": 0.0983, "step": 187500 }, { "epoch": 8.51063829787234, "grad_norm": 0.29446855187416077, "learning_rate": 1.5844273426889998e-06, "loss": 0.1015, "step": 188000 }, { "epoch": 8.53327297419647, "grad_norm": 13.635686874389648, "learning_rate": 1.5603478997909904e-06, "loss": 0.0902, "step": 188500 }, { "epoch": 8.555907650520597, "grad_norm": 0.05709734186530113, "learning_rate": 1.5362684568929815e-06, "loss": 0.0972, "step": 189000 }, { "epoch": 8.578542326844726, "grad_norm": 0.023464586585760117, "learning_rate": 1.5121890139949725e-06, "loss": 0.1121, "step": 189500 }, { "epoch": 8.601177003168855, "grad_norm": 0.10528367012739182, "learning_rate": 1.4881095710969631e-06, "loss": 0.1001, "step": 190000 }, { "epoch": 8.623811679492983, "grad_norm": 0.4939417243003845, "learning_rate": 1.4640301281989542e-06, "loss": 0.098, "step": 190500 }, { "epoch": 8.646446355817112, "grad_norm": 12.716562271118164, "learning_rate": 1.4399506853009452e-06, "loss": 0.0945, "step": 191000 }, { "epoch": 8.669081032141241, "grad_norm": 145.86587524414062, "learning_rate": 1.4158712424029358e-06, "loss": 0.0888, "step": 191500 }, { "epoch": 8.691715708465368, "grad_norm": 0.03197444975376129, "learning_rate": 1.3917917995049269e-06, "loss": 0.0858, "step": 192000 }, { "epoch": 8.714350384789498, "grad_norm": 0.11782459169626236, "learning_rate": 1.3677123566069175e-06, "loss": 0.1027, "step": 192500 }, { "epoch": 8.736985061113627, "grad_norm": 0.1515623927116394, "learning_rate": 1.3436329137089086e-06, "loss": 0.1003, "step": 193000 }, { "epoch": 8.759619737437754, "grad_norm": 0.04349144920706749, "learning_rate": 1.3195534708108996e-06, "loss": 0.0926, "step": 193500 }, { "epoch": 8.782254413761883, "grad_norm": 14.532307624816895, "learning_rate": 1.2954740279128902e-06, "loss": 0.0935, "step": 194000 }, { "epoch": 8.80488909008601, "grad_norm": 0.05790287256240845, "learning_rate": 1.2713945850148813e-06, "loss": 0.0923, "step": 194500 }, { "epoch": 8.82752376641014, "grad_norm": 81.71268463134766, "learning_rate": 1.247315142116872e-06, "loss": 0.1172, "step": 195000 }, { "epoch": 8.85015844273427, "grad_norm": 0.03535538911819458, "learning_rate": 1.223235699218863e-06, "loss": 0.0985, "step": 195500 }, { "epoch": 8.872793119058397, "grad_norm": 0.05989941582083702, "learning_rate": 1.199156256320854e-06, "loss": 0.1033, "step": 196000 }, { "epoch": 8.895427795382526, "grad_norm": 0.18569760024547577, "learning_rate": 1.1750768134228448e-06, "loss": 0.1178, "step": 196500 }, { "epoch": 8.918062471706655, "grad_norm": 0.02892606146633625, "learning_rate": 1.1509973705248357e-06, "loss": 0.0964, "step": 197000 }, { "epoch": 8.940697148030782, "grad_norm": 181.0758819580078, "learning_rate": 1.1269179276268265e-06, "loss": 0.1056, "step": 197500 }, { "epoch": 8.963331824354912, "grad_norm": 28.286996841430664, "learning_rate": 1.1028384847288175e-06, "loss": 0.1055, "step": 198000 }, { "epoch": 8.98596650067904, "grad_norm": 0.08270686864852905, "learning_rate": 1.0787590418308084e-06, "loss": 0.1083, "step": 198500 }, { "epoch": 9.0, "eval_accuracy": 0.8651931450688803, "eval_loss": 0.9116848111152649, "eval_runtime": 25.9694, "eval_samples_per_second": 1512.2, "eval_steps_per_second": 94.534, "step": 198810 }, { "epoch": 9.008601177003168, "grad_norm": 0.3910556733608246, "learning_rate": 1.0546795989327992e-06, "loss": 0.0919, "step": 199000 }, { "epoch": 9.031235853327297, "grad_norm": 39.47013854980469, "learning_rate": 1.03060015603479e-06, "loss": 0.0828, "step": 199500 }, { "epoch": 9.053870529651427, "grad_norm": 0.02229388989508152, "learning_rate": 1.006520713136781e-06, "loss": 0.0814, "step": 200000 }, { "epoch": 9.076505205975554, "grad_norm": 0.028238942846655846, "learning_rate": 9.82441270238772e-07, "loss": 0.0835, "step": 200500 }, { "epoch": 9.099139882299683, "grad_norm": 10.808701515197754, "learning_rate": 9.583618273407628e-07, "loss": 0.0923, "step": 201000 }, { "epoch": 9.121774558623812, "grad_norm": 221.81275939941406, "learning_rate": 9.342823844427536e-07, "loss": 0.0665, "step": 201500 }, { "epoch": 9.14440923494794, "grad_norm": 0.1080513447523117, "learning_rate": 9.102029415447445e-07, "loss": 0.0876, "step": 202000 }, { "epoch": 9.167043911272069, "grad_norm": 232.07345581054688, "learning_rate": 8.861234986467354e-07, "loss": 0.0863, "step": 202500 }, { "epoch": 9.189678587596198, "grad_norm": 0.04613710194826126, "learning_rate": 8.620440557487263e-07, "loss": 0.0938, "step": 203000 }, { "epoch": 9.212313263920326, "grad_norm": 0.022649744525551796, "learning_rate": 8.379646128507171e-07, "loss": 0.0815, "step": 203500 }, { "epoch": 9.234947940244455, "grad_norm": 0.0704297199845314, "learning_rate": 8.138851699527081e-07, "loss": 0.0793, "step": 204000 }, { "epoch": 9.257582616568584, "grad_norm": 169.63650512695312, "learning_rate": 7.898057270546989e-07, "loss": 0.0757, "step": 204500 }, { "epoch": 9.280217292892711, "grad_norm": 145.97549438476562, "learning_rate": 7.657262841566899e-07, "loss": 0.0818, "step": 205000 }, { "epoch": 9.30285196921684, "grad_norm": 0.3320428431034088, "learning_rate": 7.416468412586807e-07, "loss": 0.0996, "step": 205500 }, { "epoch": 9.325486645540968, "grad_norm": 2.268958806991577, "learning_rate": 7.175673983606715e-07, "loss": 0.0835, "step": 206000 }, { "epoch": 9.348121321865097, "grad_norm": 0.038560718297958374, "learning_rate": 6.934879554626625e-07, "loss": 0.0822, "step": 206500 }, { "epoch": 9.370755998189226, "grad_norm": 92.30923461914062, "learning_rate": 6.694085125646534e-07, "loss": 0.0854, "step": 207000 }, { "epoch": 9.393390674513354, "grad_norm": 0.2284342646598816, "learning_rate": 6.453290696666442e-07, "loss": 0.0807, "step": 207500 }, { "epoch": 9.416025350837483, "grad_norm": 0.8369685411453247, "learning_rate": 6.212496267686352e-07, "loss": 0.0744, "step": 208000 }, { "epoch": 9.438660027161612, "grad_norm": 0.49937498569488525, "learning_rate": 5.97170183870626e-07, "loss": 0.0942, "step": 208500 }, { "epoch": 9.46129470348574, "grad_norm": 0.020057352259755135, "learning_rate": 5.73090740972617e-07, "loss": 0.0748, "step": 209000 }, { "epoch": 9.483929379809869, "grad_norm": 196.6532745361328, "learning_rate": 5.490112980746078e-07, "loss": 0.0964, "step": 209500 }, { "epoch": 9.506564056133998, "grad_norm": 0.5909414887428284, "learning_rate": 5.249318551765987e-07, "loss": 0.0759, "step": 210000 }, { "epoch": 9.529198732458125, "grad_norm": 135.35101318359375, "learning_rate": 5.008524122785896e-07, "loss": 0.0804, "step": 210500 }, { "epoch": 9.551833408782255, "grad_norm": 0.0622185617685318, "learning_rate": 4.7677296938058045e-07, "loss": 0.0709, "step": 211000 }, { "epoch": 9.574468085106384, "grad_norm": 0.05341747775673866, "learning_rate": 4.526935264825713e-07, "loss": 0.0863, "step": 211500 }, { "epoch": 9.597102761430511, "grad_norm": 55.312278747558594, "learning_rate": 4.286140835845622e-07, "loss": 0.0782, "step": 212000 }, { "epoch": 9.61973743775464, "grad_norm": 0.06182483211159706, "learning_rate": 4.0453464068655306e-07, "loss": 0.0881, "step": 212500 }, { "epoch": 9.64237211407877, "grad_norm": 0.06101556122303009, "learning_rate": 3.80455197788544e-07, "loss": 0.0834, "step": 213000 }, { "epoch": 9.665006790402897, "grad_norm": 0.35538122057914734, "learning_rate": 3.5637575489053483e-07, "loss": 0.0776, "step": 213500 }, { "epoch": 9.687641466727026, "grad_norm": 0.08858389407396317, "learning_rate": 3.322963119925258e-07, "loss": 0.0832, "step": 214000 }, { "epoch": 9.710276143051153, "grad_norm": 0.07368449866771698, "learning_rate": 3.0821686909451666e-07, "loss": 0.0921, "step": 214500 }, { "epoch": 9.732910819375283, "grad_norm": 0.4127441644668579, "learning_rate": 2.8413742619650755e-07, "loss": 0.0846, "step": 215000 }, { "epoch": 9.755545495699412, "grad_norm": 0.06469714641571045, "learning_rate": 2.6005798329849844e-07, "loss": 0.0928, "step": 215500 }, { "epoch": 9.77818017202354, "grad_norm": 0.28226494789123535, "learning_rate": 2.3597854040048932e-07, "loss": 0.0796, "step": 216000 }, { "epoch": 9.800814848347668, "grad_norm": 250.34405517578125, "learning_rate": 2.118990975024802e-07, "loss": 0.0859, "step": 216500 }, { "epoch": 9.823449524671798, "grad_norm": 22.07843589782715, "learning_rate": 1.878196546044711e-07, "loss": 0.0846, "step": 217000 }, { "epoch": 9.846084200995925, "grad_norm": 14.63901138305664, "learning_rate": 1.6374021170646199e-07, "loss": 0.0788, "step": 217500 }, { "epoch": 9.868718877320054, "grad_norm": 80.16069793701172, "learning_rate": 1.3966076880845285e-07, "loss": 0.095, "step": 218000 }, { "epoch": 9.891353553644183, "grad_norm": 0.07123162597417831, "learning_rate": 1.1558132591044375e-07, "loss": 0.0935, "step": 218500 }, { "epoch": 9.91398822996831, "grad_norm": 17.373336791992188, "learning_rate": 9.150188301243464e-08, "loss": 0.0839, "step": 219000 }, { "epoch": 9.93662290629244, "grad_norm": 0.0160963274538517, "learning_rate": 6.742244011442552e-08, "loss": 0.0746, "step": 219500 }, { "epoch": 9.95925758261657, "grad_norm": 0.07339876890182495, "learning_rate": 4.3342997216416404e-08, "loss": 0.0755, "step": 220000 }, { "epoch": 9.981892258940697, "grad_norm": 0.06764261424541473, "learning_rate": 1.926355431840729e-08, "loss": 0.0835, "step": 220500 }, { "epoch": 10.0, "eval_accuracy": 0.8659570675562119, "eval_loss": 0.9559618830680847, "eval_runtime": 26.0025, "eval_samples_per_second": 1510.279, "eval_steps_per_second": 94.414, "step": 220900 }, { "epoch": 10.0, "step": 220900, "total_flos": 1.4387669402147813e+17, "train_loss": 0.2225576283846467, "train_runtime": 11995.4266, "train_samples_per_second": 294.638, "train_steps_per_second": 18.415 } ], "logging_steps": 500, "max_steps": 220900, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.4387669402147813e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }