| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.9999002576676919, |
| "eval_steps": 10000, |
| "global_step": 30076, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.006649488820546921, |
| "grad_norm": 0.2602408528327942, |
| "learning_rate": 3.324468085106383e-06, |
| "loss": 0.818, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.013298977641093841, |
| "grad_norm": 0.14098776876926422, |
| "learning_rate": 6.648936170212766e-06, |
| "loss": 0.5412, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.019948466461640762, |
| "grad_norm": 0.4028046429157257, |
| "learning_rate": 9.973404255319148e-06, |
| "loss": 0.5292, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.026597955282187682, |
| "grad_norm": 0.15940147638320923, |
| "learning_rate": 1.3297872340425532e-05, |
| "loss": 0.5203, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.0332474441027346, |
| "grad_norm": 0.32640570402145386, |
| "learning_rate": 1.6622340425531915e-05, |
| "loss": 0.5078, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.039896932923281524, |
| "grad_norm": 0.24910107254981995, |
| "learning_rate": 1.9946808510638297e-05, |
| "loss": 0.4875, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.04654642174382844, |
| "grad_norm": 0.43389075994491577, |
| "learning_rate": 2.3271276595744682e-05, |
| "loss": 0.4978, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.053195910564375365, |
| "grad_norm": 0.2782938778400421, |
| "learning_rate": 2.6595744680851064e-05, |
| "loss": 0.4807, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.05984539938492228, |
| "grad_norm": 0.3213005065917969, |
| "learning_rate": 2.9920212765957452e-05, |
| "loss": 0.5007, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.0664948882054692, |
| "grad_norm": 0.7565635442733765, |
| "learning_rate": 3.324468085106383e-05, |
| "loss": 0.4754, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.07314437702601613, |
| "grad_norm": 0.5631841421127319, |
| "learning_rate": 3.6569148936170215e-05, |
| "loss": 0.4943, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.07979386584656305, |
| "grad_norm": 0.40041548013687134, |
| "learning_rate": 3.9893617021276594e-05, |
| "loss": 0.4928, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.08644335466710996, |
| "grad_norm": 0.43981432914733887, |
| "learning_rate": 4.3218085106382986e-05, |
| "loss": 0.4829, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.09309284348765688, |
| "grad_norm": 0.49957290291786194, |
| "learning_rate": 4.6542553191489364e-05, |
| "loss": 0.4797, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.09974233230820381, |
| "grad_norm": 1.0348299741744995, |
| "learning_rate": 4.986702127659575e-05, |
| "loss": 0.5052, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.10639182112875073, |
| "grad_norm": 0.7030391097068787, |
| "learning_rate": 5.319148936170213e-05, |
| "loss": 0.4693, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.11304130994929765, |
| "grad_norm": 0.7447443008422852, |
| "learning_rate": 5.651595744680851e-05, |
| "loss": 0.4767, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.11969079876984456, |
| "grad_norm": 0.5579438209533691, |
| "learning_rate": 5.9840425531914904e-05, |
| "loss": 0.4823, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.12634028759039148, |
| "grad_norm": 0.5548936724662781, |
| "learning_rate": 6.316489361702128e-05, |
| "loss": 0.4631, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.1329897764109384, |
| "grad_norm": 0.9034917950630188, |
| "learning_rate": 6.648936170212766e-05, |
| "loss": 0.5277, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.13963926523148534, |
| "grad_norm": 0.7092494964599609, |
| "learning_rate": 6.981382978723405e-05, |
| "loss": 0.4813, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.14628875405203226, |
| "grad_norm": 1.0150775909423828, |
| "learning_rate": 7.313829787234043e-05, |
| "loss": 0.5026, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.15293824287257918, |
| "grad_norm": 1.1101422309875488, |
| "learning_rate": 7.646276595744682e-05, |
| "loss": 0.5226, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.1595877316931261, |
| "grad_norm": 0.5893256664276123, |
| "learning_rate": 7.978723404255319e-05, |
| "loss": 0.5285, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.166237220513673, |
| "grad_norm": 0.9117974042892456, |
| "learning_rate": 8.311170212765957e-05, |
| "loss": 0.5183, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.17288670933421993, |
| "grad_norm": 1.3572144508361816, |
| "learning_rate": 8.643617021276597e-05, |
| "loss": 0.5129, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.17953619815476685, |
| "grad_norm": 0.890178382396698, |
| "learning_rate": 8.976063829787234e-05, |
| "loss": 0.5081, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.18618568697531376, |
| "grad_norm": 2.725203275680542, |
| "learning_rate": 9.308510638297873e-05, |
| "loss": 0.5042, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.19283517579586068, |
| "grad_norm": 0.962457001209259, |
| "learning_rate": 9.640957446808511e-05, |
| "loss": 0.5357, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.19948466461640763, |
| "grad_norm": 2.529735803604126, |
| "learning_rate": 9.97340425531915e-05, |
| "loss": 0.5407, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.20613415343695454, |
| "grad_norm": 1.643441081047058, |
| "learning_rate": 9.999714964504067e-05, |
| "loss": 0.5244, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.21278364225750146, |
| "grad_norm": 0.9840033054351807, |
| "learning_rate": 9.998758599534463e-05, |
| "loss": 0.5562, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.21943313107804838, |
| "grad_norm": 1.0612396001815796, |
| "learning_rate": 9.997128878619808e-05, |
| "loss": 0.5273, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.2260826198985953, |
| "grad_norm": 1.1562029123306274, |
| "learning_rate": 9.994826021291056e-05, |
| "loss": 0.5412, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.2327321087191422, |
| "grad_norm": 1.5558404922485352, |
| "learning_rate": 9.991850337753762e-05, |
| "loss": 0.5405, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.23938159753968913, |
| "grad_norm": 2.0897717475891113, |
| "learning_rate": 9.988202228846291e-05, |
| "loss": 0.5049, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.24603108636023605, |
| "grad_norm": 1.031879186630249, |
| "learning_rate": 9.983882185985808e-05, |
| "loss": 0.5785, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.25268057518078296, |
| "grad_norm": 1.3407084941864014, |
| "learning_rate": 9.978890791102109e-05, |
| "loss": 0.5487, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.2593300640013299, |
| "grad_norm": 1.261853814125061, |
| "learning_rate": 9.973228716559209e-05, |
| "loss": 0.533, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.2659795528218768, |
| "grad_norm": 1.1546040773391724, |
| "learning_rate": 9.966896725064786e-05, |
| "loss": 0.5166, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.27262904164242374, |
| "grad_norm": 1.1395092010498047, |
| "learning_rate": 9.959895669567435e-05, |
| "loss": 0.5366, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.2792785304629707, |
| "grad_norm": 1.5521488189697266, |
| "learning_rate": 9.952226493141765e-05, |
| "loss": 0.5339, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.2859280192835176, |
| "grad_norm": 1.9154760837554932, |
| "learning_rate": 9.943890228861383e-05, |
| "loss": 0.4976, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.2925775081040645, |
| "grad_norm": 1.273290753364563, |
| "learning_rate": 9.934887999659707e-05, |
| "loss": 0.5331, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.2992269969246114, |
| "grad_norm": 1.6757972240447998, |
| "learning_rate": 9.925221018178728e-05, |
| "loss": 0.5212, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.30587648574515836, |
| "grad_norm": 1.3594906330108643, |
| "learning_rate": 9.914890586605638e-05, |
| "loss": 0.4991, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.31252597456570524, |
| "grad_norm": 1.622048258781433, |
| "learning_rate": 9.903898096497441e-05, |
| "loss": 0.569, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.3191754633862522, |
| "grad_norm": 1.6205965280532837, |
| "learning_rate": 9.892245028593483e-05, |
| "loss": 0.5436, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.3258249522067991, |
| "grad_norm": 2.7058024406433105, |
| "learning_rate": 9.879932952616009e-05, |
| "loss": 0.5028, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.332474441027346, |
| "grad_norm": 1.8897464275360107, |
| "learning_rate": 9.8669635270587e-05, |
| "loss": 0.7608, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.33912392984789297, |
| "grad_norm": 0.9735029935836792, |
| "learning_rate": 9.853338498963272e-05, |
| "loss": 0.523, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.34577341866843986, |
| "grad_norm": 0.9950180053710938, |
| "learning_rate": 9.839059703684139e-05, |
| "loss": 0.5426, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.3524229074889868, |
| "grad_norm": 0.7262745499610901, |
| "learning_rate": 9.82412906464119e-05, |
| "loss": 0.5387, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.3590723963095337, |
| "grad_norm": 1.133631706237793, |
| "learning_rate": 9.808548593060681e-05, |
| "loss": 0.5259, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.36572188513008064, |
| "grad_norm": 1.8873803615570068, |
| "learning_rate": 9.792320387704328e-05, |
| "loss": 0.5363, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.3723713739506275, |
| "grad_norm": 2.6183230876922607, |
| "learning_rate": 9.775446634586584e-05, |
| "loss": 0.542, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.37902086277117447, |
| "grad_norm": 1.7073196172714233, |
| "learning_rate": 9.757929606680181e-05, |
| "loss": 0.4947, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.38567035159172136, |
| "grad_norm": 1.0039314031600952, |
| "learning_rate": 9.73977166360994e-05, |
| "loss": 0.5584, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.3923198404122683, |
| "grad_norm": 1.0683766603469849, |
| "learning_rate": 9.720975251334929e-05, |
| "loss": 0.5318, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.39896932923281525, |
| "grad_norm": 1.4205585718154907, |
| "learning_rate": 9.701542901818974e-05, |
| "loss": 0.5352, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.40561881805336214, |
| "grad_norm": 2.596090316772461, |
| "learning_rate": 9.681477232689596e-05, |
| "loss": 0.5229, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.4122683068739091, |
| "grad_norm": 1.1745169162750244, |
| "learning_rate": 9.660780946885397e-05, |
| "loss": 0.5419, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.418917795694456, |
| "grad_norm": 2.7921059131622314, |
| "learning_rate": 9.639456832291974e-05, |
| "loss": 0.5428, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.4255672845150029, |
| "grad_norm": 3.064069986343384, |
| "learning_rate": 9.617507761366367e-05, |
| "loss": 0.5256, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.4322167733355498, |
| "grad_norm": 1.322499394416809, |
| "learning_rate": 9.594936690750129e-05, |
| "loss": 0.5633, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.43886626215609675, |
| "grad_norm": 1.6360127925872803, |
| "learning_rate": 9.571746660871058e-05, |
| "loss": 0.5269, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.44551575097664364, |
| "grad_norm": 1.442179799079895, |
| "learning_rate": 9.547940795533627e-05, |
| "loss": 0.5105, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.4521652397971906, |
| "grad_norm": 1.1154464483261108, |
| "learning_rate": 9.523522301498202e-05, |
| "loss": 0.5301, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.45881472861773753, |
| "grad_norm": 1.2995010614395142, |
| "learning_rate": 9.498494468049072e-05, |
| "loss": 0.5173, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.4654642174382844, |
| "grad_norm": 1.0487501621246338, |
| "learning_rate": 9.472860666551369e-05, |
| "loss": 0.5284, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.47211370625883137, |
| "grad_norm": 1.8207134008407593, |
| "learning_rate": 9.446624349996929e-05, |
| "loss": 0.5317, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.47876319507937826, |
| "grad_norm": 1.7722837924957275, |
| "learning_rate": 9.419789052539157e-05, |
| "loss": 0.5195, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.4854126838999252, |
| "grad_norm": 2.749199390411377, |
| "learning_rate": 9.392358389016961e-05, |
| "loss": 0.5274, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.4920621727204721, |
| "grad_norm": 1.3200327157974243, |
| "learning_rate": 9.364336054467819e-05, |
| "loss": 0.523, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.49871166154101904, |
| "grad_norm": 2.9674599170684814, |
| "learning_rate": 9.335725823630035e-05, |
| "loss": 0.5294, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.5053611503615659, |
| "grad_norm": 2.8731188774108887, |
| "learning_rate": 9.306531550434268e-05, |
| "loss": 0.5087, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.5120106391821129, |
| "grad_norm": 2.215608596801758, |
| "learning_rate": 9.276757167484389e-05, |
| "loss": 0.5219, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.5186601280026598, |
| "grad_norm": 1.5501729249954224, |
| "learning_rate": 9.246406685527739e-05, |
| "loss": 0.5582, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.5253096168232068, |
| "grad_norm": 1.8816163539886475, |
| "learning_rate": 9.21548419291487e-05, |
| "loss": 0.5394, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.5319591056437536, |
| "grad_norm": 1.338460087776184, |
| "learning_rate": 9.183993855048811e-05, |
| "loss": 0.5299, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.5386085944643005, |
| "grad_norm": 1.7432674169540405, |
| "learning_rate": 9.151939913823988e-05, |
| "loss": 0.5307, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.5452580832848475, |
| "grad_norm": 3.2521355152130127, |
| "learning_rate": 9.119326687054802e-05, |
| "loss": 0.5004, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.5519075721053944, |
| "grad_norm": 0.9576961994171143, |
| "learning_rate": 9.086158567894013e-05, |
| "loss": 0.5453, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.5585570609259414, |
| "grad_norm": 1.3667104244232178, |
| "learning_rate": 9.052440024240956e-05, |
| "loss": 0.5003, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.5652065497464882, |
| "grad_norm": 1.9807591438293457, |
| "learning_rate": 9.018175598139696e-05, |
| "loss": 0.508, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.5718560385670352, |
| "grad_norm": 2.281437397003174, |
| "learning_rate": 8.983369905167191e-05, |
| "loss": 0.4788, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.5785055273875821, |
| "grad_norm": 2.3274779319763184, |
| "learning_rate": 8.948027633811557e-05, |
| "loss": 0.4994, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.585155016208129, |
| "grad_norm": 3.7350873947143555, |
| "learning_rate": 8.912153544840507e-05, |
| "loss": 0.546, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.5918045050286759, |
| "grad_norm": 1.8497984409332275, |
| "learning_rate": 8.875752470660043e-05, |
| "loss": 0.4939, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.5984539938492228, |
| "grad_norm": 2.2370221614837646, |
| "learning_rate": 8.838829314663522e-05, |
| "loss": 0.5288, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.6051034826697698, |
| "grad_norm": 1.4384081363677979, |
| "learning_rate": 8.80138905057114e-05, |
| "loss": 0.5176, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.6117529714903167, |
| "grad_norm": 3.588320255279541, |
| "learning_rate": 8.763436721759952e-05, |
| "loss": 0.4898, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.6184024603108637, |
| "grad_norm": 2.6353256702423096, |
| "learning_rate": 8.724977440584497e-05, |
| "loss": 0.5011, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.6250519491314105, |
| "grad_norm": 1.7550867795944214, |
| "learning_rate": 8.686016387688153e-05, |
| "loss": 0.5101, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.6317014379519574, |
| "grad_norm": 1.9632377624511719, |
| "learning_rate": 8.646558811305268e-05, |
| "loss": 0.5023, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.6383509267725044, |
| "grad_norm": 2.103327989578247, |
| "learning_rate": 8.60661002655421e-05, |
| "loss": 0.5055, |
| "step": 9600 |
| }, |
| { |
| "epoch": 0.6450004155930513, |
| "grad_norm": 2.518740653991699, |
| "learning_rate": 8.566175414721384e-05, |
| "loss": 0.4997, |
| "step": 9700 |
| }, |
| { |
| "epoch": 0.6516499044135982, |
| "grad_norm": 2.6153292655944824, |
| "learning_rate": 8.525260422536358e-05, |
| "loss": 0.5104, |
| "step": 9800 |
| }, |
| { |
| "epoch": 0.6582993932341451, |
| "grad_norm": 1.0710101127624512, |
| "learning_rate": 8.483870561438161e-05, |
| "loss": 0.5281, |
| "step": 9900 |
| }, |
| { |
| "epoch": 0.664948882054692, |
| "grad_norm": 3.5133893489837646, |
| "learning_rate": 8.442011406832859e-05, |
| "loss": 0.5147, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.664948882054692, |
| "eval_loss": 0.4313640594482422, |
| "eval_runtime": 4340.2869, |
| "eval_samples_per_second": 1.54, |
| "eval_steps_per_second": 1.54, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.671598370875239, |
| "grad_norm": 1.416164755821228, |
| "learning_rate": 8.399688597342535e-05, |
| "loss": 0.5229, |
| "step": 10100 |
| }, |
| { |
| "epoch": 0.6782478596957859, |
| "grad_norm": 4.202746391296387, |
| "learning_rate": 8.356907834045726e-05, |
| "loss": 0.5008, |
| "step": 10200 |
| }, |
| { |
| "epoch": 0.6848973485163328, |
| "grad_norm": 1.6467583179473877, |
| "learning_rate": 8.313674879709475e-05, |
| "loss": 0.5084, |
| "step": 10300 |
| }, |
| { |
| "epoch": 0.6915468373368797, |
| "grad_norm": 0.9791653752326965, |
| "learning_rate": 8.269995558013049e-05, |
| "loss": 0.5049, |
| "step": 10400 |
| }, |
| { |
| "epoch": 0.6981963261574267, |
| "grad_norm": 1.2445042133331299, |
| "learning_rate": 8.225875752763468e-05, |
| "loss": 0.4711, |
| "step": 10500 |
| }, |
| { |
| "epoch": 0.7048458149779736, |
| "grad_norm": 3.4758522510528564, |
| "learning_rate": 8.181321407102929e-05, |
| "loss": 0.4835, |
| "step": 10600 |
| }, |
| { |
| "epoch": 0.7114953037985204, |
| "grad_norm": 1.5787672996520996, |
| "learning_rate": 8.136338522708233e-05, |
| "loss": 0.4849, |
| "step": 10700 |
| }, |
| { |
| "epoch": 0.7181447926190674, |
| "grad_norm": 2.1171529293060303, |
| "learning_rate": 8.090933158982338e-05, |
| "loss": 0.4881, |
| "step": 10800 |
| }, |
| { |
| "epoch": 0.7247942814396143, |
| "grad_norm": 3.9836864471435547, |
| "learning_rate": 8.045111432238121e-05, |
| "loss": 0.4699, |
| "step": 10900 |
| }, |
| { |
| "epoch": 0.7314437702601613, |
| "grad_norm": 3.3204433917999268, |
| "learning_rate": 7.998879514874491e-05, |
| "loss": 0.5022, |
| "step": 11000 |
| }, |
| { |
| "epoch": 0.7380932590807082, |
| "grad_norm": 1.5986028909683228, |
| "learning_rate": 7.952243634544929e-05, |
| "loss": 0.4933, |
| "step": 11100 |
| }, |
| { |
| "epoch": 0.744742747901255, |
| "grad_norm": 3.2928895950317383, |
| "learning_rate": 7.905210073318605e-05, |
| "loss": 0.4677, |
| "step": 11200 |
| }, |
| { |
| "epoch": 0.751392236721802, |
| "grad_norm": 2.922384738922119, |
| "learning_rate": 7.857785166834144e-05, |
| "loss": 0.4757, |
| "step": 11300 |
| }, |
| { |
| "epoch": 0.7580417255423489, |
| "grad_norm": 1.8758389949798584, |
| "learning_rate": 7.809975303446195e-05, |
| "loss": 0.4938, |
| "step": 11400 |
| }, |
| { |
| "epoch": 0.7646912143628959, |
| "grad_norm": 2.5382237434387207, |
| "learning_rate": 7.761786923364878e-05, |
| "loss": 0.4733, |
| "step": 11500 |
| }, |
| { |
| "epoch": 0.7713407031834427, |
| "grad_norm": 3.0387401580810547, |
| "learning_rate": 7.713226517788275e-05, |
| "loss": 0.4655, |
| "step": 11600 |
| }, |
| { |
| "epoch": 0.7779901920039897, |
| "grad_norm": 3.491082191467285, |
| "learning_rate": 7.664300628028017e-05, |
| "loss": 0.4933, |
| "step": 11700 |
| }, |
| { |
| "epoch": 0.7846396808245366, |
| "grad_norm": 2.891706943511963, |
| "learning_rate": 7.615015844628157e-05, |
| "loss": 0.4699, |
| "step": 11800 |
| }, |
| { |
| "epoch": 0.7912891696450836, |
| "grad_norm": 1.315558671951294, |
| "learning_rate": 7.565378806477377e-05, |
| "loss": 0.4514, |
| "step": 11900 |
| }, |
| { |
| "epoch": 0.7979386584656305, |
| "grad_norm": 5.145135402679443, |
| "learning_rate": 7.515396199914708e-05, |
| "loss": 0.4532, |
| "step": 12000 |
| }, |
| { |
| "epoch": 0.8045881472861773, |
| "grad_norm": 2.2307186126708984, |
| "learning_rate": 7.46507475782885e-05, |
| "loss": 0.4749, |
| "step": 12100 |
| }, |
| { |
| "epoch": 0.8112376361067243, |
| "grad_norm": 1.4436883926391602, |
| "learning_rate": 7.414421258751212e-05, |
| "loss": 0.4977, |
| "step": 12200 |
| }, |
| { |
| "epoch": 0.8178871249272712, |
| "grad_norm": 2.0418126583099365, |
| "learning_rate": 7.363442525942826e-05, |
| "loss": 0.4544, |
| "step": 12300 |
| }, |
| { |
| "epoch": 0.8245366137478182, |
| "grad_norm": 1.2745349407196045, |
| "learning_rate": 7.31214542647521e-05, |
| "loss": 0.5053, |
| "step": 12400 |
| }, |
| { |
| "epoch": 0.831186102568365, |
| "grad_norm": 3.4680609703063965, |
| "learning_rate": 7.260536870305347e-05, |
| "loss": 0.4648, |
| "step": 12500 |
| }, |
| { |
| "epoch": 0.837835591388912, |
| "grad_norm": 0.97030109167099, |
| "learning_rate": 7.208623809344879e-05, |
| "loss": 0.4313, |
| "step": 12600 |
| }, |
| { |
| "epoch": 0.8444850802094589, |
| "grad_norm": 3.7414844036102295, |
| "learning_rate": 7.156413236523656e-05, |
| "loss": 0.4816, |
| "step": 12700 |
| }, |
| { |
| "epoch": 0.8511345690300058, |
| "grad_norm": 1.8031582832336426, |
| "learning_rate": 7.103912184847757e-05, |
| "loss": 0.4534, |
| "step": 12800 |
| }, |
| { |
| "epoch": 0.8577840578505528, |
| "grad_norm": 1.6116315126419067, |
| "learning_rate": 7.051127726452102e-05, |
| "loss": 0.4304, |
| "step": 12900 |
| }, |
| { |
| "epoch": 0.8644335466710996, |
| "grad_norm": 2.230163335800171, |
| "learning_rate": 6.998066971647817e-05, |
| "loss": 0.4511, |
| "step": 13000 |
| }, |
| { |
| "epoch": 0.8710830354916466, |
| "grad_norm": 1.854151964187622, |
| "learning_rate": 6.944737067964429e-05, |
| "loss": 0.4412, |
| "step": 13100 |
| }, |
| { |
| "epoch": 0.8777325243121935, |
| "grad_norm": 2.757798194885254, |
| "learning_rate": 6.891145199187065e-05, |
| "loss": 0.4455, |
| "step": 13200 |
| }, |
| { |
| "epoch": 0.8843820131327405, |
| "grad_norm": 2.430884838104248, |
| "learning_rate": 6.837298584388771e-05, |
| "loss": 0.4598, |
| "step": 13300 |
| }, |
| { |
| "epoch": 0.8910315019532873, |
| "grad_norm": 1.6644638776779175, |
| "learning_rate": 6.783204476958058e-05, |
| "loss": 0.4561, |
| "step": 13400 |
| }, |
| { |
| "epoch": 0.8976809907738342, |
| "grad_norm": 2.7078986167907715, |
| "learning_rate": 6.728870163621836e-05, |
| "loss": 0.4818, |
| "step": 13500 |
| }, |
| { |
| "epoch": 0.9043304795943812, |
| "grad_norm": 1.8221591711044312, |
| "learning_rate": 6.674302963463876e-05, |
| "loss": 0.4322, |
| "step": 13600 |
| }, |
| { |
| "epoch": 0.9109799684149281, |
| "grad_norm": 1.0504103899002075, |
| "learning_rate": 6.61951022693887e-05, |
| "loss": 0.4601, |
| "step": 13700 |
| }, |
| { |
| "epoch": 0.9176294572354751, |
| "grad_norm": 1.788719892501831, |
| "learning_rate": 6.564499334882312e-05, |
| "loss": 0.4787, |
| "step": 13800 |
| }, |
| { |
| "epoch": 0.9242789460560219, |
| "grad_norm": 1.5331158638000488, |
| "learning_rate": 6.509277697516255e-05, |
| "loss": 0.4766, |
| "step": 13900 |
| }, |
| { |
| "epoch": 0.9309284348765688, |
| "grad_norm": 2.4608702659606934, |
| "learning_rate": 6.453852753451119e-05, |
| "loss": 0.4614, |
| "step": 14000 |
| }, |
| { |
| "epoch": 0.9375779236971158, |
| "grad_norm": 1.432626485824585, |
| "learning_rate": 6.39823196868368e-05, |
| "loss": 0.48, |
| "step": 14100 |
| }, |
| { |
| "epoch": 0.9442274125176627, |
| "grad_norm": 3.4951376914978027, |
| "learning_rate": 6.342422835591368e-05, |
| "loss": 0.4717, |
| "step": 14200 |
| }, |
| { |
| "epoch": 0.9508769013382097, |
| "grad_norm": 1.7661138772964478, |
| "learning_rate": 6.286432871923e-05, |
| "loss": 0.4609, |
| "step": 14300 |
| }, |
| { |
| "epoch": 0.9575263901587565, |
| "grad_norm": 2.387974500656128, |
| "learning_rate": 6.230269619786111e-05, |
| "loss": 0.4233, |
| "step": 14400 |
| }, |
| { |
| "epoch": 0.9641758789793035, |
| "grad_norm": 1.5833674669265747, |
| "learning_rate": 6.173940644630996e-05, |
| "loss": 0.4368, |
| "step": 14500 |
| }, |
| { |
| "epoch": 0.9708253677998504, |
| "grad_norm": 2.6246190071105957, |
| "learning_rate": 6.117453534231606e-05, |
| "loss": 0.4205, |
| "step": 14600 |
| }, |
| { |
| "epoch": 0.9774748566203973, |
| "grad_norm": 1.6440967321395874, |
| "learning_rate": 6.060815897663447e-05, |
| "loss": 0.4247, |
| "step": 14700 |
| }, |
| { |
| "epoch": 0.9841243454409442, |
| "grad_norm": 2.0472023487091064, |
| "learning_rate": 6.004035364278593e-05, |
| "loss": 0.4338, |
| "step": 14800 |
| }, |
| { |
| "epoch": 0.9907738342614911, |
| "grad_norm": 3.3487770557403564, |
| "learning_rate": 5.9471195826779834e-05, |
| "loss": 0.4114, |
| "step": 14900 |
| }, |
| { |
| "epoch": 0.9974233230820381, |
| "grad_norm": 2.1973326206207275, |
| "learning_rate": 5.8900762196811175e-05, |
| "loss": 0.4209, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.004072811902585, |
| "grad_norm": 2.8700783252716064, |
| "learning_rate": 5.832912959293304e-05, |
| "loss": 0.3562, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.0107223007231319, |
| "grad_norm": 0.8708151578903198, |
| "learning_rate": 5.775637501670579e-05, |
| "loss": 0.3388, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.017371789543679, |
| "grad_norm": 0.9684904217720032, |
| "learning_rate": 5.718257562082471e-05, |
| "loss": 0.3572, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.0240212783642257, |
| "grad_norm": 1.5389324426651, |
| "learning_rate": 5.660780869872711e-05, |
| "loss": 0.3377, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.0306707671847726, |
| "grad_norm": 1.646532416343689, |
| "learning_rate": 5.6032151674180575e-05, |
| "loss": 0.3561, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.0373202560053196, |
| "grad_norm": 2.062817096710205, |
| "learning_rate": 5.5455682090853624e-05, |
| "loss": 0.3698, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.0439697448258665, |
| "grad_norm": 2.0568647384643555, |
| "learning_rate": 5.4878477601870194e-05, |
| "loss": 0.3223, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.0506192336464135, |
| "grad_norm": 2.101365566253662, |
| "learning_rate": 5.430061595934941e-05, |
| "loss": 0.3483, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.0572687224669604, |
| "grad_norm": 1.6456376314163208, |
| "learning_rate": 5.372217500393205e-05, |
| "loss": 0.3642, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.0639182112875072, |
| "grad_norm": 2.0329535007476807, |
| "learning_rate": 5.314323265429501e-05, |
| "loss": 0.3234, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.0705677001080542, |
| "grad_norm": 2.085597276687622, |
| "learning_rate": 5.2563866896655275e-05, |
| "loss": 0.3448, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.077217188928601, |
| "grad_norm": 2.2339513301849365, |
| "learning_rate": 5.198415577426493e-05, |
| "loss": 0.3621, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.0838666777491481, |
| "grad_norm": 2.7302517890930176, |
| "learning_rate": 5.140417737689822e-05, |
| "loss": 0.3532, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.090516166569695, |
| "grad_norm": 2.6171867847442627, |
| "learning_rate": 5.0824009830332606e-05, |
| "loss": 0.3424, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.0971656553902418, |
| "grad_norm": 1.7419236898422241, |
| "learning_rate": 5.02437312858248e-05, |
| "loss": 0.3465, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.1038151442107889, |
| "grad_norm": 1.149902582168579, |
| "learning_rate": 4.966341990958347e-05, |
| "loss": 0.3471, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.1104646330313357, |
| "grad_norm": 2.0462777614593506, |
| "learning_rate": 4.908315387223985e-05, |
| "loss": 0.3408, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.1171141218518827, |
| "grad_norm": 1.4826865196228027, |
| "learning_rate": 4.850301133831786e-05, |
| "loss": 0.3351, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.1237636106724296, |
| "grad_norm": 1.297524094581604, |
| "learning_rate": 4.792307045570486e-05, |
| "loss": 0.3531, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.1304130994929764, |
| "grad_norm": 0.9654291272163391, |
| "learning_rate": 4.734340934512492e-05, |
| "loss": 0.3171, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.1370625883135235, |
| "grad_norm": 1.5802520513534546, |
| "learning_rate": 4.6764106089615454e-05, |
| "loss": 0.3388, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.1437120771340703, |
| "grad_norm": 3.1234793663024902, |
| "learning_rate": 4.618523872400921e-05, |
| "loss": 0.3356, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.1503615659546171, |
| "grad_norm": 4.1535444259643555, |
| "learning_rate": 4.56068852244225e-05, |
| "loss": 0.3483, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.1570110547751642, |
| "grad_norm": 2.166757583618164, |
| "learning_rate": 4.5029123497751514e-05, |
| "loss": 0.3223, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.163660543595711, |
| "grad_norm": 1.9507828950881958, |
| "learning_rate": 4.445203137117788e-05, |
| "loss": 0.3181, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.170310032416258, |
| "grad_norm": 2.319566011428833, |
| "learning_rate": 4.3875686581685e-05, |
| "loss": 0.3416, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.176959521236805, |
| "grad_norm": 2.692155599594116, |
| "learning_rate": 4.330016676558651e-05, |
| "loss": 0.3269, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.1836090100573518, |
| "grad_norm": 1.3827095031738281, |
| "learning_rate": 4.272554944806831e-05, |
| "loss": 0.3166, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.1902584988778988, |
| "grad_norm": 1.681174635887146, |
| "learning_rate": 4.2151912032745547e-05, |
| "loss": 0.3354, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.1969079876984456, |
| "grad_norm": 1.9821159839630127, |
| "learning_rate": 4.157933179123599e-05, |
| "loss": 0.3066, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.2035574765189927, |
| "grad_norm": 1.9395629167556763, |
| "learning_rate": 4.100788585275125e-05, |
| "loss": 0.3427, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.2102069653395395, |
| "grad_norm": 1.5355767011642456, |
| "learning_rate": 4.043765119370699e-05, |
| "loss": 0.3074, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.2168564541600864, |
| "grad_norm": 2.809375047683716, |
| "learning_rate": 3.9868704627354e-05, |
| "loss": 0.3172, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.2235059429806334, |
| "grad_norm": 2.0123348236083984, |
| "learning_rate": 3.930112279343094e-05, |
| "loss": 0.3088, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.2301554318011803, |
| "grad_norm": 3.793424367904663, |
| "learning_rate": 3.8734982147840756e-05, |
| "loss": 0.3053, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.236804920621727, |
| "grad_norm": 3.0886518955230713, |
| "learning_rate": 3.817035895235159e-05, |
| "loss": 0.3178, |
| "step": 18600 |
| }, |
| { |
| "epoch": 1.2434544094422741, |
| "grad_norm": 3.0276455879211426, |
| "learning_rate": 3.760732926432407e-05, |
| "loss": 0.2876, |
| "step": 18700 |
| }, |
| { |
| "epoch": 1.250103898262821, |
| "grad_norm": 2.8517942428588867, |
| "learning_rate": 3.704596892646593e-05, |
| "loss": 0.3264, |
| "step": 18800 |
| }, |
| { |
| "epoch": 1.256753387083368, |
| "grad_norm": 3.9238364696502686, |
| "learning_rate": 3.648635355661577e-05, |
| "loss": 0.3151, |
| "step": 18900 |
| }, |
| { |
| "epoch": 1.2634028759039149, |
| "grad_norm": 2.554077625274658, |
| "learning_rate": 3.5928558537556895e-05, |
| "loss": 0.3086, |
| "step": 19000 |
| }, |
| { |
| "epoch": 1.270052364724462, |
| "grad_norm": 1.959659457206726, |
| "learning_rate": 3.537265900686286e-05, |
| "loss": 0.2933, |
| "step": 19100 |
| }, |
| { |
| "epoch": 1.2767018535450088, |
| "grad_norm": 2.120922327041626, |
| "learning_rate": 3.4818729846776254e-05, |
| "loss": 0.2867, |
| "step": 19200 |
| }, |
| { |
| "epoch": 1.2833513423655556, |
| "grad_norm": 2.423426389694214, |
| "learning_rate": 3.426684567412153e-05, |
| "loss": 0.2956, |
| "step": 19300 |
| }, |
| { |
| "epoch": 1.2900008311861026, |
| "grad_norm": 2.462615489959717, |
| "learning_rate": 3.371708083025392e-05, |
| "loss": 0.2944, |
| "step": 19400 |
| }, |
| { |
| "epoch": 1.2966503200066495, |
| "grad_norm": 1.9235759973526, |
| "learning_rate": 3.316950937104518e-05, |
| "loss": 0.2868, |
| "step": 19500 |
| }, |
| { |
| "epoch": 1.3032998088271963, |
| "grad_norm": 3.4119069576263428, |
| "learning_rate": 3.26242050569081e-05, |
| "loss": 0.306, |
| "step": 19600 |
| }, |
| { |
| "epoch": 1.3099492976477434, |
| "grad_norm": 5.942444801330566, |
| "learning_rate": 3.208124134286038e-05, |
| "loss": 0.294, |
| "step": 19700 |
| }, |
| { |
| "epoch": 1.3165987864682902, |
| "grad_norm": 1.254133701324463, |
| "learning_rate": 3.1540691368630185e-05, |
| "loss": 0.3237, |
| "step": 19800 |
| }, |
| { |
| "epoch": 1.323248275288837, |
| "grad_norm": 3.514387369155884, |
| "learning_rate": 3.100262794880363e-05, |
| "loss": 0.3111, |
| "step": 19900 |
| }, |
| { |
| "epoch": 1.329897764109384, |
| "grad_norm": 7.018650054931641, |
| "learning_rate": 3.0467123563016513e-05, |
| "loss": 0.3106, |
| "step": 20000 |
| }, |
| { |
| "epoch": 1.329897764109384, |
| "eval_loss": 0.31190410256385803, |
| "eval_runtime": 4327.0173, |
| "eval_samples_per_second": 1.545, |
| "eval_steps_per_second": 1.545, |
| "step": 20000 |
| }, |
| { |
| "epoch": 1.336547252929931, |
| "grad_norm": 0.9992861747741699, |
| "learning_rate": 2.9934250346190818e-05, |
| "loss": 0.2899, |
| "step": 20100 |
| }, |
| { |
| "epoch": 1.343196741750478, |
| "grad_norm": 2.0007855892181396, |
| "learning_rate": 2.9404080078817924e-05, |
| "loss": 0.284, |
| "step": 20200 |
| }, |
| { |
| "epoch": 1.3498462305710248, |
| "grad_norm": 1.0622339248657227, |
| "learning_rate": 2.8876684177289404e-05, |
| "loss": 0.2886, |
| "step": 20300 |
| }, |
| { |
| "epoch": 1.3564957193915719, |
| "grad_norm": 1.351544737815857, |
| "learning_rate": 2.8352133684276853e-05, |
| "loss": 0.2738, |
| "step": 20400 |
| }, |
| { |
| "epoch": 1.3631452082121187, |
| "grad_norm": 1.9642298221588135, |
| "learning_rate": 2.7830499259162213e-05, |
| "loss": 0.2981, |
| "step": 20500 |
| }, |
| { |
| "epoch": 1.3697946970326655, |
| "grad_norm": 1.769395351409912, |
| "learning_rate": 2.7311851168519496e-05, |
| "loss": 0.3159, |
| "step": 20600 |
| }, |
| { |
| "epoch": 1.3764441858532126, |
| "grad_norm": 3.744534492492676, |
| "learning_rate": 2.6796259276649693e-05, |
| "loss": 0.3019, |
| "step": 20700 |
| }, |
| { |
| "epoch": 1.3830936746737594, |
| "grad_norm": 0.7485169768333435, |
| "learning_rate": 2.6283793036169603e-05, |
| "loss": 0.3029, |
| "step": 20800 |
| }, |
| { |
| "epoch": 1.3897431634943063, |
| "grad_norm": 1.2193599939346313, |
| "learning_rate": 2.5774521478656343e-05, |
| "loss": 0.2799, |
| "step": 20900 |
| }, |
| { |
| "epoch": 1.3963926523148533, |
| "grad_norm": 3.9265201091766357, |
| "learning_rate": 2.52685132053484e-05, |
| "loss": 0.2783, |
| "step": 21000 |
| }, |
| { |
| "epoch": 1.4030421411354002, |
| "grad_norm": 1.0576444864273071, |
| "learning_rate": 2.4765836377904787e-05, |
| "loss": 0.3043, |
| "step": 21100 |
| }, |
| { |
| "epoch": 1.4096916299559472, |
| "grad_norm": 1.704397201538086, |
| "learning_rate": 2.4266558709223293e-05, |
| "loss": 0.2823, |
| "step": 21200 |
| }, |
| { |
| "epoch": 1.416341118776494, |
| "grad_norm": 5.58804178237915, |
| "learning_rate": 2.377074745431931e-05, |
| "loss": 0.2899, |
| "step": 21300 |
| }, |
| { |
| "epoch": 1.422990607597041, |
| "grad_norm": 2.1619229316711426, |
| "learning_rate": 2.3278469401266178e-05, |
| "loss": 0.259, |
| "step": 21400 |
| }, |
| { |
| "epoch": 1.429640096417588, |
| "grad_norm": 3.6991875171661377, |
| "learning_rate": 2.2789790862198628e-05, |
| "loss": 0.2508, |
| "step": 21500 |
| }, |
| { |
| "epoch": 1.4362895852381348, |
| "grad_norm": 0.9216620922088623, |
| "learning_rate": 2.2304777664380176e-05, |
| "loss": 0.2701, |
| "step": 21600 |
| }, |
| { |
| "epoch": 1.4429390740586818, |
| "grad_norm": 1.0544836521148682, |
| "learning_rate": 2.182349514133583e-05, |
| "loss": 0.2741, |
| "step": 21700 |
| }, |
| { |
| "epoch": 1.4495885628792287, |
| "grad_norm": 1.3279638290405273, |
| "learning_rate": 2.134600812405151e-05, |
| "loss": 0.2608, |
| "step": 21800 |
| }, |
| { |
| "epoch": 1.4562380516997755, |
| "grad_norm": 2.69641375541687, |
| "learning_rate": 2.0872380932240832e-05, |
| "loss": 0.2527, |
| "step": 21900 |
| }, |
| { |
| "epoch": 1.4628875405203225, |
| "grad_norm": 1.3837428092956543, |
| "learning_rate": 2.0402677365681112e-05, |
| "loss": 0.2739, |
| "step": 22000 |
| }, |
| { |
| "epoch": 1.4695370293408694, |
| "grad_norm": 2.859133720397949, |
| "learning_rate": 1.99369606956191e-05, |
| "loss": 0.2823, |
| "step": 22100 |
| }, |
| { |
| "epoch": 1.4761865181614162, |
| "grad_norm": 2.0277671813964844, |
| "learning_rate": 1.9475293656248182e-05, |
| "loss": 0.2753, |
| "step": 22200 |
| }, |
| { |
| "epoch": 1.4828360069819633, |
| "grad_norm": 1.3588882684707642, |
| "learning_rate": 1.9017738436257655e-05, |
| "loss": 0.2806, |
| "step": 22300 |
| }, |
| { |
| "epoch": 1.48948549580251, |
| "grad_norm": 0.9804733991622925, |
| "learning_rate": 1.8564356670455767e-05, |
| "loss": 0.2749, |
| "step": 22400 |
| }, |
| { |
| "epoch": 1.4961349846230572, |
| "grad_norm": 1.8412333726882935, |
| "learning_rate": 1.8115209431467074e-05, |
| "loss": 0.265, |
| "step": 22500 |
| }, |
| { |
| "epoch": 1.502784473443604, |
| "grad_norm": 0.9384289979934692, |
| "learning_rate": 1.767035722150582e-05, |
| "loss": 0.264, |
| "step": 22600 |
| }, |
| { |
| "epoch": 1.509433962264151, |
| "grad_norm": 1.8639055490493774, |
| "learning_rate": 1.7229859964225868e-05, |
| "loss": 0.2587, |
| "step": 22700 |
| }, |
| { |
| "epoch": 1.5160834510846979, |
| "grad_norm": 1.7392961978912354, |
| "learning_rate": 1.679377699664884e-05, |
| "loss": 0.2638, |
| "step": 22800 |
| }, |
| { |
| "epoch": 1.5227329399052447, |
| "grad_norm": 2.0196003913879395, |
| "learning_rate": 1.6362167061171063e-05, |
| "loss": 0.2718, |
| "step": 22900 |
| }, |
| { |
| "epoch": 1.5293824287257918, |
| "grad_norm": 1.6492582559585571, |
| "learning_rate": 1.5935088297650674e-05, |
| "loss": 0.2553, |
| "step": 23000 |
| }, |
| { |
| "epoch": 1.5360319175463386, |
| "grad_norm": 1.1431901454925537, |
| "learning_rate": 1.551259823557602e-05, |
| "loss": 0.2569, |
| "step": 23100 |
| }, |
| { |
| "epoch": 1.5426814063668854, |
| "grad_norm": 1.3804975748062134, |
| "learning_rate": 1.509475378631603e-05, |
| "loss": 0.2474, |
| "step": 23200 |
| }, |
| { |
| "epoch": 1.5493308951874325, |
| "grad_norm": 1.6143451929092407, |
| "learning_rate": 1.468161123545413e-05, |
| "loss": 0.2618, |
| "step": 23300 |
| }, |
| { |
| "epoch": 1.5559803840079793, |
| "grad_norm": 2.3701083660125732, |
| "learning_rate": 1.4273226235206178e-05, |
| "loss": 0.2412, |
| "step": 23400 |
| }, |
| { |
| "epoch": 1.5626298728285262, |
| "grad_norm": 0.5906326174736023, |
| "learning_rate": 1.3869653796923993e-05, |
| "loss": 0.234, |
| "step": 23500 |
| }, |
| { |
| "epoch": 1.5692793616490732, |
| "grad_norm": 0.9088567495346069, |
| "learning_rate": 1.3470948283684925e-05, |
| "loss": 0.2529, |
| "step": 23600 |
| }, |
| { |
| "epoch": 1.5759288504696203, |
| "grad_norm": 2.244845151901245, |
| "learning_rate": 1.307716340296904e-05, |
| "loss": 0.2444, |
| "step": 23700 |
| }, |
| { |
| "epoch": 1.5825783392901671, |
| "grad_norm": 4.36794376373291, |
| "learning_rate": 1.268835219942433e-05, |
| "loss": 0.2533, |
| "step": 23800 |
| }, |
| { |
| "epoch": 1.589227828110714, |
| "grad_norm": 2.707139730453491, |
| "learning_rate": 1.23045670477215e-05, |
| "loss": 0.2467, |
| "step": 23900 |
| }, |
| { |
| "epoch": 1.595877316931261, |
| "grad_norm": 1.5950795412063599, |
| "learning_rate": 1.1925859645498722e-05, |
| "loss": 0.2349, |
| "step": 24000 |
| }, |
| { |
| "epoch": 1.6025268057518078, |
| "grad_norm": 1.1506330966949463, |
| "learning_rate": 1.1552281006397819e-05, |
| "loss": 0.2518, |
| "step": 24100 |
| }, |
| { |
| "epoch": 1.6091762945723547, |
| "grad_norm": 2.356943130493164, |
| "learning_rate": 1.1183881453192479e-05, |
| "loss": 0.2376, |
| "step": 24200 |
| }, |
| { |
| "epoch": 1.6158257833929017, |
| "grad_norm": 1.1734323501586914, |
| "learning_rate": 1.082071061100945e-05, |
| "loss": 0.2277, |
| "step": 24300 |
| }, |
| { |
| "epoch": 1.6224752722134486, |
| "grad_norm": 3.4108047485351562, |
| "learning_rate": 1.0462817400643959e-05, |
| "loss": 0.2313, |
| "step": 24400 |
| }, |
| { |
| "epoch": 1.6291247610339954, |
| "grad_norm": 2.890157699584961, |
| "learning_rate": 1.0110250031969709e-05, |
| "loss": 0.2364, |
| "step": 24500 |
| }, |
| { |
| "epoch": 1.6357742498545425, |
| "grad_norm": 1.4353581666946411, |
| "learning_rate": 9.763055997444897e-06, |
| "loss": 0.2401, |
| "step": 24600 |
| }, |
| { |
| "epoch": 1.6424237386750895, |
| "grad_norm": 0.6841472387313843, |
| "learning_rate": 9.421282065714676e-06, |
| "loss": 0.2344, |
| "step": 24700 |
| }, |
| { |
| "epoch": 1.6490732274956361, |
| "grad_norm": 2.0864779949188232, |
| "learning_rate": 9.08497427531128e-06, |
| "loss": 0.2186, |
| "step": 24800 |
| }, |
| { |
| "epoch": 1.6557227163161832, |
| "grad_norm": 1.5929943323135376, |
| "learning_rate": 8.754177928452328e-06, |
| "loss": 0.2307, |
| "step": 24900 |
| }, |
| { |
| "epoch": 1.6623722051367302, |
| "grad_norm": 0.5336441993713379, |
| "learning_rate": 8.428937584938496e-06, |
| "loss": 0.2401, |
| "step": 25000 |
| }, |
| { |
| "epoch": 1.669021693957277, |
| "grad_norm": 2.829430103302002, |
| "learning_rate": 8.109297056151067e-06, |
| "loss": 0.2285, |
| "step": 25100 |
| }, |
| { |
| "epoch": 1.675671182777824, |
| "grad_norm": 1.822135090827942, |
| "learning_rate": 7.79529939915029e-06, |
| "loss": 0.238, |
| "step": 25200 |
| }, |
| { |
| "epoch": 1.682320671598371, |
| "grad_norm": 3.5365030765533447, |
| "learning_rate": 7.486986910875499e-06, |
| "loss": 0.2245, |
| "step": 25300 |
| }, |
| { |
| "epoch": 1.6889701604189178, |
| "grad_norm": 3.5611343383789062, |
| "learning_rate": 7.184401122447398e-06, |
| "loss": 0.2344, |
| "step": 25400 |
| }, |
| { |
| "epoch": 1.6956196492394646, |
| "grad_norm": 3.247079372406006, |
| "learning_rate": 6.887582793573727e-06, |
| "loss": 0.2393, |
| "step": 25500 |
| }, |
| { |
| "epoch": 1.7022691380600117, |
| "grad_norm": 2.570157766342163, |
| "learning_rate": 6.596571907058707e-06, |
| "loss": 0.2238, |
| "step": 25600 |
| }, |
| { |
| "epoch": 1.7089186268805585, |
| "grad_norm": 1.488516092300415, |
| "learning_rate": 6.31140766341713e-06, |
| "loss": 0.2326, |
| "step": 25700 |
| }, |
| { |
| "epoch": 1.7155681157011053, |
| "grad_norm": 1.449952483177185, |
| "learning_rate": 6.032128475593924e-06, |
| "loss": 0.2196, |
| "step": 25800 |
| }, |
| { |
| "epoch": 1.7222176045216524, |
| "grad_norm": 2.537747621536255, |
| "learning_rate": 5.758771963789722e-06, |
| "loss": 0.2065, |
| "step": 25900 |
| }, |
| { |
| "epoch": 1.7288670933421995, |
| "grad_norm": 1.783013105392456, |
| "learning_rate": 5.4913749503932575e-06, |
| "loss": 0.2001, |
| "step": 26000 |
| }, |
| { |
| "epoch": 1.7355165821627463, |
| "grad_norm": 2.2559077739715576, |
| "learning_rate": 5.229973455021231e-06, |
| "loss": 0.2183, |
| "step": 26100 |
| }, |
| { |
| "epoch": 1.7421660709832931, |
| "grad_norm": 2.048455238342285, |
| "learning_rate": 4.974602689666252e-06, |
| "loss": 0.2106, |
| "step": 26200 |
| }, |
| { |
| "epoch": 1.7488155598038402, |
| "grad_norm": 1.1111422777175903, |
| "learning_rate": 4.725297053953692e-06, |
| "loss": 0.2397, |
| "step": 26300 |
| }, |
| { |
| "epoch": 1.755465048624387, |
| "grad_norm": 1.549456238746643, |
| "learning_rate": 4.48209013050781e-06, |
| "loss": 0.2209, |
| "step": 26400 |
| }, |
| { |
| "epoch": 1.7621145374449338, |
| "grad_norm": 2.2229185104370117, |
| "learning_rate": 4.245014680428117e-06, |
| "loss": 0.211, |
| "step": 26500 |
| }, |
| { |
| "epoch": 1.768764026265481, |
| "grad_norm": 3.065992593765259, |
| "learning_rate": 4.014102638876205e-06, |
| "loss": 0.2134, |
| "step": 26600 |
| }, |
| { |
| "epoch": 1.7754135150860277, |
| "grad_norm": 1.1763139963150024, |
| "learning_rate": 3.789385110774013e-06, |
| "loss": 0.2095, |
| "step": 26700 |
| }, |
| { |
| "epoch": 1.7820630039065746, |
| "grad_norm": 2.860395908355713, |
| "learning_rate": 3.5708923666137927e-06, |
| "loss": 0.238, |
| "step": 26800 |
| }, |
| { |
| "epoch": 1.7887124927271216, |
| "grad_norm": 2.1108574867248535, |
| "learning_rate": 3.358653838380571e-06, |
| "loss": 0.2087, |
| "step": 26900 |
| }, |
| { |
| "epoch": 1.7953619815476687, |
| "grad_norm": 0.8311635255813599, |
| "learning_rate": 3.1526981155875156e-06, |
| "loss": 0.2164, |
| "step": 27000 |
| }, |
| { |
| "epoch": 1.8020114703682153, |
| "grad_norm": 0.8638597130775452, |
| "learning_rate": 2.9530529414247608e-06, |
| "loss": 0.2143, |
| "step": 27100 |
| }, |
| { |
| "epoch": 1.8086609591887624, |
| "grad_norm": 3.8020806312561035, |
| "learning_rate": 2.7597452090223354e-06, |
| "loss": 0.2037, |
| "step": 27200 |
| }, |
| { |
| "epoch": 1.8153104480093094, |
| "grad_norm": 4.387086868286133, |
| "learning_rate": 2.572800957827476e-06, |
| "loss": 0.2156, |
| "step": 27300 |
| }, |
| { |
| "epoch": 1.8219599368298562, |
| "grad_norm": 1.4160668849945068, |
| "learning_rate": 2.3922453700970295e-06, |
| "loss": 0.2182, |
| "step": 27400 |
| }, |
| { |
| "epoch": 1.828609425650403, |
| "grad_norm": 1.636501669883728, |
| "learning_rate": 2.2181027675052534e-06, |
| "loss": 0.2238, |
| "step": 27500 |
| }, |
| { |
| "epoch": 1.8352589144709501, |
| "grad_norm": 2.2405593395233154, |
| "learning_rate": 2.0503966078676217e-06, |
| "loss": 0.2179, |
| "step": 27600 |
| }, |
| { |
| "epoch": 1.841908403291497, |
| "grad_norm": 1.6398348808288574, |
| "learning_rate": 1.8891494819808841e-06, |
| "loss": 0.2149, |
| "step": 27700 |
| }, |
| { |
| "epoch": 1.8485578921120438, |
| "grad_norm": 3.231600522994995, |
| "learning_rate": 1.7343831105800511e-06, |
| "loss": 0.2071, |
| "step": 27800 |
| }, |
| { |
| "epoch": 1.8552073809325909, |
| "grad_norm": 2.70027232170105, |
| "learning_rate": 1.5861183414124403e-06, |
| "loss": 0.2132, |
| "step": 27900 |
| }, |
| { |
| "epoch": 1.8618568697531377, |
| "grad_norm": 2.990867853164673, |
| "learning_rate": 1.4443751464294664e-06, |
| "loss": 0.2097, |
| "step": 28000 |
| }, |
| { |
| "epoch": 1.8685063585736845, |
| "grad_norm": 1.75977623462677, |
| "learning_rate": 1.3091726190962329e-06, |
| "loss": 0.1923, |
| "step": 28100 |
| }, |
| { |
| "epoch": 1.8751558473942316, |
| "grad_norm": 0.49737560749053955, |
| "learning_rate": 1.18052897181965e-06, |
| "loss": 0.189, |
| "step": 28200 |
| }, |
| { |
| "epoch": 1.8818053362147786, |
| "grad_norm": 1.1605116128921509, |
| "learning_rate": 1.0584615334950643e-06, |
| "loss": 0.2089, |
| "step": 28300 |
| }, |
| { |
| "epoch": 1.8884548250353252, |
| "grad_norm": 2.284116268157959, |
| "learning_rate": 9.429867471720255e-07, |
| "loss": 0.2045, |
| "step": 28400 |
| }, |
| { |
| "epoch": 1.8951043138558723, |
| "grad_norm": 2.2018344402313232, |
| "learning_rate": 8.341201678392974e-07, |
| "loss": 0.2235, |
| "step": 28500 |
| }, |
| { |
| "epoch": 1.9017538026764194, |
| "grad_norm": 1.3855736255645752, |
| "learning_rate": 7.318764603295447e-07, |
| "loss": 0.2037, |
| "step": 28600 |
| }, |
| { |
| "epoch": 1.9084032914969662, |
| "grad_norm": 2.4615190029144287, |
| "learning_rate": 6.362693973439193e-07, |
| "loss": 0.2108, |
| "step": 28700 |
| }, |
| { |
| "epoch": 1.915052780317513, |
| "grad_norm": 2.775005578994751, |
| "learning_rate": 5.47311857596794e-07, |
| "loss": 0.1984, |
| "step": 28800 |
| }, |
| { |
| "epoch": 1.92170226913806, |
| "grad_norm": 3.7345826625823975, |
| "learning_rate": 4.6501582408096657e-07, |
| "loss": 0.2124, |
| "step": 28900 |
| }, |
| { |
| "epoch": 1.928351757958607, |
| "grad_norm": 0.46711668372154236, |
| "learning_rate": 3.893923824534629e-07, |
| "loss": 0.1897, |
| "step": 29000 |
| }, |
| { |
| "epoch": 1.9350012467791537, |
| "grad_norm": 2.135247230529785, |
| "learning_rate": 3.204517195422696e-07, |
| "loss": 0.2267, |
| "step": 29100 |
| }, |
| { |
| "epoch": 1.9416507355997008, |
| "grad_norm": 0.9539075493812561, |
| "learning_rate": 2.5820312197411543e-07, |
| "loss": 0.2104, |
| "step": 29200 |
| }, |
| { |
| "epoch": 1.9483002244202476, |
| "grad_norm": 3.765820026397705, |
| "learning_rate": 2.0265497492352735e-07, |
| "loss": 0.2072, |
| "step": 29300 |
| }, |
| { |
| "epoch": 1.9549497132407945, |
| "grad_norm": 2.848856210708618, |
| "learning_rate": 1.538147609832896e-07, |
| "loss": 0.2018, |
| "step": 29400 |
| }, |
| { |
| "epoch": 1.9615992020613415, |
| "grad_norm": 1.4174009561538696, |
| "learning_rate": 1.1168905915652228e-07, |
| "loss": 0.1963, |
| "step": 29500 |
| }, |
| { |
| "epoch": 1.9682486908818886, |
| "grad_norm": 3.067931652069092, |
| "learning_rate": 7.628354397045123e-08, |
| "loss": 0.2062, |
| "step": 29600 |
| }, |
| { |
| "epoch": 1.9748981797024354, |
| "grad_norm": 2.1927988529205322, |
| "learning_rate": 4.760298471201963e-08, |
| "loss": 0.2007, |
| "step": 29700 |
| }, |
| { |
| "epoch": 1.9815476685229823, |
| "grad_norm": 2.1186487674713135, |
| "learning_rate": 2.565124478545733e-08, |
| "loss": 0.2119, |
| "step": 29800 |
| }, |
| { |
| "epoch": 1.9881971573435293, |
| "grad_norm": 1.58423912525177, |
| "learning_rate": 1.043128119184167e-08, |
| "loss": 0.2192, |
| "step": 29900 |
| }, |
| { |
| "epoch": 1.9948466461640761, |
| "grad_norm": 1.9130926132202148, |
| "learning_rate": 1.945144130788279e-09, |
| "loss": 0.2238, |
| "step": 30000 |
| }, |
| { |
| "epoch": 1.9948466461640761, |
| "eval_loss": 0.23713438212871552, |
| "eval_runtime": 4319.7371, |
| "eval_samples_per_second": 1.547, |
| "eval_steps_per_second": 1.547, |
| "step": 30000 |
| }, |
| { |
| "epoch": 1.9999002576676919, |
| "step": 30076, |
| "total_flos": 1.2039888836550205e+19, |
| "train_loss": 0.385877049085006, |
| "train_runtime": 246042.9971, |
| "train_samples_per_second": 0.489, |
| "train_steps_per_second": 0.122 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 30076, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.2039888836550205e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|