diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7111 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.05060984867655246, + "eval_steps": 200, + "global_step": 2000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 5.060984867655246e-05, + "grad_norm": 5.005204677581787, + "learning_rate": 4.999915650252206e-05, + "loss": 0.722, + "step": 2 + }, + { + "epoch": 0.00010121969735310492, + "grad_norm": 4.560483932495117, + "learning_rate": 4.9998313005044116e-05, + "loss": 0.3625, + "step": 4 + }, + { + "epoch": 0.00015182954602965736, + "grad_norm": 9.806407928466797, + "learning_rate": 4.9997469507566176e-05, + "loss": 0.786, + "step": 6 + }, + { + "epoch": 0.00020243939470620983, + "grad_norm": 11.181573867797852, + "learning_rate": 4.999662601008823e-05, + "loss": 0.6063, + "step": 8 + }, + { + "epoch": 0.0002530492433827623, + "grad_norm": 3.7916452884674072, + "learning_rate": 4.999578251261029e-05, + "loss": 0.4681, + "step": 10 + }, + { + "epoch": 0.0003036590920593147, + "grad_norm": 10.196318626403809, + "learning_rate": 4.999493901513235e-05, + "loss": 0.8192, + "step": 12 + }, + { + "epoch": 0.0003542689407358672, + "grad_norm": 8.210321426391602, + "learning_rate": 4.9994095517654403e-05, + "loss": 0.7953, + "step": 14 + }, + { + "epoch": 0.00040487878941241966, + "grad_norm": 8.648380279541016, + "learning_rate": 4.9993252020176464e-05, + "loss": 0.5404, + "step": 16 + }, + { + "epoch": 0.0004554886380889721, + "grad_norm": NaN, + "learning_rate": 4.999283027143749e-05, + "loss": 0.7557, + "step": 18 + }, + { + "epoch": 0.0005060984867655246, + "grad_norm": 3.2002220153808594, + "learning_rate": 4.999198677395955e-05, + "loss": 0.5948, + "step": 20 + }, + { + "epoch": 0.000556708335442077, + "grad_norm": 6.236138343811035, + "learning_rate": 4.9991143276481604e-05, + "loss": 1.1657, + "step": 22 + }, + { + "epoch": 0.0006073181841186294, + "grad_norm": 8.113347053527832, + "learning_rate": 4.9990299779003664e-05, + "loss": 0.6126, + "step": 24 + }, + { + "epoch": 0.0006579280327951819, + "grad_norm": 6.6634979248046875, + "learning_rate": 4.998945628152572e-05, + "loss": 0.5794, + "step": 26 + }, + { + "epoch": 0.0007085378814717344, + "grad_norm": 7.940774440765381, + "learning_rate": 4.998861278404778e-05, + "loss": 0.3316, + "step": 28 + }, + { + "epoch": 0.0007591477301482868, + "grad_norm": 7.296051025390625, + "learning_rate": 4.998776928656984e-05, + "loss": 0.8218, + "step": 30 + }, + { + "epoch": 0.0008097575788248393, + "grad_norm": 6.318060398101807, + "learning_rate": 4.998692578909189e-05, + "loss": 0.7266, + "step": 32 + }, + { + "epoch": 0.0008603674275013917, + "grad_norm": 13.293371200561523, + "learning_rate": 4.998608229161395e-05, + "loss": 0.96, + "step": 34 + }, + { + "epoch": 0.0009109772761779442, + "grad_norm": NaN, + "learning_rate": 4.998566054287498e-05, + "loss": 0.9009, + "step": 36 + }, + { + "epoch": 0.0009615871248544967, + "grad_norm": 4.929239749908447, + "learning_rate": 4.998481704539704e-05, + "loss": 0.7153, + "step": 38 + }, + { + "epoch": 0.001012196973531049, + "grad_norm": 8.825164794921875, + "learning_rate": 4.998397354791909e-05, + "loss": 0.4486, + "step": 40 + }, + { + "epoch": 0.0010628068222076016, + "grad_norm": 8.393811225891113, + "learning_rate": 4.998313005044115e-05, + "loss": 0.7293, + "step": 42 + }, + { + "epoch": 0.001113416670884154, + "grad_norm": 4.109868049621582, + "learning_rate": 4.9982286552963206e-05, + "loss": 0.5638, + "step": 44 + }, + { + "epoch": 0.0011640265195607066, + "grad_norm": 13.631953239440918, + "learning_rate": 4.9981443055485266e-05, + "loss": 1.0084, + "step": 46 + }, + { + "epoch": 0.0012146363682372589, + "grad_norm": 9.624829292297363, + "learning_rate": 4.9980599558007326e-05, + "loss": 0.9545, + "step": 48 + }, + { + "epoch": 0.0012652462169138114, + "grad_norm": 5.674628257751465, + "learning_rate": 4.997975606052938e-05, + "loss": 0.5188, + "step": 50 + }, + { + "epoch": 0.0013158560655903639, + "grad_norm": 6.339033603668213, + "learning_rate": 4.997891256305144e-05, + "loss": 0.5308, + "step": 52 + }, + { + "epoch": 0.0013664659142669164, + "grad_norm": 3.895756244659424, + "learning_rate": 4.997806906557349e-05, + "loss": 0.6042, + "step": 54 + }, + { + "epoch": 0.0014170757629434689, + "grad_norm": 4.451659679412842, + "learning_rate": 4.997722556809555e-05, + "loss": 0.4957, + "step": 56 + }, + { + "epoch": 0.0014676856116200212, + "grad_norm": 5.246840476989746, + "learning_rate": 4.9976382070617613e-05, + "loss": 0.7031, + "step": 58 + }, + { + "epoch": 0.0015182954602965737, + "grad_norm": 7.467830657958984, + "learning_rate": 4.997553857313967e-05, + "loss": 0.8567, + "step": 60 + }, + { + "epoch": 0.0015689053089731262, + "grad_norm": 6.088858127593994, + "learning_rate": 4.997469507566173e-05, + "loss": 0.5998, + "step": 62 + }, + { + "epoch": 0.0016195151576496787, + "grad_norm": 10.719757080078125, + "learning_rate": 4.997385157818378e-05, + "loss": 0.8178, + "step": 64 + }, + { + "epoch": 0.0016701250063262312, + "grad_norm": 5.923684120178223, + "learning_rate": 4.997300808070584e-05, + "loss": 0.9103, + "step": 66 + }, + { + "epoch": 0.0017207348550027834, + "grad_norm": 8.114771842956543, + "learning_rate": 4.99721645832279e-05, + "loss": 0.4154, + "step": 68 + }, + { + "epoch": 0.001771344703679336, + "grad_norm": 9.67381477355957, + "learning_rate": 4.9971321085749954e-05, + "loss": 0.6067, + "step": 70 + }, + { + "epoch": 0.0018219545523558884, + "grad_norm": 7.255108833312988, + "learning_rate": 4.9970477588272014e-05, + "loss": 0.4104, + "step": 72 + }, + { + "epoch": 0.001872564401032441, + "grad_norm": 6.330812454223633, + "learning_rate": 4.996963409079407e-05, + "loss": 0.7361, + "step": 74 + }, + { + "epoch": 0.0019231742497089934, + "grad_norm": 10.727275848388672, + "learning_rate": 4.996879059331613e-05, + "loss": 0.8339, + "step": 76 + }, + { + "epoch": 0.001973784098385546, + "grad_norm": 5.6855950355529785, + "learning_rate": 4.996794709583819e-05, + "loss": 0.6211, + "step": 78 + }, + { + "epoch": 0.002024393947062098, + "grad_norm": 9.52706527709961, + "learning_rate": 4.996710359836024e-05, + "loss": 0.6736, + "step": 80 + }, + { + "epoch": 0.002075003795738651, + "grad_norm": 9.556534767150879, + "learning_rate": 4.99662601008823e-05, + "loss": 0.5151, + "step": 82 + }, + { + "epoch": 0.002125613644415203, + "grad_norm": 4.778113842010498, + "learning_rate": 4.9965416603404355e-05, + "loss": 0.5348, + "step": 84 + }, + { + "epoch": 0.0021762234930917555, + "grad_norm": 10.250826835632324, + "learning_rate": 4.9964573105926416e-05, + "loss": 0.5674, + "step": 86 + }, + { + "epoch": 0.002226833341768308, + "grad_norm": 6.03010892868042, + "learning_rate": 4.9963729608448476e-05, + "loss": 0.4641, + "step": 88 + }, + { + "epoch": 0.0022774431904448605, + "grad_norm": 10.152463912963867, + "learning_rate": 4.996288611097053e-05, + "loss": 1.0214, + "step": 90 + }, + { + "epoch": 0.002328053039121413, + "grad_norm": 8.930192947387695, + "learning_rate": 4.996204261349259e-05, + "loss": 0.7772, + "step": 92 + }, + { + "epoch": 0.0023786628877979655, + "grad_norm": 10.251153945922852, + "learning_rate": 4.996119911601464e-05, + "loss": 0.9428, + "step": 94 + }, + { + "epoch": 0.0024292727364745178, + "grad_norm": 6.223593711853027, + "learning_rate": 4.99603556185367e-05, + "loss": 0.6318, + "step": 96 + }, + { + "epoch": 0.0024798825851510705, + "grad_norm": 2.4227373600006104, + "learning_rate": 4.995951212105876e-05, + "loss": 0.436, + "step": 98 + }, + { + "epoch": 0.0025304924338276228, + "grad_norm": 6.225770473480225, + "learning_rate": 4.9958668623580817e-05, + "loss": 0.7753, + "step": 100 + }, + { + "epoch": 0.0025811022825041755, + "grad_norm": 5.370884895324707, + "learning_rate": 4.995782512610288e-05, + "loss": 0.5137, + "step": 102 + }, + { + "epoch": 0.0026317121311807278, + "grad_norm": 5.137220859527588, + "learning_rate": 4.995698162862493e-05, + "loss": 0.5953, + "step": 104 + }, + { + "epoch": 0.00268232197985728, + "grad_norm": 5.909286975860596, + "learning_rate": 4.995613813114699e-05, + "loss": 0.8687, + "step": 106 + }, + { + "epoch": 0.0027329318285338328, + "grad_norm": 6.003445148468018, + "learning_rate": 4.995529463366905e-05, + "loss": 0.53, + "step": 108 + }, + { + "epoch": 0.002783541677210385, + "grad_norm": 3.8742566108703613, + "learning_rate": 4.9954451136191104e-05, + "loss": 0.5512, + "step": 110 + }, + { + "epoch": 0.0028341515258869378, + "grad_norm": 3.5072569847106934, + "learning_rate": 4.9953607638713164e-05, + "loss": 0.3383, + "step": 112 + }, + { + "epoch": 0.00288476137456349, + "grad_norm": 5.679248809814453, + "learning_rate": 4.995276414123522e-05, + "loss": 0.7801, + "step": 114 + }, + { + "epoch": 0.0029353712232400423, + "grad_norm": 5.291722297668457, + "learning_rate": 4.995192064375728e-05, + "loss": 0.6538, + "step": 116 + }, + { + "epoch": 0.002985981071916595, + "grad_norm": 4.748198986053467, + "learning_rate": 4.995107714627934e-05, + "loss": 0.398, + "step": 118 + }, + { + "epoch": 0.0030365909205931473, + "grad_norm": 4.336450576782227, + "learning_rate": 4.995023364880139e-05, + "loss": 0.3351, + "step": 120 + }, + { + "epoch": 0.0030872007692697, + "grad_norm": 5.251417636871338, + "learning_rate": 4.994939015132345e-05, + "loss": 0.4014, + "step": 122 + }, + { + "epoch": 0.0031378106179462523, + "grad_norm": 7.911617755889893, + "learning_rate": 4.9948546653845505e-05, + "loss": 0.3151, + "step": 124 + }, + { + "epoch": 0.0031884204666228046, + "grad_norm": 6.031822681427002, + "learning_rate": 4.9947703156367565e-05, + "loss": 0.9222, + "step": 126 + }, + { + "epoch": 0.0032390303152993573, + "grad_norm": 10.08216381072998, + "learning_rate": 4.994685965888962e-05, + "loss": 0.6676, + "step": 128 + }, + { + "epoch": 0.0032896401639759096, + "grad_norm": 3.475123167037964, + "learning_rate": 4.994601616141168e-05, + "loss": 0.2461, + "step": 130 + }, + { + "epoch": 0.0033402500126524623, + "grad_norm": 4.408476829528809, + "learning_rate": 4.994517266393374e-05, + "loss": 0.26, + "step": 132 + }, + { + "epoch": 0.0033908598613290146, + "grad_norm": 6.929357528686523, + "learning_rate": 4.994432916645579e-05, + "loss": 0.9818, + "step": 134 + }, + { + "epoch": 0.003441469710005567, + "grad_norm": 6.456033229827881, + "learning_rate": 4.994348566897785e-05, + "loss": 0.4414, + "step": 136 + }, + { + "epoch": 0.0034920795586821196, + "grad_norm": 26.87461280822754, + "learning_rate": 4.9942642171499906e-05, + "loss": 0.7951, + "step": 138 + }, + { + "epoch": 0.003542689407358672, + "grad_norm": 7.420943737030029, + "learning_rate": 4.9941798674021966e-05, + "loss": 0.5827, + "step": 140 + }, + { + "epoch": 0.0035932992560352246, + "grad_norm": 5.9366583824157715, + "learning_rate": 4.9940955176544027e-05, + "loss": 0.6941, + "step": 142 + }, + { + "epoch": 0.003643909104711777, + "grad_norm": 5.8037543296813965, + "learning_rate": 4.994011167906608e-05, + "loss": 0.6301, + "step": 144 + }, + { + "epoch": 0.003694518953388329, + "grad_norm": 4.243365287780762, + "learning_rate": 4.993926818158814e-05, + "loss": 0.6995, + "step": 146 + }, + { + "epoch": 0.003745128802064882, + "grad_norm": 7.5515522956848145, + "learning_rate": 4.9938424684110194e-05, + "loss": 0.5537, + "step": 148 + }, + { + "epoch": 0.003795738650741434, + "grad_norm": 8.966812133789062, + "learning_rate": 4.9937581186632254e-05, + "loss": 0.5416, + "step": 150 + }, + { + "epoch": 0.003846348499417987, + "grad_norm": 7.70654296875, + "learning_rate": 4.9936737689154314e-05, + "loss": 0.6907, + "step": 152 + }, + { + "epoch": 0.003896958348094539, + "grad_norm": 6.371860504150391, + "learning_rate": 4.993589419167637e-05, + "loss": 0.468, + "step": 154 + }, + { + "epoch": 0.003947568196771092, + "grad_norm": 6.1000542640686035, + "learning_rate": 4.993505069419843e-05, + "loss": 0.6523, + "step": 156 + }, + { + "epoch": 0.003998178045447644, + "grad_norm": 5.390421390533447, + "learning_rate": 4.993420719672048e-05, + "loss": 1.0799, + "step": 158 + }, + { + "epoch": 0.004048787894124196, + "grad_norm": 7.502729415893555, + "learning_rate": 4.993336369924254e-05, + "loss": 0.7921, + "step": 160 + }, + { + "epoch": 0.004099397742800749, + "grad_norm": 3.19514536857605, + "learning_rate": 4.99325202017646e-05, + "loss": 0.2293, + "step": 162 + }, + { + "epoch": 0.004150007591477302, + "grad_norm": 7.2197585105896, + "learning_rate": 4.9931676704286655e-05, + "loss": 0.6778, + "step": 164 + }, + { + "epoch": 0.004200617440153854, + "grad_norm": 4.662430286407471, + "learning_rate": 4.9930833206808715e-05, + "loss": 0.4956, + "step": 166 + }, + { + "epoch": 0.004251227288830406, + "grad_norm": 5.545823574066162, + "learning_rate": 4.992998970933077e-05, + "loss": 0.4587, + "step": 168 + }, + { + "epoch": 0.004301837137506959, + "grad_norm": 8.578537940979004, + "learning_rate": 4.992914621185283e-05, + "loss": 0.2751, + "step": 170 + }, + { + "epoch": 0.004352446986183511, + "grad_norm": 4.3413472175598145, + "learning_rate": 4.992830271437489e-05, + "loss": 0.3712, + "step": 172 + }, + { + "epoch": 0.004403056834860064, + "grad_norm": 3.3616743087768555, + "learning_rate": 4.992745921689694e-05, + "loss": 0.4654, + "step": 174 + }, + { + "epoch": 0.004453666683536616, + "grad_norm": 3.7586441040039062, + "learning_rate": 4.9926615719419e-05, + "loss": 0.5338, + "step": 176 + }, + { + "epoch": 0.004504276532213168, + "grad_norm": 4.674256801605225, + "learning_rate": 4.9925772221941056e-05, + "loss": 0.3885, + "step": 178 + }, + { + "epoch": 0.004554886380889721, + "grad_norm": 6.164566993713379, + "learning_rate": 4.9924928724463116e-05, + "loss": 0.3949, + "step": 180 + }, + { + "epoch": 0.004605496229566274, + "grad_norm": 12.5003023147583, + "learning_rate": 4.9924085226985176e-05, + "loss": 0.6829, + "step": 182 + }, + { + "epoch": 0.004656106078242826, + "grad_norm": 8.045877456665039, + "learning_rate": 4.992324172950723e-05, + "loss": 0.7711, + "step": 184 + }, + { + "epoch": 0.004706715926919378, + "grad_norm": 5.4890570640563965, + "learning_rate": 4.992239823202929e-05, + "loss": 0.6287, + "step": 186 + }, + { + "epoch": 0.004757325775595931, + "grad_norm": 3.246534585952759, + "learning_rate": 4.992155473455134e-05, + "loss": 0.3895, + "step": 188 + }, + { + "epoch": 0.004807935624272484, + "grad_norm": 7.648909568786621, + "learning_rate": 4.9920711237073403e-05, + "loss": 0.7327, + "step": 190 + }, + { + "epoch": 0.0048585454729490355, + "grad_norm": 5.417934894561768, + "learning_rate": 4.9919867739595464e-05, + "loss": 0.5529, + "step": 192 + }, + { + "epoch": 0.004909155321625588, + "grad_norm": 18.588844299316406, + "learning_rate": 4.991902424211752e-05, + "loss": 1.1181, + "step": 194 + }, + { + "epoch": 0.004959765170302141, + "grad_norm": 11.765092849731445, + "learning_rate": 4.991818074463958e-05, + "loss": 0.7821, + "step": 196 + }, + { + "epoch": 0.005010375018978693, + "grad_norm": 6.446319580078125, + "learning_rate": 4.991733724716163e-05, + "loss": 0.7071, + "step": 198 + }, + { + "epoch": 0.0050609848676552455, + "grad_norm": 7.347968101501465, + "learning_rate": 4.991649374968369e-05, + "loss": 0.4164, + "step": 200 + }, + { + "epoch": 0.0050609848676552455, + "eval_cer": 0.23587196123692475, + "eval_loss": 0.37637796998023987, + "eval_runtime": 2775.7084, + "eval_samples_per_second": 5.647, + "eval_steps_per_second": 0.353, + "step": 200 + }, + { + "epoch": 0.005111594716331798, + "grad_norm": 10.311556816101074, + "learning_rate": 4.991565025220575e-05, + "loss": 1.1246, + "step": 202 + }, + { + "epoch": 0.005162204565008351, + "grad_norm": 10.519822120666504, + "learning_rate": 4.9914806754727805e-05, + "loss": 0.6127, + "step": 204 + }, + { + "epoch": 0.005212814413684903, + "grad_norm": 6.434566974639893, + "learning_rate": 4.9913963257249865e-05, + "loss": 0.5269, + "step": 206 + }, + { + "epoch": 0.0052634242623614555, + "grad_norm": 8.08187198638916, + "learning_rate": 4.991311975977192e-05, + "loss": 1.0978, + "step": 208 + }, + { + "epoch": 0.005314034111038008, + "grad_norm": 7.574239730834961, + "learning_rate": 4.991227626229398e-05, + "loss": 0.7327, + "step": 210 + }, + { + "epoch": 0.00536464395971456, + "grad_norm": 7.921819686889648, + "learning_rate": 4.991143276481603e-05, + "loss": 1.194, + "step": 212 + }, + { + "epoch": 0.005415253808391113, + "grad_norm": 9.017950057983398, + "learning_rate": 4.991058926733809e-05, + "loss": 0.9906, + "step": 214 + }, + { + "epoch": 0.0054658636570676655, + "grad_norm": 6.131129741668701, + "learning_rate": 4.990974576986015e-05, + "loss": 0.8681, + "step": 216 + }, + { + "epoch": 0.005516473505744218, + "grad_norm": 6.411858081817627, + "learning_rate": 4.9908902272382206e-05, + "loss": 0.6271, + "step": 218 + }, + { + "epoch": 0.00556708335442077, + "grad_norm": 6.6365227699279785, + "learning_rate": 4.9908058774904266e-05, + "loss": 0.9051, + "step": 220 + }, + { + "epoch": 0.005617693203097323, + "grad_norm": 7.575653553009033, + "learning_rate": 4.990721527742632e-05, + "loss": 0.9199, + "step": 222 + }, + { + "epoch": 0.0056683030517738755, + "grad_norm": 8.839277267456055, + "learning_rate": 4.990637177994838e-05, + "loss": 0.909, + "step": 224 + }, + { + "epoch": 0.005718912900450427, + "grad_norm": 8.077840805053711, + "learning_rate": 4.990552828247044e-05, + "loss": 0.6063, + "step": 226 + }, + { + "epoch": 0.00576952274912698, + "grad_norm": 8.73996639251709, + "learning_rate": 4.990468478499249e-05, + "loss": 0.9447, + "step": 228 + }, + { + "epoch": 0.005820132597803533, + "grad_norm": 6.638965129852295, + "learning_rate": 4.990384128751455e-05, + "loss": 0.6174, + "step": 230 + }, + { + "epoch": 0.005870742446480085, + "grad_norm": 5.351879119873047, + "learning_rate": 4.990299779003661e-05, + "loss": 0.4632, + "step": 232 + }, + { + "epoch": 0.005921352295156637, + "grad_norm": 6.497000694274902, + "learning_rate": 4.990215429255867e-05, + "loss": 0.5266, + "step": 234 + }, + { + "epoch": 0.00597196214383319, + "grad_norm": 10.24178695678711, + "learning_rate": 4.990131079508073e-05, + "loss": 0.8037, + "step": 236 + }, + { + "epoch": 0.006022571992509743, + "grad_norm": 5.579001426696777, + "learning_rate": 4.990046729760278e-05, + "loss": 0.6132, + "step": 238 + }, + { + "epoch": 0.006073181841186295, + "grad_norm": 3.613511562347412, + "learning_rate": 4.989962380012484e-05, + "loss": 0.3918, + "step": 240 + }, + { + "epoch": 0.006123791689862847, + "grad_norm": 9.699933052062988, + "learning_rate": 4.9898780302646894e-05, + "loss": 0.8956, + "step": 242 + }, + { + "epoch": 0.0061744015385394, + "grad_norm": 10.87297534942627, + "learning_rate": 4.9897936805168954e-05, + "loss": 1.1579, + "step": 244 + }, + { + "epoch": 0.006225011387215952, + "grad_norm": 6.29569673538208, + "learning_rate": 4.9897093307691014e-05, + "loss": 0.4237, + "step": 246 + }, + { + "epoch": 0.006275621235892505, + "grad_norm": 9.018596649169922, + "learning_rate": 4.989624981021307e-05, + "loss": 0.7854, + "step": 248 + }, + { + "epoch": 0.006326231084569057, + "grad_norm": 9.707599639892578, + "learning_rate": 4.989540631273513e-05, + "loss": 1.3102, + "step": 250 + }, + { + "epoch": 0.006376840933245609, + "grad_norm": 5.1023430824279785, + "learning_rate": 4.989456281525718e-05, + "loss": 0.5906, + "step": 252 + }, + { + "epoch": 0.006427450781922162, + "grad_norm": 2.3029062747955322, + "learning_rate": 4.989371931777924e-05, + "loss": 0.7183, + "step": 254 + }, + { + "epoch": 0.006478060630598715, + "grad_norm": 11.459485054016113, + "learning_rate": 4.98928758203013e-05, + "loss": 0.5505, + "step": 256 + }, + { + "epoch": 0.006528670479275267, + "grad_norm": 7.620695114135742, + "learning_rate": 4.9892032322823355e-05, + "loss": 0.9374, + "step": 258 + }, + { + "epoch": 0.006579280327951819, + "grad_norm": 5.305224895477295, + "learning_rate": 4.9891188825345416e-05, + "loss": 0.7362, + "step": 260 + }, + { + "epoch": 0.006629890176628372, + "grad_norm": 7.901217460632324, + "learning_rate": 4.989034532786747e-05, + "loss": 0.9518, + "step": 262 + }, + { + "epoch": 0.006680500025304925, + "grad_norm": 7.463654041290283, + "learning_rate": 4.988950183038953e-05, + "loss": 0.9789, + "step": 264 + }, + { + "epoch": 0.0067311098739814765, + "grad_norm": 4.86021089553833, + "learning_rate": 4.988865833291159e-05, + "loss": 0.8441, + "step": 266 + }, + { + "epoch": 0.006781719722658029, + "grad_norm": 5.242839813232422, + "learning_rate": 4.988781483543364e-05, + "loss": 0.8258, + "step": 268 + }, + { + "epoch": 0.006832329571334582, + "grad_norm": 3.524228572845459, + "learning_rate": 4.98869713379557e-05, + "loss": 0.3924, + "step": 270 + }, + { + "epoch": 0.006882939420011134, + "grad_norm": 3.8527095317840576, + "learning_rate": 4.9886127840477756e-05, + "loss": 0.5253, + "step": 272 + }, + { + "epoch": 0.0069335492686876865, + "grad_norm": 3.4836299419403076, + "learning_rate": 4.9885284342999817e-05, + "loss": 0.5859, + "step": 274 + }, + { + "epoch": 0.006984159117364239, + "grad_norm": 9.650673866271973, + "learning_rate": 4.988444084552188e-05, + "loss": 0.7122, + "step": 276 + }, + { + "epoch": 0.007034768966040792, + "grad_norm": 5.5393829345703125, + "learning_rate": 4.988359734804393e-05, + "loss": 0.5434, + "step": 278 + }, + { + "epoch": 0.007085378814717344, + "grad_norm": 11.345093727111816, + "learning_rate": 4.988275385056599e-05, + "loss": 0.9166, + "step": 280 + }, + { + "epoch": 0.0071359886633938965, + "grad_norm": 4.247881889343262, + "learning_rate": 4.9881910353088044e-05, + "loss": 0.6162, + "step": 282 + }, + { + "epoch": 0.007186598512070449, + "grad_norm": 5.472718238830566, + "learning_rate": 4.9881066855610104e-05, + "loss": 0.8378, + "step": 284 + }, + { + "epoch": 0.007237208360747001, + "grad_norm": 6.520363807678223, + "learning_rate": 4.9880223358132164e-05, + "loss": 0.6258, + "step": 286 + }, + { + "epoch": 0.007287818209423554, + "grad_norm": 4.629631996154785, + "learning_rate": 4.987937986065422e-05, + "loss": 0.5166, + "step": 288 + }, + { + "epoch": 0.0073384280581001065, + "grad_norm": 9.533284187316895, + "learning_rate": 4.987853636317628e-05, + "loss": 1.1145, + "step": 290 + }, + { + "epoch": 0.007389037906776658, + "grad_norm": 7.794739246368408, + "learning_rate": 4.987769286569833e-05, + "loss": 1.0134, + "step": 292 + }, + { + "epoch": 0.007439647755453211, + "grad_norm": 8.504875183105469, + "learning_rate": 4.987684936822039e-05, + "loss": 0.8261, + "step": 294 + }, + { + "epoch": 0.007490257604129764, + "grad_norm": 6.671535015106201, + "learning_rate": 4.9876005870742445e-05, + "loss": 0.7141, + "step": 296 + }, + { + "epoch": 0.0075408674528063165, + "grad_norm": 11.959641456604004, + "learning_rate": 4.9875162373264505e-05, + "loss": 0.8247, + "step": 298 + }, + { + "epoch": 0.007591477301482868, + "grad_norm": 4.4155378341674805, + "learning_rate": 4.9874318875786565e-05, + "loss": 0.348, + "step": 300 + }, + { + "epoch": 0.007642087150159421, + "grad_norm": 5.849681377410889, + "learning_rate": 4.987347537830862e-05, + "loss": 0.8355, + "step": 302 + }, + { + "epoch": 0.007692696998835974, + "grad_norm": 10.244682312011719, + "learning_rate": 4.987263188083068e-05, + "loss": 0.7634, + "step": 304 + }, + { + "epoch": 0.007743306847512526, + "grad_norm": 3.5256447792053223, + "learning_rate": 4.987178838335273e-05, + "loss": 0.451, + "step": 306 + }, + { + "epoch": 0.007793916696189078, + "grad_norm": 7.318663120269775, + "learning_rate": 4.987094488587479e-05, + "loss": 0.7009, + "step": 308 + }, + { + "epoch": 0.007844526544865631, + "grad_norm": 5.343123912811279, + "learning_rate": 4.987010138839685e-05, + "loss": 0.4291, + "step": 310 + }, + { + "epoch": 0.007895136393542184, + "grad_norm": 6.702011585235596, + "learning_rate": 4.9869257890918906e-05, + "loss": 0.9008, + "step": 312 + }, + { + "epoch": 0.007945746242218736, + "grad_norm": 9.040878295898438, + "learning_rate": 4.9868414393440966e-05, + "loss": 1.0295, + "step": 314 + }, + { + "epoch": 0.007996356090895287, + "grad_norm": 5.2994585037231445, + "learning_rate": 4.986757089596302e-05, + "loss": 0.6427, + "step": 316 + }, + { + "epoch": 0.00804696593957184, + "grad_norm": 8.381704330444336, + "learning_rate": 4.986672739848508e-05, + "loss": 0.8675, + "step": 318 + }, + { + "epoch": 0.008097575788248393, + "grad_norm": 9.162700653076172, + "learning_rate": 4.986588390100714e-05, + "loss": 1.1891, + "step": 320 + }, + { + "epoch": 0.008148185636924946, + "grad_norm": 10.618518829345703, + "learning_rate": 4.9865040403529194e-05, + "loss": 0.7131, + "step": 322 + }, + { + "epoch": 0.008198795485601498, + "grad_norm": 7.559556484222412, + "learning_rate": 4.9864196906051254e-05, + "loss": 0.7327, + "step": 324 + }, + { + "epoch": 0.008249405334278051, + "grad_norm": 7.372714519500732, + "learning_rate": 4.986335340857331e-05, + "loss": 0.5496, + "step": 326 + }, + { + "epoch": 0.008300015182954604, + "grad_norm": 5.887473106384277, + "learning_rate": 4.986250991109537e-05, + "loss": 0.4605, + "step": 328 + }, + { + "epoch": 0.008350625031631155, + "grad_norm": 3.8664021492004395, + "learning_rate": 4.986166641361743e-05, + "loss": 0.8836, + "step": 330 + }, + { + "epoch": 0.008401234880307707, + "grad_norm": 6.713327407836914, + "learning_rate": 4.986082291613948e-05, + "loss": 0.7846, + "step": 332 + }, + { + "epoch": 0.00845184472898426, + "grad_norm": 6.036464691162109, + "learning_rate": 4.985997941866154e-05, + "loss": 0.8055, + "step": 334 + }, + { + "epoch": 0.008502454577660813, + "grad_norm": 4.087986946105957, + "learning_rate": 4.9859135921183595e-05, + "loss": 0.7829, + "step": 336 + }, + { + "epoch": 0.008553064426337366, + "grad_norm": 9.335679054260254, + "learning_rate": 4.9858292423705655e-05, + "loss": 1.279, + "step": 338 + }, + { + "epoch": 0.008603674275013918, + "grad_norm": 6.914140701293945, + "learning_rate": 4.9857448926227715e-05, + "loss": 0.8773, + "step": 340 + }, + { + "epoch": 0.008654284123690471, + "grad_norm": 5.106595993041992, + "learning_rate": 4.985660542874977e-05, + "loss": 0.7357, + "step": 342 + }, + { + "epoch": 0.008704893972367022, + "grad_norm": 5.217001914978027, + "learning_rate": 4.985576193127183e-05, + "loss": 1.0202, + "step": 344 + }, + { + "epoch": 0.008755503821043575, + "grad_norm": 7.956677436828613, + "learning_rate": 4.985491843379388e-05, + "loss": 0.8517, + "step": 346 + }, + { + "epoch": 0.008806113669720127, + "grad_norm": 6.432283878326416, + "learning_rate": 4.985407493631594e-05, + "loss": 0.7963, + "step": 348 + }, + { + "epoch": 0.00885672351839668, + "grad_norm": 25.808292388916016, + "learning_rate": 4.9853231438838e-05, + "loss": 0.5176, + "step": 350 + }, + { + "epoch": 0.008907333367073233, + "grad_norm": 6.786406517028809, + "learning_rate": 4.9852387941360056e-05, + "loss": 0.9981, + "step": 352 + }, + { + "epoch": 0.008957943215749786, + "grad_norm": 13.001777648925781, + "learning_rate": 4.9851544443882116e-05, + "loss": 0.6546, + "step": 354 + }, + { + "epoch": 0.009008553064426337, + "grad_norm": 5.085973262786865, + "learning_rate": 4.985070094640417e-05, + "loss": 0.4445, + "step": 356 + }, + { + "epoch": 0.00905916291310289, + "grad_norm": 10.445619583129883, + "learning_rate": 4.984985744892623e-05, + "loss": 0.519, + "step": 358 + }, + { + "epoch": 0.009109772761779442, + "grad_norm": 7.5755510330200195, + "learning_rate": 4.984901395144829e-05, + "loss": 0.9253, + "step": 360 + }, + { + "epoch": 0.009160382610455995, + "grad_norm": 12.85741901397705, + "learning_rate": 4.984817045397034e-05, + "loss": 0.7553, + "step": 362 + }, + { + "epoch": 0.009210992459132547, + "grad_norm": 4.291009426116943, + "learning_rate": 4.9847326956492404e-05, + "loss": 0.7912, + "step": 364 + }, + { + "epoch": 0.0092616023078091, + "grad_norm": 9.602625846862793, + "learning_rate": 4.984648345901446e-05, + "loss": 0.7954, + "step": 366 + }, + { + "epoch": 0.009312212156485653, + "grad_norm": 10.099987030029297, + "learning_rate": 4.984563996153652e-05, + "loss": 0.6241, + "step": 368 + }, + { + "epoch": 0.009362822005162204, + "grad_norm": 9.353877067565918, + "learning_rate": 4.984479646405858e-05, + "loss": 0.7845, + "step": 370 + }, + { + "epoch": 0.009413431853838757, + "grad_norm": 5.5436506271362305, + "learning_rate": 4.984395296658063e-05, + "loss": 0.6337, + "step": 372 + }, + { + "epoch": 0.00946404170251531, + "grad_norm": 6.538369655609131, + "learning_rate": 4.984310946910269e-05, + "loss": 0.4818, + "step": 374 + }, + { + "epoch": 0.009514651551191862, + "grad_norm": 11.956756591796875, + "learning_rate": 4.9842265971624744e-05, + "loss": 0.9237, + "step": 376 + }, + { + "epoch": 0.009565261399868415, + "grad_norm": 11.44477367401123, + "learning_rate": 4.9841422474146805e-05, + "loss": 1.0572, + "step": 378 + }, + { + "epoch": 0.009615871248544967, + "grad_norm": 9.988565444946289, + "learning_rate": 4.984057897666886e-05, + "loss": 0.6858, + "step": 380 + }, + { + "epoch": 0.00966648109722152, + "grad_norm": 4.8977813720703125, + "learning_rate": 4.983973547919092e-05, + "loss": 0.5965, + "step": 382 + }, + { + "epoch": 0.009717090945898071, + "grad_norm": 6.308709144592285, + "learning_rate": 4.983889198171298e-05, + "loss": 0.4817, + "step": 384 + }, + { + "epoch": 0.009767700794574624, + "grad_norm": 5.661224842071533, + "learning_rate": 4.983804848423503e-05, + "loss": 0.5738, + "step": 386 + }, + { + "epoch": 0.009818310643251177, + "grad_norm": 5.926614284515381, + "learning_rate": 4.983720498675709e-05, + "loss": 0.5391, + "step": 388 + }, + { + "epoch": 0.00986892049192773, + "grad_norm": 4.607276916503906, + "learning_rate": 4.9836361489279145e-05, + "loss": 0.4223, + "step": 390 + }, + { + "epoch": 0.009919530340604282, + "grad_norm": 14.739141464233398, + "learning_rate": 4.9835517991801206e-05, + "loss": 1.0047, + "step": 392 + }, + { + "epoch": 0.009970140189280835, + "grad_norm": 4.335879325866699, + "learning_rate": 4.9834674494323266e-05, + "loss": 0.7172, + "step": 394 + }, + { + "epoch": 0.010020750037957386, + "grad_norm": 8.950626373291016, + "learning_rate": 4.983383099684532e-05, + "loss": 0.8787, + "step": 396 + }, + { + "epoch": 0.010071359886633938, + "grad_norm": 5.630717754364014, + "learning_rate": 4.983298749936738e-05, + "loss": 0.5813, + "step": 398 + }, + { + "epoch": 0.010121969735310491, + "grad_norm": 9.189420700073242, + "learning_rate": 4.983214400188943e-05, + "loss": 0.9012, + "step": 400 + }, + { + "epoch": 0.010121969735310491, + "eval_cer": 0.22362332855033584, + "eval_loss": 0.3546978533267975, + "eval_runtime": 2758.4684, + "eval_samples_per_second": 5.682, + "eval_steps_per_second": 0.355, + "step": 400 + }, + { + "epoch": 0.010172579583987044, + "grad_norm": 3.5027172565460205, + "learning_rate": 4.983130050441149e-05, + "loss": 0.5343, + "step": 402 + }, + { + "epoch": 0.010223189432663597, + "grad_norm": 7.1599836349487305, + "learning_rate": 4.983045700693355e-05, + "loss": 0.64, + "step": 404 + }, + { + "epoch": 0.01027379928134015, + "grad_norm": 6.179046630859375, + "learning_rate": 4.982961350945561e-05, + "loss": 0.6958, + "step": 406 + }, + { + "epoch": 0.010324409130016702, + "grad_norm": 6.452561855316162, + "learning_rate": 4.982877001197767e-05, + "loss": 0.9579, + "step": 408 + }, + { + "epoch": 0.010375018978693253, + "grad_norm": 6.45066499710083, + "learning_rate": 4.982792651449972e-05, + "loss": 0.8967, + "step": 410 + }, + { + "epoch": 0.010425628827369806, + "grad_norm": 8.554861068725586, + "learning_rate": 4.982708301702178e-05, + "loss": 1.1857, + "step": 412 + }, + { + "epoch": 0.010476238676046358, + "grad_norm": 3.6959705352783203, + "learning_rate": 4.982623951954384e-05, + "loss": 0.4889, + "step": 414 + }, + { + "epoch": 0.010526848524722911, + "grad_norm": 12.848132133483887, + "learning_rate": 4.9825396022065894e-05, + "loss": 0.64, + "step": 416 + }, + { + "epoch": 0.010577458373399464, + "grad_norm": 6.260450839996338, + "learning_rate": 4.9824552524587954e-05, + "loss": 1.0134, + "step": 418 + }, + { + "epoch": 0.010628068222076016, + "grad_norm": 6.700595378875732, + "learning_rate": 4.982370902711001e-05, + "loss": 0.8566, + "step": 420 + }, + { + "epoch": 0.01067867807075257, + "grad_norm": 6.211639404296875, + "learning_rate": 4.982286552963207e-05, + "loss": 0.9721, + "step": 422 + }, + { + "epoch": 0.01072928791942912, + "grad_norm": 7.872774600982666, + "learning_rate": 4.982202203215413e-05, + "loss": 0.4064, + "step": 424 + }, + { + "epoch": 0.010779897768105673, + "grad_norm": 9.730685234069824, + "learning_rate": 4.982117853467618e-05, + "loss": 1.044, + "step": 426 + }, + { + "epoch": 0.010830507616782226, + "grad_norm": 4.908615589141846, + "learning_rate": 4.982033503719824e-05, + "loss": 0.6847, + "step": 428 + }, + { + "epoch": 0.010881117465458778, + "grad_norm": 7.32842493057251, + "learning_rate": 4.9819491539720295e-05, + "loss": 0.8148, + "step": 430 + }, + { + "epoch": 0.010931727314135331, + "grad_norm": 4.537167072296143, + "learning_rate": 4.9818648042242355e-05, + "loss": 0.4907, + "step": 432 + }, + { + "epoch": 0.010982337162811884, + "grad_norm": 5.290032863616943, + "learning_rate": 4.9817804544764416e-05, + "loss": 0.6806, + "step": 434 + }, + { + "epoch": 0.011032947011488436, + "grad_norm": 6.21295166015625, + "learning_rate": 4.981696104728647e-05, + "loss": 0.6636, + "step": 436 + }, + { + "epoch": 0.011083556860164987, + "grad_norm": 5.210526466369629, + "learning_rate": 4.981611754980853e-05, + "loss": 0.7971, + "step": 438 + }, + { + "epoch": 0.01113416670884154, + "grad_norm": 40.482147216796875, + "learning_rate": 4.981527405233058e-05, + "loss": 0.7819, + "step": 440 + }, + { + "epoch": 0.011184776557518093, + "grad_norm": 6.940312385559082, + "learning_rate": 4.981443055485264e-05, + "loss": 0.8548, + "step": 442 + }, + { + "epoch": 0.011235386406194646, + "grad_norm": 10.026724815368652, + "learning_rate": 4.98135870573747e-05, + "loss": 0.5857, + "step": 444 + }, + { + "epoch": 0.011285996254871198, + "grad_norm": 10.645378112792969, + "learning_rate": 4.9812743559896756e-05, + "loss": 0.7426, + "step": 446 + }, + { + "epoch": 0.011336606103547751, + "grad_norm": 17.70381736755371, + "learning_rate": 4.981190006241882e-05, + "loss": 1.0465, + "step": 448 + }, + { + "epoch": 0.011387215952224302, + "grad_norm": 4.017563819885254, + "learning_rate": 4.981105656494087e-05, + "loss": 0.6183, + "step": 450 + }, + { + "epoch": 0.011437825800900855, + "grad_norm": 7.969531059265137, + "learning_rate": 4.981021306746293e-05, + "loss": 0.8661, + "step": 452 + }, + { + "epoch": 0.011488435649577407, + "grad_norm": 12.08060359954834, + "learning_rate": 4.980936956998499e-05, + "loss": 0.7461, + "step": 454 + }, + { + "epoch": 0.01153904549825396, + "grad_norm": 6.960525989532471, + "learning_rate": 4.9808526072507044e-05, + "loss": 0.5767, + "step": 456 + }, + { + "epoch": 0.011589655346930513, + "grad_norm": 10.034902572631836, + "learning_rate": 4.9807682575029104e-05, + "loss": 0.7012, + "step": 458 + }, + { + "epoch": 0.011640265195607066, + "grad_norm": 11.939443588256836, + "learning_rate": 4.980683907755116e-05, + "loss": 0.6601, + "step": 460 + }, + { + "epoch": 0.011690875044283618, + "grad_norm": 5.776790618896484, + "learning_rate": 4.980599558007322e-05, + "loss": 0.7514, + "step": 462 + }, + { + "epoch": 0.01174148489296017, + "grad_norm": 9.15820598602295, + "learning_rate": 4.980515208259527e-05, + "loss": 0.8779, + "step": 464 + }, + { + "epoch": 0.011792094741636722, + "grad_norm": 7.228218078613281, + "learning_rate": 4.980430858511733e-05, + "loss": 1.0643, + "step": 466 + }, + { + "epoch": 0.011842704590313275, + "grad_norm": 4.174966812133789, + "learning_rate": 4.980346508763939e-05, + "loss": 0.3511, + "step": 468 + }, + { + "epoch": 0.011893314438989827, + "grad_norm": 9.501602172851562, + "learning_rate": 4.9802621590161445e-05, + "loss": 0.8323, + "step": 470 + }, + { + "epoch": 0.01194392428766638, + "grad_norm": 4.438223361968994, + "learning_rate": 4.9801778092683505e-05, + "loss": 0.3112, + "step": 472 + }, + { + "epoch": 0.011994534136342933, + "grad_norm": 8.708006858825684, + "learning_rate": 4.980093459520556e-05, + "loss": 0.6483, + "step": 474 + }, + { + "epoch": 0.012045143985019486, + "grad_norm": 4.928430557250977, + "learning_rate": 4.980009109772762e-05, + "loss": 0.7263, + "step": 476 + }, + { + "epoch": 0.012095753833696037, + "grad_norm": 11.870718955993652, + "learning_rate": 4.979924760024968e-05, + "loss": 0.9143, + "step": 478 + }, + { + "epoch": 0.01214636368237259, + "grad_norm": 6.654867649078369, + "learning_rate": 4.979840410277173e-05, + "loss": 0.5257, + "step": 480 + }, + { + "epoch": 0.012196973531049142, + "grad_norm": 4.725414752960205, + "learning_rate": 4.979756060529379e-05, + "loss": 0.5919, + "step": 482 + }, + { + "epoch": 0.012247583379725695, + "grad_norm": 5.607127666473389, + "learning_rate": 4.9796717107815846e-05, + "loss": 0.5198, + "step": 484 + }, + { + "epoch": 0.012298193228402247, + "grad_norm": 5.8716864585876465, + "learning_rate": 4.9795873610337906e-05, + "loss": 0.7554, + "step": 486 + }, + { + "epoch": 0.0123488030770788, + "grad_norm": 7.410449504852295, + "learning_rate": 4.9795030112859966e-05, + "loss": 0.8967, + "step": 488 + }, + { + "epoch": 0.012399412925755351, + "grad_norm": 10.435539245605469, + "learning_rate": 4.979418661538202e-05, + "loss": 0.7446, + "step": 490 + }, + { + "epoch": 0.012450022774431904, + "grad_norm": 4.085461139678955, + "learning_rate": 4.979334311790408e-05, + "loss": 0.56, + "step": 492 + }, + { + "epoch": 0.012500632623108457, + "grad_norm": 7.522024154663086, + "learning_rate": 4.9792499620426133e-05, + "loss": 0.7217, + "step": 494 + }, + { + "epoch": 0.01255124247178501, + "grad_norm": 6.347813606262207, + "learning_rate": 4.9791656122948194e-05, + "loss": 0.5247, + "step": 496 + }, + { + "epoch": 0.012601852320461562, + "grad_norm": 8.447410583496094, + "learning_rate": 4.9790812625470254e-05, + "loss": 0.74, + "step": 498 + }, + { + "epoch": 0.012652462169138115, + "grad_norm": 5.143301486968994, + "learning_rate": 4.978996912799231e-05, + "loss": 0.5553, + "step": 500 + }, + { + "epoch": 0.012703072017814667, + "grad_norm": 5.69724702835083, + "learning_rate": 4.978912563051437e-05, + "loss": 0.9869, + "step": 502 + }, + { + "epoch": 0.012753681866491218, + "grad_norm": 3.5676660537719727, + "learning_rate": 4.978828213303642e-05, + "loss": 0.423, + "step": 504 + }, + { + "epoch": 0.012804291715167771, + "grad_norm": 5.91968297958374, + "learning_rate": 4.978743863555848e-05, + "loss": 0.7037, + "step": 506 + }, + { + "epoch": 0.012854901563844324, + "grad_norm": 9.412429809570312, + "learning_rate": 4.978659513808054e-05, + "loss": 1.1408, + "step": 508 + }, + { + "epoch": 0.012905511412520877, + "grad_norm": 8.342994689941406, + "learning_rate": 4.9785751640602595e-05, + "loss": 0.8661, + "step": 510 + }, + { + "epoch": 0.01295612126119743, + "grad_norm": 7.621118068695068, + "learning_rate": 4.9784908143124655e-05, + "loss": 0.677, + "step": 512 + }, + { + "epoch": 0.013006731109873982, + "grad_norm": 4.431066036224365, + "learning_rate": 4.978406464564671e-05, + "loss": 0.637, + "step": 514 + }, + { + "epoch": 0.013057340958550535, + "grad_norm": 9.607247352600098, + "learning_rate": 4.978322114816877e-05, + "loss": 0.5467, + "step": 516 + }, + { + "epoch": 0.013107950807227086, + "grad_norm": 8.333694458007812, + "learning_rate": 4.978237765069083e-05, + "loss": 1.0044, + "step": 518 + }, + { + "epoch": 0.013158560655903638, + "grad_norm": 3.5092146396636963, + "learning_rate": 4.978153415321288e-05, + "loss": 0.3496, + "step": 520 + }, + { + "epoch": 0.013209170504580191, + "grad_norm": 5.619466781616211, + "learning_rate": 4.978069065573494e-05, + "loss": 0.5608, + "step": 522 + }, + { + "epoch": 0.013259780353256744, + "grad_norm": 10.299610137939453, + "learning_rate": 4.9779847158256996e-05, + "loss": 0.7437, + "step": 524 + }, + { + "epoch": 0.013310390201933297, + "grad_norm": 9.14527416229248, + "learning_rate": 4.9779003660779056e-05, + "loss": 0.7961, + "step": 526 + }, + { + "epoch": 0.01336100005060985, + "grad_norm": 4.311985015869141, + "learning_rate": 4.9778160163301116e-05, + "loss": 0.5069, + "step": 528 + }, + { + "epoch": 0.013411609899286402, + "grad_norm": 7.6014862060546875, + "learning_rate": 4.977731666582317e-05, + "loss": 0.8133, + "step": 530 + }, + { + "epoch": 0.013462219747962953, + "grad_norm": 5.695495128631592, + "learning_rate": 4.977647316834523e-05, + "loss": 0.4957, + "step": 532 + }, + { + "epoch": 0.013512829596639506, + "grad_norm": 8.789469718933105, + "learning_rate": 4.977562967086728e-05, + "loss": 0.9009, + "step": 534 + }, + { + "epoch": 0.013563439445316058, + "grad_norm": 6.521892547607422, + "learning_rate": 4.977478617338934e-05, + "loss": 0.4392, + "step": 536 + }, + { + "epoch": 0.013614049293992611, + "grad_norm": 8.529374122619629, + "learning_rate": 4.9773942675911404e-05, + "loss": 0.507, + "step": 538 + }, + { + "epoch": 0.013664659142669164, + "grad_norm": 5.722716808319092, + "learning_rate": 4.977309917843346e-05, + "loss": 0.5071, + "step": 540 + }, + { + "epoch": 0.013715268991345717, + "grad_norm": 4.3265156745910645, + "learning_rate": 4.977225568095552e-05, + "loss": 0.7658, + "step": 542 + }, + { + "epoch": 0.013765878840022268, + "grad_norm": 5.346278190612793, + "learning_rate": 4.977141218347757e-05, + "loss": 0.5658, + "step": 544 + }, + { + "epoch": 0.01381648868869882, + "grad_norm": 7.274371147155762, + "learning_rate": 4.977056868599963e-05, + "loss": 0.8856, + "step": 546 + }, + { + "epoch": 0.013867098537375373, + "grad_norm": 5.354006767272949, + "learning_rate": 4.9769725188521684e-05, + "loss": 0.5224, + "step": 548 + }, + { + "epoch": 0.013917708386051926, + "grad_norm": 8.813762664794922, + "learning_rate": 4.9768881691043744e-05, + "loss": 1.1558, + "step": 550 + }, + { + "epoch": 0.013968318234728478, + "grad_norm": 6.957085132598877, + "learning_rate": 4.9768038193565805e-05, + "loss": 0.4142, + "step": 552 + }, + { + "epoch": 0.014018928083405031, + "grad_norm": 12.203534126281738, + "learning_rate": 4.976719469608786e-05, + "loss": 0.5945, + "step": 554 + }, + { + "epoch": 0.014069537932081584, + "grad_norm": 6.845212459564209, + "learning_rate": 4.976635119860992e-05, + "loss": 0.4359, + "step": 556 + }, + { + "epoch": 0.014120147780758135, + "grad_norm": 6.074872970581055, + "learning_rate": 4.976550770113197e-05, + "loss": 0.7704, + "step": 558 + }, + { + "epoch": 0.014170757629434688, + "grad_norm": 5.776919364929199, + "learning_rate": 4.976466420365403e-05, + "loss": 0.7799, + "step": 560 + }, + { + "epoch": 0.01422136747811124, + "grad_norm": 11.90626335144043, + "learning_rate": 4.976382070617609e-05, + "loss": 0.9279, + "step": 562 + }, + { + "epoch": 0.014271977326787793, + "grad_norm": 4.0126848220825195, + "learning_rate": 4.9762977208698145e-05, + "loss": 0.8148, + "step": 564 + }, + { + "epoch": 0.014322587175464346, + "grad_norm": 5.540651798248291, + "learning_rate": 4.9762133711220206e-05, + "loss": 0.528, + "step": 566 + }, + { + "epoch": 0.014373197024140898, + "grad_norm": 4.119876861572266, + "learning_rate": 4.976129021374226e-05, + "loss": 0.8705, + "step": 568 + }, + { + "epoch": 0.014423806872817451, + "grad_norm": 11.188005447387695, + "learning_rate": 4.976044671626432e-05, + "loss": 0.467, + "step": 570 + }, + { + "epoch": 0.014474416721494002, + "grad_norm": 6.426743030548096, + "learning_rate": 4.975960321878638e-05, + "loss": 0.5936, + "step": 572 + }, + { + "epoch": 0.014525026570170555, + "grad_norm": 6.039766788482666, + "learning_rate": 4.975875972130843e-05, + "loss": 0.7222, + "step": 574 + }, + { + "epoch": 0.014575636418847107, + "grad_norm": 12.856761932373047, + "learning_rate": 4.975791622383049e-05, + "loss": 0.501, + "step": 576 + }, + { + "epoch": 0.01462624626752366, + "grad_norm": 13.36010456085205, + "learning_rate": 4.9757072726352547e-05, + "loss": 0.473, + "step": 578 + }, + { + "epoch": 0.014676856116200213, + "grad_norm": 13.758210182189941, + "learning_rate": 4.975622922887461e-05, + "loss": 1.1569, + "step": 580 + }, + { + "epoch": 0.014727465964876766, + "grad_norm": 5.978826522827148, + "learning_rate": 4.975538573139667e-05, + "loss": 0.5853, + "step": 582 + }, + { + "epoch": 0.014778075813553317, + "grad_norm": 4.91432523727417, + "learning_rate": 4.975454223391872e-05, + "loss": 0.5073, + "step": 584 + }, + { + "epoch": 0.01482868566222987, + "grad_norm": 4.914419174194336, + "learning_rate": 4.975369873644078e-05, + "loss": 0.5638, + "step": 586 + }, + { + "epoch": 0.014879295510906422, + "grad_norm": 6.7935967445373535, + "learning_rate": 4.9752855238962834e-05, + "loss": 0.773, + "step": 588 + }, + { + "epoch": 0.014929905359582975, + "grad_norm": 5.500904083251953, + "learning_rate": 4.9752011741484894e-05, + "loss": 0.6985, + "step": 590 + }, + { + "epoch": 0.014980515208259527, + "grad_norm": 10.455467224121094, + "learning_rate": 4.9751168244006954e-05, + "loss": 0.5856, + "step": 592 + }, + { + "epoch": 0.01503112505693608, + "grad_norm": 14.375707626342773, + "learning_rate": 4.975032474652901e-05, + "loss": 0.8323, + "step": 594 + }, + { + "epoch": 0.015081734905612633, + "grad_norm": 8.137022972106934, + "learning_rate": 4.974948124905107e-05, + "loss": 0.7872, + "step": 596 + }, + { + "epoch": 0.015132344754289184, + "grad_norm": 4.684526443481445, + "learning_rate": 4.974863775157312e-05, + "loss": 0.7216, + "step": 598 + }, + { + "epoch": 0.015182954602965737, + "grad_norm": 6.151386737823486, + "learning_rate": 4.974779425409518e-05, + "loss": 0.7315, + "step": 600 + }, + { + "epoch": 0.015182954602965737, + "eval_cer": 0.21082101191590244, + "eval_loss": 0.33135300874710083, + "eval_runtime": 2751.8944, + "eval_samples_per_second": 5.696, + "eval_steps_per_second": 0.356, + "step": 600 + }, + { + "epoch": 0.01523356445164229, + "grad_norm": 7.473769664764404, + "learning_rate": 4.974695075661724e-05, + "loss": 0.3884, + "step": 602 + }, + { + "epoch": 0.015284174300318842, + "grad_norm": 6.925198078155518, + "learning_rate": 4.9746107259139295e-05, + "loss": 0.6214, + "step": 604 + }, + { + "epoch": 0.015334784148995395, + "grad_norm": 7.6508684158325195, + "learning_rate": 4.9745263761661355e-05, + "loss": 0.3224, + "step": 606 + }, + { + "epoch": 0.015385393997671947, + "grad_norm": 8.625561714172363, + "learning_rate": 4.974442026418341e-05, + "loss": 1.1576, + "step": 608 + }, + { + "epoch": 0.0154360038463485, + "grad_norm": 4.799185276031494, + "learning_rate": 4.974357676670547e-05, + "loss": 0.4953, + "step": 610 + }, + { + "epoch": 0.015486613695025051, + "grad_norm": 6.022134780883789, + "learning_rate": 4.974273326922753e-05, + "loss": 0.417, + "step": 612 + }, + { + "epoch": 0.015537223543701604, + "grad_norm": 6.1436333656311035, + "learning_rate": 4.974188977174958e-05, + "loss": 0.7479, + "step": 614 + }, + { + "epoch": 0.015587833392378157, + "grad_norm": 11.182157516479492, + "learning_rate": 4.974104627427164e-05, + "loss": 1.1078, + "step": 616 + }, + { + "epoch": 0.01563844324105471, + "grad_norm": 7.718019485473633, + "learning_rate": 4.9740202776793696e-05, + "loss": 0.7953, + "step": 618 + }, + { + "epoch": 0.015689053089731262, + "grad_norm": 6.890071392059326, + "learning_rate": 4.9739359279315756e-05, + "loss": 0.8054, + "step": 620 + }, + { + "epoch": 0.015739662938407815, + "grad_norm": 9.046427726745605, + "learning_rate": 4.973851578183782e-05, + "loss": 0.8913, + "step": 622 + }, + { + "epoch": 0.015790272787084367, + "grad_norm": 7.701432704925537, + "learning_rate": 4.973767228435987e-05, + "loss": 1.2825, + "step": 624 + }, + { + "epoch": 0.01584088263576092, + "grad_norm": 6.6972808837890625, + "learning_rate": 4.973682878688193e-05, + "loss": 0.7063, + "step": 626 + }, + { + "epoch": 0.015891492484437473, + "grad_norm": 7.926259517669678, + "learning_rate": 4.9735985289403984e-05, + "loss": 0.6431, + "step": 628 + }, + { + "epoch": 0.015942102333114026, + "grad_norm": 6.506617546081543, + "learning_rate": 4.9735141791926044e-05, + "loss": 0.7283, + "step": 630 + }, + { + "epoch": 0.015992712181790575, + "grad_norm": 5.657668590545654, + "learning_rate": 4.97342982944481e-05, + "loss": 0.7906, + "step": 632 + }, + { + "epoch": 0.016043322030467128, + "grad_norm": 5.814841270446777, + "learning_rate": 4.973345479697016e-05, + "loss": 0.7872, + "step": 634 + }, + { + "epoch": 0.01609393187914368, + "grad_norm": 4.063594818115234, + "learning_rate": 4.973261129949222e-05, + "loss": 0.7104, + "step": 636 + }, + { + "epoch": 0.016144541727820233, + "grad_norm": 2.756787061691284, + "learning_rate": 4.973176780201427e-05, + "loss": 0.3609, + "step": 638 + }, + { + "epoch": 0.016195151576496786, + "grad_norm": 6.163934707641602, + "learning_rate": 4.973092430453633e-05, + "loss": 0.7657, + "step": 640 + }, + { + "epoch": 0.01624576142517334, + "grad_norm": 7.708083629608154, + "learning_rate": 4.9730080807058385e-05, + "loss": 0.7486, + "step": 642 + }, + { + "epoch": 0.01629637127384989, + "grad_norm": 6.693125247955322, + "learning_rate": 4.9729237309580445e-05, + "loss": 0.799, + "step": 644 + }, + { + "epoch": 0.016346981122526444, + "grad_norm": 8.108495712280273, + "learning_rate": 4.9728393812102505e-05, + "loss": 0.8542, + "step": 646 + }, + { + "epoch": 0.016397590971202997, + "grad_norm": 13.466230392456055, + "learning_rate": 4.972755031462456e-05, + "loss": 0.9259, + "step": 648 + }, + { + "epoch": 0.01644820081987955, + "grad_norm": 7.1850056648254395, + "learning_rate": 4.972670681714662e-05, + "loss": 0.4547, + "step": 650 + }, + { + "epoch": 0.016498810668556102, + "grad_norm": 10.240496635437012, + "learning_rate": 4.972586331966867e-05, + "loss": 0.7534, + "step": 652 + }, + { + "epoch": 0.016549420517232655, + "grad_norm": 7.284756183624268, + "learning_rate": 4.972501982219073e-05, + "loss": 0.7111, + "step": 654 + }, + { + "epoch": 0.016600030365909207, + "grad_norm": 4.935502052307129, + "learning_rate": 4.972417632471279e-05, + "loss": 0.587, + "step": 656 + }, + { + "epoch": 0.016650640214585757, + "grad_norm": 9.42276382446289, + "learning_rate": 4.9723332827234846e-05, + "loss": 0.6914, + "step": 658 + }, + { + "epoch": 0.01670125006326231, + "grad_norm": 8.303751945495605, + "learning_rate": 4.9722489329756906e-05, + "loss": 0.7424, + "step": 660 + }, + { + "epoch": 0.016751859911938862, + "grad_norm": 5.425510883331299, + "learning_rate": 4.972164583227896e-05, + "loss": 0.4959, + "step": 662 + }, + { + "epoch": 0.016802469760615415, + "grad_norm": 8.581243515014648, + "learning_rate": 4.972080233480102e-05, + "loss": 0.5844, + "step": 664 + }, + { + "epoch": 0.016853079609291968, + "grad_norm": 7.702110290527344, + "learning_rate": 4.971995883732308e-05, + "loss": 0.6744, + "step": 666 + }, + { + "epoch": 0.01690368945796852, + "grad_norm": 5.569732189178467, + "learning_rate": 4.9719115339845133e-05, + "loss": 0.5237, + "step": 668 + }, + { + "epoch": 0.016954299306645073, + "grad_norm": 7.927607536315918, + "learning_rate": 4.9718271842367194e-05, + "loss": 0.6969, + "step": 670 + }, + { + "epoch": 0.017004909155321626, + "grad_norm": 6.792113304138184, + "learning_rate": 4.971742834488925e-05, + "loss": 0.4953, + "step": 672 + }, + { + "epoch": 0.01705551900399818, + "grad_norm": 41.776954650878906, + "learning_rate": 4.971658484741131e-05, + "loss": 0.5495, + "step": 674 + }, + { + "epoch": 0.01710612885267473, + "grad_norm": 5.085512638092041, + "learning_rate": 4.971574134993337e-05, + "loss": 0.6303, + "step": 676 + }, + { + "epoch": 0.017156738701351284, + "grad_norm": 5.323431491851807, + "learning_rate": 4.971489785245542e-05, + "loss": 0.7973, + "step": 678 + }, + { + "epoch": 0.017207348550027837, + "grad_norm": 5.844650745391846, + "learning_rate": 4.971405435497748e-05, + "loss": 1.0397, + "step": 680 + }, + { + "epoch": 0.01725795839870439, + "grad_norm": 1.8864692449569702, + "learning_rate": 4.9713210857499535e-05, + "loss": 0.3274, + "step": 682 + }, + { + "epoch": 0.017308568247380942, + "grad_norm": 3.4474141597747803, + "learning_rate": 4.9712367360021595e-05, + "loss": 0.5992, + "step": 684 + }, + { + "epoch": 0.01735917809605749, + "grad_norm": 5.180785655975342, + "learning_rate": 4.9711523862543655e-05, + "loss": 0.6462, + "step": 686 + }, + { + "epoch": 0.017409787944734044, + "grad_norm": 5.768479347229004, + "learning_rate": 4.971068036506571e-05, + "loss": 0.6288, + "step": 688 + }, + { + "epoch": 0.017460397793410597, + "grad_norm": 5.524656772613525, + "learning_rate": 4.970983686758777e-05, + "loss": 0.6045, + "step": 690 + }, + { + "epoch": 0.01751100764208715, + "grad_norm": 6.222484111785889, + "learning_rate": 4.970899337010982e-05, + "loss": 0.4882, + "step": 692 + }, + { + "epoch": 0.017561617490763702, + "grad_norm": 2.960340738296509, + "learning_rate": 4.970814987263188e-05, + "loss": 0.3803, + "step": 694 + }, + { + "epoch": 0.017612227339440255, + "grad_norm": 3.2236385345458984, + "learning_rate": 4.970730637515394e-05, + "loss": 0.5617, + "step": 696 + }, + { + "epoch": 0.017662837188116808, + "grad_norm": 8.282496452331543, + "learning_rate": 4.9706462877675996e-05, + "loss": 0.6835, + "step": 698 + }, + { + "epoch": 0.01771344703679336, + "grad_norm": 7.472006797790527, + "learning_rate": 4.9705619380198056e-05, + "loss": 0.4747, + "step": 700 + }, + { + "epoch": 0.017764056885469913, + "grad_norm": 6.8331780433654785, + "learning_rate": 4.970477588272011e-05, + "loss": 0.4765, + "step": 702 + }, + { + "epoch": 0.017814666734146466, + "grad_norm": 4.116718292236328, + "learning_rate": 4.970393238524217e-05, + "loss": 0.4177, + "step": 704 + }, + { + "epoch": 0.01786527658282302, + "grad_norm": 14.779850959777832, + "learning_rate": 4.970308888776423e-05, + "loss": 0.5939, + "step": 706 + }, + { + "epoch": 0.01791588643149957, + "grad_norm": 10.066930770874023, + "learning_rate": 4.970224539028628e-05, + "loss": 0.7392, + "step": 708 + }, + { + "epoch": 0.017966496280176124, + "grad_norm": 5.495025634765625, + "learning_rate": 4.9701401892808343e-05, + "loss": 0.6957, + "step": 710 + }, + { + "epoch": 0.018017106128852673, + "grad_norm": 9.583708763122559, + "learning_rate": 4.97005583953304e-05, + "loss": 0.8032, + "step": 712 + }, + { + "epoch": 0.018067715977529226, + "grad_norm": 6.409923076629639, + "learning_rate": 4.969971489785246e-05, + "loss": 0.9505, + "step": 714 + }, + { + "epoch": 0.01811832582620578, + "grad_norm": 8.952066421508789, + "learning_rate": 4.969887140037451e-05, + "loss": 0.7739, + "step": 716 + }, + { + "epoch": 0.01816893567488233, + "grad_norm": 6.334309101104736, + "learning_rate": 4.969802790289657e-05, + "loss": 1.0117, + "step": 718 + }, + { + "epoch": 0.018219545523558884, + "grad_norm": 8.889698028564453, + "learning_rate": 4.969718440541863e-05, + "loss": 0.6535, + "step": 720 + }, + { + "epoch": 0.018270155372235437, + "grad_norm": 2.9195971488952637, + "learning_rate": 4.9696340907940684e-05, + "loss": 0.3021, + "step": 722 + }, + { + "epoch": 0.01832076522091199, + "grad_norm": 8.154908180236816, + "learning_rate": 4.9695497410462744e-05, + "loss": 0.8477, + "step": 724 + }, + { + "epoch": 0.018371375069588542, + "grad_norm": 3.179515838623047, + "learning_rate": 4.96946539129848e-05, + "loss": 0.5856, + "step": 726 + }, + { + "epoch": 0.018421984918265095, + "grad_norm": 7.491313457489014, + "learning_rate": 4.969381041550686e-05, + "loss": 0.8113, + "step": 728 + }, + { + "epoch": 0.018472594766941648, + "grad_norm": 6.968503475189209, + "learning_rate": 4.969296691802892e-05, + "loss": 0.6095, + "step": 730 + }, + { + "epoch": 0.0185232046156182, + "grad_norm": 7.4488348960876465, + "learning_rate": 4.969212342055097e-05, + "loss": 1.0515, + "step": 732 + }, + { + "epoch": 0.018573814464294753, + "grad_norm": 2.959810733795166, + "learning_rate": 4.969127992307303e-05, + "loss": 0.5783, + "step": 734 + }, + { + "epoch": 0.018624424312971306, + "grad_norm": 9.092787742614746, + "learning_rate": 4.9690436425595085e-05, + "loss": 0.878, + "step": 736 + }, + { + "epoch": 0.01867503416164786, + "grad_norm": 7.564940929412842, + "learning_rate": 4.9689592928117146e-05, + "loss": 0.8122, + "step": 738 + }, + { + "epoch": 0.018725644010324408, + "grad_norm": 7.550844192504883, + "learning_rate": 4.9688749430639206e-05, + "loss": 0.4964, + "step": 740 + }, + { + "epoch": 0.01877625385900096, + "grad_norm": 11.271805763244629, + "learning_rate": 4.968790593316126e-05, + "loss": 0.9631, + "step": 742 + }, + { + "epoch": 0.018826863707677513, + "grad_norm": 13.20101547241211, + "learning_rate": 4.968706243568332e-05, + "loss": 0.6274, + "step": 744 + }, + { + "epoch": 0.018877473556354066, + "grad_norm": 5.18681526184082, + "learning_rate": 4.968621893820537e-05, + "loss": 0.6049, + "step": 746 + }, + { + "epoch": 0.01892808340503062, + "grad_norm": 3.5832290649414062, + "learning_rate": 4.968537544072743e-05, + "loss": 0.363, + "step": 748 + }, + { + "epoch": 0.01897869325370717, + "grad_norm": 4.764203071594238, + "learning_rate": 4.968453194324949e-05, + "loss": 0.4824, + "step": 750 + }, + { + "epoch": 0.019029303102383724, + "grad_norm": 3.9694879055023193, + "learning_rate": 4.9683688445771547e-05, + "loss": 0.6074, + "step": 752 + }, + { + "epoch": 0.019079912951060277, + "grad_norm": 4.219832897186279, + "learning_rate": 4.968284494829361e-05, + "loss": 0.3649, + "step": 754 + }, + { + "epoch": 0.01913052279973683, + "grad_norm": 25.03742027282715, + "learning_rate": 4.968200145081566e-05, + "loss": 0.742, + "step": 756 + }, + { + "epoch": 0.019181132648413382, + "grad_norm": 8.844772338867188, + "learning_rate": 4.968115795333772e-05, + "loss": 0.6859, + "step": 758 + }, + { + "epoch": 0.019231742497089935, + "grad_norm": 7.3545918464660645, + "learning_rate": 4.968031445585978e-05, + "loss": 0.5395, + "step": 760 + }, + { + "epoch": 0.019282352345766487, + "grad_norm": 3.499608278274536, + "learning_rate": 4.9679470958381834e-05, + "loss": 0.613, + "step": 762 + }, + { + "epoch": 0.01933296219444304, + "grad_norm": 4.7259440422058105, + "learning_rate": 4.9678627460903894e-05, + "loss": 0.7538, + "step": 764 + }, + { + "epoch": 0.01938357204311959, + "grad_norm": 5.475869178771973, + "learning_rate": 4.967778396342595e-05, + "loss": 0.7714, + "step": 766 + }, + { + "epoch": 0.019434181891796142, + "grad_norm": 18.313730239868164, + "learning_rate": 4.967694046594801e-05, + "loss": 0.9697, + "step": 768 + }, + { + "epoch": 0.019484791740472695, + "grad_norm": 8.159904479980469, + "learning_rate": 4.967609696847007e-05, + "loss": 0.5248, + "step": 770 + }, + { + "epoch": 0.019535401589149248, + "grad_norm": 5.177513122558594, + "learning_rate": 4.967525347099212e-05, + "loss": 0.4911, + "step": 772 + }, + { + "epoch": 0.0195860114378258, + "grad_norm": 4.758183479309082, + "learning_rate": 4.967440997351418e-05, + "loss": 0.7151, + "step": 774 + }, + { + "epoch": 0.019636621286502353, + "grad_norm": 8.288613319396973, + "learning_rate": 4.9673566476036235e-05, + "loss": 0.49, + "step": 776 + }, + { + "epoch": 0.019687231135178906, + "grad_norm": 8.448434829711914, + "learning_rate": 4.9672722978558295e-05, + "loss": 0.7352, + "step": 778 + }, + { + "epoch": 0.01973784098385546, + "grad_norm": 11.22861385345459, + "learning_rate": 4.9671879481080355e-05, + "loss": 0.4209, + "step": 780 + }, + { + "epoch": 0.01978845083253201, + "grad_norm": 8.725863456726074, + "learning_rate": 4.967103598360241e-05, + "loss": 0.5673, + "step": 782 + }, + { + "epoch": 0.019839060681208564, + "grad_norm": 6.322774410247803, + "learning_rate": 4.967019248612447e-05, + "loss": 0.6288, + "step": 784 + }, + { + "epoch": 0.019889670529885117, + "grad_norm": 3.711097240447998, + "learning_rate": 4.966934898864652e-05, + "loss": 0.6343, + "step": 786 + }, + { + "epoch": 0.01994028037856167, + "grad_norm": 7.519350528717041, + "learning_rate": 4.966850549116858e-05, + "loss": 0.7308, + "step": 788 + }, + { + "epoch": 0.019990890227238222, + "grad_norm": 7.28798246383667, + "learning_rate": 4.966766199369064e-05, + "loss": 0.5802, + "step": 790 + }, + { + "epoch": 0.02004150007591477, + "grad_norm": 5.8284783363342285, + "learning_rate": 4.9666818496212696e-05, + "loss": 0.5121, + "step": 792 + }, + { + "epoch": 0.020092109924591324, + "grad_norm": 6.361229419708252, + "learning_rate": 4.9665974998734757e-05, + "loss": 0.7215, + "step": 794 + }, + { + "epoch": 0.020142719773267877, + "grad_norm": 5.134431838989258, + "learning_rate": 4.966513150125681e-05, + "loss": 0.4415, + "step": 796 + }, + { + "epoch": 0.02019332962194443, + "grad_norm": 6.047237873077393, + "learning_rate": 4.966428800377887e-05, + "loss": 0.8233, + "step": 798 + }, + { + "epoch": 0.020243939470620982, + "grad_norm": 3.95216965675354, + "learning_rate": 4.9663444506300924e-05, + "loss": 0.3292, + "step": 800 + }, + { + "epoch": 0.020243939470620982, + "eval_cer": 0.21470199846112256, + "eval_loss": 0.33481982350349426, + "eval_runtime": 2771.1883, + "eval_samples_per_second": 5.656, + "eval_steps_per_second": 0.354, + "step": 800 + }, + { + "epoch": 0.020294549319297535, + "grad_norm": 11.878438949584961, + "learning_rate": 4.9662601008822984e-05, + "loss": 0.6688, + "step": 802 + }, + { + "epoch": 0.020345159167974088, + "grad_norm": 4.3031392097473145, + "learning_rate": 4.9661757511345044e-05, + "loss": 0.5529, + "step": 804 + }, + { + "epoch": 0.02039576901665064, + "grad_norm": 3.877115249633789, + "learning_rate": 4.96609140138671e-05, + "loss": 0.4497, + "step": 806 + }, + { + "epoch": 0.020446378865327193, + "grad_norm": 4.481441497802734, + "learning_rate": 4.966007051638916e-05, + "loss": 0.5141, + "step": 808 + }, + { + "epoch": 0.020496988714003746, + "grad_norm": 8.904651641845703, + "learning_rate": 4.965922701891121e-05, + "loss": 0.5108, + "step": 810 + }, + { + "epoch": 0.0205475985626803, + "grad_norm": 7.119351863861084, + "learning_rate": 4.965838352143327e-05, + "loss": 1.0192, + "step": 812 + }, + { + "epoch": 0.02059820841135685, + "grad_norm": 9.75696849822998, + "learning_rate": 4.965754002395533e-05, + "loss": 0.8766, + "step": 814 + }, + { + "epoch": 0.020648818260033404, + "grad_norm": 3.441471576690674, + "learning_rate": 4.9656696526477385e-05, + "loss": 0.4397, + "step": 816 + }, + { + "epoch": 0.020699428108709957, + "grad_norm": 5.399097442626953, + "learning_rate": 4.9655853028999445e-05, + "loss": 0.8645, + "step": 818 + }, + { + "epoch": 0.020750037957386506, + "grad_norm": 8.58830738067627, + "learning_rate": 4.96550095315215e-05, + "loss": 0.6172, + "step": 820 + }, + { + "epoch": 0.02080064780606306, + "grad_norm": 4.000862121582031, + "learning_rate": 4.965416603404356e-05, + "loss": 0.4244, + "step": 822 + }, + { + "epoch": 0.02085125765473961, + "grad_norm": 5.518575668334961, + "learning_rate": 4.965332253656562e-05, + "loss": 0.7934, + "step": 824 + }, + { + "epoch": 0.020901867503416164, + "grad_norm": 5.466125011444092, + "learning_rate": 4.965247903908767e-05, + "loss": 0.5228, + "step": 826 + }, + { + "epoch": 0.020952477352092717, + "grad_norm": 8.060519218444824, + "learning_rate": 4.965163554160973e-05, + "loss": 0.7465, + "step": 828 + }, + { + "epoch": 0.02100308720076927, + "grad_norm": 7.969659805297852, + "learning_rate": 4.9650792044131786e-05, + "loss": 0.7079, + "step": 830 + }, + { + "epoch": 0.021053697049445822, + "grad_norm": 5.721604347229004, + "learning_rate": 4.9649948546653846e-05, + "loss": 0.5805, + "step": 832 + }, + { + "epoch": 0.021104306898122375, + "grad_norm": 5.849686145782471, + "learning_rate": 4.9649105049175906e-05, + "loss": 0.7729, + "step": 834 + }, + { + "epoch": 0.021154916746798928, + "grad_norm": 3.3855998516082764, + "learning_rate": 4.964826155169796e-05, + "loss": 0.7193, + "step": 836 + }, + { + "epoch": 0.02120552659547548, + "grad_norm": 3.7108068466186523, + "learning_rate": 4.964741805422002e-05, + "loss": 0.5297, + "step": 838 + }, + { + "epoch": 0.021256136444152033, + "grad_norm": 7.068669319152832, + "learning_rate": 4.964657455674207e-05, + "loss": 0.8856, + "step": 840 + }, + { + "epoch": 0.021306746292828586, + "grad_norm": 3.230989694595337, + "learning_rate": 4.9645731059264133e-05, + "loss": 0.4015, + "step": 842 + }, + { + "epoch": 0.02135735614150514, + "grad_norm": 3.4090418815612793, + "learning_rate": 4.9644887561786194e-05, + "loss": 0.2911, + "step": 844 + }, + { + "epoch": 0.021407965990181688, + "grad_norm": 4.298464298248291, + "learning_rate": 4.964404406430825e-05, + "loss": 0.6512, + "step": 846 + }, + { + "epoch": 0.02145857583885824, + "grad_norm": 2.939181327819824, + "learning_rate": 4.964320056683031e-05, + "loss": 0.4267, + "step": 848 + }, + { + "epoch": 0.021509185687534793, + "grad_norm": 6.4090728759765625, + "learning_rate": 4.964235706935236e-05, + "loss": 0.5932, + "step": 850 + }, + { + "epoch": 0.021559795536211346, + "grad_norm": 10.887248992919922, + "learning_rate": 4.964151357187442e-05, + "loss": 0.8047, + "step": 852 + }, + { + "epoch": 0.0216104053848879, + "grad_norm": 5.774937629699707, + "learning_rate": 4.964067007439648e-05, + "loss": 0.7478, + "step": 854 + }, + { + "epoch": 0.02166101523356445, + "grad_norm": 4.135079860687256, + "learning_rate": 4.9639826576918535e-05, + "loss": 0.5991, + "step": 856 + }, + { + "epoch": 0.021711625082241004, + "grad_norm": 5.668390274047852, + "learning_rate": 4.9638983079440595e-05, + "loss": 0.4716, + "step": 858 + }, + { + "epoch": 0.021762234930917557, + "grad_norm": 2.0646257400512695, + "learning_rate": 4.963813958196265e-05, + "loss": 0.3783, + "step": 860 + }, + { + "epoch": 0.02181284477959411, + "grad_norm": 5.5543532371521, + "learning_rate": 4.963729608448471e-05, + "loss": 0.8937, + "step": 862 + }, + { + "epoch": 0.021863454628270662, + "grad_norm": 4.02618408203125, + "learning_rate": 4.963645258700677e-05, + "loss": 0.4268, + "step": 864 + }, + { + "epoch": 0.021914064476947215, + "grad_norm": 5.774693012237549, + "learning_rate": 4.963560908952882e-05, + "loss": 0.7588, + "step": 866 + }, + { + "epoch": 0.021964674325623768, + "grad_norm": 3.130143880844116, + "learning_rate": 4.963476559205088e-05, + "loss": 0.4623, + "step": 868 + }, + { + "epoch": 0.02201528417430032, + "grad_norm": 7.7838521003723145, + "learning_rate": 4.9633922094572936e-05, + "loss": 0.8172, + "step": 870 + }, + { + "epoch": 0.022065894022976873, + "grad_norm": 7.278140544891357, + "learning_rate": 4.9633078597094996e-05, + "loss": 0.4937, + "step": 872 + }, + { + "epoch": 0.022116503871653422, + "grad_norm": 3.3574812412261963, + "learning_rate": 4.9632235099617056e-05, + "loss": 0.4263, + "step": 874 + }, + { + "epoch": 0.022167113720329975, + "grad_norm": 5.792145729064941, + "learning_rate": 4.963139160213911e-05, + "loss": 0.6564, + "step": 876 + }, + { + "epoch": 0.022217723569006528, + "grad_norm": 4.801455020904541, + "learning_rate": 4.963054810466117e-05, + "loss": 0.4392, + "step": 878 + }, + { + "epoch": 0.02226833341768308, + "grad_norm": 5.278634548187256, + "learning_rate": 4.962970460718322e-05, + "loss": 0.6453, + "step": 880 + }, + { + "epoch": 0.022318943266359633, + "grad_norm": 4.173251628875732, + "learning_rate": 4.962886110970528e-05, + "loss": 0.5477, + "step": 882 + }, + { + "epoch": 0.022369553115036186, + "grad_norm": 3.603672981262207, + "learning_rate": 4.962801761222734e-05, + "loss": 0.8842, + "step": 884 + }, + { + "epoch": 0.02242016296371274, + "grad_norm": 18.358938217163086, + "learning_rate": 4.96271741147494e-05, + "loss": 0.8393, + "step": 886 + }, + { + "epoch": 0.02247077281238929, + "grad_norm": 6.532278537750244, + "learning_rate": 4.962633061727146e-05, + "loss": 0.5507, + "step": 888 + }, + { + "epoch": 0.022521382661065844, + "grad_norm": 6.95924711227417, + "learning_rate": 4.962548711979351e-05, + "loss": 0.6536, + "step": 890 + }, + { + "epoch": 0.022571992509742397, + "grad_norm": 3.4727678298950195, + "learning_rate": 4.962464362231557e-05, + "loss": 0.435, + "step": 892 + }, + { + "epoch": 0.02262260235841895, + "grad_norm": 5.473514080047607, + "learning_rate": 4.9623800124837624e-05, + "loss": 0.3681, + "step": 894 + }, + { + "epoch": 0.022673212207095502, + "grad_norm": 7.168368339538574, + "learning_rate": 4.9622956627359684e-05, + "loss": 0.6088, + "step": 896 + }, + { + "epoch": 0.022723822055772055, + "grad_norm": 9.777496337890625, + "learning_rate": 4.9622113129881744e-05, + "loss": 0.8791, + "step": 898 + }, + { + "epoch": 0.022774431904448604, + "grad_norm": 4.331769943237305, + "learning_rate": 4.96212696324038e-05, + "loss": 0.4703, + "step": 900 + }, + { + "epoch": 0.022825041753125157, + "grad_norm": 11.051033973693848, + "learning_rate": 4.962042613492586e-05, + "loss": 0.6771, + "step": 902 + }, + { + "epoch": 0.02287565160180171, + "grad_norm": 4.339256763458252, + "learning_rate": 4.961958263744791e-05, + "loss": 0.3235, + "step": 904 + }, + { + "epoch": 0.022926261450478262, + "grad_norm": 9.73657512664795, + "learning_rate": 4.961873913996997e-05, + "loss": 0.8238, + "step": 906 + }, + { + "epoch": 0.022976871299154815, + "grad_norm": 8.267867088317871, + "learning_rate": 4.961789564249203e-05, + "loss": 0.4584, + "step": 908 + }, + { + "epoch": 0.023027481147831368, + "grad_norm": 4.065835952758789, + "learning_rate": 4.9617052145014085e-05, + "loss": 0.5364, + "step": 910 + }, + { + "epoch": 0.02307809099650792, + "grad_norm": 9.213961601257324, + "learning_rate": 4.9616208647536146e-05, + "loss": 0.7232, + "step": 912 + }, + { + "epoch": 0.023128700845184473, + "grad_norm": 3.7316653728485107, + "learning_rate": 4.96153651500582e-05, + "loss": 0.417, + "step": 914 + }, + { + "epoch": 0.023179310693861026, + "grad_norm": 4.829885959625244, + "learning_rate": 4.961452165258026e-05, + "loss": 0.4171, + "step": 916 + }, + { + "epoch": 0.02322992054253758, + "grad_norm": 13.912676811218262, + "learning_rate": 4.961367815510232e-05, + "loss": 0.5439, + "step": 918 + }, + { + "epoch": 0.02328053039121413, + "grad_norm": 6.918741226196289, + "learning_rate": 4.961283465762437e-05, + "loss": 0.6434, + "step": 920 + }, + { + "epoch": 0.023331140239890684, + "grad_norm": 7.197675704956055, + "learning_rate": 4.961199116014643e-05, + "loss": 0.7114, + "step": 922 + }, + { + "epoch": 0.023381750088567237, + "grad_norm": 6.958329200744629, + "learning_rate": 4.9611147662668486e-05, + "loss": 0.6549, + "step": 924 + }, + { + "epoch": 0.02343235993724379, + "grad_norm": 7.052434921264648, + "learning_rate": 4.9610304165190547e-05, + "loss": 0.8423, + "step": 926 + }, + { + "epoch": 0.02348296978592034, + "grad_norm": 4.447729110717773, + "learning_rate": 4.960946066771261e-05, + "loss": 0.4996, + "step": 928 + }, + { + "epoch": 0.02353357963459689, + "grad_norm": 8.922036170959473, + "learning_rate": 4.960861717023466e-05, + "loss": 0.7087, + "step": 930 + }, + { + "epoch": 0.023584189483273444, + "grad_norm": 5.072062969207764, + "learning_rate": 4.960777367275672e-05, + "loss": 0.6203, + "step": 932 + }, + { + "epoch": 0.023634799331949997, + "grad_norm": 2.5875844955444336, + "learning_rate": 4.9606930175278774e-05, + "loss": 0.3663, + "step": 934 + }, + { + "epoch": 0.02368540918062655, + "grad_norm": 5.000091552734375, + "learning_rate": 4.9606086677800834e-05, + "loss": 0.2929, + "step": 936 + }, + { + "epoch": 0.023736019029303102, + "grad_norm": 5.237270355224609, + "learning_rate": 4.9605243180322894e-05, + "loss": 0.632, + "step": 938 + }, + { + "epoch": 0.023786628877979655, + "grad_norm": 9.747302055358887, + "learning_rate": 4.960439968284495e-05, + "loss": 0.7382, + "step": 940 + }, + { + "epoch": 0.023837238726656208, + "grad_norm": 7.886781215667725, + "learning_rate": 4.960355618536701e-05, + "loss": 0.6935, + "step": 942 + }, + { + "epoch": 0.02388784857533276, + "grad_norm": 6.9037885665893555, + "learning_rate": 4.960271268788906e-05, + "loss": 0.631, + "step": 944 + }, + { + "epoch": 0.023938458424009313, + "grad_norm": 4.556064128875732, + "learning_rate": 4.960186919041112e-05, + "loss": 0.5162, + "step": 946 + }, + { + "epoch": 0.023989068272685866, + "grad_norm": 6.1615519523620605, + "learning_rate": 4.960102569293318e-05, + "loss": 0.6383, + "step": 948 + }, + { + "epoch": 0.02403967812136242, + "grad_norm": 142.0865020751953, + "learning_rate": 4.9600182195455235e-05, + "loss": 0.7586, + "step": 950 + }, + { + "epoch": 0.02409028797003897, + "grad_norm": 10.337366104125977, + "learning_rate": 4.9599338697977295e-05, + "loss": 0.5753, + "step": 952 + }, + { + "epoch": 0.02414089781871552, + "grad_norm": 6.216306686401367, + "learning_rate": 4.959849520049935e-05, + "loss": 0.6553, + "step": 954 + }, + { + "epoch": 0.024191507667392073, + "grad_norm": 2.8385775089263916, + "learning_rate": 4.959765170302141e-05, + "loss": 0.3002, + "step": 956 + }, + { + "epoch": 0.024242117516068626, + "grad_norm": 5.520168304443359, + "learning_rate": 4.959680820554347e-05, + "loss": 0.6907, + "step": 958 + }, + { + "epoch": 0.02429272736474518, + "grad_norm": 3.9114444255828857, + "learning_rate": 4.959596470806552e-05, + "loss": 0.419, + "step": 960 + }, + { + "epoch": 0.02434333721342173, + "grad_norm": 6.165211200714111, + "learning_rate": 4.959512121058758e-05, + "loss": 0.6344, + "step": 962 + }, + { + "epoch": 0.024393947062098284, + "grad_norm": 4.263137340545654, + "learning_rate": 4.9594277713109636e-05, + "loss": 0.5867, + "step": 964 + }, + { + "epoch": 0.024444556910774837, + "grad_norm": 2.720306873321533, + "learning_rate": 4.9593434215631696e-05, + "loss": 0.3657, + "step": 966 + }, + { + "epoch": 0.02449516675945139, + "grad_norm": 6.873605251312256, + "learning_rate": 4.959259071815375e-05, + "loss": 0.6407, + "step": 968 + }, + { + "epoch": 0.024545776608127942, + "grad_norm": 16.427352905273438, + "learning_rate": 4.959174722067581e-05, + "loss": 0.8508, + "step": 970 + }, + { + "epoch": 0.024596386456804495, + "grad_norm": 8.866124153137207, + "learning_rate": 4.959090372319787e-05, + "loss": 0.803, + "step": 972 + }, + { + "epoch": 0.024646996305481048, + "grad_norm": 5.541032791137695, + "learning_rate": 4.9590060225719924e-05, + "loss": 0.6485, + "step": 974 + }, + { + "epoch": 0.0246976061541576, + "grad_norm": 2.8217546939849854, + "learning_rate": 4.9589216728241984e-05, + "loss": 0.3261, + "step": 976 + }, + { + "epoch": 0.024748216002834153, + "grad_norm": 6.076355457305908, + "learning_rate": 4.958837323076404e-05, + "loss": 0.5189, + "step": 978 + }, + { + "epoch": 0.024798825851510702, + "grad_norm": 2.687714099884033, + "learning_rate": 4.95875297332861e-05, + "loss": 0.3242, + "step": 980 + }, + { + "epoch": 0.024849435700187255, + "grad_norm": 6.32763671875, + "learning_rate": 4.958668623580816e-05, + "loss": 0.6136, + "step": 982 + }, + { + "epoch": 0.024900045548863808, + "grad_norm": 11.270010948181152, + "learning_rate": 4.958584273833021e-05, + "loss": 0.8674, + "step": 984 + }, + { + "epoch": 0.02495065539754036, + "grad_norm": 5.883991718292236, + "learning_rate": 4.958499924085227e-05, + "loss": 0.5242, + "step": 986 + }, + { + "epoch": 0.025001265246216913, + "grad_norm": 6.999844551086426, + "learning_rate": 4.9584155743374325e-05, + "loss": 0.5351, + "step": 988 + }, + { + "epoch": 0.025051875094893466, + "grad_norm": 6.340963363647461, + "learning_rate": 4.9583312245896385e-05, + "loss": 0.5624, + "step": 990 + }, + { + "epoch": 0.02510248494357002, + "grad_norm": 7.818021774291992, + "learning_rate": 4.9582468748418445e-05, + "loss": 0.9044, + "step": 992 + }, + { + "epoch": 0.02515309479224657, + "grad_norm": 6.447050094604492, + "learning_rate": 4.95816252509405e-05, + "loss": 0.7352, + "step": 994 + }, + { + "epoch": 0.025203704640923124, + "grad_norm": 4.902888774871826, + "learning_rate": 4.958078175346256e-05, + "loss": 0.4516, + "step": 996 + }, + { + "epoch": 0.025254314489599677, + "grad_norm": 71.4312973022461, + "learning_rate": 4.957993825598461e-05, + "loss": 0.6593, + "step": 998 + }, + { + "epoch": 0.02530492433827623, + "grad_norm": 5.481997489929199, + "learning_rate": 4.957909475850667e-05, + "loss": 0.4801, + "step": 1000 + }, + { + "epoch": 0.02530492433827623, + "eval_cer": 0.20989040697069894, + "eval_loss": 0.3260483741760254, + "eval_runtime": 2649.7137, + "eval_samples_per_second": 5.915, + "eval_steps_per_second": 0.37, + "step": 1000 + }, + { + "epoch": 0.025355534186952782, + "grad_norm": 13.54730224609375, + "learning_rate": 4.957825126102873e-05, + "loss": 0.5685, + "step": 1002 + }, + { + "epoch": 0.025406144035629335, + "grad_norm": 8.101643562316895, + "learning_rate": 4.9577407763550786e-05, + "loss": 0.6376, + "step": 1004 + }, + { + "epoch": 0.025456753884305888, + "grad_norm": 5.483541965484619, + "learning_rate": 4.9576564266072846e-05, + "loss": 0.5417, + "step": 1006 + }, + { + "epoch": 0.025507363732982437, + "grad_norm": 5.6926493644714355, + "learning_rate": 4.95757207685949e-05, + "loss": 0.5145, + "step": 1008 + }, + { + "epoch": 0.02555797358165899, + "grad_norm": 4.287508487701416, + "learning_rate": 4.957487727111696e-05, + "loss": 0.5817, + "step": 1010 + }, + { + "epoch": 0.025608583430335542, + "grad_norm": 3.4692742824554443, + "learning_rate": 4.957403377363902e-05, + "loss": 0.4781, + "step": 1012 + }, + { + "epoch": 0.025659193279012095, + "grad_norm": 6.066647052764893, + "learning_rate": 4.957319027616107e-05, + "loss": 0.5475, + "step": 1014 + }, + { + "epoch": 0.025709803127688648, + "grad_norm": 3.1936872005462646, + "learning_rate": 4.9572346778683134e-05, + "loss": 0.6294, + "step": 1016 + }, + { + "epoch": 0.0257604129763652, + "grad_norm": 10.364377975463867, + "learning_rate": 4.957150328120519e-05, + "loss": 0.7561, + "step": 1018 + }, + { + "epoch": 0.025811022825041753, + "grad_norm": 5.162211894989014, + "learning_rate": 4.957065978372725e-05, + "loss": 0.5821, + "step": 1020 + }, + { + "epoch": 0.025861632673718306, + "grad_norm": 8.685132026672363, + "learning_rate": 4.956981628624931e-05, + "loss": 0.7048, + "step": 1022 + }, + { + "epoch": 0.02591224252239486, + "grad_norm": 4.4980316162109375, + "learning_rate": 4.956897278877136e-05, + "loss": 0.5405, + "step": 1024 + }, + { + "epoch": 0.02596285237107141, + "grad_norm": 5.45166540145874, + "learning_rate": 4.956812929129342e-05, + "loss": 0.3885, + "step": 1026 + }, + { + "epoch": 0.026013462219747964, + "grad_norm": 7.379059314727783, + "learning_rate": 4.9567285793815474e-05, + "loss": 0.5148, + "step": 1028 + }, + { + "epoch": 0.026064072068424517, + "grad_norm": 4.766157150268555, + "learning_rate": 4.9566442296337535e-05, + "loss": 0.7287, + "step": 1030 + }, + { + "epoch": 0.02611468191710107, + "grad_norm": 7.131243705749512, + "learning_rate": 4.9565598798859595e-05, + "loss": 0.6868, + "step": 1032 + }, + { + "epoch": 0.02616529176577762, + "grad_norm": 9.707527160644531, + "learning_rate": 4.956475530138165e-05, + "loss": 0.6592, + "step": 1034 + }, + { + "epoch": 0.02621590161445417, + "grad_norm": 5.630888938903809, + "learning_rate": 4.956391180390371e-05, + "loss": 0.5028, + "step": 1036 + }, + { + "epoch": 0.026266511463130724, + "grad_norm": 3.7485404014587402, + "learning_rate": 4.956306830642576e-05, + "loss": 0.3452, + "step": 1038 + }, + { + "epoch": 0.026317121311807277, + "grad_norm": 6.3416643142700195, + "learning_rate": 4.956222480894782e-05, + "loss": 0.4316, + "step": 1040 + }, + { + "epoch": 0.02636773116048383, + "grad_norm": 8.458013534545898, + "learning_rate": 4.956138131146988e-05, + "loss": 0.8142, + "step": 1042 + }, + { + "epoch": 0.026418341009160382, + "grad_norm": 9.704322814941406, + "learning_rate": 4.9560537813991936e-05, + "loss": 0.7614, + "step": 1044 + }, + { + "epoch": 0.026468950857836935, + "grad_norm": 3.311298370361328, + "learning_rate": 4.9559694316513996e-05, + "loss": 0.5417, + "step": 1046 + }, + { + "epoch": 0.026519560706513488, + "grad_norm": 7.961212635040283, + "learning_rate": 4.955885081903605e-05, + "loss": 0.6401, + "step": 1048 + }, + { + "epoch": 0.02657017055519004, + "grad_norm": 6.631721496582031, + "learning_rate": 4.955800732155811e-05, + "loss": 0.5357, + "step": 1050 + }, + { + "epoch": 0.026620780403866593, + "grad_norm": 5.143334865570068, + "learning_rate": 4.955716382408016e-05, + "loss": 0.545, + "step": 1052 + }, + { + "epoch": 0.026671390252543146, + "grad_norm": 8.77175235748291, + "learning_rate": 4.955632032660222e-05, + "loss": 0.5938, + "step": 1054 + }, + { + "epoch": 0.0267220001012197, + "grad_norm": 10.350188255310059, + "learning_rate": 4.955547682912428e-05, + "loss": 0.9662, + "step": 1056 + }, + { + "epoch": 0.02677260994989625, + "grad_norm": 6.271733283996582, + "learning_rate": 4.955463333164634e-05, + "loss": 0.5715, + "step": 1058 + }, + { + "epoch": 0.026823219798572804, + "grad_norm": 5.548452854156494, + "learning_rate": 4.95537898341684e-05, + "loss": 0.5213, + "step": 1060 + }, + { + "epoch": 0.026873829647249353, + "grad_norm": 5.460413455963135, + "learning_rate": 4.955294633669045e-05, + "loss": 0.3746, + "step": 1062 + }, + { + "epoch": 0.026924439495925906, + "grad_norm": 10.801025390625, + "learning_rate": 4.955210283921251e-05, + "loss": 0.66, + "step": 1064 + }, + { + "epoch": 0.02697504934460246, + "grad_norm": 35.335445404052734, + "learning_rate": 4.955125934173457e-05, + "loss": 0.5122, + "step": 1066 + }, + { + "epoch": 0.02702565919327901, + "grad_norm": 3.974865436553955, + "learning_rate": 4.9550415844256624e-05, + "loss": 0.3642, + "step": 1068 + }, + { + "epoch": 0.027076269041955564, + "grad_norm": 6.508527755737305, + "learning_rate": 4.9549572346778684e-05, + "loss": 0.6126, + "step": 1070 + }, + { + "epoch": 0.027126878890632117, + "grad_norm": 11.70407772064209, + "learning_rate": 4.954872884930074e-05, + "loss": 0.4939, + "step": 1072 + }, + { + "epoch": 0.02717748873930867, + "grad_norm": 9.648119926452637, + "learning_rate": 4.95478853518228e-05, + "loss": 0.6462, + "step": 1074 + }, + { + "epoch": 0.027228098587985222, + "grad_norm": 8.660693168640137, + "learning_rate": 4.954704185434486e-05, + "loss": 0.5923, + "step": 1076 + }, + { + "epoch": 0.027278708436661775, + "grad_norm": 17.504438400268555, + "learning_rate": 4.954619835686691e-05, + "loss": 0.9972, + "step": 1078 + }, + { + "epoch": 0.027329318285338328, + "grad_norm": 6.019506454467773, + "learning_rate": 4.954535485938897e-05, + "loss": 0.5884, + "step": 1080 + }, + { + "epoch": 0.02737992813401488, + "grad_norm": 11.403207778930664, + "learning_rate": 4.9544511361911025e-05, + "loss": 0.7651, + "step": 1082 + }, + { + "epoch": 0.027430537982691433, + "grad_norm": 3.842545747756958, + "learning_rate": 4.9543667864433085e-05, + "loss": 0.3572, + "step": 1084 + }, + { + "epoch": 0.027481147831367986, + "grad_norm": 4.691946029663086, + "learning_rate": 4.9542824366955146e-05, + "loss": 0.556, + "step": 1086 + }, + { + "epoch": 0.027531757680044535, + "grad_norm": 8.138900756835938, + "learning_rate": 4.95419808694772e-05, + "loss": 0.545, + "step": 1088 + }, + { + "epoch": 0.027582367528721088, + "grad_norm": 7.7400431632995605, + "learning_rate": 4.954113737199926e-05, + "loss": 0.6399, + "step": 1090 + }, + { + "epoch": 0.02763297737739764, + "grad_norm": 18.444286346435547, + "learning_rate": 4.954029387452131e-05, + "loss": 0.9819, + "step": 1092 + }, + { + "epoch": 0.027683587226074193, + "grad_norm": 4.818946838378906, + "learning_rate": 4.953945037704337e-05, + "loss": 0.5091, + "step": 1094 + }, + { + "epoch": 0.027734197074750746, + "grad_norm": 6.969218730926514, + "learning_rate": 4.953860687956543e-05, + "loss": 0.6144, + "step": 1096 + }, + { + "epoch": 0.0277848069234273, + "grad_norm": 5.701696395874023, + "learning_rate": 4.9537763382087486e-05, + "loss": 0.6378, + "step": 1098 + }, + { + "epoch": 0.02783541677210385, + "grad_norm": 4.948043346405029, + "learning_rate": 4.953691988460955e-05, + "loss": 0.61, + "step": 1100 + }, + { + "epoch": 0.027886026620780404, + "grad_norm": 6.133516788482666, + "learning_rate": 4.95360763871316e-05, + "loss": 0.5396, + "step": 1102 + }, + { + "epoch": 0.027936636469456957, + "grad_norm": 4.206554889678955, + "learning_rate": 4.953523288965366e-05, + "loss": 0.3066, + "step": 1104 + }, + { + "epoch": 0.02798724631813351, + "grad_norm": 3.7985496520996094, + "learning_rate": 4.953438939217572e-05, + "loss": 0.5389, + "step": 1106 + }, + { + "epoch": 0.028037856166810062, + "grad_norm": 6.991200923919678, + "learning_rate": 4.9533545894697774e-05, + "loss": 0.7457, + "step": 1108 + }, + { + "epoch": 0.028088466015486615, + "grad_norm": 5.731369495391846, + "learning_rate": 4.9532702397219834e-05, + "loss": 0.3976, + "step": 1110 + }, + { + "epoch": 0.028139075864163168, + "grad_norm": 7.29412317276001, + "learning_rate": 4.953185889974189e-05, + "loss": 0.56, + "step": 1112 + }, + { + "epoch": 0.02818968571283972, + "grad_norm": 10.055305480957031, + "learning_rate": 4.953101540226395e-05, + "loss": 0.8518, + "step": 1114 + }, + { + "epoch": 0.02824029556151627, + "grad_norm": 5.1682047843933105, + "learning_rate": 4.953017190478601e-05, + "loss": 0.3921, + "step": 1116 + }, + { + "epoch": 0.028290905410192822, + "grad_norm": 7.5293049812316895, + "learning_rate": 4.952932840730806e-05, + "loss": 0.732, + "step": 1118 + }, + { + "epoch": 0.028341515258869375, + "grad_norm": 7.173330783843994, + "learning_rate": 4.952848490983012e-05, + "loss": 0.6786, + "step": 1120 + }, + { + "epoch": 0.028392125107545928, + "grad_norm": 7.453824520111084, + "learning_rate": 4.9527641412352175e-05, + "loss": 0.6836, + "step": 1122 + }, + { + "epoch": 0.02844273495622248, + "grad_norm": 16.70603370666504, + "learning_rate": 4.9526797914874235e-05, + "loss": 0.8831, + "step": 1124 + }, + { + "epoch": 0.028493344804899033, + "grad_norm": 8.2377290725708, + "learning_rate": 4.952595441739629e-05, + "loss": 0.8214, + "step": 1126 + }, + { + "epoch": 0.028543954653575586, + "grad_norm": 4.513237953186035, + "learning_rate": 4.952511091991835e-05, + "loss": 0.7878, + "step": 1128 + }, + { + "epoch": 0.02859456450225214, + "grad_norm": 6.024347305297852, + "learning_rate": 4.952426742244041e-05, + "loss": 0.2515, + "step": 1130 + }, + { + "epoch": 0.02864517435092869, + "grad_norm": 7.672776699066162, + "learning_rate": 4.952342392496246e-05, + "loss": 0.9578, + "step": 1132 + }, + { + "epoch": 0.028695784199605244, + "grad_norm": 11.622359275817871, + "learning_rate": 4.952258042748452e-05, + "loss": 0.6932, + "step": 1134 + }, + { + "epoch": 0.028746394048281797, + "grad_norm": 6.2994232177734375, + "learning_rate": 4.9521736930006576e-05, + "loss": 0.5317, + "step": 1136 + }, + { + "epoch": 0.02879700389695835, + "grad_norm": 8.886155128479004, + "learning_rate": 4.9520893432528636e-05, + "loss": 0.9717, + "step": 1138 + }, + { + "epoch": 0.028847613745634902, + "grad_norm": 3.2561750411987305, + "learning_rate": 4.9520049935050696e-05, + "loss": 0.3471, + "step": 1140 + }, + { + "epoch": 0.02889822359431145, + "grad_norm": 6.805208683013916, + "learning_rate": 4.951920643757275e-05, + "loss": 0.4706, + "step": 1142 + }, + { + "epoch": 0.028948833442988004, + "grad_norm": 6.713986396789551, + "learning_rate": 4.951836294009481e-05, + "loss": 0.7055, + "step": 1144 + }, + { + "epoch": 0.028999443291664557, + "grad_norm": 5.203835964202881, + "learning_rate": 4.9517519442616863e-05, + "loss": 0.6585, + "step": 1146 + }, + { + "epoch": 0.02905005314034111, + "grad_norm": 3.168962001800537, + "learning_rate": 4.9516675945138924e-05, + "loss": 0.4794, + "step": 1148 + }, + { + "epoch": 0.029100662989017662, + "grad_norm": 11.134737014770508, + "learning_rate": 4.9515832447660984e-05, + "loss": 0.899, + "step": 1150 + }, + { + "epoch": 0.029151272837694215, + "grad_norm": 5.522861480712891, + "learning_rate": 4.951498895018304e-05, + "loss": 0.4002, + "step": 1152 + }, + { + "epoch": 0.029201882686370768, + "grad_norm": 2.908127784729004, + "learning_rate": 4.95141454527051e-05, + "loss": 0.4008, + "step": 1154 + }, + { + "epoch": 0.02925249253504732, + "grad_norm": 11.76309871673584, + "learning_rate": 4.951330195522715e-05, + "loss": 0.3493, + "step": 1156 + }, + { + "epoch": 0.029303102383723873, + "grad_norm": 4.505433082580566, + "learning_rate": 4.951245845774921e-05, + "loss": 0.3181, + "step": 1158 + }, + { + "epoch": 0.029353712232400426, + "grad_norm": 4.662505149841309, + "learning_rate": 4.951161496027127e-05, + "loss": 0.6274, + "step": 1160 + }, + { + "epoch": 0.02940432208107698, + "grad_norm": 4.870098114013672, + "learning_rate": 4.9510771462793325e-05, + "loss": 0.5994, + "step": 1162 + }, + { + "epoch": 0.02945493192975353, + "grad_norm": 4.153433322906494, + "learning_rate": 4.9509927965315385e-05, + "loss": 0.3769, + "step": 1164 + }, + { + "epoch": 0.029505541778430084, + "grad_norm": 8.398072242736816, + "learning_rate": 4.950908446783744e-05, + "loss": 0.8386, + "step": 1166 + }, + { + "epoch": 0.029556151627106633, + "grad_norm": 5.86764669418335, + "learning_rate": 4.95082409703595e-05, + "loss": 0.3906, + "step": 1168 + }, + { + "epoch": 0.029606761475783186, + "grad_norm": 4.7525200843811035, + "learning_rate": 4.950739747288156e-05, + "loss": 0.5805, + "step": 1170 + }, + { + "epoch": 0.02965737132445974, + "grad_norm": 14.163674354553223, + "learning_rate": 4.950655397540361e-05, + "loss": 0.714, + "step": 1172 + }, + { + "epoch": 0.02970798117313629, + "grad_norm": 6.36476469039917, + "learning_rate": 4.950571047792567e-05, + "loss": 0.7714, + "step": 1174 + }, + { + "epoch": 0.029758591021812844, + "grad_norm": 3.8334286212921143, + "learning_rate": 4.9504866980447726e-05, + "loss": 0.4451, + "step": 1176 + }, + { + "epoch": 0.029809200870489397, + "grad_norm": 3.1083927154541016, + "learning_rate": 4.9504023482969786e-05, + "loss": 0.3391, + "step": 1178 + }, + { + "epoch": 0.02985981071916595, + "grad_norm": 11.239181518554688, + "learning_rate": 4.9503179985491846e-05, + "loss": 0.6345, + "step": 1180 + }, + { + "epoch": 0.029910420567842502, + "grad_norm": 6.135115623474121, + "learning_rate": 4.95023364880139e-05, + "loss": 0.8366, + "step": 1182 + }, + { + "epoch": 0.029961030416519055, + "grad_norm": 3.8036446571350098, + "learning_rate": 4.950149299053596e-05, + "loss": 0.4256, + "step": 1184 + }, + { + "epoch": 0.030011640265195608, + "grad_norm": 11.396409034729004, + "learning_rate": 4.950064949305801e-05, + "loss": 0.5988, + "step": 1186 + }, + { + "epoch": 0.03006225011387216, + "grad_norm": 6.251826286315918, + "learning_rate": 4.949980599558007e-05, + "loss": 0.4033, + "step": 1188 + }, + { + "epoch": 0.030112859962548713, + "grad_norm": 6.115148544311523, + "learning_rate": 4.9498962498102134e-05, + "loss": 0.4699, + "step": 1190 + }, + { + "epoch": 0.030163469811225266, + "grad_norm": 8.004677772521973, + "learning_rate": 4.949811900062419e-05, + "loss": 0.7723, + "step": 1192 + }, + { + "epoch": 0.03021407965990182, + "grad_norm": 8.397205352783203, + "learning_rate": 4.949727550314625e-05, + "loss": 0.4565, + "step": 1194 + }, + { + "epoch": 0.030264689508578368, + "grad_norm": 7.252643585205078, + "learning_rate": 4.94964320056683e-05, + "loss": 0.7132, + "step": 1196 + }, + { + "epoch": 0.03031529935725492, + "grad_norm": 5.881059169769287, + "learning_rate": 4.949558850819036e-05, + "loss": 0.7198, + "step": 1198 + }, + { + "epoch": 0.030365909205931473, + "grad_norm": 11.146910667419434, + "learning_rate": 4.949474501071242e-05, + "loss": 0.5668, + "step": 1200 + }, + { + "epoch": 0.030365909205931473, + "eval_cer": 0.2020166573086281, + "eval_loss": 0.3112943768501282, + "eval_runtime": 2732.7913, + "eval_samples_per_second": 5.736, + "eval_steps_per_second": 0.359, + "step": 1200 + }, + { + "epoch": 0.030416519054608026, + "grad_norm": 5.239530563354492, + "learning_rate": 4.9493901513234474e-05, + "loss": 0.8572, + "step": 1202 + }, + { + "epoch": 0.03046712890328458, + "grad_norm": 6.763473987579346, + "learning_rate": 4.9493058015756535e-05, + "loss": 0.4804, + "step": 1204 + }, + { + "epoch": 0.03051773875196113, + "grad_norm": 15.63022232055664, + "learning_rate": 4.949221451827859e-05, + "loss": 0.6519, + "step": 1206 + }, + { + "epoch": 0.030568348600637684, + "grad_norm": 6.332169055938721, + "learning_rate": 4.949137102080065e-05, + "loss": 0.4457, + "step": 1208 + }, + { + "epoch": 0.030618958449314237, + "grad_norm": 8.737252235412598, + "learning_rate": 4.94905275233227e-05, + "loss": 0.6122, + "step": 1210 + }, + { + "epoch": 0.03066956829799079, + "grad_norm": 9.297114372253418, + "learning_rate": 4.948968402584476e-05, + "loss": 0.7155, + "step": 1212 + }, + { + "epoch": 0.030720178146667342, + "grad_norm": 7.846271991729736, + "learning_rate": 4.948884052836682e-05, + "loss": 0.7046, + "step": 1214 + }, + { + "epoch": 0.030770787995343895, + "grad_norm": 9.104436874389648, + "learning_rate": 4.9487997030888875e-05, + "loss": 0.5831, + "step": 1216 + }, + { + "epoch": 0.030821397844020448, + "grad_norm": 4.265629768371582, + "learning_rate": 4.9487153533410936e-05, + "loss": 0.2498, + "step": 1218 + }, + { + "epoch": 0.030872007692697, + "grad_norm": 3.7431282997131348, + "learning_rate": 4.948631003593299e-05, + "loss": 0.6634, + "step": 1220 + }, + { + "epoch": 0.03092261754137355, + "grad_norm": 5.563877105712891, + "learning_rate": 4.948546653845505e-05, + "loss": 0.7814, + "step": 1222 + }, + { + "epoch": 0.030973227390050102, + "grad_norm": 5.606101036071777, + "learning_rate": 4.948462304097711e-05, + "loss": 0.7189, + "step": 1224 + }, + { + "epoch": 0.031023837238726655, + "grad_norm": 9.97654914855957, + "learning_rate": 4.948377954349916e-05, + "loss": 0.7406, + "step": 1226 + }, + { + "epoch": 0.031074447087403208, + "grad_norm": 10.628059387207031, + "learning_rate": 4.948293604602122e-05, + "loss": 0.904, + "step": 1228 + }, + { + "epoch": 0.03112505693607976, + "grad_norm": 2.7047297954559326, + "learning_rate": 4.9482092548543277e-05, + "loss": 0.3665, + "step": 1230 + }, + { + "epoch": 0.031175666784756313, + "grad_norm": 8.546875953674316, + "learning_rate": 4.948124905106534e-05, + "loss": 0.5705, + "step": 1232 + }, + { + "epoch": 0.031226276633432866, + "grad_norm": 6.664468765258789, + "learning_rate": 4.94804055535874e-05, + "loss": 0.518, + "step": 1234 + }, + { + "epoch": 0.03127688648210942, + "grad_norm": 8.334696769714355, + "learning_rate": 4.947956205610945e-05, + "loss": 0.5282, + "step": 1236 + }, + { + "epoch": 0.03132749633078597, + "grad_norm": 7.652597427368164, + "learning_rate": 4.947871855863151e-05, + "loss": 0.6678, + "step": 1238 + }, + { + "epoch": 0.031378106179462524, + "grad_norm": 5.162440776824951, + "learning_rate": 4.9477875061153564e-05, + "loss": 0.4045, + "step": 1240 + }, + { + "epoch": 0.03142871602813908, + "grad_norm": 6.496342658996582, + "learning_rate": 4.9477031563675624e-05, + "loss": 0.4663, + "step": 1242 + }, + { + "epoch": 0.03147932587681563, + "grad_norm": 7.366580963134766, + "learning_rate": 4.9476188066197684e-05, + "loss": 0.6894, + "step": 1244 + }, + { + "epoch": 0.03152993572549218, + "grad_norm": 13.762931823730469, + "learning_rate": 4.947534456871974e-05, + "loss": 0.7944, + "step": 1246 + }, + { + "epoch": 0.031580545574168735, + "grad_norm": 12.880459785461426, + "learning_rate": 4.94745010712418e-05, + "loss": 0.7428, + "step": 1248 + }, + { + "epoch": 0.03163115542284529, + "grad_norm": 6.380914211273193, + "learning_rate": 4.947365757376385e-05, + "loss": 0.6661, + "step": 1250 + }, + { + "epoch": 0.03168176527152184, + "grad_norm": 7.456797122955322, + "learning_rate": 4.947281407628591e-05, + "loss": 0.7024, + "step": 1252 + }, + { + "epoch": 0.03173237512019839, + "grad_norm": 4.948566913604736, + "learning_rate": 4.947197057880797e-05, + "loss": 0.5951, + "step": 1254 + }, + { + "epoch": 0.031782984968874946, + "grad_norm": 7.60888147354126, + "learning_rate": 4.9471127081330025e-05, + "loss": 0.4111, + "step": 1256 + }, + { + "epoch": 0.0318335948175515, + "grad_norm": 5.659879684448242, + "learning_rate": 4.9470283583852085e-05, + "loss": 0.6262, + "step": 1258 + }, + { + "epoch": 0.03188420466622805, + "grad_norm": 6.134880542755127, + "learning_rate": 4.946944008637414e-05, + "loss": 0.5249, + "step": 1260 + }, + { + "epoch": 0.0319348145149046, + "grad_norm": 6.531559467315674, + "learning_rate": 4.94685965888962e-05, + "loss": 0.5775, + "step": 1262 + }, + { + "epoch": 0.03198542436358115, + "grad_norm": 12.462188720703125, + "learning_rate": 4.946775309141826e-05, + "loss": 0.6286, + "step": 1264 + }, + { + "epoch": 0.0320360342122577, + "grad_norm": 2.809241533279419, + "learning_rate": 4.946690959394031e-05, + "loss": 0.3842, + "step": 1266 + }, + { + "epoch": 0.032086644060934255, + "grad_norm": 5.521834850311279, + "learning_rate": 4.946606609646237e-05, + "loss": 0.4794, + "step": 1268 + }, + { + "epoch": 0.03213725390961081, + "grad_norm": 5.101696491241455, + "learning_rate": 4.9465222598984426e-05, + "loss": 0.4713, + "step": 1270 + }, + { + "epoch": 0.03218786375828736, + "grad_norm": 6.392932891845703, + "learning_rate": 4.9464379101506486e-05, + "loss": 0.6618, + "step": 1272 + }, + { + "epoch": 0.03223847360696391, + "grad_norm": 4.8280768394470215, + "learning_rate": 4.946353560402855e-05, + "loss": 0.3516, + "step": 1274 + }, + { + "epoch": 0.032289083455640466, + "grad_norm": 3.3262429237365723, + "learning_rate": 4.94626921065506e-05, + "loss": 0.5898, + "step": 1276 + }, + { + "epoch": 0.03233969330431702, + "grad_norm": 5.624804496765137, + "learning_rate": 4.946184860907266e-05, + "loss": 0.5552, + "step": 1278 + }, + { + "epoch": 0.03239030315299357, + "grad_norm": 3.1129558086395264, + "learning_rate": 4.9461005111594714e-05, + "loss": 0.6172, + "step": 1280 + }, + { + "epoch": 0.032440913001670124, + "grad_norm": 3.5634653568267822, + "learning_rate": 4.9460161614116774e-05, + "loss": 0.3858, + "step": 1282 + }, + { + "epoch": 0.03249152285034668, + "grad_norm": 7.236937046051025, + "learning_rate": 4.9459318116638834e-05, + "loss": 0.6032, + "step": 1284 + }, + { + "epoch": 0.03254213269902323, + "grad_norm": 3.344967842102051, + "learning_rate": 4.945847461916089e-05, + "loss": 0.5774, + "step": 1286 + }, + { + "epoch": 0.03259274254769978, + "grad_norm": 3.4330480098724365, + "learning_rate": 4.945763112168295e-05, + "loss": 0.7142, + "step": 1288 + }, + { + "epoch": 0.032643352396376335, + "grad_norm": 5.767812728881836, + "learning_rate": 4.9456787624205e-05, + "loss": 0.5658, + "step": 1290 + }, + { + "epoch": 0.03269396224505289, + "grad_norm": 5.187238693237305, + "learning_rate": 4.945594412672707e-05, + "loss": 0.346, + "step": 1292 + }, + { + "epoch": 0.03274457209372944, + "grad_norm": 5.828567028045654, + "learning_rate": 4.945510062924912e-05, + "loss": 0.3802, + "step": 1294 + }, + { + "epoch": 0.03279518194240599, + "grad_norm": 6.394260406494141, + "learning_rate": 4.945425713177118e-05, + "loss": 1.0168, + "step": 1296 + }, + { + "epoch": 0.032845791791082546, + "grad_norm": 4.123929977416992, + "learning_rate": 4.945341363429324e-05, + "loss": 0.4564, + "step": 1298 + }, + { + "epoch": 0.0328964016397591, + "grad_norm": 6.843530178070068, + "learning_rate": 4.9452570136815295e-05, + "loss": 0.6159, + "step": 1300 + }, + { + "epoch": 0.03294701148843565, + "grad_norm": 10.560795783996582, + "learning_rate": 4.9451726639337356e-05, + "loss": 0.517, + "step": 1302 + }, + { + "epoch": 0.032997621337112204, + "grad_norm": 2.8675217628479004, + "learning_rate": 4.945088314185941e-05, + "loss": 0.6848, + "step": 1304 + }, + { + "epoch": 0.03304823118578876, + "grad_norm": 3.9155211448669434, + "learning_rate": 4.945003964438147e-05, + "loss": 0.6188, + "step": 1306 + }, + { + "epoch": 0.03309884103446531, + "grad_norm": 6.229773998260498, + "learning_rate": 4.944919614690352e-05, + "loss": 0.4295, + "step": 1308 + }, + { + "epoch": 0.03314945088314186, + "grad_norm": 4.002541542053223, + "learning_rate": 4.944835264942558e-05, + "loss": 0.5816, + "step": 1310 + }, + { + "epoch": 0.033200060731818415, + "grad_norm": 6.393809795379639, + "learning_rate": 4.944750915194764e-05, + "loss": 0.8652, + "step": 1312 + }, + { + "epoch": 0.03325067058049497, + "grad_norm": 3.2906885147094727, + "learning_rate": 4.9446665654469696e-05, + "loss": 0.6938, + "step": 1314 + }, + { + "epoch": 0.03330128042917151, + "grad_norm": 2.870098352432251, + "learning_rate": 4.944582215699176e-05, + "loss": 0.2623, + "step": 1316 + }, + { + "epoch": 0.033351890277848066, + "grad_norm": 5.652560710906982, + "learning_rate": 4.944497865951381e-05, + "loss": 0.5474, + "step": 1318 + }, + { + "epoch": 0.03340250012652462, + "grad_norm": 12.31924819946289, + "learning_rate": 4.944413516203587e-05, + "loss": 0.7774, + "step": 1320 + }, + { + "epoch": 0.03345310997520117, + "grad_norm": 4.865198135375977, + "learning_rate": 4.944329166455793e-05, + "loss": 0.2116, + "step": 1322 + }, + { + "epoch": 0.033503719823877724, + "grad_norm": 4.536335468292236, + "learning_rate": 4.9442448167079984e-05, + "loss": 0.541, + "step": 1324 + }, + { + "epoch": 0.03355432967255428, + "grad_norm": 14.479270935058594, + "learning_rate": 4.9441604669602044e-05, + "loss": 0.3704, + "step": 1326 + }, + { + "epoch": 0.03360493952123083, + "grad_norm": 7.071706295013428, + "learning_rate": 4.94407611721241e-05, + "loss": 0.5868, + "step": 1328 + }, + { + "epoch": 0.03365554936990738, + "grad_norm": 12.812637329101562, + "learning_rate": 4.943991767464616e-05, + "loss": 0.7385, + "step": 1330 + }, + { + "epoch": 0.033706159218583935, + "grad_norm": 10.205668449401855, + "learning_rate": 4.943907417716822e-05, + "loss": 0.5933, + "step": 1332 + }, + { + "epoch": 0.03375676906726049, + "grad_norm": 6.436938762664795, + "learning_rate": 4.943823067969027e-05, + "loss": 1.0516, + "step": 1334 + }, + { + "epoch": 0.03380737891593704, + "grad_norm": 4.715915679931641, + "learning_rate": 4.943738718221233e-05, + "loss": 0.5055, + "step": 1336 + }, + { + "epoch": 0.03385798876461359, + "grad_norm": 5.58711051940918, + "learning_rate": 4.9436543684734385e-05, + "loss": 0.4184, + "step": 1338 + }, + { + "epoch": 0.033908598613290146, + "grad_norm": 16.581960678100586, + "learning_rate": 4.9435700187256445e-05, + "loss": 0.478, + "step": 1340 + }, + { + "epoch": 0.0339592084619667, + "grad_norm": 7.326168060302734, + "learning_rate": 4.9434856689778505e-05, + "loss": 0.7105, + "step": 1342 + }, + { + "epoch": 0.03400981831064325, + "grad_norm": 4.85835075378418, + "learning_rate": 4.943401319230056e-05, + "loss": 0.7862, + "step": 1344 + }, + { + "epoch": 0.034060428159319804, + "grad_norm": 3.9227964878082275, + "learning_rate": 4.943316969482262e-05, + "loss": 0.5298, + "step": 1346 + }, + { + "epoch": 0.03411103800799636, + "grad_norm": 6.2520904541015625, + "learning_rate": 4.943232619734467e-05, + "loss": 0.5795, + "step": 1348 + }, + { + "epoch": 0.03416164785667291, + "grad_norm": 6.502024173736572, + "learning_rate": 4.943148269986673e-05, + "loss": 0.4504, + "step": 1350 + }, + { + "epoch": 0.03421225770534946, + "grad_norm": 6.299930095672607, + "learning_rate": 4.943063920238879e-05, + "loss": 0.8499, + "step": 1352 + }, + { + "epoch": 0.034262867554026015, + "grad_norm": 3.0065736770629883, + "learning_rate": 4.9429795704910846e-05, + "loss": 0.3785, + "step": 1354 + }, + { + "epoch": 0.03431347740270257, + "grad_norm": 6.053264617919922, + "learning_rate": 4.9428952207432906e-05, + "loss": 0.4405, + "step": 1356 + }, + { + "epoch": 0.03436408725137912, + "grad_norm": 2.8478128910064697, + "learning_rate": 4.942810870995496e-05, + "loss": 0.3576, + "step": 1358 + }, + { + "epoch": 0.03441469710005567, + "grad_norm": 4.3910369873046875, + "learning_rate": 4.942726521247702e-05, + "loss": 0.6643, + "step": 1360 + }, + { + "epoch": 0.034465306948732226, + "grad_norm": 8.032602310180664, + "learning_rate": 4.942642171499908e-05, + "loss": 0.3172, + "step": 1362 + }, + { + "epoch": 0.03451591679740878, + "grad_norm": 4.423160552978516, + "learning_rate": 4.9425578217521134e-05, + "loss": 0.5107, + "step": 1364 + }, + { + "epoch": 0.03456652664608533, + "grad_norm": 4.310211181640625, + "learning_rate": 4.9424734720043194e-05, + "loss": 0.7117, + "step": 1366 + }, + { + "epoch": 0.034617136494761884, + "grad_norm": 7.594289779663086, + "learning_rate": 4.942389122256525e-05, + "loss": 0.3758, + "step": 1368 + }, + { + "epoch": 0.03466774634343843, + "grad_norm": 10.414362907409668, + "learning_rate": 4.942304772508731e-05, + "loss": 0.9048, + "step": 1370 + }, + { + "epoch": 0.03471835619211498, + "grad_norm": 5.809612274169922, + "learning_rate": 4.942220422760937e-05, + "loss": 0.4148, + "step": 1372 + }, + { + "epoch": 0.034768966040791535, + "grad_norm": 7.8530449867248535, + "learning_rate": 4.942136073013142e-05, + "loss": 0.6252, + "step": 1374 + }, + { + "epoch": 0.03481957588946809, + "grad_norm": 9.151339530944824, + "learning_rate": 4.942051723265348e-05, + "loss": 0.6158, + "step": 1376 + }, + { + "epoch": 0.03487018573814464, + "grad_norm": 20.164710998535156, + "learning_rate": 4.9419673735175535e-05, + "loss": 0.6702, + "step": 1378 + }, + { + "epoch": 0.03492079558682119, + "grad_norm": 6.756007194519043, + "learning_rate": 4.9418830237697595e-05, + "loss": 0.2371, + "step": 1380 + }, + { + "epoch": 0.034971405435497746, + "grad_norm": 6.241485595703125, + "learning_rate": 4.9417986740219655e-05, + "loss": 0.7635, + "step": 1382 + }, + { + "epoch": 0.0350220152841743, + "grad_norm": 10.091927528381348, + "learning_rate": 4.941714324274171e-05, + "loss": 0.8506, + "step": 1384 + }, + { + "epoch": 0.03507262513285085, + "grad_norm": 7.4213080406188965, + "learning_rate": 4.941629974526377e-05, + "loss": 0.5124, + "step": 1386 + }, + { + "epoch": 0.035123234981527404, + "grad_norm": 11.214619636535645, + "learning_rate": 4.941545624778582e-05, + "loss": 0.4202, + "step": 1388 + }, + { + "epoch": 0.03517384483020396, + "grad_norm": 8.220820426940918, + "learning_rate": 4.941461275030788e-05, + "loss": 0.9393, + "step": 1390 + }, + { + "epoch": 0.03522445467888051, + "grad_norm": 8.862229347229004, + "learning_rate": 4.9413769252829936e-05, + "loss": 0.813, + "step": 1392 + }, + { + "epoch": 0.03527506452755706, + "grad_norm": 3.892012596130371, + "learning_rate": 4.9412925755351996e-05, + "loss": 0.4279, + "step": 1394 + }, + { + "epoch": 0.035325674376233615, + "grad_norm": 13.525033950805664, + "learning_rate": 4.9412082257874056e-05, + "loss": 0.4542, + "step": 1396 + }, + { + "epoch": 0.03537628422491017, + "grad_norm": 7.098147869110107, + "learning_rate": 4.941123876039611e-05, + "loss": 0.613, + "step": 1398 + }, + { + "epoch": 0.03542689407358672, + "grad_norm": 6.280513286590576, + "learning_rate": 4.941039526291817e-05, + "loss": 0.8402, + "step": 1400 + }, + { + "epoch": 0.03542689407358672, + "eval_cer": 0.19522168153555014, + "eval_loss": 0.2994668483734131, + "eval_runtime": 2672.9422, + "eval_samples_per_second": 5.864, + "eval_steps_per_second": 0.367, + "step": 1400 + }, + { + "epoch": 0.03547750392226327, + "grad_norm": 20.038259506225586, + "learning_rate": 4.940955176544022e-05, + "loss": 0.4221, + "step": 1402 + }, + { + "epoch": 0.035528113770939826, + "grad_norm": 4.178603649139404, + "learning_rate": 4.940870826796228e-05, + "loss": 0.5002, + "step": 1404 + }, + { + "epoch": 0.03557872361961638, + "grad_norm": 4.7133469581604, + "learning_rate": 4.9407864770484344e-05, + "loss": 0.7616, + "step": 1406 + }, + { + "epoch": 0.03562933346829293, + "grad_norm": 5.610054969787598, + "learning_rate": 4.94070212730064e-05, + "loss": 0.5404, + "step": 1408 + }, + { + "epoch": 0.035679943316969484, + "grad_norm": 12.473624229431152, + "learning_rate": 4.940617777552846e-05, + "loss": 0.4731, + "step": 1410 + }, + { + "epoch": 0.03573055316564604, + "grad_norm": 5.456740856170654, + "learning_rate": 4.940533427805051e-05, + "loss": 0.3353, + "step": 1412 + }, + { + "epoch": 0.03578116301432259, + "grad_norm": 8.414679527282715, + "learning_rate": 4.940449078057257e-05, + "loss": 0.5136, + "step": 1414 + }, + { + "epoch": 0.03583177286299914, + "grad_norm": 6.136534214019775, + "learning_rate": 4.940364728309463e-05, + "loss": 0.4605, + "step": 1416 + }, + { + "epoch": 0.035882382711675695, + "grad_norm": 6.461178779602051, + "learning_rate": 4.9402803785616684e-05, + "loss": 0.395, + "step": 1418 + }, + { + "epoch": 0.03593299256035225, + "grad_norm": 8.947699546813965, + "learning_rate": 4.9401960288138745e-05, + "loss": 0.8262, + "step": 1420 + }, + { + "epoch": 0.0359836024090288, + "grad_norm": 7.892368793487549, + "learning_rate": 4.94011167906608e-05, + "loss": 0.529, + "step": 1422 + }, + { + "epoch": 0.036034212257705346, + "grad_norm": 6.520657062530518, + "learning_rate": 4.940027329318286e-05, + "loss": 0.6126, + "step": 1424 + }, + { + "epoch": 0.0360848221063819, + "grad_norm": 17.974655151367188, + "learning_rate": 4.939942979570492e-05, + "loss": 0.748, + "step": 1426 + }, + { + "epoch": 0.03613543195505845, + "grad_norm": 3.9267520904541016, + "learning_rate": 4.939858629822697e-05, + "loss": 0.3001, + "step": 1428 + }, + { + "epoch": 0.036186041803735004, + "grad_norm": 14.933813095092773, + "learning_rate": 4.939774280074903e-05, + "loss": 0.766, + "step": 1430 + }, + { + "epoch": 0.03623665165241156, + "grad_norm": 9.554229736328125, + "learning_rate": 4.9396899303271085e-05, + "loss": 0.5409, + "step": 1432 + }, + { + "epoch": 0.03628726150108811, + "grad_norm": 8.939488410949707, + "learning_rate": 4.9396055805793146e-05, + "loss": 0.4732, + "step": 1434 + }, + { + "epoch": 0.03633787134976466, + "grad_norm": 3.313821315765381, + "learning_rate": 4.9395212308315206e-05, + "loss": 0.3792, + "step": 1436 + }, + { + "epoch": 0.036388481198441215, + "grad_norm": 15.510714530944824, + "learning_rate": 4.939436881083726e-05, + "loss": 0.7043, + "step": 1438 + }, + { + "epoch": 0.03643909104711777, + "grad_norm": 8.56005573272705, + "learning_rate": 4.939352531335932e-05, + "loss": 0.4035, + "step": 1440 + }, + { + "epoch": 0.03648970089579432, + "grad_norm": 5.8778252601623535, + "learning_rate": 4.939268181588137e-05, + "loss": 0.3744, + "step": 1442 + }, + { + "epoch": 0.03654031074447087, + "grad_norm": 6.815622806549072, + "learning_rate": 4.939183831840343e-05, + "loss": 0.5826, + "step": 1444 + }, + { + "epoch": 0.036590920593147426, + "grad_norm": 7.3577680587768555, + "learning_rate": 4.939099482092549e-05, + "loss": 0.505, + "step": 1446 + }, + { + "epoch": 0.03664153044182398, + "grad_norm": 8.071341514587402, + "learning_rate": 4.939015132344755e-05, + "loss": 0.7999, + "step": 1448 + }, + { + "epoch": 0.03669214029050053, + "grad_norm": 10.3243989944458, + "learning_rate": 4.938930782596961e-05, + "loss": 0.6247, + "step": 1450 + }, + { + "epoch": 0.036742750139177084, + "grad_norm": 4.544474124908447, + "learning_rate": 4.938846432849166e-05, + "loss": 0.5807, + "step": 1452 + }, + { + "epoch": 0.03679335998785364, + "grad_norm": 9.284850120544434, + "learning_rate": 4.938762083101372e-05, + "loss": 0.4313, + "step": 1454 + }, + { + "epoch": 0.03684396983653019, + "grad_norm": 7.1159749031066895, + "learning_rate": 4.938677733353578e-05, + "loss": 0.4875, + "step": 1456 + }, + { + "epoch": 0.03689457968520674, + "grad_norm": 7.782135009765625, + "learning_rate": 4.9385933836057834e-05, + "loss": 0.8014, + "step": 1458 + }, + { + "epoch": 0.036945189533883295, + "grad_norm": 9.787236213684082, + "learning_rate": 4.9385090338579894e-05, + "loss": 0.9414, + "step": 1460 + }, + { + "epoch": 0.03699579938255985, + "grad_norm": 5.5786848068237305, + "learning_rate": 4.938424684110195e-05, + "loss": 0.6361, + "step": 1462 + }, + { + "epoch": 0.0370464092312364, + "grad_norm": 4.597846508026123, + "learning_rate": 4.938340334362401e-05, + "loss": 0.2795, + "step": 1464 + }, + { + "epoch": 0.03709701907991295, + "grad_norm": 5.9186553955078125, + "learning_rate": 4.938255984614607e-05, + "loss": 0.4051, + "step": 1466 + }, + { + "epoch": 0.037147628928589506, + "grad_norm": 5.021423816680908, + "learning_rate": 4.938171634866812e-05, + "loss": 0.4298, + "step": 1468 + }, + { + "epoch": 0.03719823877726606, + "grad_norm": 3.111738681793213, + "learning_rate": 4.938087285119018e-05, + "loss": 0.3101, + "step": 1470 + }, + { + "epoch": 0.03724884862594261, + "grad_norm": 6.111041069030762, + "learning_rate": 4.9380029353712235e-05, + "loss": 0.6118, + "step": 1472 + }, + { + "epoch": 0.037299458474619164, + "grad_norm": 30.16288185119629, + "learning_rate": 4.9379185856234295e-05, + "loss": 0.6206, + "step": 1474 + }, + { + "epoch": 0.03735006832329572, + "grad_norm": 9.460359573364258, + "learning_rate": 4.937834235875635e-05, + "loss": 0.3322, + "step": 1476 + }, + { + "epoch": 0.03740067817197226, + "grad_norm": 7.554260730743408, + "learning_rate": 4.937749886127841e-05, + "loss": 0.5655, + "step": 1478 + }, + { + "epoch": 0.037451288020648815, + "grad_norm": 8.392956733703613, + "learning_rate": 4.937665536380047e-05, + "loss": 0.6669, + "step": 1480 + }, + { + "epoch": 0.03750189786932537, + "grad_norm": 7.463123321533203, + "learning_rate": 4.937581186632252e-05, + "loss": 0.6082, + "step": 1482 + }, + { + "epoch": 0.03755250771800192, + "grad_norm": 2.9462671279907227, + "learning_rate": 4.937496836884458e-05, + "loss": 0.7547, + "step": 1484 + }, + { + "epoch": 0.03760311756667847, + "grad_norm": 9.454707145690918, + "learning_rate": 4.9374124871366636e-05, + "loss": 0.5665, + "step": 1486 + }, + { + "epoch": 0.037653727415355026, + "grad_norm": 2.5308687686920166, + "learning_rate": 4.9373281373888696e-05, + "loss": 0.4683, + "step": 1488 + }, + { + "epoch": 0.03770433726403158, + "grad_norm": 3.8949882984161377, + "learning_rate": 4.937243787641076e-05, + "loss": 0.3981, + "step": 1490 + }, + { + "epoch": 0.03775494711270813, + "grad_norm": 3.255460739135742, + "learning_rate": 4.937159437893281e-05, + "loss": 0.3254, + "step": 1492 + }, + { + "epoch": 0.037805556961384684, + "grad_norm": 10.862367630004883, + "learning_rate": 4.937075088145487e-05, + "loss": 0.8337, + "step": 1494 + }, + { + "epoch": 0.03785616681006124, + "grad_norm": 12.926139831542969, + "learning_rate": 4.9369907383976924e-05, + "loss": 0.8218, + "step": 1496 + }, + { + "epoch": 0.03790677665873779, + "grad_norm": 5.084683418273926, + "learning_rate": 4.9369063886498984e-05, + "loss": 0.4073, + "step": 1498 + }, + { + "epoch": 0.03795738650741434, + "grad_norm": 5.375326633453369, + "learning_rate": 4.9368220389021044e-05, + "loss": 0.4468, + "step": 1500 + }, + { + "epoch": 0.038007996356090895, + "grad_norm": 2.891730785369873, + "learning_rate": 4.93673768915431e-05, + "loss": 0.4952, + "step": 1502 + }, + { + "epoch": 0.03805860620476745, + "grad_norm": 6.7384161949157715, + "learning_rate": 4.936653339406516e-05, + "loss": 0.6817, + "step": 1504 + }, + { + "epoch": 0.038109216053444, + "grad_norm": 4.95111608505249, + "learning_rate": 4.936568989658721e-05, + "loss": 0.7322, + "step": 1506 + }, + { + "epoch": 0.03815982590212055, + "grad_norm": 4.689101219177246, + "learning_rate": 4.936484639910927e-05, + "loss": 0.6041, + "step": 1508 + }, + { + "epoch": 0.038210435750797106, + "grad_norm": 6.9700798988342285, + "learning_rate": 4.936400290163133e-05, + "loss": 0.4521, + "step": 1510 + }, + { + "epoch": 0.03826104559947366, + "grad_norm": 5.602224826812744, + "learning_rate": 4.9363159404153385e-05, + "loss": 0.6519, + "step": 1512 + }, + { + "epoch": 0.03831165544815021, + "grad_norm": 5.881453037261963, + "learning_rate": 4.9362315906675445e-05, + "loss": 0.6069, + "step": 1514 + }, + { + "epoch": 0.038362265296826764, + "grad_norm": 5.395936012268066, + "learning_rate": 4.93614724091975e-05, + "loss": 0.3001, + "step": 1516 + }, + { + "epoch": 0.03841287514550332, + "grad_norm": 5.4586663246154785, + "learning_rate": 4.936062891171956e-05, + "loss": 0.6286, + "step": 1518 + }, + { + "epoch": 0.03846348499417987, + "grad_norm": 5.338792324066162, + "learning_rate": 4.935978541424162e-05, + "loss": 0.4526, + "step": 1520 + }, + { + "epoch": 0.03851409484285642, + "grad_norm": 2.985135555267334, + "learning_rate": 4.935894191676367e-05, + "loss": 0.2444, + "step": 1522 + }, + { + "epoch": 0.038564704691532975, + "grad_norm": 7.952456474304199, + "learning_rate": 4.935809841928573e-05, + "loss": 0.5033, + "step": 1524 + }, + { + "epoch": 0.03861531454020953, + "grad_norm": 7.913976192474365, + "learning_rate": 4.9357254921807786e-05, + "loss": 0.5823, + "step": 1526 + }, + { + "epoch": 0.03866592438888608, + "grad_norm": 7.781038284301758, + "learning_rate": 4.9356411424329846e-05, + "loss": 0.6915, + "step": 1528 + }, + { + "epoch": 0.03871653423756263, + "grad_norm": 5.074882984161377, + "learning_rate": 4.9355567926851906e-05, + "loss": 0.8664, + "step": 1530 + }, + { + "epoch": 0.03876714408623918, + "grad_norm": 3.08144474029541, + "learning_rate": 4.935472442937396e-05, + "loss": 0.3723, + "step": 1532 + }, + { + "epoch": 0.03881775393491573, + "grad_norm": 6.625777244567871, + "learning_rate": 4.935388093189602e-05, + "loss": 0.5908, + "step": 1534 + }, + { + "epoch": 0.038868363783592284, + "grad_norm": 8.451767921447754, + "learning_rate": 4.9353037434418073e-05, + "loss": 0.8605, + "step": 1536 + }, + { + "epoch": 0.03891897363226884, + "grad_norm": 5.070049285888672, + "learning_rate": 4.9352193936940134e-05, + "loss": 0.3042, + "step": 1538 + }, + { + "epoch": 0.03896958348094539, + "grad_norm": 7.37528133392334, + "learning_rate": 4.9351350439462194e-05, + "loss": 0.7376, + "step": 1540 + }, + { + "epoch": 0.03902019332962194, + "grad_norm": 3.3460285663604736, + "learning_rate": 4.935050694198425e-05, + "loss": 0.6662, + "step": 1542 + }, + { + "epoch": 0.039070803178298495, + "grad_norm": 3.999098062515259, + "learning_rate": 4.934966344450631e-05, + "loss": 0.3612, + "step": 1544 + }, + { + "epoch": 0.03912141302697505, + "grad_norm": 6.4491868019104, + "learning_rate": 4.934881994702836e-05, + "loss": 0.7708, + "step": 1546 + }, + { + "epoch": 0.0391720228756516, + "grad_norm": 4.037872314453125, + "learning_rate": 4.934797644955042e-05, + "loss": 0.4529, + "step": 1548 + }, + { + "epoch": 0.03922263272432815, + "grad_norm": 7.757660865783691, + "learning_rate": 4.9347132952072475e-05, + "loss": 0.6572, + "step": 1550 + }, + { + "epoch": 0.039273242573004706, + "grad_norm": 10.399569511413574, + "learning_rate": 4.9346289454594535e-05, + "loss": 0.971, + "step": 1552 + }, + { + "epoch": 0.03932385242168126, + "grad_norm": 4.098683834075928, + "learning_rate": 4.9345445957116595e-05, + "loss": 0.5687, + "step": 1554 + }, + { + "epoch": 0.03937446227035781, + "grad_norm": 4.029207706451416, + "learning_rate": 4.934460245963865e-05, + "loss": 0.4164, + "step": 1556 + }, + { + "epoch": 0.039425072119034364, + "grad_norm": 9.3286771774292, + "learning_rate": 4.934375896216071e-05, + "loss": 0.8485, + "step": 1558 + }, + { + "epoch": 0.03947568196771092, + "grad_norm": 8.822758674621582, + "learning_rate": 4.934291546468276e-05, + "loss": 0.5747, + "step": 1560 + }, + { + "epoch": 0.03952629181638747, + "grad_norm": 6.740304946899414, + "learning_rate": 4.934207196720482e-05, + "loss": 0.5014, + "step": 1562 + }, + { + "epoch": 0.03957690166506402, + "grad_norm": 4.9694929122924805, + "learning_rate": 4.934122846972688e-05, + "loss": 0.5403, + "step": 1564 + }, + { + "epoch": 0.039627511513740575, + "grad_norm": 7.032242298126221, + "learning_rate": 4.9340384972248936e-05, + "loss": 0.3896, + "step": 1566 + }, + { + "epoch": 0.03967812136241713, + "grad_norm": 3.1985256671905518, + "learning_rate": 4.9339541474770996e-05, + "loss": 0.4138, + "step": 1568 + }, + { + "epoch": 0.03972873121109368, + "grad_norm": 4.911308288574219, + "learning_rate": 4.933869797729305e-05, + "loss": 0.6613, + "step": 1570 + }, + { + "epoch": 0.03977934105977023, + "grad_norm": 10.836835861206055, + "learning_rate": 4.933785447981511e-05, + "loss": 0.368, + "step": 1572 + }, + { + "epoch": 0.039829950908446786, + "grad_norm": 7.5843377113342285, + "learning_rate": 4.933701098233717e-05, + "loss": 0.8429, + "step": 1574 + }, + { + "epoch": 0.03988056075712334, + "grad_norm": 4.915452003479004, + "learning_rate": 4.933616748485922e-05, + "loss": 0.4179, + "step": 1576 + }, + { + "epoch": 0.03993117060579989, + "grad_norm": 4.324033260345459, + "learning_rate": 4.9335323987381283e-05, + "loss": 0.5588, + "step": 1578 + }, + { + "epoch": 0.039981780454476444, + "grad_norm": 5.227353096008301, + "learning_rate": 4.933448048990334e-05, + "loss": 0.4188, + "step": 1580 + }, + { + "epoch": 0.040032390303153, + "grad_norm": 6.144440650939941, + "learning_rate": 4.93336369924254e-05, + "loss": 0.7091, + "step": 1582 + }, + { + "epoch": 0.04008300015182954, + "grad_norm": 9.614253044128418, + "learning_rate": 4.933279349494746e-05, + "loss": 0.5816, + "step": 1584 + }, + { + "epoch": 0.040133610000506095, + "grad_norm": 5.539832592010498, + "learning_rate": 4.933194999746951e-05, + "loss": 0.2897, + "step": 1586 + }, + { + "epoch": 0.04018421984918265, + "grad_norm": 4.285924911499023, + "learning_rate": 4.933110649999157e-05, + "loss": 0.4647, + "step": 1588 + }, + { + "epoch": 0.0402348296978592, + "grad_norm": 5.418667316436768, + "learning_rate": 4.9330263002513624e-05, + "loss": 0.377, + "step": 1590 + }, + { + "epoch": 0.04028543954653575, + "grad_norm": 7.248344421386719, + "learning_rate": 4.9329419505035684e-05, + "loss": 0.5599, + "step": 1592 + }, + { + "epoch": 0.040336049395212306, + "grad_norm": 7.9554643630981445, + "learning_rate": 4.9328576007557745e-05, + "loss": 0.8153, + "step": 1594 + }, + { + "epoch": 0.04038665924388886, + "grad_norm": 8.028340339660645, + "learning_rate": 4.93277325100798e-05, + "loss": 0.5364, + "step": 1596 + }, + { + "epoch": 0.04043726909256541, + "grad_norm": 7.2708964347839355, + "learning_rate": 4.932688901260186e-05, + "loss": 0.8303, + "step": 1598 + }, + { + "epoch": 0.040487878941241964, + "grad_norm": 5.855560779571533, + "learning_rate": 4.932604551512391e-05, + "loss": 0.4462, + "step": 1600 + }, + { + "epoch": 0.040487878941241964, + "eval_cer": 0.18224000249547695, + "eval_loss": 0.28801780939102173, + "eval_runtime": 2605.2542, + "eval_samples_per_second": 6.016, + "eval_steps_per_second": 0.376, + "step": 1600 + }, + { + "epoch": 0.04053848878991852, + "grad_norm": 10.296344757080078, + "learning_rate": 4.932520201764597e-05, + "loss": 0.5008, + "step": 1602 + }, + { + "epoch": 0.04058909863859507, + "grad_norm": 8.924449920654297, + "learning_rate": 4.932435852016803e-05, + "loss": 0.3729, + "step": 1604 + }, + { + "epoch": 0.04063970848727162, + "grad_norm": 5.110000133514404, + "learning_rate": 4.9323515022690085e-05, + "loss": 0.5821, + "step": 1606 + }, + { + "epoch": 0.040690318335948175, + "grad_norm": 5.40463399887085, + "learning_rate": 4.9322671525212146e-05, + "loss": 0.6681, + "step": 1608 + }, + { + "epoch": 0.04074092818462473, + "grad_norm": 10.994048118591309, + "learning_rate": 4.93218280277342e-05, + "loss": 0.453, + "step": 1610 + }, + { + "epoch": 0.04079153803330128, + "grad_norm": 6.396229267120361, + "learning_rate": 4.932098453025626e-05, + "loss": 0.4864, + "step": 1612 + }, + { + "epoch": 0.04084214788197783, + "grad_norm": 6.884303569793701, + "learning_rate": 4.932014103277832e-05, + "loss": 0.9021, + "step": 1614 + }, + { + "epoch": 0.040892757730654386, + "grad_norm": 15.925917625427246, + "learning_rate": 4.931929753530037e-05, + "loss": 0.6345, + "step": 1616 + }, + { + "epoch": 0.04094336757933094, + "grad_norm": 3.3225505352020264, + "learning_rate": 4.931845403782243e-05, + "loss": 0.5429, + "step": 1618 + }, + { + "epoch": 0.04099397742800749, + "grad_norm": 5.548943519592285, + "learning_rate": 4.9317610540344487e-05, + "loss": 0.5081, + "step": 1620 + }, + { + "epoch": 0.041044587276684044, + "grad_norm": 8.253005027770996, + "learning_rate": 4.931676704286655e-05, + "loss": 0.4432, + "step": 1622 + }, + { + "epoch": 0.0410951971253606, + "grad_norm": 10.242237091064453, + "learning_rate": 4.931592354538861e-05, + "loss": 0.6442, + "step": 1624 + }, + { + "epoch": 0.04114580697403715, + "grad_norm": 4.957180976867676, + "learning_rate": 4.931508004791066e-05, + "loss": 0.4146, + "step": 1626 + }, + { + "epoch": 0.0411964168227137, + "grad_norm": 7.067080497741699, + "learning_rate": 4.931423655043272e-05, + "loss": 0.424, + "step": 1628 + }, + { + "epoch": 0.041247026671390255, + "grad_norm": 5.9971842765808105, + "learning_rate": 4.9313393052954774e-05, + "loss": 0.4941, + "step": 1630 + }, + { + "epoch": 0.04129763652006681, + "grad_norm": 5.412482261657715, + "learning_rate": 4.9312549555476834e-05, + "loss": 0.4065, + "step": 1632 + }, + { + "epoch": 0.04134824636874336, + "grad_norm": 13.345954895019531, + "learning_rate": 4.931170605799889e-05, + "loss": 0.8215, + "step": 1634 + }, + { + "epoch": 0.04139885621741991, + "grad_norm": 9.993549346923828, + "learning_rate": 4.931086256052095e-05, + "loss": 0.8707, + "step": 1636 + }, + { + "epoch": 0.04144946606609646, + "grad_norm": 27.219308853149414, + "learning_rate": 4.931001906304301e-05, + "loss": 0.5455, + "step": 1638 + }, + { + "epoch": 0.04150007591477301, + "grad_norm": 5.13199520111084, + "learning_rate": 4.930917556556506e-05, + "loss": 0.5479, + "step": 1640 + }, + { + "epoch": 0.041550685763449564, + "grad_norm": 4.877132892608643, + "learning_rate": 4.930833206808712e-05, + "loss": 0.5966, + "step": 1642 + }, + { + "epoch": 0.04160129561212612, + "grad_norm": 3.1169495582580566, + "learning_rate": 4.9307488570609175e-05, + "loss": 0.5371, + "step": 1644 + }, + { + "epoch": 0.04165190546080267, + "grad_norm": 4.428645133972168, + "learning_rate": 4.9306645073131235e-05, + "loss": 0.3974, + "step": 1646 + }, + { + "epoch": 0.04170251530947922, + "grad_norm": 12.36322021484375, + "learning_rate": 4.9305801575653295e-05, + "loss": 0.6203, + "step": 1648 + }, + { + "epoch": 0.041753125158155775, + "grad_norm": 4.944615840911865, + "learning_rate": 4.930495807817535e-05, + "loss": 0.6994, + "step": 1650 + }, + { + "epoch": 0.04180373500683233, + "grad_norm": 3.7057061195373535, + "learning_rate": 4.930411458069741e-05, + "loss": 0.2836, + "step": 1652 + }, + { + "epoch": 0.04185434485550888, + "grad_norm": 5.559707164764404, + "learning_rate": 4.930327108321946e-05, + "loss": 0.7595, + "step": 1654 + }, + { + "epoch": 0.04190495470418543, + "grad_norm": 6.805633544921875, + "learning_rate": 4.930242758574152e-05, + "loss": 0.497, + "step": 1656 + }, + { + "epoch": 0.041955564552861986, + "grad_norm": 6.29421329498291, + "learning_rate": 4.930158408826358e-05, + "loss": 0.5615, + "step": 1658 + }, + { + "epoch": 0.04200617440153854, + "grad_norm": 3.3329832553863525, + "learning_rate": 4.9300740590785636e-05, + "loss": 0.2914, + "step": 1660 + }, + { + "epoch": 0.04205678425021509, + "grad_norm": 5.0592827796936035, + "learning_rate": 4.9299897093307696e-05, + "loss": 0.3799, + "step": 1662 + }, + { + "epoch": 0.042107394098891644, + "grad_norm": 4.176607131958008, + "learning_rate": 4.929905359582975e-05, + "loss": 0.3987, + "step": 1664 + }, + { + "epoch": 0.0421580039475682, + "grad_norm": 6.9350175857543945, + "learning_rate": 4.929821009835181e-05, + "loss": 0.337, + "step": 1666 + }, + { + "epoch": 0.04220861379624475, + "grad_norm": 12.642129898071289, + "learning_rate": 4.929736660087387e-05, + "loss": 0.5035, + "step": 1668 + }, + { + "epoch": 0.0422592236449213, + "grad_norm": 3.18143892288208, + "learning_rate": 4.9296523103395924e-05, + "loss": 0.2602, + "step": 1670 + }, + { + "epoch": 0.042309833493597855, + "grad_norm": 7.448511600494385, + "learning_rate": 4.9295679605917984e-05, + "loss": 0.6351, + "step": 1672 + }, + { + "epoch": 0.04236044334227441, + "grad_norm": 9.632100105285645, + "learning_rate": 4.929483610844004e-05, + "loss": 0.6639, + "step": 1674 + }, + { + "epoch": 0.04241105319095096, + "grad_norm": 4.993129730224609, + "learning_rate": 4.92939926109621e-05, + "loss": 0.5902, + "step": 1676 + }, + { + "epoch": 0.04246166303962751, + "grad_norm": 8.809192657470703, + "learning_rate": 4.929314911348416e-05, + "loss": 0.7136, + "step": 1678 + }, + { + "epoch": 0.042512272888304066, + "grad_norm": 4.831273078918457, + "learning_rate": 4.929230561600621e-05, + "loss": 0.702, + "step": 1680 + }, + { + "epoch": 0.04256288273698062, + "grad_norm": 3.406562089920044, + "learning_rate": 4.929146211852827e-05, + "loss": 0.7114, + "step": 1682 + }, + { + "epoch": 0.04261349258565717, + "grad_norm": 4.17376184463501, + "learning_rate": 4.9290618621050325e-05, + "loss": 0.5236, + "step": 1684 + }, + { + "epoch": 0.042664102434333724, + "grad_norm": 3.0841195583343506, + "learning_rate": 4.9289775123572385e-05, + "loss": 0.4595, + "step": 1686 + }, + { + "epoch": 0.04271471228301028, + "grad_norm": 6.880139350891113, + "learning_rate": 4.9288931626094445e-05, + "loss": 0.6339, + "step": 1688 + }, + { + "epoch": 0.04276532213168683, + "grad_norm": 3.7621781826019287, + "learning_rate": 4.92880881286165e-05, + "loss": 0.4469, + "step": 1690 + }, + { + "epoch": 0.042815931980363375, + "grad_norm": 9.83583927154541, + "learning_rate": 4.928724463113856e-05, + "loss": 0.7447, + "step": 1692 + }, + { + "epoch": 0.04286654182903993, + "grad_norm": 3.6700632572174072, + "learning_rate": 4.928640113366061e-05, + "loss": 0.3185, + "step": 1694 + }, + { + "epoch": 0.04291715167771648, + "grad_norm": 5.93882417678833, + "learning_rate": 4.928555763618267e-05, + "loss": 0.5691, + "step": 1696 + }, + { + "epoch": 0.042967761526393033, + "grad_norm": 2.9145689010620117, + "learning_rate": 4.928471413870473e-05, + "loss": 0.3396, + "step": 1698 + }, + { + "epoch": 0.043018371375069586, + "grad_norm": 4.243858814239502, + "learning_rate": 4.9283870641226786e-05, + "loss": 0.3018, + "step": 1700 + }, + { + "epoch": 0.04306898122374614, + "grad_norm": 4.957646369934082, + "learning_rate": 4.9283027143748846e-05, + "loss": 0.4555, + "step": 1702 + }, + { + "epoch": 0.04311959107242269, + "grad_norm": 5.5672831535339355, + "learning_rate": 4.92821836462709e-05, + "loss": 0.4606, + "step": 1704 + }, + { + "epoch": 0.043170200921099244, + "grad_norm": 4.047713756561279, + "learning_rate": 4.928134014879296e-05, + "loss": 0.3793, + "step": 1706 + }, + { + "epoch": 0.0432208107697758, + "grad_norm": 7.607228755950928, + "learning_rate": 4.928049665131502e-05, + "loss": 0.6637, + "step": 1708 + }, + { + "epoch": 0.04327142061845235, + "grad_norm": 11.504854202270508, + "learning_rate": 4.9279653153837073e-05, + "loss": 0.6826, + "step": 1710 + }, + { + "epoch": 0.0433220304671289, + "grad_norm": 5.851675510406494, + "learning_rate": 4.9278809656359134e-05, + "loss": 0.526, + "step": 1712 + }, + { + "epoch": 0.043372640315805455, + "grad_norm": 14.779943466186523, + "learning_rate": 4.927796615888119e-05, + "loss": 0.6321, + "step": 1714 + }, + { + "epoch": 0.04342325016448201, + "grad_norm": 9.94005012512207, + "learning_rate": 4.927712266140325e-05, + "loss": 0.7401, + "step": 1716 + }, + { + "epoch": 0.04347386001315856, + "grad_norm": 4.635720729827881, + "learning_rate": 4.92762791639253e-05, + "loss": 0.5283, + "step": 1718 + }, + { + "epoch": 0.04352446986183511, + "grad_norm": 3.568844795227051, + "learning_rate": 4.927543566644736e-05, + "loss": 0.4053, + "step": 1720 + }, + { + "epoch": 0.043575079710511666, + "grad_norm": 6.748814582824707, + "learning_rate": 4.927459216896942e-05, + "loss": 0.5624, + "step": 1722 + }, + { + "epoch": 0.04362568955918822, + "grad_norm": 7.1615424156188965, + "learning_rate": 4.9273748671491475e-05, + "loss": 0.8786, + "step": 1724 + }, + { + "epoch": 0.04367629940786477, + "grad_norm": 5.897001266479492, + "learning_rate": 4.9272905174013535e-05, + "loss": 0.7447, + "step": 1726 + }, + { + "epoch": 0.043726909256541324, + "grad_norm": 6.555644512176514, + "learning_rate": 4.927206167653559e-05, + "loss": 0.4951, + "step": 1728 + }, + { + "epoch": 0.04377751910521788, + "grad_norm": 8.615190505981445, + "learning_rate": 4.927121817905765e-05, + "loss": 0.6687, + "step": 1730 + }, + { + "epoch": 0.04382812895389443, + "grad_norm": 6.572555065155029, + "learning_rate": 4.927037468157971e-05, + "loss": 0.4269, + "step": 1732 + }, + { + "epoch": 0.04387873880257098, + "grad_norm": 6.60528564453125, + "learning_rate": 4.926953118410176e-05, + "loss": 0.4845, + "step": 1734 + }, + { + "epoch": 0.043929348651247535, + "grad_norm": 5.96574592590332, + "learning_rate": 4.926868768662382e-05, + "loss": 0.6266, + "step": 1736 + }, + { + "epoch": 0.04397995849992409, + "grad_norm": 2.911107063293457, + "learning_rate": 4.9267844189145876e-05, + "loss": 0.312, + "step": 1738 + }, + { + "epoch": 0.04403056834860064, + "grad_norm": 6.024972438812256, + "learning_rate": 4.9267000691667936e-05, + "loss": 0.5205, + "step": 1740 + }, + { + "epoch": 0.04408117819727719, + "grad_norm": 12.015813827514648, + "learning_rate": 4.9266157194189996e-05, + "loss": 0.621, + "step": 1742 + }, + { + "epoch": 0.044131788045953746, + "grad_norm": 17.426437377929688, + "learning_rate": 4.926531369671205e-05, + "loss": 0.7822, + "step": 1744 + }, + { + "epoch": 0.04418239789463029, + "grad_norm": 8.361766815185547, + "learning_rate": 4.926447019923411e-05, + "loss": 0.8237, + "step": 1746 + }, + { + "epoch": 0.044233007743306844, + "grad_norm": 6.699423789978027, + "learning_rate": 4.926362670175616e-05, + "loss": 0.5673, + "step": 1748 + }, + { + "epoch": 0.0442836175919834, + "grad_norm": 3.933328151702881, + "learning_rate": 4.926278320427822e-05, + "loss": 0.6946, + "step": 1750 + }, + { + "epoch": 0.04433422744065995, + "grad_norm": 5.092313766479492, + "learning_rate": 4.9261939706800283e-05, + "loss": 0.6024, + "step": 1752 + }, + { + "epoch": 0.0443848372893365, + "grad_norm": 16.1398868560791, + "learning_rate": 4.926109620932234e-05, + "loss": 0.7255, + "step": 1754 + }, + { + "epoch": 0.044435447138013055, + "grad_norm": 4.958022117614746, + "learning_rate": 4.92602527118444e-05, + "loss": 0.4381, + "step": 1756 + }, + { + "epoch": 0.04448605698668961, + "grad_norm": 6.70586633682251, + "learning_rate": 4.925940921436645e-05, + "loss": 0.5228, + "step": 1758 + }, + { + "epoch": 0.04453666683536616, + "grad_norm": 3.8433001041412354, + "learning_rate": 4.925856571688851e-05, + "loss": 0.3635, + "step": 1760 + }, + { + "epoch": 0.04458727668404271, + "grad_norm": 3.4007515907287598, + "learning_rate": 4.925772221941057e-05, + "loss": 0.2397, + "step": 1762 + }, + { + "epoch": 0.044637886532719266, + "grad_norm": 12.362958908081055, + "learning_rate": 4.9256878721932624e-05, + "loss": 0.5092, + "step": 1764 + }, + { + "epoch": 0.04468849638139582, + "grad_norm": 3.942413568496704, + "learning_rate": 4.9256035224454684e-05, + "loss": 0.4432, + "step": 1766 + }, + { + "epoch": 0.04473910623007237, + "grad_norm": 5.850658893585205, + "learning_rate": 4.925519172697674e-05, + "loss": 0.8487, + "step": 1768 + }, + { + "epoch": 0.044789716078748924, + "grad_norm": 3.8962783813476562, + "learning_rate": 4.92543482294988e-05, + "loss": 0.6375, + "step": 1770 + }, + { + "epoch": 0.04484032592742548, + "grad_norm": 10.159829139709473, + "learning_rate": 4.925350473202086e-05, + "loss": 0.5135, + "step": 1772 + }, + { + "epoch": 0.04489093577610203, + "grad_norm": 2.924187183380127, + "learning_rate": 4.925266123454291e-05, + "loss": 0.2508, + "step": 1774 + }, + { + "epoch": 0.04494154562477858, + "grad_norm": 6.0558342933654785, + "learning_rate": 4.925181773706497e-05, + "loss": 0.3182, + "step": 1776 + }, + { + "epoch": 0.044992155473455135, + "grad_norm": 5.556451320648193, + "learning_rate": 4.9250974239587025e-05, + "loss": 0.492, + "step": 1778 + }, + { + "epoch": 0.04504276532213169, + "grad_norm": 4.185290813446045, + "learning_rate": 4.9250130742109086e-05, + "loss": 0.4214, + "step": 1780 + }, + { + "epoch": 0.04509337517080824, + "grad_norm": 5.0443196296691895, + "learning_rate": 4.9249287244631146e-05, + "loss": 0.6803, + "step": 1782 + }, + { + "epoch": 0.04514398501948479, + "grad_norm": 10.363567352294922, + "learning_rate": 4.92484437471532e-05, + "loss": 0.5701, + "step": 1784 + }, + { + "epoch": 0.045194594868161346, + "grad_norm": 6.5281243324279785, + "learning_rate": 4.924760024967526e-05, + "loss": 0.417, + "step": 1786 + }, + { + "epoch": 0.0452452047168379, + "grad_norm": 9.699246406555176, + "learning_rate": 4.924675675219731e-05, + "loss": 0.7565, + "step": 1788 + }, + { + "epoch": 0.04529581456551445, + "grad_norm": 13.578716278076172, + "learning_rate": 4.924591325471937e-05, + "loss": 0.5528, + "step": 1790 + }, + { + "epoch": 0.045346424414191004, + "grad_norm": 6.151464462280273, + "learning_rate": 4.924506975724143e-05, + "loss": 0.655, + "step": 1792 + }, + { + "epoch": 0.04539703426286756, + "grad_norm": 11.576047897338867, + "learning_rate": 4.9244226259763487e-05, + "loss": 0.8649, + "step": 1794 + }, + { + "epoch": 0.04544764411154411, + "grad_norm": 6.639153003692627, + "learning_rate": 4.924338276228555e-05, + "loss": 0.553, + "step": 1796 + }, + { + "epoch": 0.04549825396022066, + "grad_norm": 4.5320563316345215, + "learning_rate": 4.92425392648076e-05, + "loss": 0.4832, + "step": 1798 + }, + { + "epoch": 0.04554886380889721, + "grad_norm": 3.708400249481201, + "learning_rate": 4.924169576732966e-05, + "loss": 0.5646, + "step": 1800 + }, + { + "epoch": 0.04554886380889721, + "eval_cer": 0.19456141992638343, + "eval_loss": 0.29710084199905396, + "eval_runtime": 2630.6631, + "eval_samples_per_second": 5.958, + "eval_steps_per_second": 0.373, + "step": 1800 + }, + { + "epoch": 0.04559947365757376, + "grad_norm": 4.805381774902344, + "learning_rate": 4.9240852269851714e-05, + "loss": 0.5711, + "step": 1802 + }, + { + "epoch": 0.045650083506250314, + "grad_norm": 3.130221366882324, + "learning_rate": 4.9240008772373774e-05, + "loss": 0.4231, + "step": 1804 + }, + { + "epoch": 0.045700693354926866, + "grad_norm": 4.455588340759277, + "learning_rate": 4.9239165274895834e-05, + "loss": 0.3281, + "step": 1806 + }, + { + "epoch": 0.04575130320360342, + "grad_norm": 2.951422929763794, + "learning_rate": 4.923832177741789e-05, + "loss": 0.4477, + "step": 1808 + }, + { + "epoch": 0.04580191305227997, + "grad_norm": 5.397090435028076, + "learning_rate": 4.923747827993995e-05, + "loss": 0.4032, + "step": 1810 + }, + { + "epoch": 0.045852522900956524, + "grad_norm": 5.767248630523682, + "learning_rate": 4.9236634782462e-05, + "loss": 0.6842, + "step": 1812 + }, + { + "epoch": 0.04590313274963308, + "grad_norm": 3.4613723754882812, + "learning_rate": 4.923579128498406e-05, + "loss": 0.2943, + "step": 1814 + }, + { + "epoch": 0.04595374259830963, + "grad_norm": 4.715557098388672, + "learning_rate": 4.923494778750612e-05, + "loss": 0.3851, + "step": 1816 + }, + { + "epoch": 0.04600435244698618, + "grad_norm": 4.709720611572266, + "learning_rate": 4.9234104290028175e-05, + "loss": 0.2846, + "step": 1818 + }, + { + "epoch": 0.046054962295662735, + "grad_norm": 2.6926629543304443, + "learning_rate": 4.9233260792550235e-05, + "loss": 0.474, + "step": 1820 + }, + { + "epoch": 0.04610557214433929, + "grad_norm": 7.4501752853393555, + "learning_rate": 4.923241729507229e-05, + "loss": 0.6805, + "step": 1822 + }, + { + "epoch": 0.04615618199301584, + "grad_norm": 4.941989421844482, + "learning_rate": 4.923157379759435e-05, + "loss": 0.4549, + "step": 1824 + }, + { + "epoch": 0.04620679184169239, + "grad_norm": 7.671848297119141, + "learning_rate": 4.923073030011641e-05, + "loss": 0.4181, + "step": 1826 + }, + { + "epoch": 0.046257401690368946, + "grad_norm": 1.92340886592865, + "learning_rate": 4.922988680263846e-05, + "loss": 0.2794, + "step": 1828 + }, + { + "epoch": 0.0463080115390455, + "grad_norm": 4.5162200927734375, + "learning_rate": 4.922904330516052e-05, + "loss": 0.4803, + "step": 1830 + }, + { + "epoch": 0.04635862138772205, + "grad_norm": 6.522688388824463, + "learning_rate": 4.9228199807682576e-05, + "loss": 0.5851, + "step": 1832 + }, + { + "epoch": 0.046409231236398604, + "grad_norm": 3.49273419380188, + "learning_rate": 4.9227356310204636e-05, + "loss": 0.5751, + "step": 1834 + }, + { + "epoch": 0.04645984108507516, + "grad_norm": 14.866829872131348, + "learning_rate": 4.9226512812726697e-05, + "loss": 0.5211, + "step": 1836 + }, + { + "epoch": 0.04651045093375171, + "grad_norm": 6.658441066741943, + "learning_rate": 4.922566931524875e-05, + "loss": 0.5325, + "step": 1838 + }, + { + "epoch": 0.04656106078242826, + "grad_norm": 3.8592264652252197, + "learning_rate": 4.922482581777081e-05, + "loss": 0.5917, + "step": 1840 + }, + { + "epoch": 0.046611670631104815, + "grad_norm": 10.734745025634766, + "learning_rate": 4.9223982320292864e-05, + "loss": 0.3665, + "step": 1842 + }, + { + "epoch": 0.04666228047978137, + "grad_norm": 7.1689534187316895, + "learning_rate": 4.9223138822814924e-05, + "loss": 0.386, + "step": 1844 + }, + { + "epoch": 0.04671289032845792, + "grad_norm": 4.963681697845459, + "learning_rate": 4.9222295325336984e-05, + "loss": 0.4437, + "step": 1846 + }, + { + "epoch": 0.04676350017713447, + "grad_norm": 2.5440425872802734, + "learning_rate": 4.922145182785904e-05, + "loss": 0.3268, + "step": 1848 + }, + { + "epoch": 0.046814110025811026, + "grad_norm": 14.03139877319336, + "learning_rate": 4.92206083303811e-05, + "loss": 0.7161, + "step": 1850 + }, + { + "epoch": 0.04686471987448758, + "grad_norm": 7.766716480255127, + "learning_rate": 4.921976483290315e-05, + "loss": 0.4935, + "step": 1852 + }, + { + "epoch": 0.046915329723164124, + "grad_norm": 13.12921142578125, + "learning_rate": 4.921892133542521e-05, + "loss": 0.4105, + "step": 1854 + }, + { + "epoch": 0.04696593957184068, + "grad_norm": 3.441727876663208, + "learning_rate": 4.921807783794727e-05, + "loss": 0.8493, + "step": 1856 + }, + { + "epoch": 0.04701654942051723, + "grad_norm": 6.923675060272217, + "learning_rate": 4.9217234340469325e-05, + "loss": 0.819, + "step": 1858 + }, + { + "epoch": 0.04706715926919378, + "grad_norm": 8.306994438171387, + "learning_rate": 4.9216390842991385e-05, + "loss": 0.6548, + "step": 1860 + }, + { + "epoch": 0.047117769117870335, + "grad_norm": 3.314638614654541, + "learning_rate": 4.921554734551344e-05, + "loss": 0.6389, + "step": 1862 + }, + { + "epoch": 0.04716837896654689, + "grad_norm": 8.720133781433105, + "learning_rate": 4.92147038480355e-05, + "loss": 0.5129, + "step": 1864 + }, + { + "epoch": 0.04721898881522344, + "grad_norm": 15.914939880371094, + "learning_rate": 4.921386035055756e-05, + "loss": 0.5298, + "step": 1866 + }, + { + "epoch": 0.047269598663899993, + "grad_norm": 4.567658424377441, + "learning_rate": 4.921301685307961e-05, + "loss": 0.426, + "step": 1868 + }, + { + "epoch": 0.047320208512576546, + "grad_norm": 2.2042078971862793, + "learning_rate": 4.921217335560167e-05, + "loss": 0.2168, + "step": 1870 + }, + { + "epoch": 0.0473708183612531, + "grad_norm": 2.475092649459839, + "learning_rate": 4.9211329858123726e-05, + "loss": 0.5226, + "step": 1872 + }, + { + "epoch": 0.04742142820992965, + "grad_norm": 3.972266435623169, + "learning_rate": 4.9210486360645786e-05, + "loss": 0.5745, + "step": 1874 + }, + { + "epoch": 0.047472038058606204, + "grad_norm": 4.532939910888672, + "learning_rate": 4.9209642863167846e-05, + "loss": 0.6833, + "step": 1876 + }, + { + "epoch": 0.04752264790728276, + "grad_norm": 7.014007568359375, + "learning_rate": 4.92087993656899e-05, + "loss": 0.5458, + "step": 1878 + }, + { + "epoch": 0.04757325775595931, + "grad_norm": 8.639058113098145, + "learning_rate": 4.920795586821196e-05, + "loss": 0.698, + "step": 1880 + }, + { + "epoch": 0.04762386760463586, + "grad_norm": 4.487074375152588, + "learning_rate": 4.920711237073401e-05, + "loss": 0.4481, + "step": 1882 + }, + { + "epoch": 0.047674477453312415, + "grad_norm": 7.813543319702148, + "learning_rate": 4.9206268873256074e-05, + "loss": 0.5677, + "step": 1884 + }, + { + "epoch": 0.04772508730198897, + "grad_norm": 14.59656047821045, + "learning_rate": 4.920542537577813e-05, + "loss": 0.5587, + "step": 1886 + }, + { + "epoch": 0.04777569715066552, + "grad_norm": 7.641439437866211, + "learning_rate": 4.920458187830019e-05, + "loss": 0.9166, + "step": 1888 + }, + { + "epoch": 0.04782630699934207, + "grad_norm": 11.967202186584473, + "learning_rate": 4.920373838082225e-05, + "loss": 0.6675, + "step": 1890 + }, + { + "epoch": 0.047876916848018626, + "grad_norm": 3.4140710830688477, + "learning_rate": 4.92028948833443e-05, + "loss": 0.5182, + "step": 1892 + }, + { + "epoch": 0.04792752669669518, + "grad_norm": 11.876256942749023, + "learning_rate": 4.920205138586636e-05, + "loss": 0.5383, + "step": 1894 + }, + { + "epoch": 0.04797813654537173, + "grad_norm": 3.3766562938690186, + "learning_rate": 4.9201207888388414e-05, + "loss": 0.4677, + "step": 1896 + }, + { + "epoch": 0.048028746394048284, + "grad_norm": 4.277167320251465, + "learning_rate": 4.9200364390910475e-05, + "loss": 0.4177, + "step": 1898 + }, + { + "epoch": 0.04807935624272484, + "grad_norm": 5.8467936515808105, + "learning_rate": 4.9199520893432535e-05, + "loss": 0.5603, + "step": 1900 + }, + { + "epoch": 0.04812996609140139, + "grad_norm": 5.7047038078308105, + "learning_rate": 4.919867739595459e-05, + "loss": 0.416, + "step": 1902 + }, + { + "epoch": 0.04818057594007794, + "grad_norm": 8.13704776763916, + "learning_rate": 4.919783389847665e-05, + "loss": 0.2622, + "step": 1904 + }, + { + "epoch": 0.048231185788754495, + "grad_norm": 7.6252923011779785, + "learning_rate": 4.91969904009987e-05, + "loss": 0.9887, + "step": 1906 + }, + { + "epoch": 0.04828179563743104, + "grad_norm": 6.5191497802734375, + "learning_rate": 4.919614690352076e-05, + "loss": 0.6701, + "step": 1908 + }, + { + "epoch": 0.048332405486107594, + "grad_norm": 98.16575622558594, + "learning_rate": 4.919530340604282e-05, + "loss": 1.0357, + "step": 1910 + }, + { + "epoch": 0.048383015334784146, + "grad_norm": 3.221017837524414, + "learning_rate": 4.9194459908564876e-05, + "loss": 0.396, + "step": 1912 + }, + { + "epoch": 0.0484336251834607, + "grad_norm": 4.679788112640381, + "learning_rate": 4.9193616411086936e-05, + "loss": 0.3766, + "step": 1914 + }, + { + "epoch": 0.04848423503213725, + "grad_norm": 7.324192047119141, + "learning_rate": 4.919277291360899e-05, + "loss": 0.6765, + "step": 1916 + }, + { + "epoch": 0.048534844880813804, + "grad_norm": 5.692389488220215, + "learning_rate": 4.919192941613105e-05, + "loss": 0.7648, + "step": 1918 + }, + { + "epoch": 0.04858545472949036, + "grad_norm": 8.046274185180664, + "learning_rate": 4.919108591865311e-05, + "loss": 0.717, + "step": 1920 + }, + { + "epoch": 0.04863606457816691, + "grad_norm": 8.431305885314941, + "learning_rate": 4.919024242117516e-05, + "loss": 0.5176, + "step": 1922 + }, + { + "epoch": 0.04868667442684346, + "grad_norm": 4.852190017700195, + "learning_rate": 4.918939892369722e-05, + "loss": 0.4792, + "step": 1924 + }, + { + "epoch": 0.048737284275520015, + "grad_norm": 3.7859811782836914, + "learning_rate": 4.918855542621928e-05, + "loss": 0.4534, + "step": 1926 + }, + { + "epoch": 0.04878789412419657, + "grad_norm": 3.5033421516418457, + "learning_rate": 4.918771192874134e-05, + "loss": 0.5004, + "step": 1928 + }, + { + "epoch": 0.04883850397287312, + "grad_norm": 19.82146453857422, + "learning_rate": 4.91868684312634e-05, + "loss": 0.7809, + "step": 1930 + }, + { + "epoch": 0.04888911382154967, + "grad_norm": 7.877125263214111, + "learning_rate": 4.918602493378545e-05, + "loss": 0.4266, + "step": 1932 + }, + { + "epoch": 0.048939723670226226, + "grad_norm": 9.093170166015625, + "learning_rate": 4.918518143630751e-05, + "loss": 0.7279, + "step": 1934 + }, + { + "epoch": 0.04899033351890278, + "grad_norm": 6.574698448181152, + "learning_rate": 4.9184337938829564e-05, + "loss": 0.4833, + "step": 1936 + }, + { + "epoch": 0.04904094336757933, + "grad_norm": 4.608344078063965, + "learning_rate": 4.9183494441351624e-05, + "loss": 0.5268, + "step": 1938 + }, + { + "epoch": 0.049091553216255884, + "grad_norm": 8.013690948486328, + "learning_rate": 4.9182650943873685e-05, + "loss": 0.5434, + "step": 1940 + }, + { + "epoch": 0.04914216306493244, + "grad_norm": 4.216137409210205, + "learning_rate": 4.918180744639574e-05, + "loss": 0.4351, + "step": 1942 + }, + { + "epoch": 0.04919277291360899, + "grad_norm": 7.3226776123046875, + "learning_rate": 4.91809639489178e-05, + "loss": 0.3895, + "step": 1944 + }, + { + "epoch": 0.04924338276228554, + "grad_norm": 4.8168840408325195, + "learning_rate": 4.918012045143985e-05, + "loss": 0.3544, + "step": 1946 + }, + { + "epoch": 0.049293992610962095, + "grad_norm": 8.331216812133789, + "learning_rate": 4.917927695396191e-05, + "loss": 0.2721, + "step": 1948 + }, + { + "epoch": 0.04934460245963865, + "grad_norm": 10.902067184448242, + "learning_rate": 4.917843345648397e-05, + "loss": 0.4973, + "step": 1950 + }, + { + "epoch": 0.0493952123083152, + "grad_norm": 5.606088161468506, + "learning_rate": 4.9177589959006025e-05, + "loss": 0.5057, + "step": 1952 + }, + { + "epoch": 0.04944582215699175, + "grad_norm": 14.297886848449707, + "learning_rate": 4.9176746461528086e-05, + "loss": 0.6632, + "step": 1954 + }, + { + "epoch": 0.049496432005668306, + "grad_norm": 5.292407989501953, + "learning_rate": 4.917590296405014e-05, + "loss": 0.8784, + "step": 1956 + }, + { + "epoch": 0.04954704185434486, + "grad_norm": 2.2982068061828613, + "learning_rate": 4.91750594665722e-05, + "loss": 0.3005, + "step": 1958 + }, + { + "epoch": 0.049597651703021405, + "grad_norm": 5.021894931793213, + "learning_rate": 4.917421596909426e-05, + "loss": 0.5095, + "step": 1960 + }, + { + "epoch": 0.04964826155169796, + "grad_norm": 6.863503932952881, + "learning_rate": 4.917337247161631e-05, + "loss": 0.5115, + "step": 1962 + }, + { + "epoch": 0.04969887140037451, + "grad_norm": 8.457030296325684, + "learning_rate": 4.917252897413837e-05, + "loss": 0.6833, + "step": 1964 + }, + { + "epoch": 0.04974948124905106, + "grad_norm": 7.3646674156188965, + "learning_rate": 4.9171685476660426e-05, + "loss": 0.6812, + "step": 1966 + }, + { + "epoch": 0.049800091097727615, + "grad_norm": 4.478794574737549, + "learning_rate": 4.917084197918249e-05, + "loss": 0.3277, + "step": 1968 + }, + { + "epoch": 0.04985070094640417, + "grad_norm": 8.332141876220703, + "learning_rate": 4.916999848170454e-05, + "loss": 0.3486, + "step": 1970 + }, + { + "epoch": 0.04990131079508072, + "grad_norm": 5.418234348297119, + "learning_rate": 4.91691549842266e-05, + "loss": 0.5811, + "step": 1972 + }, + { + "epoch": 0.049951920643757274, + "grad_norm": 7.849460124969482, + "learning_rate": 4.916831148674866e-05, + "loss": 0.9629, + "step": 1974 + }, + { + "epoch": 0.050002530492433826, + "grad_norm": 4.812062740325928, + "learning_rate": 4.9167467989270714e-05, + "loss": 0.4392, + "step": 1976 + }, + { + "epoch": 0.05005314034111038, + "grad_norm": 7.80704402923584, + "learning_rate": 4.9166624491792774e-05, + "loss": 0.9341, + "step": 1978 + }, + { + "epoch": 0.05010375018978693, + "grad_norm": 4.2226243019104, + "learning_rate": 4.916578099431483e-05, + "loss": 0.8323, + "step": 1980 + }, + { + "epoch": 0.050154360038463484, + "grad_norm": 5.1408867835998535, + "learning_rate": 4.916493749683689e-05, + "loss": 0.3631, + "step": 1982 + }, + { + "epoch": 0.05020496988714004, + "grad_norm": 6.943533897399902, + "learning_rate": 4.916409399935895e-05, + "loss": 0.4674, + "step": 1984 + }, + { + "epoch": 0.05025557973581659, + "grad_norm": 2.5346226692199707, + "learning_rate": 4.9163250501881e-05, + "loss": 0.3056, + "step": 1986 + }, + { + "epoch": 0.05030618958449314, + "grad_norm": 7.4098944664001465, + "learning_rate": 4.916240700440306e-05, + "loss": 0.7281, + "step": 1988 + }, + { + "epoch": 0.050356799433169695, + "grad_norm": 5.00679349899292, + "learning_rate": 4.9161563506925115e-05, + "loss": 0.5875, + "step": 1990 + }, + { + "epoch": 0.05040740928184625, + "grad_norm": 7.140769004821777, + "learning_rate": 4.9160720009447175e-05, + "loss": 0.4304, + "step": 1992 + }, + { + "epoch": 0.0504580191305228, + "grad_norm": 4.408178806304932, + "learning_rate": 4.9159876511969235e-05, + "loss": 0.3655, + "step": 1994 + }, + { + "epoch": 0.05050862897919935, + "grad_norm": 6.3958353996276855, + "learning_rate": 4.915903301449129e-05, + "loss": 0.5034, + "step": 1996 + }, + { + "epoch": 0.050559238827875906, + "grad_norm": 11.404850959777832, + "learning_rate": 4.915818951701335e-05, + "loss": 0.6704, + "step": 1998 + }, + { + "epoch": 0.05060984867655246, + "grad_norm": 22.483257293701172, + "learning_rate": 4.91573460195354e-05, + "loss": 0.5152, + "step": 2000 + }, + { + "epoch": 0.05060984867655246, + "eval_cer": 0.1947745752490278, + "eval_loss": 0.306193083524704, + "eval_runtime": 2641.4126, + "eval_samples_per_second": 5.934, + "eval_steps_per_second": 0.371, + "step": 2000 + } + ], + "logging_steps": 2, + "max_steps": 118554, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 1000, + "total_flos": 4.299636354947482e+19, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}