{ "best_global_step": 6000, "best_metric": 21.781282824708335, "best_model_checkpoint": "./SALAMA_C7/checkpoint-6000", "epoch": 2.44408229782033, "eval_steps": 2000, "global_step": 6000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008148299042574863, "grad_norm": 15.915007591247559, "learning_rate": 1.9000000000000002e-06, "loss": 1.0767, "step": 20 }, { "epoch": 0.016296598085149726, "grad_norm": 10.589001655578613, "learning_rate": 3.900000000000001e-06, "loss": 0.9639, "step": 40 }, { "epoch": 0.024444897127724588, "grad_norm": 11.770127296447754, "learning_rate": 5.9e-06, "loss": 0.9679, "step": 60 }, { "epoch": 0.03259319617029945, "grad_norm": 11.00736141204834, "learning_rate": 7.9e-06, "loss": 0.9264, "step": 80 }, { "epoch": 0.04074149521287431, "grad_norm": 8.219990730285645, "learning_rate": 9.9e-06, "loss": 0.7982, "step": 100 }, { "epoch": 0.048889794255449176, "grad_norm": 7.805874824523926, "learning_rate": 9.973847212663455e-06, "loss": 0.7226, "step": 120 }, { "epoch": 0.057038093298024034, "grad_norm": 6.7659430503845215, "learning_rate": 9.946317962835514e-06, "loss": 0.7378, "step": 140 }, { "epoch": 0.0651863923405989, "grad_norm": 6.5482282638549805, "learning_rate": 9.91878871300757e-06, "loss": 0.716, "step": 160 }, { "epoch": 0.07333469138317376, "grad_norm": 6.729607105255127, "learning_rate": 9.891259463179629e-06, "loss": 0.7048, "step": 180 }, { "epoch": 0.08148299042574862, "grad_norm": 8.517411231994629, "learning_rate": 9.863730213351686e-06, "loss": 0.7016, "step": 200 }, { "epoch": 0.0896312894683235, "grad_norm": 6.785459041595459, "learning_rate": 9.836200963523744e-06, "loss": 0.7395, "step": 220 }, { "epoch": 0.09777958851089835, "grad_norm": 6.437356472015381, "learning_rate": 9.808671713695803e-06, "loss": 0.6579, "step": 240 }, { "epoch": 0.10592788755347321, "grad_norm": 6.3125762939453125, "learning_rate": 9.781142463867861e-06, "loss": 0.6969, "step": 260 }, { "epoch": 0.11407618659604807, "grad_norm": 6.493209362030029, "learning_rate": 9.753613214039918e-06, "loss": 0.6338, "step": 280 }, { "epoch": 0.12222448563862294, "grad_norm": 8.855860710144043, "learning_rate": 9.726083964211977e-06, "loss": 0.6657, "step": 300 }, { "epoch": 0.1303727846811978, "grad_norm": 5.335328578948975, "learning_rate": 9.698554714384033e-06, "loss": 0.6906, "step": 320 }, { "epoch": 0.13852108372377267, "grad_norm": 6.506489276885986, "learning_rate": 9.671025464556092e-06, "loss": 0.6485, "step": 340 }, { "epoch": 0.14666938276634753, "grad_norm": 5.023345947265625, "learning_rate": 9.64349621472815e-06, "loss": 0.6291, "step": 360 }, { "epoch": 0.15481768180892239, "grad_norm": 8.132065773010254, "learning_rate": 9.615966964900207e-06, "loss": 0.6553, "step": 380 }, { "epoch": 0.16296598085149724, "grad_norm": 7.615749359130859, "learning_rate": 9.588437715072266e-06, "loss": 0.6635, "step": 400 }, { "epoch": 0.1711142798940721, "grad_norm": 6.487279415130615, "learning_rate": 9.560908465244322e-06, "loss": 0.7554, "step": 420 }, { "epoch": 0.179262578936647, "grad_norm": 6.409642219543457, "learning_rate": 9.53337921541638e-06, "loss": 0.5839, "step": 440 }, { "epoch": 0.18741087797922185, "grad_norm": 5.239986419677734, "learning_rate": 9.505849965588438e-06, "loss": 0.7271, "step": 460 }, { "epoch": 0.1955591770217967, "grad_norm": 5.452706336975098, "learning_rate": 9.478320715760496e-06, "loss": 0.6758, "step": 480 }, { "epoch": 0.20370747606437156, "grad_norm": 4.827511787414551, "learning_rate": 9.450791465932555e-06, "loss": 0.6178, "step": 500 }, { "epoch": 0.21185577510694642, "grad_norm": 6.320280075073242, "learning_rate": 9.423262216104611e-06, "loss": 0.6781, "step": 520 }, { "epoch": 0.22000407414952128, "grad_norm": 5.648658275604248, "learning_rate": 9.39573296627667e-06, "loss": 0.6448, "step": 540 }, { "epoch": 0.22815237319209614, "grad_norm": 6.806399822235107, "learning_rate": 9.368203716448727e-06, "loss": 0.6206, "step": 560 }, { "epoch": 0.23630067223467102, "grad_norm": 6.556358337402344, "learning_rate": 9.340674466620785e-06, "loss": 0.6105, "step": 580 }, { "epoch": 0.24444897127724588, "grad_norm": 5.21808385848999, "learning_rate": 9.313145216792842e-06, "loss": 0.651, "step": 600 }, { "epoch": 0.2525972703198207, "grad_norm": 6.145657539367676, "learning_rate": 9.2856159669649e-06, "loss": 0.6504, "step": 620 }, { "epoch": 0.2607455693623956, "grad_norm": 4.201127529144287, "learning_rate": 9.258086717136959e-06, "loss": 0.5885, "step": 640 }, { "epoch": 0.2688938684049705, "grad_norm": 5.38640022277832, "learning_rate": 9.230557467309017e-06, "loss": 0.5888, "step": 660 }, { "epoch": 0.27704216744754534, "grad_norm": 5.217910289764404, "learning_rate": 9.203028217481074e-06, "loss": 0.5901, "step": 680 }, { "epoch": 0.2851904664901202, "grad_norm": 5.227973461151123, "learning_rate": 9.175498967653133e-06, "loss": 0.7178, "step": 700 }, { "epoch": 0.29333876553269506, "grad_norm": 6.235045909881592, "learning_rate": 9.147969717825191e-06, "loss": 0.6711, "step": 720 }, { "epoch": 0.3014870645752699, "grad_norm": 5.588578224182129, "learning_rate": 9.120440467997248e-06, "loss": 0.522, "step": 740 }, { "epoch": 0.30963536361784477, "grad_norm": 5.49254035949707, "learning_rate": 9.092911218169306e-06, "loss": 0.5962, "step": 760 }, { "epoch": 0.31778366266041963, "grad_norm": 5.4190239906311035, "learning_rate": 9.065381968341363e-06, "loss": 0.5981, "step": 780 }, { "epoch": 0.3259319617029945, "grad_norm": 6.350463390350342, "learning_rate": 9.037852718513422e-06, "loss": 0.6114, "step": 800 }, { "epoch": 0.33408026074556935, "grad_norm": 6.231777191162109, "learning_rate": 9.010323468685478e-06, "loss": 0.6729, "step": 820 }, { "epoch": 0.3422285597881442, "grad_norm": 6.590610027313232, "learning_rate": 8.982794218857537e-06, "loss": 0.6124, "step": 840 }, { "epoch": 0.35037685883071906, "grad_norm": 5.504190444946289, "learning_rate": 8.955264969029595e-06, "loss": 0.6555, "step": 860 }, { "epoch": 0.358525157873294, "grad_norm": 4.821358680725098, "learning_rate": 8.927735719201652e-06, "loss": 0.5743, "step": 880 }, { "epoch": 0.36667345691586883, "grad_norm": 6.006564140319824, "learning_rate": 8.90020646937371e-06, "loss": 0.6157, "step": 900 }, { "epoch": 0.3748217559584437, "grad_norm": 5.9578471183776855, "learning_rate": 8.872677219545767e-06, "loss": 0.6485, "step": 920 }, { "epoch": 0.38297005500101855, "grad_norm": 5.149762153625488, "learning_rate": 8.845147969717826e-06, "loss": 0.5951, "step": 940 }, { "epoch": 0.3911183540435934, "grad_norm": 6.143222332000732, "learning_rate": 8.817618719889883e-06, "loss": 0.5805, "step": 960 }, { "epoch": 0.39926665308616827, "grad_norm": 5.496342658996582, "learning_rate": 8.790089470061941e-06, "loss": 0.6158, "step": 980 }, { "epoch": 0.4074149521287431, "grad_norm": 5.14535665512085, "learning_rate": 8.762560220234e-06, "loss": 0.6036, "step": 1000 }, { "epoch": 0.415563251171318, "grad_norm": 5.73581600189209, "learning_rate": 8.735030970406058e-06, "loss": 0.5925, "step": 1020 }, { "epoch": 0.42371155021389284, "grad_norm": 4.455753326416016, "learning_rate": 8.707501720578115e-06, "loss": 0.5894, "step": 1040 }, { "epoch": 0.4318598492564677, "grad_norm": 6.366560935974121, "learning_rate": 8.679972470750173e-06, "loss": 0.6424, "step": 1060 }, { "epoch": 0.44000814829904256, "grad_norm": 5.11430025100708, "learning_rate": 8.652443220922232e-06, "loss": 0.5731, "step": 1080 }, { "epoch": 0.4481564473416174, "grad_norm": 6.0588812828063965, "learning_rate": 8.624913971094288e-06, "loss": 0.6257, "step": 1100 }, { "epoch": 0.4563047463841923, "grad_norm": 6.879603385925293, "learning_rate": 8.597384721266347e-06, "loss": 0.6427, "step": 1120 }, { "epoch": 0.4644530454267672, "grad_norm": 4.6673502922058105, "learning_rate": 8.569855471438404e-06, "loss": 0.5763, "step": 1140 }, { "epoch": 0.47260134446934204, "grad_norm": 5.316527843475342, "learning_rate": 8.542326221610462e-06, "loss": 0.6207, "step": 1160 }, { "epoch": 0.4807496435119169, "grad_norm": 5.4205780029296875, "learning_rate": 8.514796971782519e-06, "loss": 0.5435, "step": 1180 }, { "epoch": 0.48889794255449176, "grad_norm": 4.579510688781738, "learning_rate": 8.487267721954577e-06, "loss": 0.5937, "step": 1200 }, { "epoch": 0.4970462415970666, "grad_norm": 6.327101707458496, "learning_rate": 8.459738472126636e-06, "loss": 0.6186, "step": 1220 }, { "epoch": 0.5051945406396414, "grad_norm": 5.347973346710205, "learning_rate": 8.432209222298693e-06, "loss": 0.564, "step": 1240 }, { "epoch": 0.5133428396822164, "grad_norm": 5.44881010055542, "learning_rate": 8.404679972470751e-06, "loss": 0.5221, "step": 1260 }, { "epoch": 0.5214911387247912, "grad_norm": 5.802955150604248, "learning_rate": 8.377150722642808e-06, "loss": 0.5893, "step": 1280 }, { "epoch": 0.5296394377673661, "grad_norm": 5.931309223175049, "learning_rate": 8.349621472814866e-06, "loss": 0.6306, "step": 1300 }, { "epoch": 0.537787736809941, "grad_norm": 5.793541431427002, "learning_rate": 8.322092222986923e-06, "loss": 0.6199, "step": 1320 }, { "epoch": 0.5459360358525158, "grad_norm": 4.972778797149658, "learning_rate": 8.294562973158982e-06, "loss": 0.4771, "step": 1340 }, { "epoch": 0.5540843348950907, "grad_norm": 5.573855876922607, "learning_rate": 8.26703372333104e-06, "loss": 0.5054, "step": 1360 }, { "epoch": 0.5622326339376655, "grad_norm": 4.7720947265625, "learning_rate": 8.239504473503099e-06, "loss": 0.5599, "step": 1380 }, { "epoch": 0.5703809329802404, "grad_norm": 6.338388442993164, "learning_rate": 8.211975223675155e-06, "loss": 0.5612, "step": 1400 }, { "epoch": 0.5785292320228153, "grad_norm": 6.1034464836120605, "learning_rate": 8.184445973847214e-06, "loss": 0.5662, "step": 1420 }, { "epoch": 0.5866775310653901, "grad_norm": 5.423385143280029, "learning_rate": 8.15691672401927e-06, "loss": 0.6246, "step": 1440 }, { "epoch": 0.594825830107965, "grad_norm": 6.507946014404297, "learning_rate": 8.12938747419133e-06, "loss": 0.5879, "step": 1460 }, { "epoch": 0.6029741291505398, "grad_norm": 5.927179336547852, "learning_rate": 8.101858224363388e-06, "loss": 0.5479, "step": 1480 }, { "epoch": 0.6111224281931147, "grad_norm": 6.775645732879639, "learning_rate": 8.074328974535444e-06, "loss": 0.5809, "step": 1500 }, { "epoch": 0.6192707272356895, "grad_norm": 4.263314247131348, "learning_rate": 8.046799724707503e-06, "loss": 0.55, "step": 1520 }, { "epoch": 0.6274190262782644, "grad_norm": 4.671892166137695, "learning_rate": 8.01927047487956e-06, "loss": 0.529, "step": 1540 }, { "epoch": 0.6355673253208393, "grad_norm": 5.7395501136779785, "learning_rate": 7.991741225051618e-06, "loss": 0.5349, "step": 1560 }, { "epoch": 0.6437156243634141, "grad_norm": 5.88397216796875, "learning_rate": 7.964211975223675e-06, "loss": 0.5675, "step": 1580 }, { "epoch": 0.651863923405989, "grad_norm": 5.037132263183594, "learning_rate": 7.936682725395733e-06, "loss": 0.5838, "step": 1600 }, { "epoch": 0.6600122224485638, "grad_norm": 6.442205429077148, "learning_rate": 7.909153475567792e-06, "loss": 0.5275, "step": 1620 }, { "epoch": 0.6681605214911387, "grad_norm": 4.626034259796143, "learning_rate": 7.881624225739849e-06, "loss": 0.5251, "step": 1640 }, { "epoch": 0.6763088205337136, "grad_norm": 6.96089506149292, "learning_rate": 7.854094975911907e-06, "loss": 0.589, "step": 1660 }, { "epoch": 0.6844571195762884, "grad_norm": 5.303503036499023, "learning_rate": 7.826565726083964e-06, "loss": 0.6233, "step": 1680 }, { "epoch": 0.6926054186188633, "grad_norm": 6.015293121337891, "learning_rate": 7.799036476256022e-06, "loss": 0.5517, "step": 1700 }, { "epoch": 0.7007537176614381, "grad_norm": 5.348663806915283, "learning_rate": 7.77150722642808e-06, "loss": 0.5403, "step": 1720 }, { "epoch": 0.708902016704013, "grad_norm": 4.189513206481934, "learning_rate": 7.743977976600138e-06, "loss": 0.5817, "step": 1740 }, { "epoch": 0.717050315746588, "grad_norm": 7.446457386016846, "learning_rate": 7.716448726772196e-06, "loss": 0.5437, "step": 1760 }, { "epoch": 0.7251986147891628, "grad_norm": 4.759880542755127, "learning_rate": 7.688919476944255e-06, "loss": 0.5648, "step": 1780 }, { "epoch": 0.7333469138317377, "grad_norm": 5.6719584465026855, "learning_rate": 7.661390227116311e-06, "loss": 0.6118, "step": 1800 }, { "epoch": 0.7414952128743125, "grad_norm": 4.943815231323242, "learning_rate": 7.63386097728837e-06, "loss": 0.5326, "step": 1820 }, { "epoch": 0.7496435119168874, "grad_norm": 3.6254968643188477, "learning_rate": 7.6063317274604275e-06, "loss": 0.5087, "step": 1840 }, { "epoch": 0.7577918109594622, "grad_norm": 5.9594526290893555, "learning_rate": 7.578802477632485e-06, "loss": 0.532, "step": 1860 }, { "epoch": 0.7659401100020371, "grad_norm": 5.103460311889648, "learning_rate": 7.551273227804543e-06, "loss": 0.553, "step": 1880 }, { "epoch": 0.774088409044612, "grad_norm": 6.327749729156494, "learning_rate": 7.5237439779766004e-06, "loss": 0.5767, "step": 1900 }, { "epoch": 0.7822367080871868, "grad_norm": 5.8910088539123535, "learning_rate": 7.496214728148659e-06, "loss": 0.5948, "step": 1920 }, { "epoch": 0.7903850071297617, "grad_norm": 5.571481227874756, "learning_rate": 7.4686854783207165e-06, "loss": 0.5199, "step": 1940 }, { "epoch": 0.7985333061723365, "grad_norm": 4.894184589385986, "learning_rate": 7.441156228492774e-06, "loss": 0.5577, "step": 1960 }, { "epoch": 0.8066816052149114, "grad_norm": 5.402436256408691, "learning_rate": 7.413626978664832e-06, "loss": 0.5395, "step": 1980 }, { "epoch": 0.8148299042574862, "grad_norm": 5.212170600891113, "learning_rate": 7.3860977288368894e-06, "loss": 0.5319, "step": 2000 }, { "epoch": 0.8148299042574862, "eval_loss": 0.4680393934249878, "eval_runtime": 10851.5751, "eval_samples_per_second": 1.81, "eval_steps_per_second": 0.226, "eval_wer": 38.607193210094984, "step": 2000 }, { "epoch": 0.8229782033000611, "grad_norm": 5.673357963562012, "learning_rate": 7.358568479008947e-06, "loss": 0.5641, "step": 2020 }, { "epoch": 0.831126502342636, "grad_norm": 4.529426097869873, "learning_rate": 7.331039229181005e-06, "loss": 0.489, "step": 2040 }, { "epoch": 0.8392748013852108, "grad_norm": 4.871108055114746, "learning_rate": 7.303509979353063e-06, "loss": 0.5413, "step": 2060 }, { "epoch": 0.8474231004277857, "grad_norm": 4.407359600067139, "learning_rate": 7.275980729525121e-06, "loss": 0.4993, "step": 2080 }, { "epoch": 0.8555713994703605, "grad_norm": 4.965864181518555, "learning_rate": 7.248451479697178e-06, "loss": 0.5825, "step": 2100 }, { "epoch": 0.8637196985129354, "grad_norm": 6.825056552886963, "learning_rate": 7.220922229869237e-06, "loss": 0.5672, "step": 2120 }, { "epoch": 0.8718679975555103, "grad_norm": 5.5648040771484375, "learning_rate": 7.1933929800412945e-06, "loss": 0.5188, "step": 2140 }, { "epoch": 0.8800162965980851, "grad_norm": 6.5213799476623535, "learning_rate": 7.165863730213353e-06, "loss": 0.5806, "step": 2160 }, { "epoch": 0.88816459564066, "grad_norm": 5.0149993896484375, "learning_rate": 7.138334480385411e-06, "loss": 0.5729, "step": 2180 }, { "epoch": 0.8963128946832348, "grad_norm": 4.783459186553955, "learning_rate": 7.110805230557468e-06, "loss": 0.4991, "step": 2200 }, { "epoch": 0.9044611937258097, "grad_norm": 5.186647891998291, "learning_rate": 7.083275980729526e-06, "loss": 0.6553, "step": 2220 }, { "epoch": 0.9126094927683845, "grad_norm": 4.135760307312012, "learning_rate": 7.0557467309015835e-06, "loss": 0.5947, "step": 2240 }, { "epoch": 0.9207577918109595, "grad_norm": 5.20528507232666, "learning_rate": 7.028217481073641e-06, "loss": 0.5454, "step": 2260 }, { "epoch": 0.9289060908535344, "grad_norm": 5.948816299438477, "learning_rate": 7.000688231245699e-06, "loss": 0.5475, "step": 2280 }, { "epoch": 0.9370543898961092, "grad_norm": 5.592283248901367, "learning_rate": 6.973158981417757e-06, "loss": 0.5153, "step": 2300 }, { "epoch": 0.9452026889386841, "grad_norm": 5.162081241607666, "learning_rate": 6.945629731589815e-06, "loss": 0.4637, "step": 2320 }, { "epoch": 0.953350987981259, "grad_norm": 5.888420104980469, "learning_rate": 6.9181004817618725e-06, "loss": 0.4956, "step": 2340 }, { "epoch": 0.9614992870238338, "grad_norm": 6.354236602783203, "learning_rate": 6.89057123193393e-06, "loss": 0.4912, "step": 2360 }, { "epoch": 0.9696475860664087, "grad_norm": 4.269996166229248, "learning_rate": 6.863041982105988e-06, "loss": 0.489, "step": 2380 }, { "epoch": 0.9777958851089835, "grad_norm": 5.822803974151611, "learning_rate": 6.835512732278045e-06, "loss": 0.55, "step": 2400 }, { "epoch": 0.9859441841515584, "grad_norm": 4.092251777648926, "learning_rate": 6.807983482450103e-06, "loss": 0.5625, "step": 2420 }, { "epoch": 0.9940924831941332, "grad_norm": 6.759193420410156, "learning_rate": 6.7804542326221615e-06, "loss": 0.5363, "step": 2440 }, { "epoch": 1.0020370747606437, "grad_norm": 5.126402378082275, "learning_rate": 6.752924982794219e-06, "loss": 0.535, "step": 2460 }, { "epoch": 1.0101853738032185, "grad_norm": 5.261082172393799, "learning_rate": 6.725395732966277e-06, "loss": 0.4262, "step": 2480 }, { "epoch": 1.0183336728457935, "grad_norm": 4.407663345336914, "learning_rate": 6.697866483138335e-06, "loss": 0.4076, "step": 2500 }, { "epoch": 1.0264819718883682, "grad_norm": 5.785089492797852, "learning_rate": 6.670337233310394e-06, "loss": 0.4255, "step": 2520 }, { "epoch": 1.0346302709309432, "grad_norm": 4.659370422363281, "learning_rate": 6.642807983482451e-06, "loss": 0.4116, "step": 2540 }, { "epoch": 1.042778569973518, "grad_norm": 5.880095958709717, "learning_rate": 6.615278733654509e-06, "loss": 0.4943, "step": 2560 }, { "epoch": 1.050926869016093, "grad_norm": 4.94119930267334, "learning_rate": 6.587749483826567e-06, "loss": 0.3904, "step": 2580 }, { "epoch": 1.0590751680586679, "grad_norm": 4.008668422698975, "learning_rate": 6.560220233998624e-06, "loss": 0.403, "step": 2600 }, { "epoch": 1.0672234671012426, "grad_norm": 6.079473495483398, "learning_rate": 6.532690984170682e-06, "loss": 0.4833, "step": 2620 }, { "epoch": 1.0753717661438176, "grad_norm": 5.263147830963135, "learning_rate": 6.5051617343427395e-06, "loss": 0.4423, "step": 2640 }, { "epoch": 1.0835200651863923, "grad_norm": 4.1324543952941895, "learning_rate": 6.477632484514798e-06, "loss": 0.4631, "step": 2660 }, { "epoch": 1.0916683642289673, "grad_norm": 3.648576259613037, "learning_rate": 6.450103234686856e-06, "loss": 0.4002, "step": 2680 }, { "epoch": 1.099816663271542, "grad_norm": 5.458010673522949, "learning_rate": 6.422573984858913e-06, "loss": 0.4572, "step": 2700 }, { "epoch": 1.107964962314117, "grad_norm": 5.379840850830078, "learning_rate": 6.395044735030971e-06, "loss": 0.4283, "step": 2720 }, { "epoch": 1.1161132613566918, "grad_norm": 5.2092108726501465, "learning_rate": 6.3675154852030285e-06, "loss": 0.4384, "step": 2740 }, { "epoch": 1.1242615603992667, "grad_norm": 5.582503795623779, "learning_rate": 6.339986235375086e-06, "loss": 0.4202, "step": 2760 }, { "epoch": 1.1324098594418415, "grad_norm": 4.186822891235352, "learning_rate": 6.312456985547144e-06, "loss": 0.3972, "step": 2780 }, { "epoch": 1.1405581584844164, "grad_norm": 5.029235363006592, "learning_rate": 6.284927735719201e-06, "loss": 0.4368, "step": 2800 }, { "epoch": 1.1487064575269912, "grad_norm": 4.080201148986816, "learning_rate": 6.25739848589126e-06, "loss": 0.4561, "step": 2820 }, { "epoch": 1.1568547565695662, "grad_norm": 5.171266078948975, "learning_rate": 6.2298692360633175e-06, "loss": 0.4594, "step": 2840 }, { "epoch": 1.165003055612141, "grad_norm": 6.01788854598999, "learning_rate": 6.202339986235376e-06, "loss": 0.363, "step": 2860 }, { "epoch": 1.1731513546547159, "grad_norm": 5.374858379364014, "learning_rate": 6.1748107364074336e-06, "loss": 0.4782, "step": 2880 }, { "epoch": 1.1812996536972906, "grad_norm": 4.696728706359863, "learning_rate": 6.147281486579492e-06, "loss": 0.4598, "step": 2900 }, { "epoch": 1.1894479527398656, "grad_norm": 4.785463333129883, "learning_rate": 6.11975223675155e-06, "loss": 0.4502, "step": 2920 }, { "epoch": 1.1975962517824403, "grad_norm": 4.824402809143066, "learning_rate": 6.092222986923607e-06, "loss": 0.4216, "step": 2940 }, { "epoch": 1.2057445508250153, "grad_norm": 4.554792404174805, "learning_rate": 6.064693737095665e-06, "loss": 0.39, "step": 2960 }, { "epoch": 1.21389284986759, "grad_norm": 5.259772300720215, "learning_rate": 6.0371644872677226e-06, "loss": 0.4199, "step": 2980 }, { "epoch": 1.222041148910165, "grad_norm": 4.8776140213012695, "learning_rate": 6.00963523743978e-06, "loss": 0.4742, "step": 3000 }, { "epoch": 1.23018944795274, "grad_norm": 5.5702033042907715, "learning_rate": 5.982105987611838e-06, "loss": 0.3934, "step": 3020 }, { "epoch": 1.2383377469953147, "grad_norm": 5.325311660766602, "learning_rate": 5.954576737783896e-06, "loss": 0.3655, "step": 3040 }, { "epoch": 1.2464860460378895, "grad_norm": 3.9275801181793213, "learning_rate": 5.927047487955954e-06, "loss": 0.3965, "step": 3060 }, { "epoch": 1.2546343450804645, "grad_norm": 4.71364688873291, "learning_rate": 5.8995182381280116e-06, "loss": 0.4026, "step": 3080 }, { "epoch": 1.2627826441230394, "grad_norm": 4.507187843322754, "learning_rate": 5.871988988300069e-06, "loss": 0.4004, "step": 3100 }, { "epoch": 1.2709309431656142, "grad_norm": 4.939427375793457, "learning_rate": 5.844459738472127e-06, "loss": 0.407, "step": 3120 }, { "epoch": 1.279079242208189, "grad_norm": 5.642360210418701, "learning_rate": 5.8169304886441844e-06, "loss": 0.4273, "step": 3140 }, { "epoch": 1.2872275412507639, "grad_norm": 4.531342506408691, "learning_rate": 5.789401238816242e-06, "loss": 0.4455, "step": 3160 }, { "epoch": 1.2953758402933389, "grad_norm": 5.926376819610596, "learning_rate": 5.7618719889883005e-06, "loss": 0.4627, "step": 3180 }, { "epoch": 1.3035241393359136, "grad_norm": 3.818305730819702, "learning_rate": 5.734342739160358e-06, "loss": 0.3957, "step": 3200 }, { "epoch": 1.3116724383784886, "grad_norm": 4.200014114379883, "learning_rate": 5.706813489332416e-06, "loss": 0.4516, "step": 3220 }, { "epoch": 1.3198207374210633, "grad_norm": 5.380039691925049, "learning_rate": 5.679284239504474e-06, "loss": 0.4775, "step": 3240 }, { "epoch": 1.3279690364636383, "grad_norm": 6.382922172546387, "learning_rate": 5.651754989676532e-06, "loss": 0.3792, "step": 3260 }, { "epoch": 1.336117335506213, "grad_norm": 5.670558929443359, "learning_rate": 5.62422573984859e-06, "loss": 0.4301, "step": 3280 }, { "epoch": 1.344265634548788, "grad_norm": 4.901285171508789, "learning_rate": 5.596696490020648e-06, "loss": 0.4095, "step": 3300 }, { "epoch": 1.3524139335913627, "grad_norm": 5.272710800170898, "learning_rate": 5.569167240192706e-06, "loss": 0.3585, "step": 3320 }, { "epoch": 1.3605622326339377, "grad_norm": 5.442037582397461, "learning_rate": 5.541637990364763e-06, "loss": 0.4019, "step": 3340 }, { "epoch": 1.3687105316765125, "grad_norm": 4.817733287811279, "learning_rate": 5.514108740536821e-06, "loss": 0.368, "step": 3360 }, { "epoch": 1.3768588307190874, "grad_norm": 4.459897994995117, "learning_rate": 5.4865794907088785e-06, "loss": 0.3646, "step": 3380 }, { "epoch": 1.3850071297616622, "grad_norm": 5.927896976470947, "learning_rate": 5.459050240880936e-06, "loss": 0.4448, "step": 3400 }, { "epoch": 1.3931554288042372, "grad_norm": 5.564925193786621, "learning_rate": 5.431520991052995e-06, "loss": 0.3794, "step": 3420 }, { "epoch": 1.401303727846812, "grad_norm": 4.68287467956543, "learning_rate": 5.403991741225052e-06, "loss": 0.3931, "step": 3440 }, { "epoch": 1.4094520268893869, "grad_norm": 5.004433631896973, "learning_rate": 5.37646249139711e-06, "loss": 0.432, "step": 3460 }, { "epoch": 1.4176003259319616, "grad_norm": 5.418098449707031, "learning_rate": 5.3489332415691675e-06, "loss": 0.4533, "step": 3480 }, { "epoch": 1.4257486249745366, "grad_norm": 4.825730800628662, "learning_rate": 5.321403991741225e-06, "loss": 0.4159, "step": 3500 }, { "epoch": 1.4338969240171116, "grad_norm": 5.892017364501953, "learning_rate": 5.293874741913283e-06, "loss": 0.394, "step": 3520 }, { "epoch": 1.4420452230596863, "grad_norm": 5.600419521331787, "learning_rate": 5.26634549208534e-06, "loss": 0.4357, "step": 3540 }, { "epoch": 1.450193522102261, "grad_norm": 4.615939617156982, "learning_rate": 5.238816242257399e-06, "loss": 0.4029, "step": 3560 }, { "epoch": 1.458341821144836, "grad_norm": 4.384542465209961, "learning_rate": 5.2112869924294565e-06, "loss": 0.4042, "step": 3580 }, { "epoch": 1.466490120187411, "grad_norm": 5.981232643127441, "learning_rate": 5.183757742601514e-06, "loss": 0.3923, "step": 3600 }, { "epoch": 1.4746384192299857, "grad_norm": 5.077232837677002, "learning_rate": 5.156228492773573e-06, "loss": 0.4995, "step": 3620 }, { "epoch": 1.4827867182725605, "grad_norm": 5.000918865203857, "learning_rate": 5.12869924294563e-06, "loss": 0.4167, "step": 3640 }, { "epoch": 1.4909350173151354, "grad_norm": 2.953489303588867, "learning_rate": 5.101169993117689e-06, "loss": 0.3798, "step": 3660 }, { "epoch": 1.4990833163577104, "grad_norm": 5.880030632019043, "learning_rate": 5.073640743289746e-06, "loss": 0.4349, "step": 3680 }, { "epoch": 1.5072316154002852, "grad_norm": 5.726639270782471, "learning_rate": 5.046111493461804e-06, "loss": 0.3809, "step": 3700 }, { "epoch": 1.51537991444286, "grad_norm": 4.755875587463379, "learning_rate": 5.018582243633862e-06, "loss": 0.4345, "step": 3720 }, { "epoch": 1.5235282134854349, "grad_norm": 3.825904130935669, "learning_rate": 4.991052993805919e-06, "loss": 0.4304, "step": 3740 }, { "epoch": 1.5316765125280098, "grad_norm": 6.1705217361450195, "learning_rate": 4.963523743977977e-06, "loss": 0.4224, "step": 3760 }, { "epoch": 1.5398248115705846, "grad_norm": 5.098936557769775, "learning_rate": 4.9359944941500345e-06, "loss": 0.4254, "step": 3780 }, { "epoch": 1.5479731106131593, "grad_norm": 4.633610248565674, "learning_rate": 4.908465244322093e-06, "loss": 0.3542, "step": 3800 }, { "epoch": 1.5561214096557343, "grad_norm": 5.469082832336426, "learning_rate": 4.880935994494151e-06, "loss": 0.3577, "step": 3820 }, { "epoch": 1.5642697086983093, "grad_norm": 4.476110458374023, "learning_rate": 4.853406744666208e-06, "loss": 0.3931, "step": 3840 }, { "epoch": 1.5724180077408842, "grad_norm": 4.693641185760498, "learning_rate": 4.825877494838266e-06, "loss": 0.3793, "step": 3860 }, { "epoch": 1.580566306783459, "grad_norm": 5.025706768035889, "learning_rate": 4.7983482450103235e-06, "loss": 0.4152, "step": 3880 }, { "epoch": 1.5887146058260337, "grad_norm": 5.075003147125244, "learning_rate": 4.770818995182382e-06, "loss": 0.3813, "step": 3900 }, { "epoch": 1.5968629048686087, "grad_norm": 4.8126115798950195, "learning_rate": 4.74328974535444e-06, "loss": 0.408, "step": 3920 }, { "epoch": 1.6050112039111837, "grad_norm": 4.3969340324401855, "learning_rate": 4.715760495526497e-06, "loss": 0.3667, "step": 3940 }, { "epoch": 1.6131595029537584, "grad_norm": 4.514830589294434, "learning_rate": 4.688231245698555e-06, "loss": 0.4022, "step": 3960 }, { "epoch": 1.6213078019963332, "grad_norm": 5.9185309410095215, "learning_rate": 4.6607019958706125e-06, "loss": 0.4277, "step": 3980 }, { "epoch": 1.6294561010389081, "grad_norm": 4.446202754974365, "learning_rate": 4.633172746042671e-06, "loss": 0.4179, "step": 4000 }, { "epoch": 1.6294561010389081, "eval_loss": 0.3749663531780243, "eval_runtime": 13864.4556, "eval_samples_per_second": 1.416, "eval_steps_per_second": 0.177, "eval_wer": 60.07223542231883, "step": 4000 }, { "epoch": 1.6376044000814831, "grad_norm": 5.807653903961182, "learning_rate": 4.605643496214729e-06, "loss": 0.4408, "step": 4020 }, { "epoch": 1.6457526991240579, "grad_norm": 7.353831768035889, "learning_rate": 4.578114246386786e-06, "loss": 0.4145, "step": 4040 }, { "epoch": 1.6539009981666326, "grad_norm": 3.9637539386749268, "learning_rate": 4.550584996558844e-06, "loss": 0.4446, "step": 4060 }, { "epoch": 1.6620492972092076, "grad_norm": 5.770019054412842, "learning_rate": 4.523055746730902e-06, "loss": 0.428, "step": 4080 }, { "epoch": 1.6701975962517825, "grad_norm": 6.5047688484191895, "learning_rate": 4.49552649690296e-06, "loss": 0.4178, "step": 4100 }, { "epoch": 1.6783458952943573, "grad_norm": 6.010311603546143, "learning_rate": 4.467997247075018e-06, "loss": 0.4164, "step": 4120 }, { "epoch": 1.686494194336932, "grad_norm": 5.805934906005859, "learning_rate": 4.440467997247075e-06, "loss": 0.388, "step": 4140 }, { "epoch": 1.694642493379507, "grad_norm": 5.374007701873779, "learning_rate": 4.412938747419133e-06, "loss": 0.4259, "step": 4160 }, { "epoch": 1.702790792422082, "grad_norm": 5.766719341278076, "learning_rate": 4.385409497591191e-06, "loss": 0.4388, "step": 4180 }, { "epoch": 1.7109390914646567, "grad_norm": 3.3628714084625244, "learning_rate": 4.357880247763249e-06, "loss": 0.3568, "step": 4200 }, { "epoch": 1.7190873905072315, "grad_norm": 3.830449342727661, "learning_rate": 4.3303509979353066e-06, "loss": 0.391, "step": 4220 }, { "epoch": 1.7272356895498064, "grad_norm": 5.824588298797607, "learning_rate": 4.302821748107364e-06, "loss": 0.4416, "step": 4240 }, { "epoch": 1.7353839885923814, "grad_norm": 5.445583343505859, "learning_rate": 4.275292498279423e-06, "loss": 0.3935, "step": 4260 }, { "epoch": 1.7435322876349562, "grad_norm": 3.97123646736145, "learning_rate": 4.24776324845148e-06, "loss": 0.3596, "step": 4280 }, { "epoch": 1.751680586677531, "grad_norm": 3.1131341457366943, "learning_rate": 4.220233998623538e-06, "loss": 0.4498, "step": 4300 }, { "epoch": 1.7598288857201059, "grad_norm": 3.785813093185425, "learning_rate": 4.1927047487955956e-06, "loss": 0.389, "step": 4320 }, { "epoch": 1.7679771847626808, "grad_norm": 4.826812744140625, "learning_rate": 4.165175498967653e-06, "loss": 0.4157, "step": 4340 }, { "epoch": 1.7761254838052558, "grad_norm": 5.719616889953613, "learning_rate": 4.137646249139711e-06, "loss": 0.333, "step": 4360 }, { "epoch": 1.7842737828478306, "grad_norm": 5.590183734893799, "learning_rate": 4.110116999311769e-06, "loss": 0.4176, "step": 4380 }, { "epoch": 1.7924220818904053, "grad_norm": 7.032625675201416, "learning_rate": 4.082587749483827e-06, "loss": 0.3992, "step": 4400 }, { "epoch": 1.8005703809329803, "grad_norm": 5.545759677886963, "learning_rate": 4.0550584996558846e-06, "loss": 0.4189, "step": 4420 }, { "epoch": 1.8087186799755552, "grad_norm": 6.759115219116211, "learning_rate": 4.027529249827942e-06, "loss": 0.413, "step": 4440 }, { "epoch": 1.81686697901813, "grad_norm": 5.781651973724365, "learning_rate": 4.000000000000001e-06, "loss": 0.4273, "step": 4460 }, { "epoch": 1.8250152780607047, "grad_norm": 4.690906524658203, "learning_rate": 3.972470750172058e-06, "loss": 0.3943, "step": 4480 }, { "epoch": 1.8331635771032797, "grad_norm": 5.967350006103516, "learning_rate": 3.944941500344116e-06, "loss": 0.3424, "step": 4500 }, { "epoch": 1.8413118761458547, "grad_norm": 5.927072048187256, "learning_rate": 3.9174122505161736e-06, "loss": 0.3857, "step": 4520 }, { "epoch": 1.8494601751884294, "grad_norm": 3.1506669521331787, "learning_rate": 3.889883000688231e-06, "loss": 0.3751, "step": 4540 }, { "epoch": 1.8576084742310042, "grad_norm": 3.858947992324829, "learning_rate": 3.86235375086029e-06, "loss": 0.3735, "step": 4560 }, { "epoch": 1.8657567732735791, "grad_norm": 6.204343318939209, "learning_rate": 3.834824501032347e-06, "loss": 0.4406, "step": 4580 }, { "epoch": 1.873905072316154, "grad_norm": 5.989670276641846, "learning_rate": 3.807295251204405e-06, "loss": 0.4468, "step": 4600 }, { "epoch": 1.8820533713587289, "grad_norm": 5.673979759216309, "learning_rate": 3.7797660013764625e-06, "loss": 0.4066, "step": 4620 }, { "epoch": 1.8902016704013036, "grad_norm": 4.2719197273254395, "learning_rate": 3.752236751548521e-06, "loss": 0.4463, "step": 4640 }, { "epoch": 1.8983499694438786, "grad_norm": 4.865420818328857, "learning_rate": 3.7247075017205787e-06, "loss": 0.4532, "step": 4660 }, { "epoch": 1.9064982684864535, "grad_norm": 4.259848117828369, "learning_rate": 3.6971782518926363e-06, "loss": 0.462, "step": 4680 }, { "epoch": 1.9146465675290283, "grad_norm": 3.7624459266662598, "learning_rate": 3.669649002064694e-06, "loss": 0.4175, "step": 4700 }, { "epoch": 1.922794866571603, "grad_norm": 4.538738250732422, "learning_rate": 3.642119752236752e-06, "loss": 0.3381, "step": 4720 }, { "epoch": 1.930943165614178, "grad_norm": 3.968751907348633, "learning_rate": 3.6145905024088096e-06, "loss": 0.3636, "step": 4740 }, { "epoch": 1.939091464656753, "grad_norm": 4.724824905395508, "learning_rate": 3.5870612525808672e-06, "loss": 0.3928, "step": 4760 }, { "epoch": 1.9472397636993277, "grad_norm": 4.30732536315918, "learning_rate": 3.5595320027529253e-06, "loss": 0.4176, "step": 4780 }, { "epoch": 1.9553880627419025, "grad_norm": 5.967071533203125, "learning_rate": 3.532002752924983e-06, "loss": 0.3979, "step": 4800 }, { "epoch": 1.9635363617844774, "grad_norm": 5.40061092376709, "learning_rate": 3.504473503097041e-06, "loss": 0.4085, "step": 4820 }, { "epoch": 1.9716846608270524, "grad_norm": 6.015532970428467, "learning_rate": 3.476944253269099e-06, "loss": 0.3589, "step": 4840 }, { "epoch": 1.9798329598696274, "grad_norm": 4.844973087310791, "learning_rate": 3.4494150034411566e-06, "loss": 0.4034, "step": 4860 }, { "epoch": 1.9879812589122021, "grad_norm": 3.5679476261138916, "learning_rate": 3.4218857536132143e-06, "loss": 0.3458, "step": 4880 }, { "epoch": 1.9961295579547769, "grad_norm": 5.0959153175354, "learning_rate": 3.3943565037852723e-06, "loss": 0.3818, "step": 4900 }, { "epoch": 2.0040741495212875, "grad_norm": 4.379068851470947, "learning_rate": 3.36682725395733e-06, "loss": 0.3527, "step": 4920 }, { "epoch": 2.0122224485638625, "grad_norm": 4.350337982177734, "learning_rate": 3.3392980041293876e-06, "loss": 0.348, "step": 4940 }, { "epoch": 2.020370747606437, "grad_norm": 3.3158302307128906, "learning_rate": 3.311768754301445e-06, "loss": 0.2506, "step": 4960 }, { "epoch": 2.028519046649012, "grad_norm": 5.536191463470459, "learning_rate": 3.2842395044735033e-06, "loss": 0.3206, "step": 4980 }, { "epoch": 2.036667345691587, "grad_norm": 7.019426345825195, "learning_rate": 3.256710254645561e-06, "loss": 0.3724, "step": 5000 }, { "epoch": 2.044815644734162, "grad_norm": 4.025884628295898, "learning_rate": 3.2291810048176194e-06, "loss": 0.3062, "step": 5020 }, { "epoch": 2.0529639437767364, "grad_norm": 4.956037521362305, "learning_rate": 3.201651754989677e-06, "loss": 0.3124, "step": 5040 }, { "epoch": 2.0611122428193114, "grad_norm": 6.105999946594238, "learning_rate": 3.1741225051617346e-06, "loss": 0.3047, "step": 5060 }, { "epoch": 2.0692605418618863, "grad_norm": 5.956209182739258, "learning_rate": 3.1465932553337922e-06, "loss": 0.3221, "step": 5080 }, { "epoch": 2.0774088409044613, "grad_norm": 5.465188026428223, "learning_rate": 3.1190640055058503e-06, "loss": 0.2867, "step": 5100 }, { "epoch": 2.085557139947036, "grad_norm": 5.349979877471924, "learning_rate": 3.091534755677908e-06, "loss": 0.323, "step": 5120 }, { "epoch": 2.093705438989611, "grad_norm": 4.8694305419921875, "learning_rate": 3.0640055058499656e-06, "loss": 0.3486, "step": 5140 }, { "epoch": 2.101853738032186, "grad_norm": 4.890730857849121, "learning_rate": 3.0364762560220236e-06, "loss": 0.3654, "step": 5160 }, { "epoch": 2.1100020370747608, "grad_norm": 4.631341457366943, "learning_rate": 3.0089470061940812e-06, "loss": 0.2903, "step": 5180 }, { "epoch": 2.1181503361173357, "grad_norm": 3.8383870124816895, "learning_rate": 2.9814177563661393e-06, "loss": 0.2885, "step": 5200 }, { "epoch": 2.1262986351599102, "grad_norm": 4.418496131896973, "learning_rate": 2.9538885065381973e-06, "loss": 0.3318, "step": 5220 }, { "epoch": 2.134446934202485, "grad_norm": 5.5654401779174805, "learning_rate": 2.926359256710255e-06, "loss": 0.3022, "step": 5240 }, { "epoch": 2.14259523324506, "grad_norm": 3.4770777225494385, "learning_rate": 2.8988300068823126e-06, "loss": 0.3232, "step": 5260 }, { "epoch": 2.150743532287635, "grad_norm": 3.5590999126434326, "learning_rate": 2.8713007570543707e-06, "loss": 0.2824, "step": 5280 }, { "epoch": 2.1588918313302097, "grad_norm": 5.644620895385742, "learning_rate": 2.8437715072264283e-06, "loss": 0.3407, "step": 5300 }, { "epoch": 2.1670401303727846, "grad_norm": 4.308077335357666, "learning_rate": 2.816242257398486e-06, "loss": 0.2948, "step": 5320 }, { "epoch": 2.1751884294153596, "grad_norm": 4.851817607879639, "learning_rate": 2.7887130075705435e-06, "loss": 0.3457, "step": 5340 }, { "epoch": 2.1833367284579346, "grad_norm": 4.964145660400391, "learning_rate": 2.7611837577426016e-06, "loss": 0.318, "step": 5360 }, { "epoch": 2.191485027500509, "grad_norm": 3.544668197631836, "learning_rate": 2.7336545079146596e-06, "loss": 0.3017, "step": 5380 }, { "epoch": 2.199633326543084, "grad_norm": 5.297823905944824, "learning_rate": 2.7061252580867177e-06, "loss": 0.2955, "step": 5400 }, { "epoch": 2.207781625585659, "grad_norm": 4.36570930480957, "learning_rate": 2.6785960082587753e-06, "loss": 0.3451, "step": 5420 }, { "epoch": 2.215929924628234, "grad_norm": 4.8854475021362305, "learning_rate": 2.651066758430833e-06, "loss": 0.3115, "step": 5440 }, { "epoch": 2.2240782236708085, "grad_norm": 5.848690986633301, "learning_rate": 2.623537508602891e-06, "loss": 0.3196, "step": 5460 }, { "epoch": 2.2322265227133835, "grad_norm": 6.296275615692139, "learning_rate": 2.5960082587749486e-06, "loss": 0.3577, "step": 5480 }, { "epoch": 2.2403748217559585, "grad_norm": 6.340885162353516, "learning_rate": 2.5684790089470063e-06, "loss": 0.3347, "step": 5500 }, { "epoch": 2.2485231207985334, "grad_norm": 5.3483195304870605, "learning_rate": 2.540949759119064e-06, "loss": 0.2829, "step": 5520 }, { "epoch": 2.2566714198411084, "grad_norm": 3.8228771686553955, "learning_rate": 2.513420509291122e-06, "loss": 0.3002, "step": 5540 }, { "epoch": 2.264819718883683, "grad_norm": 5.000001907348633, "learning_rate": 2.48589125946318e-06, "loss": 0.2816, "step": 5560 }, { "epoch": 2.272968017926258, "grad_norm": 4.822083473205566, "learning_rate": 2.4583620096352376e-06, "loss": 0.2982, "step": 5580 }, { "epoch": 2.281116316968833, "grad_norm": 4.827035427093506, "learning_rate": 2.4308327598072953e-06, "loss": 0.3384, "step": 5600 }, { "epoch": 2.2892646160114074, "grad_norm": 4.659197807312012, "learning_rate": 2.4033035099793533e-06, "loss": 0.2845, "step": 5620 }, { "epoch": 2.2974129150539824, "grad_norm": 3.4446418285369873, "learning_rate": 2.375774260151411e-06, "loss": 0.3019, "step": 5640 }, { "epoch": 2.3055612140965573, "grad_norm": 4.726231575012207, "learning_rate": 2.348245010323469e-06, "loss": 0.3568, "step": 5660 }, { "epoch": 2.3137095131391323, "grad_norm": 5.017887115478516, "learning_rate": 2.3207157604955266e-06, "loss": 0.2857, "step": 5680 }, { "epoch": 2.3218578121817073, "grad_norm": 4.18927001953125, "learning_rate": 2.2931865106675847e-06, "loss": 0.3285, "step": 5700 }, { "epoch": 2.330006111224282, "grad_norm": 5.701695919036865, "learning_rate": 2.2656572608396423e-06, "loss": 0.3167, "step": 5720 }, { "epoch": 2.3381544102668568, "grad_norm": 3.6883463859558105, "learning_rate": 2.2381280110117e-06, "loss": 0.3102, "step": 5740 }, { "epoch": 2.3463027093094317, "grad_norm": 5.979183197021484, "learning_rate": 2.210598761183758e-06, "loss": 0.3405, "step": 5760 }, { "epoch": 2.3544510083520063, "grad_norm": 3.127885580062866, "learning_rate": 2.1830695113558156e-06, "loss": 0.3427, "step": 5780 }, { "epoch": 2.3625993073945812, "grad_norm": 4.842801094055176, "learning_rate": 2.1555402615278737e-06, "loss": 0.2836, "step": 5800 }, { "epoch": 2.370747606437156, "grad_norm": 4.311405658721924, "learning_rate": 2.1280110116999313e-06, "loss": 0.3005, "step": 5820 }, { "epoch": 2.378895905479731, "grad_norm": 4.682214260101318, "learning_rate": 2.1004817618719894e-06, "loss": 0.321, "step": 5840 }, { "epoch": 2.387044204522306, "grad_norm": 7.238677501678467, "learning_rate": 2.072952512044047e-06, "loss": 0.2943, "step": 5860 }, { "epoch": 2.3951925035648807, "grad_norm": 2.814465045928955, "learning_rate": 2.0454232622161046e-06, "loss": 0.3057, "step": 5880 }, { "epoch": 2.4033408026074556, "grad_norm": 5.485500812530518, "learning_rate": 2.0178940123881627e-06, "loss": 0.3054, "step": 5900 }, { "epoch": 2.4114891016500306, "grad_norm": 5.754581451416016, "learning_rate": 1.9903647625602203e-06, "loss": 0.314, "step": 5920 }, { "epoch": 2.4196374006926056, "grad_norm": 3.692474603652954, "learning_rate": 1.9628355127322783e-06, "loss": 0.3377, "step": 5940 }, { "epoch": 2.42778569973518, "grad_norm": 5.103245258331299, "learning_rate": 1.935306262904336e-06, "loss": 0.3204, "step": 5960 }, { "epoch": 2.435933998777755, "grad_norm": 3.2880680561065674, "learning_rate": 1.907777013076394e-06, "loss": 0.2942, "step": 5980 }, { "epoch": 2.44408229782033, "grad_norm": 4.239877700805664, "learning_rate": 1.8802477632484517e-06, "loss": 0.3616, "step": 6000 }, { "epoch": 2.44408229782033, "eval_loss": 0.3231365978717804, "eval_runtime": 9411.3014, "eval_samples_per_second": 2.086, "eval_steps_per_second": 0.261, "eval_wer": 21.781282824708335, "step": 6000 } ], "logging_steps": 20, "max_steps": 7365, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.53968534454272e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }